Update README.md
Browse files
README.md
CHANGED
@@ -1,5 +1,9 @@
|
|
1 |
---
|
2 |
language: en
|
|
|
|
|
|
|
|
|
3 |
license: other
|
4 |
license_name: link-attribution
|
5 |
license_link: https://dejanmarketing.com/link-attribution/
|
@@ -16,6 +20,14 @@ This model is a hierarchical text classifier designed to categorize text into a
|
|
16 |
- **Model Developers:** [DEJAN.AI](https://dejan.ai/)
|
17 |
- **Model Type:** Hierarchical Text Classification
|
18 |
- **Base Model:** [`albert/albert-base-v2`](https://huggingface.co/albert/albert-base-v2)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
19 |
- **Model Architecture:**
|
20 |
- **Level 1:** Standard sequence classification using `AlbertForSequenceClassification`.
|
21 |
- **Levels 2-7:** Custom architecture (`TaxonomyClassifier`) where the ALBERT pooled output is concatenated with a one-hot encoded representation of the predicted ID from the previous level before being fed into a linear classification layer.
|
@@ -77,7 +89,7 @@ The model was trained on a dataset of 374,521 samples. Each row in the training
|
|
77 |
Validation loss was used as the primary evaluation metric during training. The following validation loss trends were observed:
|
78 |
|
79 |
- **Level 1, 2, and 3:** Showed a relatively rapid decrease in validation loss during training.
|
80 |
-
- **Level 4:** Exhibited a slower decrease in validation loss, potentially due to the significant increase in the dimensionality of the parent ID one-hot encoding.
|
81 |
|
82 |
Further evaluation on downstream tasks is recommended to assess the model's practical performance.
|
83 |
|
|
|
1 |
---
|
2 |
language: en
|
3 |
+
tags:
|
4 |
+
- transformers
|
5 |
+
- text-classification
|
6 |
+
- taxonomy
|
7 |
license: other
|
8 |
license_name: link-attribution
|
9 |
license_link: https://dejanmarketing.com/link-attribution/
|
|
|
20 |
- **Model Developers:** [DEJAN.AI](https://dejan.ai/)
|
21 |
- **Model Type:** Hierarchical Text Classification
|
22 |
- **Base Model:** [`albert/albert-base-v2`](https://huggingface.co/albert/albert-base-v2)
|
23 |
+
- **Taxonomy Structure:** The model classifies text into a taxonomy with the following structure:
|
24 |
+
- **Level 1:** 21 unique classes
|
25 |
+
- **Level 2:** 193 unique classes
|
26 |
+
- **Level 3:** 1350 unique classes
|
27 |
+
- **Level 4:** 2205 unique classes
|
28 |
+
- **Level 5:** 1387 unique classes
|
29 |
+
- **Level 6:** 399 unique classes
|
30 |
+
- **Level 7:** 50 unique classes
|
31 |
- **Model Architecture:**
|
32 |
- **Level 1:** Standard sequence classification using `AlbertForSequenceClassification`.
|
33 |
- **Levels 2-7:** Custom architecture (`TaxonomyClassifier`) where the ALBERT pooled output is concatenated with a one-hot encoded representation of the predicted ID from the previous level before being fed into a linear classification layer.
|
|
|
89 |
Validation loss was used as the primary evaluation metric during training. The following validation loss trends were observed:
|
90 |
|
91 |
- **Level 1, 2, and 3:** Showed a relatively rapid decrease in validation loss during training.
|
92 |
+
- **Level 4:** Exhibited a slower decrease in validation loss, potentially due to the significant increase in the dimensionality of the parent ID one-hot encoding and the larger number of unique classes at this level.
|
93 |
|
94 |
Further evaluation on downstream tasks is recommended to assess the model's practical performance.
|
95 |
|