salbatarni commited on
Commit
9d36132
·
verified ·
1 Parent(s): b8bf2e9

Training in progress, step 160

Browse files
README.md CHANGED
@@ -3,20 +3,20 @@ base_model: aubmindlab/bert-base-arabertv02
3
  tags:
4
  - generated_from_trainer
5
  model-index:
6
- - name: arabert_cross_relevance_task1_fold1
7
  results: []
8
  ---
9
 
10
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
11
  should probably proofread and complete it, then remove this comment. -->
12
 
13
- # arabert_cross_relevance_task1_fold1
14
 
15
  This model is a fine-tuned version of [aubmindlab/bert-base-arabertv02](https://huggingface.co/aubmindlab/bert-base-arabertv02) on the None dataset.
16
  It achieves the following results on the evaluation set:
17
- - Loss: 0.3646
18
- - Qwk: 0.0454
19
- - Mse: 0.3647
20
 
21
  ## Model description
22
 
@@ -36,48 +36,92 @@ More information needed
36
 
37
  The following hyperparameters were used during training:
38
  - learning_rate: 2e-05
39
- - train_batch_size: 16
40
- - eval_batch_size: 16
41
  - seed: 42
42
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
  - lr_scheduler_type: linear
44
- - num_epochs: 1
45
 
46
  ### Training results
47
 
48
- | Training Loss | Epoch | Step | Validation Loss | Qwk | Mse |
49
- |:-------------:|:------:|:----:|:---------------:|:-------:|:------:|
50
- | No log | 0.0323 | 2 | 0.3884 | -0.0193 | 0.3901 |
51
- | No log | 0.0645 | 4 | 0.5676 | 0.0570 | 0.5677 |
52
- | No log | 0.0968 | 6 | 0.6160 | 0.1181 | 0.6161 |
53
- | No log | 0.1290 | 8 | 0.3914 | 0.0326 | 0.3913 |
54
- | No log | 0.1613 | 10 | 0.2970 | 0.0245 | 0.2972 |
55
- | No log | 0.1935 | 12 | 0.3016 | 0.0122 | 0.3018 |
56
- | No log | 0.2258 | 14 | 0.2905 | 0.0122 | 0.2908 |
57
- | No log | 0.2581 | 16 | 0.3013 | 0.0491 | 0.3016 |
58
- | No log | 0.2903 | 18 | 0.3427 | 0.0284 | 0.3429 |
59
- | No log | 0.3226 | 20 | 0.4170 | 0.0059 | 0.4170 |
60
- | No log | 0.3548 | 22 | 0.4971 | 0.0518 | 0.4971 |
61
- | No log | 0.3871 | 24 | 0.4934 | 0.0958 | 0.4933 |
62
- | No log | 0.4194 | 26 | 0.4203 | -0.0430 | 0.4203 |
63
- | No log | 0.4516 | 28 | 0.3668 | -0.0521 | 0.3669 |
64
- | No log | 0.4839 | 30 | 0.3462 | -0.0092 | 0.3463 |
65
- | No log | 0.5161 | 32 | 0.3500 | -0.0092 | 0.3500 |
66
- | No log | 0.5484 | 34 | 0.3623 | -0.0010 | 0.3623 |
67
- | No log | 0.5806 | 36 | 0.3911 | -0.0188 | 0.3911 |
68
- | No log | 0.6129 | 38 | 0.3635 | -0.0135 | 0.3634 |
69
- | No log | 0.6452 | 40 | 0.3500 | -0.0092 | 0.3500 |
70
- | No log | 0.6774 | 42 | 0.3369 | -0.0048 | 0.3369 |
71
- | No log | 0.7097 | 44 | 0.3357 | -0.0048 | 0.3358 |
72
- | No log | 0.7419 | 46 | 0.3574 | 0.0537 | 0.3574 |
73
- | No log | 0.7742 | 48 | 0.3901 | 0.0370 | 0.3901 |
74
- | No log | 0.8065 | 50 | 0.3986 | 0.0723 | 0.3985 |
75
- | No log | 0.8387 | 52 | 0.3875 | 0.0544 | 0.3875 |
76
- | No log | 0.8710 | 54 | 0.3792 | 0.0586 | 0.3792 |
77
- | No log | 0.9032 | 56 | 0.3727 | 0.0412 | 0.3727 |
78
- | No log | 0.9355 | 58 | 0.3667 | 0.0454 | 0.3667 |
79
- | No log | 0.9677 | 60 | 0.3648 | 0.0454 | 0.3648 |
80
- | No log | 1.0 | 62 | 0.3646 | 0.0454 | 0.3647 |
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
81
 
82
 
83
  ### Framework versions
 
3
  tags:
4
  - generated_from_trainer
5
  model-index:
6
+ - name: arabert_cross_relevance_task1_fold0
7
  results: []
8
  ---
9
 
10
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
11
  should probably proofread and complete it, then remove this comment. -->
12
 
13
+ # arabert_cross_relevance_task1_fold0
14
 
15
  This model is a fine-tuned version of [aubmindlab/bert-base-arabertv02](https://huggingface.co/aubmindlab/bert-base-arabertv02) on the None dataset.
16
  It achieves the following results on the evaluation set:
17
+ - Loss: 0.2694
18
+ - Qwk: 0.0109
19
+ - Mse: 0.2694
20
 
21
  ## Model description
22
 
 
36
 
37
  The following hyperparameters were used during training:
38
  - learning_rate: 2e-05
39
+ - train_batch_size: 64
40
+ - eval_batch_size: 64
41
  - seed: 42
42
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
  - lr_scheduler_type: linear
44
+ - num_epochs: 10
45
 
46
  ### Training results
47
 
48
+ | Training Loss | Epoch | Step | Validation Loss | Qwk | Mse |
49
+ |:-------------:|:------:|:----:|:---------------:|:------:|:------:|
50
+ | No log | 0.1333 | 2 | 2.2712 | 0.0 | 2.2712 |
51
+ | No log | 0.2667 | 4 | 0.8815 | 0.0019 | 0.8815 |
52
+ | No log | 0.4 | 6 | 0.2576 | 0.0339 | 0.2576 |
53
+ | No log | 0.5333 | 8 | 0.2104 | 0.0276 | 0.2104 |
54
+ | No log | 0.6667 | 10 | 0.3374 | 0.0345 | 0.3374 |
55
+ | No log | 0.8 | 12 | 0.2690 | 0.0323 | 0.2690 |
56
+ | No log | 0.9333 | 14 | 0.1682 | 0.0249 | 0.1682 |
57
+ | No log | 1.0667 | 16 | 0.1503 | 0.0339 | 0.1503 |
58
+ | No log | 1.2 | 18 | 0.1672 | 0.0339 | 0.1672 |
59
+ | No log | 1.3333 | 20 | 0.2563 | 0.0263 | 0.2563 |
60
+ | No log | 1.4667 | 22 | 0.3596 | 0.0519 | 0.3596 |
61
+ | No log | 1.6 | 24 | 0.3144 | 0.0507 | 0.3144 |
62
+ | No log | 1.7333 | 26 | 0.1940 | 0.0263 | 0.1940 |
63
+ | No log | 1.8667 | 28 | 0.1460 | 0.0400 | 0.1460 |
64
+ | No log | 2.0 | 30 | 0.1369 | 0.0364 | 0.1369 |
65
+ | No log | 2.1333 | 32 | 0.1324 | 0.0372 | 0.1324 |
66
+ | No log | 2.2667 | 34 | 0.1415 | 0.0339 | 0.1415 |
67
+ | No log | 2.4 | 36 | 0.1809 | 0.0339 | 0.1809 |
68
+ | No log | 2.5333 | 38 | 0.2294 | 0.0339 | 0.2294 |
69
+ | No log | 2.6667 | 40 | 0.2654 | 0.0281 | 0.2654 |
70
+ | No log | 2.8 | 42 | 0.2940 | 0.0245 | 0.2940 |
71
+ | No log | 2.9333 | 44 | 0.2676 | 0.0245 | 0.2676 |
72
+ | No log | 3.0667 | 46 | 0.2359 | 0.0245 | 0.2359 |
73
+ | No log | 3.2 | 48 | 0.2040 | 0.0263 | 0.2040 |
74
+ | No log | 3.3333 | 50 | 0.1696 | 0.0300 | 0.1696 |
75
+ | No log | 3.4667 | 52 | 0.1616 | 0.0339 | 0.1616 |
76
+ | No log | 3.6 | 54 | 0.1702 | 0.0319 | 0.1702 |
77
+ | No log | 3.7333 | 56 | 0.2010 | 0.0281 | 0.2010 |
78
+ | No log | 3.8667 | 58 | 0.2570 | 0.0228 | 0.2570 |
79
+ | No log | 4.0 | 60 | 0.3107 | 0.0327 | 0.3107 |
80
+ | No log | 4.1333 | 62 | 0.3132 | 0.0327 | 0.3132 |
81
+ | No log | 4.2667 | 64 | 0.2646 | 0.0288 | 0.2646 |
82
+ | No log | 4.4 | 66 | 0.2086 | 0.0263 | 0.2086 |
83
+ | No log | 4.5333 | 68 | 0.1708 | 0.0300 | 0.1708 |
84
+ | No log | 4.6667 | 70 | 0.1646 | 0.0179 | 0.1646 |
85
+ | No log | 4.8 | 72 | 0.1786 | 0.0144 | 0.1786 |
86
+ | No log | 4.9333 | 74 | 0.2198 | 0.0281 | 0.2198 |
87
+ | No log | 5.0667 | 76 | 0.2585 | 0.0245 | 0.2585 |
88
+ | No log | 5.2 | 78 | 0.2513 | 0.0263 | 0.2513 |
89
+ | No log | 5.3333 | 80 | 0.2441 | 0.0263 | 0.2441 |
90
+ | No log | 5.4667 | 82 | 0.2186 | 0.0197 | 0.2186 |
91
+ | No log | 5.6 | 84 | 0.2061 | 0.0197 | 0.2061 |
92
+ | No log | 5.7333 | 86 | 0.2178 | 0.0197 | 0.2178 |
93
+ | No log | 5.8667 | 88 | 0.2322 | 0.0197 | 0.2322 |
94
+ | No log | 6.0 | 90 | 0.2425 | 0.0245 | 0.2425 |
95
+ | No log | 6.1333 | 92 | 0.2585 | 0.0245 | 0.2585 |
96
+ | No log | 6.2667 | 94 | 0.2369 | 0.0195 | 0.2369 |
97
+ | No log | 6.4 | 96 | 0.2110 | 0.0158 | 0.2110 |
98
+ | No log | 6.5333 | 98 | 0.2006 | 0.0158 | 0.2006 |
99
+ | No log | 6.6667 | 100 | 0.2175 | 0.0158 | 0.2175 |
100
+ | No log | 6.8 | 102 | 0.2464 | 0.0210 | 0.2464 |
101
+ | No log | 6.9333 | 104 | 0.2615 | 0.0225 | 0.2615 |
102
+ | No log | 7.0667 | 106 | 0.2784 | 0.0288 | 0.2784 |
103
+ | No log | 7.2 | 108 | 0.2991 | 0.0327 | 0.2991 |
104
+ | No log | 7.3333 | 110 | 0.2942 | 0.0327 | 0.2942 |
105
+ | No log | 7.4667 | 112 | 0.3083 | 0.0399 | 0.3083 |
106
+ | No log | 7.6 | 114 | 0.2910 | 0.0221 | 0.2910 |
107
+ | No log | 7.7333 | 116 | 0.2561 | 0.0093 | 0.2561 |
108
+ | No log | 7.8667 | 118 | 0.2246 | 0.0125 | 0.2246 |
109
+ | No log | 8.0 | 120 | 0.2152 | 0.0156 | 0.2152 |
110
+ | No log | 8.1333 | 122 | 0.2131 | 0.0156 | 0.2131 |
111
+ | No log | 8.2667 | 124 | 0.2238 | 0.0140 | 0.2238 |
112
+ | No log | 8.4 | 126 | 0.2341 | 0.0123 | 0.2341 |
113
+ | No log | 8.5333 | 128 | 0.2459 | 0.0123 | 0.2459 |
114
+ | No log | 8.6667 | 130 | 0.2510 | 0.0123 | 0.2510 |
115
+ | No log | 8.8 | 132 | 0.2702 | 0.0109 | 0.2702 |
116
+ | No log | 8.9333 | 134 | 0.2776 | 0.0093 | 0.2776 |
117
+ | No log | 9.0667 | 136 | 0.2862 | 0.0173 | 0.2862 |
118
+ | No log | 9.2 | 138 | 0.2916 | 0.0158 | 0.2916 |
119
+ | No log | 9.3333 | 140 | 0.2853 | 0.0158 | 0.2853 |
120
+ | No log | 9.4667 | 142 | 0.2781 | 0.0173 | 0.2781 |
121
+ | No log | 9.6 | 144 | 0.2737 | 0.0093 | 0.2737 |
122
+ | No log | 9.7333 | 146 | 0.2709 | 0.0109 | 0.2709 |
123
+ | No log | 9.8667 | 148 | 0.2693 | 0.0109 | 0.2693 |
124
+ | No log | 10.0 | 150 | 0.2694 | 0.0109 | 0.2694 |
125
 
126
 
127
  ### Framework versions
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:acdc73842cacadb4ca78a5bc2f3f0fdc7adeeedf7d9df2a79e36350741c25a0c
3
  size 540799996
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d4b9f4c3fe7305809d1ce9d789c1457fe7fc9606db79e5f6d45c3a57a45dbd87
3
  size 540799996
runs/Sep02_02-03-55_a09311c9b42d/events.out.tfevents.1725242637.a09311c9b42d.22.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e6a0fc106c94b86c97514c87576eba536657bbbf37db0fced8f63570b2d75b0
3
+ size 32251
runs/Sep02_02-03-55_a09311c9b42d/events.out.tfevents.1725243376.a09311c9b42d.22.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf9640d7f42aed20bf76a1b4736e92d07381d6c1528925281e2e4d9658cd426f
3
+ size 453
runs/Sep02_02-16-18_a09311c9b42d/events.out.tfevents.1725243379.a09311c9b42d.22.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:14e3c3e7a77ec84a838120471e2c925efc236fbba8fd9eae945bd35f0a47571d
3
+ size 34076
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:95fdbdd9b3edbddf457096a98a175d68032079c43b1e60405563ec45ace8e3e3
3
  size 5240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e2a0600f0e141a3862085825ed3d0b3435afcc3f28eb9eea25149e8ebde23696
3
  size 5240