Arabic_FineTuningAraBERT_AugV4_k2_task3_organization_fold0

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.3052
  • Qwk: -0.1846
  • Mse: 1.3052
  • Rmse: 1.1425

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0357 2 4.1150 0.0307 4.1150 2.0286
No log 0.0714 4 2.6844 -0.0722 2.6844 1.6384
No log 0.1071 6 1.1465 0.0530 1.1465 1.0707
No log 0.1429 8 0.7860 0.0 0.7860 0.8866
No log 0.1786 10 0.8650 0.1852 0.8650 0.9300
No log 0.2143 12 0.8791 -0.0476 0.8791 0.9376
No log 0.25 14 0.8461 0.0 0.8461 0.9198
No log 0.2857 16 0.9045 -0.0185 0.9045 0.9511
No log 0.3214 18 0.8920 0.0 0.8920 0.9445
No log 0.3571 20 0.8531 0.0 0.8531 0.9237
No log 0.3929 22 0.8018 0.0 0.8018 0.8954
No log 0.4286 24 0.8464 0.2667 0.8464 0.9200
No log 0.4643 26 0.8666 0.1037 0.8666 0.9309
No log 0.5 28 1.0511 0.1852 1.0511 1.0252
No log 0.5357 30 1.0613 0.1852 1.0613 1.0302
No log 0.5714 32 0.9014 0.1270 0.9014 0.9494
No log 0.6071 34 0.9333 0.1270 0.9333 0.9661
No log 0.6429 36 0.8724 -0.0185 0.8724 0.9340
No log 0.6786 38 0.8238 -0.0185 0.8238 0.9076
No log 0.7143 40 0.8017 0.0 0.8017 0.8954
No log 0.75 42 0.7758 0.1538 0.7758 0.8808
No log 0.7857 44 1.1081 0.2667 1.1081 1.0527
No log 0.8214 46 1.2173 0.0 1.2173 1.1033
No log 0.8571 48 0.9246 0.2361 0.9246 0.9616
No log 0.8929 50 0.8417 -0.0185 0.8417 0.9175
No log 0.9286 52 0.8284 -0.0185 0.8284 0.9102
No log 0.9643 54 0.8194 -0.0185 0.8194 0.9052
No log 1.0 56 0.9138 -0.0185 0.9138 0.9559
No log 1.0357 58 0.8754 -0.0185 0.8754 0.9356
No log 1.0714 60 0.7813 0.0 0.7813 0.8839
No log 1.1071 62 0.8088 -0.0342 0.8088 0.8993
No log 1.1429 64 0.8078 0.1270 0.8078 0.8988
No log 1.1786 66 1.0991 0.2092 1.0991 1.0484
No log 1.2143 68 1.2379 -0.0936 1.2379 1.1126
No log 1.25 70 0.8463 0.1270 0.8463 0.9199
No log 1.2857 72 0.7914 0.1538 0.7914 0.8896
No log 1.3214 74 1.0117 -0.0694 1.0117 1.0058
No log 1.3571 76 1.0683 -0.2222 1.0683 1.0336
No log 1.3929 78 0.9433 -0.2222 0.9433 0.9712
No log 1.4286 80 0.8238 0.0 0.8238 0.9076
No log 1.4643 82 0.8412 0.0 0.8412 0.9172
No log 1.5 84 0.8224 0.0 0.8224 0.9068
No log 1.5357 86 0.9204 -0.0342 0.9204 0.9594
No log 1.5714 88 0.9734 -0.0476 0.9734 0.9866
No log 1.6071 90 0.9394 0.1037 0.9394 0.9692
No log 1.6429 92 1.0627 0.0654 1.0627 1.0309
No log 1.6786 94 1.0201 0.0654 1.0201 1.0100
No log 1.7143 96 1.0852 0.2924 1.0852 1.0417
No log 1.75 98 1.3711 0.0 1.3711 1.1709
No log 1.7857 100 1.1271 0.2924 1.1271 1.0616
No log 1.8214 102 0.9281 -0.0593 0.9281 0.9634
No log 1.8571 104 0.8916 0.1270 0.8916 0.9443
No log 1.8929 106 0.8971 -0.0342 0.8971 0.9471
No log 1.9286 108 1.0964 -0.0694 1.0964 1.0471
No log 1.9643 110 1.1433 -0.2222 1.1433 1.0693
No log 2.0 112 0.9836 -0.0342 0.9836 0.9917
No log 2.0357 114 1.0129 -0.0476 1.0129 1.0064
No log 2.0714 116 1.0887 -0.0476 1.0887 1.0434
No log 2.1071 118 1.1682 -0.0476 1.1682 1.0808
No log 2.1429 120 1.2585 -0.2367 1.2585 1.1218
No log 2.1786 122 1.3259 0.0 1.3259 1.1515
No log 2.2143 124 1.2845 -0.1846 1.2845 1.1333
No log 2.25 126 1.1547 -0.0342 1.1547 1.0746
No log 2.2857 128 1.0730 0.1538 1.0730 1.0359
No log 2.3214 130 0.9919 0.1538 0.9919 0.9959
No log 2.3571 132 0.9725 0.1270 0.9725 0.9861
No log 2.3929 134 1.0304 0.1037 1.0304 1.0151
No log 2.4286 136 0.9065 0.1270 0.9065 0.9521
No log 2.4643 138 0.8803 0.4359 0.8803 0.9382
No log 2.5 140 0.8723 0.4359 0.8723 0.9340
No log 2.5357 142 0.8649 0.1037 0.8649 0.9300
No log 2.5714 144 1.0561 -0.0694 1.0561 1.0276
No log 2.6071 146 1.1350 0.0494 1.1350 1.0654
No log 2.6429 148 0.9467 0.1037 0.9467 0.9730
No log 2.6786 150 0.8502 0.3265 0.8502 0.9221
No log 2.7143 152 0.9027 0.3265 0.9027 0.9501
No log 2.75 154 0.8354 0.3265 0.8354 0.9140
No log 2.7857 156 1.0037 0.0494 1.0037 1.0019
No log 2.8214 158 1.1951 0.2924 1.1951 1.0932
No log 2.8571 160 1.0669 -0.0784 1.0669 1.0329
No log 2.8929 162 0.8632 -0.0342 0.8632 0.9291
No log 2.9286 164 0.8158 -0.0185 0.8158 0.9032
No log 2.9643 166 0.8128 -0.0185 0.8128 0.9016
No log 3.0 168 0.8743 0.1037 0.8743 0.9350
No log 3.0357 170 0.9597 -0.0694 0.9597 0.9796
No log 3.0714 172 0.9067 0.1037 0.9067 0.9522
No log 3.1071 174 0.8626 0.1538 0.8626 0.9288
No log 3.1429 176 0.9012 0.1270 0.9012 0.9493
No log 3.1786 178 0.9905 0.1037 0.9905 0.9952
No log 3.2143 180 1.0612 -0.0694 1.0612 1.0302
No log 3.25 182 1.0330 0.1037 1.0330 1.0164
No log 3.2857 184 1.0399 0.2484 1.0399 1.0197
No log 3.3214 186 1.0706 0.2840 1.0706 1.0347
No log 3.3571 188 1.0440 0.3077 1.0440 1.0218
No log 3.3929 190 1.0384 -0.0694 1.0384 1.0190
No log 3.4286 192 1.0895 -0.0784 1.0895 1.0438
No log 3.4643 194 1.0553 -0.0784 1.0553 1.0273
No log 3.5 196 1.0473 -0.0694 1.0473 1.0234
No log 3.5357 198 1.0043 0.1270 1.0043 1.0021
No log 3.5714 200 0.9986 0.0 0.9986 0.9993
No log 3.6071 202 1.0321 0.1538 1.0321 1.0159
No log 3.6429 204 1.0737 0.1270 1.0737 1.0362
No log 3.6786 206 1.0907 0.1270 1.0907 1.0444
No log 3.7143 208 1.1261 0.0 1.1261 1.0612
No log 3.75 210 1.1870 -0.4218 1.1870 1.0895
No log 3.7857 212 1.2299 -0.4218 1.2299 1.1090
No log 3.8214 214 1.2646 -0.4218 1.2646 1.1245
No log 3.8571 216 1.2592 -0.4218 1.2592 1.1221
No log 3.8929 218 1.2586 -0.2692 1.2586 1.1219
No log 3.9286 220 1.2619 0.0330 1.2619 1.1233
No log 3.9643 222 1.2418 -0.1065 1.2418 1.1144
No log 4.0 224 1.2593 0.0330 1.2593 1.1222
No log 4.0357 226 1.1940 -0.1065 1.1940 1.0927
No log 4.0714 228 1.1292 0.1270 1.1292 1.0626
No log 4.1071 230 1.1050 0.1270 1.1050 1.0512
No log 4.1429 232 1.0522 -0.0154 1.0522 1.0258
No log 4.1786 234 1.0958 -0.1493 1.0958 1.0468
No log 4.2143 236 1.1615 -0.2384 1.1615 1.0777
No log 4.25 238 1.1193 -0.1493 1.1193 1.0580
No log 4.2857 240 1.0553 -0.0154 1.0553 1.0273
No log 4.3214 242 1.0910 0.1270 1.0910 1.0445
No log 4.3571 244 1.1026 -0.0593 1.1026 1.0501
No log 4.3929 246 1.0730 0.1270 1.0730 1.0359
No log 4.4286 248 1.1432 -0.1065 1.1432 1.0692
No log 4.4643 250 1.3491 -0.0809 1.3491 1.1615
No log 4.5 252 1.3903 -0.0809 1.3903 1.1791
No log 4.5357 254 1.3143 -0.1065 1.3143 1.1464
No log 4.5714 256 1.2829 -0.1911 1.2829 1.1326
No log 4.6071 258 1.3771 -0.0694 1.3771 1.1735
No log 4.6429 260 1.3646 -0.0694 1.3646 1.1682
No log 4.6786 262 1.2987 -0.1911 1.2987 1.1396
No log 4.7143 264 1.3046 -0.1124 1.3046 1.1422
No log 4.75 266 1.4215 -0.0809 1.4215 1.1923
No log 4.7857 268 1.5037 -0.0809 1.5037 1.2263
No log 4.8214 270 1.4364 -0.0809 1.4364 1.1985
No log 4.8571 272 1.2809 -0.2375 1.2809 1.1318
No log 4.8929 274 1.1824 -0.1818 1.1824 1.0874
No log 4.9286 276 1.1897 -0.3378 1.1897 1.0907
No log 4.9643 278 1.1809 -0.1911 1.1809 1.0867
No log 5.0 280 1.1684 -0.1818 1.1684 1.0809
No log 5.0357 282 1.2000 -0.2375 1.2000 1.0955
No log 5.0714 284 1.2247 -0.2375 1.2247 1.1067
No log 5.1071 286 1.1993 -0.1065 1.1993 1.0951
No log 5.1429 288 1.1541 -0.0288 1.1541 1.0743
No log 5.1786 290 1.1313 0.1270 1.1313 1.0636
No log 5.2143 292 1.1104 0.1538 1.1104 1.0538
No log 5.25 294 1.1022 0.1270 1.1022 1.0498
No log 5.2857 296 1.1096 0.1270 1.1096 1.0534
No log 5.3214 298 1.1104 0.1270 1.1104 1.0537
No log 5.3571 300 1.1330 -0.1493 1.1330 1.0644
No log 5.3929 302 1.2140 -0.2375 1.2140 1.1018
No log 5.4286 304 1.2435 -0.2375 1.2435 1.1151
No log 5.4643 306 1.2221 -0.2375 1.2221 1.1055
No log 5.5 308 1.1772 -0.1333 1.1772 1.0850
No log 5.5357 310 1.1527 0.1270 1.1527 1.0736
No log 5.5714 312 1.1328 0.1270 1.1328 1.0643
No log 5.6071 314 1.1492 -0.1065 1.1492 1.0720
No log 5.6429 316 1.1430 -0.1065 1.1430 1.0691
No log 5.6786 318 1.1167 0.1538 1.1167 1.0567
No log 5.7143 320 1.0993 -0.0593 1.0993 1.0485
No log 5.75 322 1.1089 -0.0593 1.1089 1.0530
No log 5.7857 324 1.1109 0.1081 1.1109 1.0540
No log 5.8214 326 1.0872 -0.0288 1.0872 1.0427
No log 5.8571 328 1.0713 -0.0288 1.0713 1.0350
No log 5.8929 330 1.0580 -0.0288 1.0580 1.0286
No log 5.9286 332 1.0690 0.1538 1.0690 1.0339
No log 5.9643 334 1.0936 0.0272 1.0936 1.0458
No log 6.0 336 1.0824 0.1316 1.0824 1.0404
No log 6.0357 338 1.0932 0.1316 1.0932 1.0455
No log 6.0714 340 1.1701 0.1538 1.1701 1.0817
No log 6.1071 342 1.2607 0.1538 1.2607 1.1228
No log 6.1429 344 1.3140 0.2840 1.3140 1.1463
No log 6.1786 346 1.3287 0.1538 1.3287 1.1527
No log 6.2143 348 1.3731 0.2840 1.3731 1.1718
No log 6.25 350 1.4328 -0.0809 1.4328 1.1970
No log 6.2857 352 1.4233 -0.0809 1.4233 1.1930
No log 6.3214 354 1.3875 -0.1818 1.3875 1.1779
No log 6.3571 356 1.3850 -0.1818 1.3850 1.1769
No log 6.3929 358 1.4204 0.0272 1.4204 1.1918
No log 6.4286 360 1.4444 -0.0645 1.4444 1.2018
No log 6.4643 362 1.4452 -0.0645 1.4452 1.2022
No log 6.5 364 1.4198 -0.2375 1.4198 1.1916
No log 6.5357 366 1.4099 -0.2375 1.4099 1.1874
No log 6.5714 368 1.4243 -0.0809 1.4243 1.1934
No log 6.6071 370 1.4383 -0.0809 1.4383 1.1993
No log 6.6429 372 1.4165 -0.0809 1.4165 1.1902
No log 6.6786 374 1.3938 -0.0809 1.3938 1.1806
No log 6.7143 376 1.3488 -0.2375 1.3488 1.1614
No log 6.75 378 1.3174 -0.1846 1.3174 1.1478
No log 6.7857 380 1.3072 -0.1846 1.3072 1.1433
No log 6.8214 382 1.2965 -0.4776 1.2965 1.1386
No log 6.8571 384 1.2809 -0.4218 1.2809 1.1318
No log 6.8929 386 1.2759 -0.2375 1.2759 1.1295
No log 6.9286 388 1.2772 -0.2375 1.2772 1.1301
No log 6.9643 390 1.2513 -0.2375 1.2513 1.1186
No log 7.0 392 1.2255 -0.1818 1.2255 1.1070
No log 7.0357 394 1.2259 -0.0154 1.2259 1.1072
No log 7.0714 396 1.2230 -0.0154 1.2230 1.1059
No log 7.1071 398 1.2117 -0.0154 1.2117 1.1008
No log 7.1429 400 1.2051 -0.1493 1.2051 1.0978
No log 7.1786 402 1.2162 -0.2375 1.2162 1.1028
No log 7.2143 404 1.2229 -0.2375 1.2229 1.1058
No log 7.25 406 1.2135 -0.1493 1.2135 1.1016
No log 7.2857 408 1.2204 -0.1846 1.2204 1.1047
No log 7.3214 410 1.2338 -0.3453 1.2338 1.1107
No log 7.3571 412 1.2300 -0.3453 1.2300 1.1091
No log 7.3929 414 1.2118 -0.1846 1.2118 1.1008
No log 7.4286 416 1.2226 -0.1818 1.2226 1.1057
No log 7.4643 418 1.2406 -0.1493 1.2406 1.1138
No log 7.5 420 1.2499 -0.1818 1.2499 1.1180
No log 7.5357 422 1.2526 -0.0288 1.2526 1.1192
No log 7.5714 424 1.2436 -0.1818 1.2436 1.1152
No log 7.6071 426 1.2416 -0.1818 1.2416 1.1143
No log 7.6429 428 1.2483 -0.1818 1.2483 1.1173
No log 7.6786 430 1.2615 -0.1818 1.2615 1.1232
No log 7.7143 432 1.2760 -0.4218 1.2760 1.1296
No log 7.75 434 1.2769 -0.1818 1.2769 1.1300
No log 7.7857 436 1.2847 -0.0288 1.2847 1.1334
No log 7.8214 438 1.2868 -0.0288 1.2868 1.1344
No log 7.8571 440 1.2849 -0.0288 1.2849 1.1335
No log 7.8929 442 1.2711 -0.1846 1.2711 1.1274
No log 7.9286 444 1.2569 -0.1846 1.2569 1.1211
No log 7.9643 446 1.2437 -0.1818 1.2437 1.1152
No log 8.0 448 1.2478 -0.1818 1.2478 1.1171
No log 8.0357 450 1.2671 -0.1818 1.2671 1.1257
No log 8.0714 452 1.2760 -0.1493 1.2760 1.1296
No log 8.1071 454 1.2735 -0.1818 1.2735 1.1285
No log 8.1429 456 1.2691 -0.1818 1.2691 1.1266
No log 8.1786 458 1.2744 -0.0154 1.2744 1.1289
No log 8.2143 460 1.2898 -0.0288 1.2898 1.1357
No log 8.25 462 1.3060 -0.0288 1.3060 1.1428
No log 8.2857 464 1.3174 -0.0288 1.3174 1.1478
No log 8.3214 466 1.3174 -0.0288 1.3174 1.1478
No log 8.3571 468 1.3135 0.0 1.3135 1.1461
No log 8.3929 470 1.3276 0.0 1.3276 1.1522
No log 8.4286 472 1.3304 0.0 1.3304 1.1534
No log 8.4643 474 1.3300 0.0 1.3300 1.1532
No log 8.5 476 1.3250 0.0 1.3250 1.1511
No log 8.5357 478 1.3223 0.0 1.3223 1.1499
No log 8.5714 480 1.3111 0.0 1.3111 1.1450
No log 8.6071 482 1.3007 -0.1493 1.3007 1.1405
No log 8.6429 484 1.2953 -0.1493 1.2953 1.1381
No log 8.6786 486 1.2902 -0.1818 1.2902 1.1359
No log 8.7143 488 1.2905 -0.1846 1.2905 1.1360
No log 8.75 490 1.2926 -0.1846 1.2926 1.1369
No log 8.7857 492 1.2931 -0.1846 1.2931 1.1371
No log 8.8214 494 1.2935 -0.1846 1.2935 1.1373
No log 8.8571 496 1.2910 -0.1846 1.2910 1.1362
No log 8.8929 498 1.2837 -0.1846 1.2837 1.1330
0.2806 8.9286 500 1.2775 -0.1846 1.2775 1.1303
0.2806 8.9643 502 1.2732 -0.1846 1.2732 1.1284
0.2806 9.0 504 1.2696 -0.1818 1.2696 1.1268
0.2806 9.0357 506 1.2647 -0.1818 1.2647 1.1246
0.2806 9.0714 508 1.2586 -0.1818 1.2586 1.1219
0.2806 9.1071 510 1.2537 -0.1818 1.2537 1.1197
0.2806 9.1429 512 1.2561 -0.1818 1.2561 1.1208
0.2806 9.1786 514 1.2622 -0.1818 1.2622 1.1235
0.2806 9.2143 516 1.2656 -0.1846 1.2656 1.1250
0.2806 9.25 518 1.2694 -0.1846 1.2694 1.1267
0.2806 9.2857 520 1.2751 -0.1846 1.2751 1.1292
0.2806 9.3214 522 1.2801 -0.1846 1.2801 1.1314
0.2806 9.3571 524 1.2831 -0.1846 1.2831 1.1328
0.2806 9.3929 526 1.2853 -0.1846 1.2853 1.1337
0.2806 9.4286 528 1.2886 -0.1846 1.2886 1.1352
0.2806 9.4643 530 1.2926 -0.1846 1.2926 1.1369
0.2806 9.5 532 1.2984 -0.1846 1.2984 1.1395
0.2806 9.5357 534 1.3013 -0.1846 1.3013 1.1407
0.2806 9.5714 536 1.3048 -0.1846 1.3048 1.1423
0.2806 9.6071 538 1.3073 -0.1846 1.3073 1.1434
0.2806 9.6429 540 1.3078 -0.1846 1.3078 1.1436
0.2806 9.6786 542 1.3075 -0.1846 1.3075 1.1434
0.2806 9.7143 544 1.3075 -0.1846 1.3075 1.1435
0.2806 9.75 546 1.3079 -0.1846 1.3079 1.1436
0.2806 9.7857 548 1.3079 -0.1846 1.3079 1.1437
0.2806 9.8214 550 1.3080 -0.1846 1.3080 1.1437
0.2806 9.8571 552 1.3071 -0.1846 1.3071 1.1433
0.2806 9.8929 554 1.3063 -0.1846 1.3063 1.1429
0.2806 9.9286 556 1.3057 -0.1846 1.3057 1.1427
0.2806 9.9643 558 1.3053 -0.1846 1.3053 1.1425
0.2806 10.0 560 1.3052 -0.1846 1.3052 1.1425

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
5
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for MayBashendy/Arabic_FineTuningAraBERT_AugV4_k2_task3_organization_fold0

Finetuned
(4222)
this model