ArabicNewSplits8_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k20_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6059
  • Qwk: 0.4525
  • Mse: 0.6059
  • Rmse: 0.7784

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0189 2 4.3978 -0.0205 4.3978 2.0971
No log 0.0377 4 2.8309 0.0258 2.8309 1.6825
No log 0.0566 6 1.6318 -0.0362 1.6318 1.2774
No log 0.0755 8 1.3628 0.0048 1.3628 1.1674
No log 0.0943 10 1.0600 0.0288 1.0600 1.0296
No log 0.1132 12 0.8890 0.1359 0.8890 0.9429
No log 0.1321 14 1.1703 -0.0786 1.1703 1.0818
No log 0.1509 16 1.2983 -0.0327 1.2983 1.1394
No log 0.1698 18 1.0438 -0.0572 1.0438 1.0217
No log 0.1887 20 0.9029 0.1532 0.9029 0.9502
No log 0.2075 22 0.9575 0.1867 0.9575 0.9785
No log 0.2264 24 1.3552 -0.0062 1.3552 1.1641
No log 0.2453 26 1.4934 0.0097 1.4934 1.2220
No log 0.2642 28 1.6124 0.0104 1.6124 1.2698
No log 0.2830 30 1.2722 0.0102 1.2722 1.1279
No log 0.3019 32 0.8988 0.0914 0.8988 0.9480
No log 0.3208 34 0.8137 0.0465 0.8137 0.9020
No log 0.3396 36 0.8177 0.1136 0.8177 0.9043
No log 0.3585 38 0.8418 0.0475 0.8418 0.9175
No log 0.3774 40 0.9760 0.0450 0.9760 0.9879
No log 0.3962 42 1.1365 0.0685 1.1365 1.0661
No log 0.4151 44 1.3171 -0.0058 1.3171 1.1476
No log 0.4340 46 1.3519 0.0205 1.3519 1.1627
No log 0.4528 48 1.2830 0.0415 1.2830 1.1327
No log 0.4717 50 1.1350 0.0903 1.1350 1.0653
No log 0.4906 52 0.8892 0.0189 0.8892 0.9430
No log 0.5094 54 0.8438 0.0733 0.8438 0.9186
No log 0.5283 56 0.8992 0.1429 0.8992 0.9483
No log 0.5472 58 1.0141 0.0687 1.0141 1.0070
No log 0.5660 60 0.9827 0.0740 0.9827 0.9913
No log 0.5849 62 0.8410 0.1837 0.8410 0.9171
No log 0.6038 64 0.7329 0.2515 0.7329 0.8561
No log 0.6226 66 0.7633 0.2114 0.7633 0.8737
No log 0.6415 68 0.7309 0.2534 0.7309 0.8549
No log 0.6604 70 0.7205 0.2295 0.7205 0.8488
No log 0.6792 72 0.7126 0.2177 0.7126 0.8442
No log 0.6981 74 0.7397 0.3030 0.7397 0.8601
No log 0.7170 76 0.7740 0.3267 0.7740 0.8798
No log 0.7358 78 0.9732 0.1540 0.9732 0.9865
No log 0.7547 80 1.0779 0.1443 1.0779 1.0382
No log 0.7736 82 1.0014 0.1400 1.0014 1.0007
No log 0.7925 84 0.7826 0.3201 0.7826 0.8847
No log 0.8113 86 0.7151 0.2686 0.7151 0.8457
No log 0.8302 88 0.7154 0.2540 0.7154 0.8458
No log 0.8491 90 0.7042 0.4275 0.7042 0.8392
No log 0.8679 92 0.7774 0.2807 0.7774 0.8817
No log 0.8868 94 0.7778 0.3375 0.7778 0.8819
No log 0.9057 96 0.7499 0.3878 0.7499 0.8660
No log 0.9245 98 0.8038 0.3880 0.8038 0.8966
No log 0.9434 100 0.7742 0.4063 0.7742 0.8799
No log 0.9623 102 0.7916 0.4114 0.7916 0.8897
No log 0.9811 104 0.9149 0.3586 0.9149 0.9565
No log 1.0 106 0.9257 0.3643 0.9257 0.9621
No log 1.0189 108 0.8858 0.3578 0.8858 0.9411
No log 1.0377 110 0.8558 0.3787 0.8558 0.9251
No log 1.0566 112 0.7615 0.4515 0.7615 0.8727
No log 1.0755 114 0.7460 0.3188 0.7460 0.8637
No log 1.0943 116 0.7714 0.2708 0.7714 0.8783
No log 1.1132 118 0.7221 0.3867 0.7221 0.8498
No log 1.1321 120 0.7967 0.3767 0.7967 0.8926
No log 1.1509 122 0.7967 0.3974 0.7967 0.8926
No log 1.1698 124 0.7358 0.3731 0.7358 0.8578
No log 1.1887 126 1.0678 0.2317 1.0678 1.0334
No log 1.2075 128 1.0470 0.2114 1.0470 1.0232
No log 1.2264 130 0.7516 0.3110 0.7516 0.8670
No log 1.2453 132 0.8413 0.3872 0.8413 0.9172
No log 1.2642 134 1.0681 0.2395 1.0681 1.0335
No log 1.2830 136 0.9986 0.2777 0.9986 0.9993
No log 1.3019 138 0.7705 0.4523 0.7705 0.8778
No log 1.3208 140 0.7253 0.4129 0.7253 0.8517
No log 1.3396 142 0.7290 0.3531 0.7290 0.8538
No log 1.3585 144 0.7059 0.4105 0.7059 0.8402
No log 1.3774 146 0.7999 0.4369 0.7999 0.8943
No log 1.3962 148 0.9173 0.2702 0.9173 0.9578
No log 1.4151 150 0.8078 0.4266 0.8078 0.8987
No log 1.4340 152 0.7186 0.4736 0.7186 0.8477
No log 1.4528 154 0.6888 0.3597 0.6888 0.8300
No log 1.4717 156 0.6948 0.3351 0.6948 0.8335
No log 1.4906 158 0.6858 0.4281 0.6858 0.8281
No log 1.5094 160 0.7062 0.4410 0.7062 0.8404
No log 1.5283 162 0.6855 0.4312 0.6855 0.8279
No log 1.5472 164 0.6773 0.3490 0.6773 0.8230
No log 1.5660 166 0.6984 0.2922 0.6984 0.8357
No log 1.5849 168 0.6612 0.3300 0.6612 0.8131
No log 1.6038 170 0.6762 0.2922 0.6762 0.8223
No log 1.6226 172 0.7355 0.3140 0.7355 0.8576
No log 1.6415 174 0.6657 0.3733 0.6657 0.8159
No log 1.6604 176 0.7289 0.4936 0.7289 0.8538
No log 1.6792 178 0.7151 0.4936 0.7151 0.8456
No log 1.6981 180 0.6459 0.4732 0.6459 0.8037
No log 1.7170 182 0.6418 0.3711 0.6418 0.8011
No log 1.7358 184 0.6352 0.4102 0.6352 0.7970
No log 1.7547 186 0.6601 0.3274 0.6601 0.8125
No log 1.7736 188 0.6848 0.4033 0.6848 0.8275
No log 1.7925 190 0.6443 0.4433 0.6443 0.8027
No log 1.8113 192 0.6091 0.4702 0.6091 0.7805
No log 1.8302 194 0.6379 0.4571 0.6379 0.7987
No log 1.8491 196 0.5976 0.4259 0.5976 0.7730
No log 1.8679 198 0.6004 0.3469 0.6004 0.7749
No log 1.8868 200 0.6501 0.3939 0.6501 0.8063
No log 1.9057 202 0.6163 0.3656 0.6163 0.7851
No log 1.9245 204 0.6096 0.3540 0.6096 0.7808
No log 1.9434 206 0.6484 0.4126 0.6484 0.8052
No log 1.9623 208 0.7264 0.4013 0.7264 0.8523
No log 1.9811 210 0.6864 0.4125 0.6864 0.8285
No log 2.0 212 0.6789 0.3703 0.6789 0.8239
No log 2.0189 214 0.6339 0.3607 0.6339 0.7962
No log 2.0377 216 0.6295 0.3740 0.6295 0.7934
No log 2.0566 218 0.6358 0.3488 0.6358 0.7974
No log 2.0755 220 0.6372 0.3439 0.6372 0.7983
No log 2.0943 222 0.6423 0.3439 0.6423 0.8014
No log 2.1132 224 0.6163 0.3564 0.6163 0.7851
No log 2.1321 226 0.6423 0.4939 0.6423 0.8015
No log 2.1509 228 0.7113 0.4600 0.7113 0.8434
No log 2.1698 230 0.6866 0.4664 0.6866 0.8286
No log 2.1887 232 0.6975 0.4492 0.6975 0.8352
No log 2.2075 234 0.6366 0.4858 0.6366 0.7979
No log 2.2264 236 0.6265 0.3293 0.6265 0.7915
No log 2.2453 238 0.6475 0.4157 0.6475 0.8047
No log 2.2642 240 0.6292 0.3904 0.6292 0.7932
No log 2.2830 242 0.6527 0.4532 0.6527 0.8079
No log 2.3019 244 0.7107 0.4406 0.7107 0.8430
No log 2.3208 246 0.6621 0.4730 0.6621 0.8137
No log 2.3396 248 0.6330 0.3819 0.6330 0.7956
No log 2.3585 250 0.7190 0.4047 0.7190 0.8479
No log 2.3774 252 0.7038 0.3947 0.7038 0.8389
No log 2.3962 254 0.6770 0.3911 0.6770 0.8228
No log 2.4151 256 0.6982 0.3946 0.6982 0.8356
No log 2.4340 258 0.6660 0.4004 0.6660 0.8161
No log 2.4528 260 0.6093 0.3841 0.6093 0.7806
No log 2.4717 262 0.6177 0.3633 0.6177 0.7859
No log 2.4906 264 0.6131 0.3708 0.6131 0.7830
No log 2.5094 266 0.6251 0.3858 0.6251 0.7907
No log 2.5283 268 0.6347 0.3764 0.6347 0.7967
No log 2.5472 270 0.6081 0.3841 0.6081 0.7798
No log 2.5660 272 0.6152 0.4153 0.6152 0.7844
No log 2.5849 274 0.6430 0.4792 0.6430 0.8019
No log 2.6038 276 0.6165 0.3785 0.6165 0.7852
No log 2.6226 278 0.6197 0.3715 0.6197 0.7872
No log 2.6415 280 0.6828 0.4038 0.6828 0.8263
No log 2.6604 282 0.6761 0.4391 0.6761 0.8223
No log 2.6792 284 0.6079 0.3894 0.6079 0.7797
No log 2.6981 286 0.5916 0.4297 0.5916 0.7692
No log 2.7170 288 0.5982 0.4498 0.5982 0.7734
No log 2.7358 290 0.6512 0.4554 0.6512 0.8070
No log 2.7547 292 0.6787 0.4555 0.6787 0.8238
No log 2.7736 294 0.6478 0.4581 0.6478 0.8049
No log 2.7925 296 0.5780 0.4167 0.5780 0.7603
No log 2.8113 298 0.5924 0.3557 0.5924 0.7697
No log 2.8302 300 0.6495 0.4149 0.6495 0.8059
No log 2.8491 302 0.6287 0.3604 0.6287 0.7929
No log 2.8679 304 0.5909 0.2860 0.5909 0.7687
No log 2.8868 306 0.6115 0.3074 0.6115 0.7820
No log 2.9057 308 0.6142 0.3252 0.6142 0.7837
No log 2.9245 310 0.6902 0.4269 0.6902 0.8308
No log 2.9434 312 0.9330 0.3756 0.9330 0.9659
No log 2.9623 314 1.0896 0.3339 1.0896 1.0439
No log 2.9811 316 1.2812 0.2599 1.2812 1.1319
No log 3.0 318 1.3116 0.2651 1.3116 1.1453
No log 3.0189 320 1.1688 0.2995 1.1688 1.0811
No log 3.0377 322 0.9415 0.4553 0.9415 0.9703
No log 3.0566 324 0.7480 0.4607 0.7480 0.8649
No log 3.0755 326 0.6241 0.4401 0.6241 0.7900
No log 3.0943 328 0.6174 0.3521 0.6174 0.7857
No log 3.1132 330 0.6040 0.4149 0.6040 0.7772
No log 3.1321 332 0.6299 0.4061 0.6299 0.7937
No log 3.1509 334 0.6755 0.4480 0.6755 0.8219
No log 3.1698 336 0.6750 0.4472 0.6750 0.8216
No log 3.1887 338 0.6441 0.4472 0.6440 0.8025
No log 3.2075 340 0.6241 0.4472 0.6241 0.7900
No log 3.2264 342 0.6760 0.4290 0.6760 0.8222
No log 3.2453 344 0.8518 0.4668 0.8518 0.9229
No log 3.2642 346 0.8617 0.4668 0.8617 0.9283
No log 3.2830 348 0.7439 0.4893 0.7439 0.8625
No log 3.3019 350 0.6381 0.4480 0.6381 0.7988
No log 3.3208 352 0.5900 0.4318 0.5900 0.7681
No log 3.3396 354 0.5697 0.4577 0.5697 0.7548
No log 3.3585 356 0.5718 0.4001 0.5718 0.7562
No log 3.3774 358 0.5769 0.4189 0.5769 0.7595
No log 3.3962 360 0.5924 0.4739 0.5924 0.7696
No log 3.4151 362 0.5771 0.4381 0.5771 0.7596
No log 3.4340 364 0.6006 0.4187 0.6006 0.7750
No log 3.4528 366 0.6092 0.4679 0.6092 0.7805
No log 3.4717 368 0.5965 0.3805 0.5965 0.7723
No log 3.4906 370 0.5995 0.4204 0.5995 0.7743
No log 3.5094 372 0.6058 0.3868 0.6058 0.7783
No log 3.5283 374 0.6708 0.4698 0.6708 0.8190
No log 3.5472 376 0.7354 0.4701 0.7354 0.8576
No log 3.5660 378 0.7085 0.4006 0.7085 0.8417
No log 3.5849 380 0.6511 0.3718 0.6511 0.8069
No log 3.6038 382 0.6324 0.3522 0.6324 0.7952
No log 3.6226 384 0.6303 0.3908 0.6303 0.7939
No log 3.6415 386 0.6732 0.4557 0.6732 0.8205
No log 3.6604 388 0.7248 0.4825 0.7248 0.8513
No log 3.6792 390 0.7755 0.4544 0.7755 0.8806
No log 3.6981 392 0.7396 0.4814 0.7396 0.8600
No log 3.7170 394 0.6452 0.4537 0.6452 0.8033
No log 3.7358 396 0.6334 0.4102 0.6334 0.7959
No log 3.7547 398 0.6490 0.4390 0.6490 0.8056
No log 3.7736 400 0.6204 0.3859 0.6204 0.7877
No log 3.7925 402 0.6103 0.3792 0.6103 0.7812
No log 3.8113 404 0.6325 0.4615 0.6325 0.7953
No log 3.8302 406 0.6270 0.4383 0.6270 0.7918
No log 3.8491 408 0.6100 0.3948 0.6100 0.7810
No log 3.8679 410 0.6973 0.4465 0.6973 0.8351
No log 3.8868 412 0.8475 0.4307 0.8475 0.9206
No log 3.9057 414 0.7824 0.4635 0.7824 0.8846
No log 3.9245 416 0.6310 0.4572 0.6310 0.7944
No log 3.9434 418 0.5905 0.4943 0.5905 0.7684
No log 3.9623 420 0.5862 0.4433 0.5862 0.7657
No log 3.9811 422 0.6199 0.4852 0.6199 0.7873
No log 4.0 424 0.6406 0.5217 0.6406 0.8004
No log 4.0189 426 0.6971 0.4884 0.6971 0.8349
No log 4.0377 428 0.7034 0.4884 0.7034 0.8387
No log 4.0566 430 0.6888 0.4857 0.6888 0.8299
No log 4.0755 432 0.6367 0.5056 0.6367 0.7979
No log 4.0943 434 0.6023 0.4612 0.6023 0.7761
No log 4.1132 436 0.6054 0.4795 0.6054 0.7781
No log 4.1321 438 0.6237 0.4807 0.6237 0.7898
No log 4.1509 440 0.6500 0.5056 0.6500 0.8062
No log 4.1698 442 0.6136 0.5144 0.6136 0.7833
No log 4.1887 444 0.5781 0.5685 0.5781 0.7603
No log 4.2075 446 0.5707 0.5427 0.5707 0.7555
No log 4.2264 448 0.5657 0.5229 0.5657 0.7522
No log 4.2453 450 0.5932 0.4500 0.5932 0.7702
No log 4.2642 452 0.6071 0.4787 0.6071 0.7792
No log 4.2830 454 0.5732 0.4499 0.5732 0.7571
No log 4.3019 456 0.5618 0.5459 0.5618 0.7495
No log 4.3208 458 0.5803 0.5345 0.5803 0.7618
No log 4.3396 460 0.5727 0.5395 0.5727 0.7568
No log 4.3585 462 0.5868 0.4965 0.5868 0.7660
No log 4.3774 464 0.6543 0.4708 0.6543 0.8089
No log 4.3962 466 0.6875 0.4813 0.6875 0.8292
No log 4.4151 468 0.6178 0.5026 0.6178 0.7860
No log 4.4340 470 0.5964 0.4902 0.5964 0.7723
No log 4.4528 472 0.5932 0.4842 0.5932 0.7702
No log 4.4717 474 0.5974 0.5111 0.5974 0.7729
No log 4.4906 476 0.6196 0.4863 0.6196 0.7871
No log 4.5094 478 0.6190 0.4769 0.6190 0.7868
No log 4.5283 480 0.6017 0.5051 0.6017 0.7757
No log 4.5472 482 0.5823 0.4960 0.5823 0.7631
No log 4.5660 484 0.5843 0.4211 0.5843 0.7644
No log 4.5849 486 0.5922 0.4365 0.5922 0.7696
No log 4.6038 488 0.6179 0.4398 0.6179 0.7861
No log 4.6226 490 0.6494 0.4163 0.6494 0.8059
No log 4.6415 492 0.6367 0.4813 0.6367 0.7980
No log 4.6604 494 0.6031 0.4965 0.6031 0.7766
No log 4.6792 496 0.6094 0.4918 0.6094 0.7806
No log 4.6981 498 0.6168 0.4773 0.6168 0.7854
0.3655 4.7170 500 0.6084 0.4773 0.6084 0.7800
0.3655 4.7358 502 0.6023 0.5045 0.6023 0.7761
0.3655 4.7547 504 0.6096 0.4806 0.6096 0.7808
0.3655 4.7736 506 0.6333 0.4503 0.6333 0.7958
0.3655 4.7925 508 0.6192 0.4442 0.6192 0.7869
0.3655 4.8113 510 0.5768 0.5011 0.5768 0.7594
0.3655 4.8302 512 0.5549 0.4977 0.5549 0.7449
0.3655 4.8491 514 0.5685 0.4749 0.5685 0.7540
0.3655 4.8679 516 0.5608 0.4647 0.5608 0.7488
0.3655 4.8868 518 0.5550 0.4764 0.5550 0.7450
0.3655 4.9057 520 0.5477 0.4935 0.5477 0.7401
0.3655 4.9245 522 0.5509 0.4890 0.5509 0.7422
0.3655 4.9434 524 0.5643 0.5158 0.5643 0.7512
0.3655 4.9623 526 0.5608 0.4991 0.5608 0.7488
0.3655 4.9811 528 0.5505 0.4839 0.5505 0.7419
0.3655 5.0 530 0.5513 0.4645 0.5513 0.7425
0.3655 5.0189 532 0.5516 0.4623 0.5516 0.7427
0.3655 5.0377 534 0.5510 0.4539 0.5510 0.7423
0.3655 5.0566 536 0.5514 0.4398 0.5514 0.7425
0.3655 5.0755 538 0.5683 0.4848 0.5683 0.7539
0.3655 5.0943 540 0.5782 0.4911 0.5782 0.7604
0.3655 5.1132 542 0.6040 0.4634 0.6040 0.7772
0.3655 5.1321 544 0.5696 0.4491 0.5696 0.7547
0.3655 5.1509 546 0.5558 0.3896 0.5558 0.7455
0.3655 5.1698 548 0.5602 0.3527 0.5602 0.7485
0.3655 5.1887 550 0.6059 0.4525 0.6059 0.7784

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
4
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for MayBashendy/ArabicNewSplits8_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k20_task2_organization

Finetuned
(4222)
this model