ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run1_AugV5_k14_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6253
  • Qwk: 0.4573
  • Mse: 0.6253
  • Rmse: 0.7907

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0274 2 4.3271 -0.0291 4.3271 2.0802
No log 0.0548 4 2.2819 0.0809 2.2819 1.5106
No log 0.0822 6 1.4941 -0.0040 1.4941 1.2223
No log 0.1096 8 1.2780 -0.0676 1.2780 1.1305
No log 0.1370 10 0.8315 0.1891 0.8315 0.9119
No log 0.1644 12 0.8947 0.1098 0.8947 0.9459
No log 0.1918 14 1.1619 0.0088 1.1619 1.0779
No log 0.2192 16 0.9975 0.0668 0.9975 0.9987
No log 0.2466 18 1.0565 0.1060 1.0565 1.0279
No log 0.2740 20 1.6473 0.1127 1.6473 1.2835
No log 0.3014 22 1.8096 0.0884 1.8096 1.3452
No log 0.3288 24 1.4479 0.1343 1.4479 1.2033
No log 0.3562 26 1.1211 0.1466 1.1211 1.0588
No log 0.3836 28 1.1420 0.1405 1.1420 1.0686
No log 0.4110 30 1.0936 0.1668 1.0936 1.0458
No log 0.4384 32 1.0018 0.2156 1.0018 1.0009
No log 0.4658 34 1.1326 0.2133 1.1326 1.0642
No log 0.4932 36 1.1395 0.1983 1.1395 1.0675
No log 0.5205 38 1.1438 0.1983 1.1438 1.0695
No log 0.5479 40 1.0000 0.1909 1.0000 1.0000
No log 0.5753 42 0.8029 0.2892 0.8029 0.8960
No log 0.6027 44 0.7251 0.3342 0.7251 0.8515
No log 0.6301 46 0.6808 0.3920 0.6808 0.8251
No log 0.6575 48 0.6675 0.4373 0.6675 0.8170
No log 0.6849 50 0.5760 0.4712 0.5760 0.7589
No log 0.7123 52 0.5771 0.4841 0.5771 0.7596
No log 0.7397 54 0.7417 0.4502 0.7417 0.8612
No log 0.7671 56 1.5108 0.2328 1.5108 1.2291
No log 0.7945 58 1.9459 0.1873 1.9459 1.3949
No log 0.8219 60 1.6518 0.2368 1.6518 1.2852
No log 0.8493 62 1.0285 0.3587 1.0285 1.0141
No log 0.8767 64 0.6277 0.5393 0.6277 0.7922
No log 0.9041 66 0.6183 0.3703 0.6183 0.7863
No log 0.9315 68 0.6096 0.3305 0.6096 0.7808
No log 0.9589 70 0.5912 0.3698 0.5912 0.7689
No log 0.9863 72 0.6507 0.3770 0.6507 0.8067
No log 1.0137 74 0.8727 0.2915 0.8727 0.9342
No log 1.0411 76 1.2279 0.3117 1.2279 1.1081
No log 1.0685 78 1.3865 0.2848 1.3865 1.1775
No log 1.0959 80 1.3746 0.2805 1.3746 1.1724
No log 1.1233 82 1.0893 0.3711 1.0893 1.0437
No log 1.1507 84 0.7849 0.3883 0.7849 0.8859
No log 1.1781 86 0.8424 0.2805 0.8424 0.9178
No log 1.2055 88 0.8964 0.2112 0.8964 0.9468
No log 1.2329 90 0.9335 0.2401 0.9335 0.9662
No log 1.2603 92 0.9142 0.2937 0.9142 0.9561
No log 1.2877 94 0.9039 0.3088 0.9039 0.9508
No log 1.3151 96 0.8543 0.3334 0.8543 0.9243
No log 1.3425 98 0.8030 0.3482 0.8030 0.8961
No log 1.3699 100 0.7527 0.3534 0.7527 0.8676
No log 1.3973 102 0.6821 0.4246 0.6821 0.8259
No log 1.4247 104 0.5765 0.5345 0.5765 0.7593
No log 1.4521 106 0.6956 0.5216 0.6956 0.8340
No log 1.4795 108 0.8476 0.4508 0.8476 0.9206
No log 1.5068 110 0.8558 0.4247 0.8558 0.9251
No log 1.5342 112 0.7919 0.4719 0.7919 0.8899
No log 1.5616 114 0.8029 0.4661 0.8029 0.8961
No log 1.5890 116 0.7163 0.4663 0.7163 0.8463
No log 1.6164 118 0.6774 0.4576 0.6774 0.8230
No log 1.6438 120 0.6743 0.4732 0.6743 0.8212
No log 1.6712 122 0.7080 0.4382 0.7080 0.8415
No log 1.6986 124 0.6411 0.4717 0.6411 0.8007
No log 1.7260 126 0.5582 0.5100 0.5582 0.7472
No log 1.7534 128 0.5798 0.5255 0.5798 0.7614
No log 1.7808 130 0.6848 0.5371 0.6848 0.8276
No log 1.8082 132 0.8832 0.5068 0.8832 0.9398
No log 1.8356 134 0.9378 0.4946 0.9378 0.9684
No log 1.8630 136 0.8240 0.5706 0.8240 0.9078
No log 1.8904 138 0.7621 0.5797 0.7621 0.8730
No log 1.9178 140 0.7260 0.5663 0.7260 0.8521
No log 1.9452 142 0.8147 0.5595 0.8147 0.9026
No log 1.9726 144 0.8990 0.5034 0.8990 0.9482
No log 2.0 146 1.0725 0.3767 1.0725 1.0356
No log 2.0274 148 0.9552 0.4407 0.9552 0.9773
No log 2.0548 150 0.6787 0.5571 0.6787 0.8238
No log 2.0822 152 0.6023 0.5147 0.6023 0.7761
No log 2.1096 154 0.6279 0.4793 0.6279 0.7924
No log 2.1370 156 0.6094 0.4843 0.6094 0.7806
No log 2.1644 158 0.5959 0.5178 0.5959 0.7720
No log 2.1918 160 0.6513 0.5660 0.6513 0.8070
No log 2.2192 162 0.7129 0.5359 0.7129 0.8443
No log 2.2466 164 0.9889 0.4043 0.9889 0.9945
No log 2.2740 166 1.0460 0.3928 1.0460 1.0227
No log 2.3014 168 0.7317 0.5373 0.7317 0.8554
No log 2.3288 170 0.6204 0.4867 0.6204 0.7877
No log 2.3562 172 0.6235 0.4705 0.6235 0.7896
No log 2.3836 174 0.5965 0.4387 0.5965 0.7724
No log 2.4110 176 0.6108 0.4262 0.6108 0.7815
No log 2.4384 178 0.6125 0.4152 0.6125 0.7826
No log 2.4658 180 0.6457 0.3887 0.6457 0.8036
No log 2.4932 182 0.6665 0.3254 0.6665 0.8164
No log 2.5205 184 0.6859 0.3092 0.6859 0.8282
No log 2.5479 186 0.6548 0.3688 0.6548 0.8092
No log 2.5753 188 0.6539 0.3998 0.6539 0.8086
No log 2.6027 190 0.6702 0.4553 0.6702 0.8186
No log 2.6301 192 0.7649 0.4022 0.7649 0.8746
No log 2.6575 194 0.7590 0.4573 0.7590 0.8712
No log 2.6849 196 0.6804 0.4843 0.6804 0.8249
No log 2.7123 198 0.6537 0.5102 0.6537 0.8085
No log 2.7397 200 0.6485 0.5309 0.6485 0.8053
No log 2.7671 202 0.6311 0.5131 0.6311 0.7944
No log 2.7945 204 0.6412 0.4811 0.6412 0.8007
No log 2.8219 206 0.6503 0.4716 0.6503 0.8064
No log 2.8493 208 0.5938 0.5276 0.5938 0.7706
No log 2.8767 210 0.5826 0.4854 0.5826 0.7633
No log 2.9041 212 0.5982 0.4587 0.5982 0.7734
No log 2.9315 214 0.6258 0.4850 0.6258 0.7911
No log 2.9589 216 0.6061 0.4991 0.6061 0.7785
No log 2.9863 218 0.5806 0.5006 0.5806 0.7620
No log 3.0137 220 0.6071 0.4863 0.6071 0.7791
No log 3.0411 222 0.5804 0.4902 0.5804 0.7618
No log 3.0685 224 0.5544 0.4893 0.5544 0.7446
No log 3.0959 226 0.6071 0.4797 0.6071 0.7792
No log 3.1233 228 0.6675 0.4910 0.6675 0.8170
No log 3.1507 230 0.6238 0.5247 0.6238 0.7898
No log 3.1781 232 0.6006 0.5374 0.6006 0.7750
No log 3.2055 234 0.5828 0.5088 0.5828 0.7634
No log 3.2329 236 0.5879 0.5390 0.5879 0.7667
No log 3.2603 238 0.6442 0.5492 0.6442 0.8026
No log 3.2877 240 0.6210 0.5541 0.6210 0.7880
No log 3.3151 242 0.5797 0.5682 0.5797 0.7614
No log 3.3425 244 0.5863 0.5563 0.5863 0.7657
No log 3.3699 246 0.5969 0.6178 0.5969 0.7726
No log 3.3973 248 0.6170 0.6102 0.6170 0.7855
No log 3.4247 250 0.5869 0.6103 0.5869 0.7661
No log 3.4521 252 0.5976 0.5919 0.5976 0.7730
No log 3.4795 254 0.5741 0.6175 0.5741 0.7577
No log 3.5068 256 0.5543 0.5688 0.5543 0.7445
No log 3.5342 258 0.5453 0.4801 0.5453 0.7385
No log 3.5616 260 0.5410 0.4818 0.5410 0.7355
No log 3.5890 262 0.5624 0.6303 0.5624 0.7499
No log 3.6164 264 0.6755 0.5077 0.6755 0.8219
No log 3.6438 266 0.6378 0.5147 0.6378 0.7986
No log 3.6712 268 0.5411 0.5501 0.5411 0.7356
No log 3.6986 270 0.5357 0.4359 0.5357 0.7319
No log 3.7260 272 0.5486 0.4728 0.5486 0.7407
No log 3.7534 274 0.5521 0.4583 0.5521 0.7430
No log 3.7808 276 0.5702 0.5943 0.5702 0.7551
No log 3.8082 278 0.6311 0.4931 0.6311 0.7944
No log 3.8356 280 0.7116 0.5008 0.7116 0.8436
No log 3.8630 282 0.6911 0.5054 0.6911 0.8313
No log 3.8904 284 0.6014 0.5339 0.6014 0.7755
No log 3.9178 286 0.5579 0.5514 0.5579 0.7469
No log 3.9452 288 0.5677 0.5293 0.5677 0.7535
No log 3.9726 290 0.5888 0.5567 0.5888 0.7673
No log 4.0 292 0.5736 0.4955 0.5736 0.7574
No log 4.0274 294 0.5666 0.4770 0.5666 0.7527
No log 4.0548 296 0.5567 0.4919 0.5567 0.7461
No log 4.0822 298 0.5664 0.4769 0.5664 0.7526
No log 4.1096 300 0.5626 0.4351 0.5626 0.7501
No log 4.1370 302 0.5618 0.4398 0.5618 0.7495
No log 4.1644 304 0.5739 0.4932 0.5739 0.7576
No log 4.1918 306 0.6049 0.4434 0.6049 0.7778
No log 4.2192 308 0.6218 0.4732 0.6218 0.7885
No log 4.2466 310 0.5749 0.4969 0.5749 0.7582
No log 4.2740 312 0.5562 0.4763 0.5562 0.7458
No log 4.3014 314 0.5892 0.4926 0.5892 0.7676
No log 4.3288 316 0.5561 0.4919 0.5561 0.7457
No log 4.3562 318 0.5546 0.4391 0.5546 0.7447
No log 4.3836 320 0.5570 0.4579 0.5570 0.7463
No log 4.4110 322 0.5660 0.5174 0.5660 0.7523
No log 4.4384 324 0.5775 0.5011 0.5775 0.7600
No log 4.4658 326 0.5844 0.5542 0.5844 0.7645
No log 4.4932 328 0.6197 0.5616 0.6197 0.7872
No log 4.5205 330 0.6341 0.5787 0.6341 0.7963
No log 4.5479 332 0.6641 0.5553 0.6641 0.8150
No log 4.5753 334 0.6147 0.5788 0.6147 0.7840
No log 4.6027 336 0.5835 0.5641 0.5835 0.7639
No log 4.6301 338 0.5900 0.5494 0.5900 0.7681
No log 4.6575 340 0.6669 0.4518 0.6669 0.8166
No log 4.6849 342 0.7451 0.4908 0.7451 0.8632
No log 4.7123 344 0.6619 0.4895 0.6619 0.8136
No log 4.7397 346 0.5985 0.5325 0.5985 0.7736
No log 4.7671 348 0.6203 0.4638 0.6203 0.7876
No log 4.7945 350 0.6591 0.4609 0.6591 0.8119
No log 4.8219 352 0.6668 0.4915 0.6668 0.8166
No log 4.8493 354 0.6654 0.5513 0.6654 0.8157
No log 4.8767 356 0.7328 0.5256 0.7328 0.8560
No log 4.9041 358 0.7304 0.5033 0.7304 0.8546
No log 4.9315 360 0.6859 0.5112 0.6859 0.8282
No log 4.9589 362 0.6646 0.5124 0.6646 0.8152
No log 4.9863 364 0.6229 0.4286 0.6229 0.7892
No log 5.0137 366 0.6304 0.4159 0.6304 0.7940
No log 5.0411 368 0.6370 0.4623 0.6370 0.7981
No log 5.0685 370 0.7233 0.4858 0.7233 0.8505
No log 5.0959 372 0.7614 0.4635 0.7614 0.8726
No log 5.1233 374 0.7559 0.4736 0.7559 0.8694
No log 5.1507 376 0.6669 0.4840 0.6669 0.8167
No log 5.1781 378 0.6583 0.3551 0.6583 0.8113
No log 5.2055 380 0.7246 0.4424 0.7246 0.8512
No log 5.2329 382 0.7489 0.4141 0.7489 0.8654
No log 5.2603 384 0.6748 0.3631 0.6748 0.8215
No log 5.2877 386 0.6625 0.3773 0.6625 0.8139
No log 5.3151 388 0.6728 0.4247 0.6728 0.8202
No log 5.3425 390 0.6546 0.3787 0.6546 0.8091
No log 5.3699 392 0.6813 0.4047 0.6813 0.8254
No log 5.3973 394 0.6811 0.4323 0.6811 0.8253
No log 5.4247 396 0.6624 0.3930 0.6624 0.8139
No log 5.4521 398 0.6406 0.4075 0.6406 0.8004
No log 5.4795 400 0.6342 0.4017 0.6342 0.7964
No log 5.5068 402 0.6546 0.4481 0.6546 0.8091
No log 5.5342 404 0.6288 0.4572 0.6288 0.7929
No log 5.5616 406 0.5963 0.3740 0.5963 0.7722
No log 5.5890 408 0.5931 0.3831 0.5931 0.7701
No log 5.6164 410 0.5837 0.4002 0.5837 0.7640
No log 5.6438 412 0.6093 0.4480 0.6093 0.7806
No log 5.6712 414 0.6756 0.4793 0.6756 0.8220
No log 5.6986 416 0.7123 0.4917 0.7123 0.8440
No log 5.7260 418 0.6974 0.4917 0.6974 0.8351
No log 5.7534 420 0.6113 0.4850 0.6113 0.7818
No log 5.7808 422 0.6042 0.4571 0.6042 0.7773
No log 5.8082 424 0.6008 0.4596 0.6008 0.7751
No log 5.8356 426 0.5902 0.4018 0.5902 0.7683
No log 5.8630 428 0.5891 0.3958 0.5891 0.7675
No log 5.8904 430 0.5797 0.4447 0.5797 0.7614
No log 5.9178 432 0.5790 0.4935 0.5790 0.7609
No log 5.9452 434 0.6028 0.4528 0.6028 0.7764
No log 5.9726 436 0.6022 0.4743 0.6022 0.7760
No log 6.0 438 0.6150 0.5304 0.6150 0.7842
No log 6.0274 440 0.6461 0.4933 0.6461 0.8038
No log 6.0548 442 0.7150 0.4929 0.7150 0.8456
No log 6.0822 444 0.6472 0.4797 0.6472 0.8045
No log 6.1096 446 0.5894 0.4531 0.5894 0.7677
No log 6.1370 448 0.6260 0.4765 0.6260 0.7912
No log 6.1644 450 0.6219 0.4932 0.6219 0.7886
No log 6.1918 452 0.5884 0.4737 0.5884 0.7671
No log 6.2192 454 0.6572 0.4976 0.6572 0.8107
No log 6.2466 456 0.7111 0.4718 0.7111 0.8432
No log 6.2740 458 0.6432 0.5259 0.6432 0.8020
No log 6.3014 460 0.5727 0.4439 0.5727 0.7568
No log 6.3288 462 0.5791 0.4291 0.5791 0.7610
No log 6.3562 464 0.5724 0.3830 0.5724 0.7566
No log 6.3836 466 0.5834 0.4253 0.5834 0.7638
No log 6.4110 468 0.6303 0.4814 0.6303 0.7939
No log 6.4384 470 0.6400 0.4803 0.6400 0.8000
No log 6.4658 472 0.6206 0.5030 0.6206 0.7878
No log 6.4932 474 0.6220 0.4835 0.6220 0.7887
No log 6.5205 476 0.5805 0.4770 0.5805 0.7619
No log 6.5479 478 0.5705 0.4779 0.5705 0.7553
No log 6.5753 480 0.5689 0.4549 0.5689 0.7543
No log 6.6027 482 0.5863 0.4460 0.5863 0.7657
No log 6.6301 484 0.6242 0.4719 0.6242 0.7901
No log 6.6575 486 0.6294 0.4883 0.6294 0.7934
No log 6.6849 488 0.6720 0.4604 0.6720 0.8198
No log 6.7123 490 0.6508 0.4846 0.6508 0.8067
No log 6.7397 492 0.5953 0.4983 0.5953 0.7716
No log 6.7671 494 0.5753 0.4856 0.5753 0.7585
No log 6.7945 496 0.5699 0.4473 0.5699 0.7549
No log 6.8219 498 0.5707 0.4756 0.5707 0.7554
0.3409 6.8493 500 0.5950 0.4753 0.5950 0.7713
0.3409 6.8767 502 0.5986 0.4648 0.5986 0.7737
0.3409 6.9041 504 0.5939 0.4481 0.5939 0.7707
0.3409 6.9315 506 0.5682 0.4250 0.5682 0.7538
0.3409 6.9589 508 0.6086 0.4880 0.6086 0.7801
0.3409 6.9863 510 0.6502 0.4580 0.6502 0.8063
0.3409 7.0137 512 0.6377 0.4792 0.6377 0.7985
0.3409 7.0411 514 0.6107 0.4901 0.6107 0.7815
0.3409 7.0685 516 0.5920 0.5215 0.5920 0.7694
0.3409 7.0959 518 0.6031 0.4813 0.6031 0.7766
0.3409 7.1233 520 0.5900 0.4756 0.5900 0.7681
0.3409 7.1507 522 0.5777 0.4747 0.5777 0.7601
0.3409 7.1781 524 0.5914 0.4441 0.5914 0.7690
0.3409 7.2055 526 0.6030 0.4441 0.6030 0.7765
0.3409 7.2329 528 0.6020 0.4784 0.6020 0.7759
0.3409 7.2603 530 0.6226 0.5616 0.6226 0.7891
0.3409 7.2877 532 0.6392 0.5719 0.6392 0.7995
0.3409 7.3151 534 0.6415 0.5098 0.6415 0.8009
0.3409 7.3425 536 0.6297 0.4807 0.6297 0.7935
0.3409 7.3699 538 0.6098 0.4155 0.6098 0.7809
0.3409 7.3973 540 0.6135 0.4579 0.6135 0.7832
0.3409 7.4247 542 0.6748 0.4331 0.6748 0.8214
0.3409 7.4521 544 0.6866 0.4962 0.6866 0.8286
0.3409 7.4795 546 0.6735 0.5291 0.6735 0.8207
0.3409 7.5068 548 0.6412 0.5587 0.6412 0.8008
0.3409 7.5342 550 0.6439 0.5598 0.6439 0.8024
0.3409 7.5616 552 0.6103 0.5783 0.6103 0.7812
0.3409 7.5890 554 0.5836 0.5040 0.5836 0.7639
0.3409 7.6164 556 0.5682 0.4514 0.5682 0.7538
0.3409 7.6438 558 0.5604 0.4436 0.5604 0.7486
0.3409 7.6712 560 0.5737 0.4105 0.5737 0.7575
0.3409 7.6986 562 0.5985 0.4339 0.5985 0.7737
0.3409 7.7260 564 0.6253 0.4573 0.6253 0.7907

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
4
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for MayBashendy/ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run1_AugV5_k14_task2_organization

Finetuned
(4222)
this model