ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run2_AugV5_k10_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5593
  • Qwk: 0.5334
  • Mse: 0.5593
  • Rmse: 0.7479

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0377 2 4.1527 -0.0404 4.1527 2.0378
No log 0.0755 4 2.4871 0.0434 2.4871 1.5770
No log 0.1132 6 1.2341 0.0356 1.2341 1.1109
No log 0.1509 8 0.9441 -0.0058 0.9441 0.9717
No log 0.1887 10 0.8149 0.2269 0.8149 0.9027
No log 0.2264 12 0.7949 0.2362 0.7949 0.8916
No log 0.2642 14 0.8553 0.2253 0.8553 0.9248
No log 0.3019 16 0.9681 0.0519 0.9681 0.9839
No log 0.3396 18 0.9890 0.0243 0.9890 0.9945
No log 0.3774 20 0.8300 0.2436 0.8300 0.9110
No log 0.4151 22 0.7580 0.2167 0.7580 0.8706
No log 0.4528 24 0.7610 0.2733 0.7610 0.8724
No log 0.4906 26 0.8545 0.2515 0.8545 0.9244
No log 0.5283 28 1.1635 0.0313 1.1635 1.0787
No log 0.5660 30 1.5260 0.0104 1.5260 1.2353
No log 0.6038 32 1.5154 0.0250 1.5154 1.2310
No log 0.6415 34 1.1349 0.0848 1.1349 1.0653
No log 0.6792 36 0.9751 0.1522 0.9751 0.9875
No log 0.7170 38 0.9758 0.1806 0.9758 0.9879
No log 0.7547 40 1.2006 0.1610 1.2006 1.0957
No log 0.7925 42 1.5482 0.1260 1.5482 1.2443
No log 0.8302 44 1.8100 0.1141 1.8100 1.3454
No log 0.8679 46 1.7220 0.1117 1.7220 1.3123
No log 0.9057 48 1.3512 0.0677 1.3512 1.1624
No log 0.9434 50 0.9244 0.1566 0.9244 0.9615
No log 0.9811 52 0.7741 0.3191 0.7741 0.8799
No log 1.0189 54 0.7661 0.3025 0.7661 0.8753
No log 1.0566 56 0.8182 0.2714 0.8182 0.9045
No log 1.0943 58 0.9831 0.1010 0.9831 0.9915
No log 1.1321 60 1.2497 0.0262 1.2497 1.1179
No log 1.1698 62 1.3611 0.1045 1.3611 1.1667
No log 1.2075 64 1.0782 0.2484 1.0782 1.0383
No log 1.2453 66 0.9136 0.2685 0.9136 0.9558
No log 1.2830 68 0.6685 0.4460 0.6685 0.8176
No log 1.3208 70 0.6296 0.4168 0.6296 0.7935
No log 1.3585 72 0.6235 0.4479 0.6235 0.7896
No log 1.3962 74 0.6388 0.4605 0.6388 0.7992
No log 1.4340 76 0.7482 0.4264 0.7482 0.8650
No log 1.4717 78 0.8236 0.3415 0.8236 0.9075
No log 1.5094 80 0.9397 0.3144 0.9397 0.9694
No log 1.5472 82 0.7928 0.3630 0.7928 0.8904
No log 1.5849 84 0.6338 0.4663 0.6338 0.7961
No log 1.6226 86 0.6975 0.4904 0.6975 0.8352
No log 1.6604 88 0.6444 0.4709 0.6444 0.8027
No log 1.6981 90 0.7076 0.4003 0.7076 0.8412
No log 1.7358 92 1.1179 0.3303 1.1179 1.0573
No log 1.7736 94 1.6207 0.2243 1.6207 1.2731
No log 1.8113 96 1.5699 0.1933 1.5699 1.2529
No log 1.8491 98 1.1673 0.1207 1.1673 1.0804
No log 1.8868 100 0.9960 0.2491 0.9960 0.9980
No log 1.9245 102 0.8339 0.2669 0.8339 0.9132
No log 1.9623 104 0.7319 0.2855 0.7319 0.8555
No log 2.0 106 0.6367 0.3612 0.6367 0.7979
No log 2.0377 108 0.5649 0.4811 0.5649 0.7516
No log 2.0755 110 0.5627 0.4653 0.5627 0.7502
No log 2.1132 112 0.5675 0.4977 0.5675 0.7533
No log 2.1509 114 0.5910 0.4736 0.5910 0.7687
No log 2.1887 116 0.7443 0.4933 0.7443 0.8627
No log 2.2264 118 0.8648 0.5245 0.8648 0.9299
No log 2.2642 120 0.8289 0.5338 0.8289 0.9105
No log 2.3019 122 0.6602 0.5545 0.6602 0.8125
No log 2.3396 124 0.6079 0.5584 0.6079 0.7797
No log 2.3774 126 0.6089 0.5134 0.6089 0.7803
No log 2.4151 128 0.6305 0.5609 0.6305 0.7941
No log 2.4528 130 0.6111 0.5520 0.6111 0.7817
No log 2.4906 132 0.6351 0.5093 0.6351 0.7969
No log 2.5283 134 0.6271 0.5211 0.6271 0.7919
No log 2.5660 136 0.6080 0.5431 0.6080 0.7797
No log 2.6038 138 0.6050 0.6158 0.6050 0.7778
No log 2.6415 140 0.7075 0.5527 0.7075 0.8411
No log 2.6792 142 0.9860 0.4360 0.9860 0.9930
No log 2.7170 144 0.9178 0.4956 0.9178 0.9580
No log 2.7547 146 0.6384 0.6082 0.6384 0.7990
No log 2.7925 148 0.6358 0.5209 0.6358 0.7974
No log 2.8302 150 0.6219 0.5416 0.6219 0.7886
No log 2.8679 152 0.5711 0.5817 0.5711 0.7557
No log 2.9057 154 0.5815 0.5481 0.5815 0.7625
No log 2.9434 156 0.6699 0.5311 0.6699 0.8185
No log 2.9811 158 0.6176 0.5180 0.6176 0.7859
No log 3.0189 160 0.6298 0.4968 0.6298 0.7936
No log 3.0566 162 0.8900 0.4417 0.8900 0.9434
No log 3.0943 164 0.8626 0.4547 0.8626 0.9287
No log 3.1321 166 0.6502 0.4636 0.6502 0.8063
No log 3.1698 168 0.5843 0.4781 0.5843 0.7644
No log 3.2075 170 0.8061 0.4296 0.8061 0.8978
No log 3.2453 172 0.8924 0.4632 0.8924 0.9446
No log 3.2830 174 0.7268 0.5109 0.7268 0.8525
No log 3.3208 176 0.6315 0.5245 0.6315 0.7947
No log 3.3585 178 0.6682 0.4943 0.6682 0.8175
No log 3.3962 180 0.6337 0.4664 0.6337 0.7960
No log 3.4340 182 0.6963 0.5845 0.6963 0.8344
No log 3.4717 184 0.8203 0.5190 0.8203 0.9057
No log 3.5094 186 0.7792 0.5502 0.7792 0.8827
No log 3.5472 188 0.6949 0.4656 0.6949 0.8336
No log 3.5849 190 0.8245 0.4670 0.8245 0.9080
No log 3.6226 192 0.8151 0.4621 0.8151 0.9028
No log 3.6604 194 0.6838 0.5047 0.6838 0.8269
No log 3.6981 196 0.7234 0.5203 0.7234 0.8505
No log 3.7358 198 0.8999 0.4467 0.8999 0.9486
No log 3.7736 200 0.9207 0.4403 0.9207 0.9595
No log 3.8113 202 0.7097 0.5181 0.7097 0.8424
No log 3.8491 204 0.6060 0.5329 0.6060 0.7784
No log 3.8868 206 0.6171 0.5362 0.6171 0.7855
No log 3.9245 208 0.6246 0.5272 0.6246 0.7903
No log 3.9623 210 0.6901 0.4989 0.6901 0.8307
No log 4.0 212 0.7187 0.4352 0.7187 0.8478
No log 4.0377 214 0.6174 0.5095 0.6174 0.7858
No log 4.0755 216 0.5955 0.5212 0.5955 0.7717
No log 4.1132 218 0.6339 0.5279 0.6339 0.7962
No log 4.1509 220 0.6496 0.5617 0.6496 0.8060
No log 4.1887 222 0.6744 0.5945 0.6744 0.8212
No log 4.2264 224 0.7436 0.5331 0.7436 0.8623
No log 4.2642 226 0.6859 0.5310 0.6859 0.8282
No log 4.3019 228 0.5818 0.5814 0.5818 0.7628
No log 4.3396 230 0.5693 0.5325 0.5693 0.7545
No log 4.3774 232 0.5764 0.4553 0.5764 0.7592
No log 4.4151 234 0.5790 0.5762 0.5790 0.7609
No log 4.4528 236 0.5795 0.6004 0.5795 0.7612
No log 4.4906 238 0.5819 0.5951 0.5819 0.7628
No log 4.5283 240 0.6029 0.5674 0.6029 0.7764
No log 4.5660 242 0.6126 0.5339 0.6126 0.7827
No log 4.6038 244 0.6763 0.5260 0.6763 0.8224
No log 4.6415 246 0.6287 0.5201 0.6287 0.7929
No log 4.6792 248 0.5628 0.5863 0.5628 0.7502
No log 4.7170 250 0.5642 0.5327 0.5642 0.7511
No log 4.7547 252 0.5664 0.5415 0.5664 0.7526
No log 4.7925 254 0.6191 0.5821 0.6191 0.7868
No log 4.8302 256 0.8312 0.4632 0.8312 0.9117
No log 4.8679 258 0.8583 0.4340 0.8583 0.9264
No log 4.9057 260 0.7139 0.5356 0.7139 0.8450
No log 4.9434 262 0.6198 0.5717 0.6198 0.7873
No log 4.9811 264 0.6162 0.5224 0.6162 0.7850
No log 5.0189 266 0.6217 0.5121 0.6217 0.7885
No log 5.0566 268 0.6495 0.5387 0.6495 0.8059
No log 5.0943 270 0.6278 0.5574 0.6278 0.7923
No log 5.1321 272 0.6246 0.5351 0.6246 0.7903
No log 5.1698 274 0.6241 0.5301 0.6241 0.7900
No log 5.2075 276 0.5991 0.4667 0.5991 0.7740
No log 5.2453 278 0.6345 0.4702 0.6345 0.7966
No log 5.2830 280 0.6000 0.4764 0.6000 0.7746
No log 5.3208 282 0.6066 0.5529 0.6066 0.7788
No log 5.3585 284 0.7885 0.5094 0.7885 0.8880
No log 5.3962 286 0.8024 0.5114 0.8024 0.8958
No log 5.4340 288 0.6875 0.5354 0.6875 0.8292
No log 5.4717 290 0.6046 0.6020 0.6046 0.7776
No log 5.5094 292 0.5876 0.5374 0.5876 0.7666
No log 5.5472 294 0.5828 0.5128 0.5828 0.7634
No log 5.5849 296 0.6235 0.5876 0.6235 0.7896
No log 5.6226 298 0.6649 0.5626 0.6649 0.8154
No log 5.6604 300 0.6440 0.5699 0.6440 0.8025
No log 5.6981 302 0.6081 0.5449 0.6081 0.7798
No log 5.7358 304 0.5981 0.5042 0.5981 0.7733
No log 5.7736 306 0.5990 0.4993 0.5990 0.7740
No log 5.8113 308 0.6099 0.5924 0.6099 0.7810
No log 5.8491 310 0.6787 0.5644 0.6787 0.8238
No log 5.8868 312 0.6876 0.5496 0.6876 0.8292
No log 5.9245 314 0.6294 0.5897 0.6294 0.7933
No log 5.9623 316 0.6114 0.4910 0.6114 0.7819
No log 6.0 318 0.6274 0.5336 0.6274 0.7921
No log 6.0377 320 0.6583 0.4511 0.6583 0.8113
No log 6.0755 322 0.6073 0.5153 0.6073 0.7793
No log 6.1132 324 0.5807 0.4539 0.5807 0.7621
No log 6.1509 326 0.5965 0.4366 0.5965 0.7723
No log 6.1887 328 0.5962 0.5242 0.5962 0.7722
No log 6.2264 330 0.6573 0.5713 0.6573 0.8108
No log 6.2642 332 0.6930 0.5311 0.6930 0.8325
No log 6.3019 334 0.7374 0.5030 0.7374 0.8587
No log 6.3396 336 0.7997 0.4945 0.7997 0.8942
No log 6.3774 338 0.7953 0.4704 0.7953 0.8918
No log 6.4151 340 0.6883 0.4985 0.6883 0.8297
No log 6.4528 342 0.6195 0.5427 0.6195 0.7871
No log 6.4906 344 0.6172 0.5607 0.6172 0.7856
No log 6.5283 346 0.6224 0.5296 0.6224 0.7889
No log 6.5660 348 0.6183 0.5475 0.6183 0.7863
No log 6.6038 350 0.6133 0.5338 0.6133 0.7831
No log 6.6415 352 0.6064 0.5271 0.6064 0.7787
No log 6.6792 354 0.5855 0.5085 0.5855 0.7652
No log 6.7170 356 0.5779 0.5175 0.5779 0.7602
No log 6.7547 358 0.5603 0.4474 0.5603 0.7486
No log 6.7925 360 0.5550 0.4798 0.5550 0.7450
No log 6.8302 362 0.5628 0.5034 0.5628 0.7502
No log 6.8679 364 0.5652 0.5130 0.5652 0.7518
No log 6.9057 366 0.5811 0.5633 0.5811 0.7623
No log 6.9434 368 0.5746 0.5205 0.5746 0.7580
No log 6.9811 370 0.5641 0.5226 0.5641 0.7511
No log 7.0189 372 0.5756 0.5137 0.5756 0.7587
No log 7.0566 374 0.5646 0.5196 0.5646 0.7514
No log 7.0943 376 0.5819 0.5242 0.5819 0.7628
No log 7.1321 378 0.5929 0.5356 0.5929 0.7700
No log 7.1698 380 0.5892 0.5342 0.5892 0.7676
No log 7.2075 382 0.5727 0.5031 0.5727 0.7568
No log 7.2453 384 0.5550 0.5246 0.5550 0.7450
No log 7.2830 386 0.5491 0.4923 0.5491 0.7410
No log 7.3208 388 0.5390 0.4686 0.5390 0.7341
No log 7.3585 390 0.5402 0.4870 0.5402 0.7350
No log 7.3962 392 0.5538 0.5357 0.5538 0.7442
No log 7.4340 394 0.5899 0.5468 0.5899 0.7681
No log 7.4717 396 0.6340 0.5473 0.6340 0.7962
No log 7.5094 398 0.5941 0.5264 0.5941 0.7708
No log 7.5472 400 0.5682 0.5372 0.5682 0.7538
No log 7.5849 402 0.5656 0.5684 0.5656 0.7521
No log 7.6226 404 0.5842 0.5408 0.5842 0.7643
No log 7.6604 406 0.6009 0.5614 0.6009 0.7752
No log 7.6981 408 0.5433 0.5394 0.5433 0.7371
No log 7.7358 410 0.5309 0.5308 0.5309 0.7286
No log 7.7736 412 0.5397 0.5158 0.5397 0.7347
No log 7.8113 414 0.5790 0.5835 0.5790 0.7609
No log 7.8491 416 0.5992 0.5869 0.5992 0.7741
No log 7.8868 418 0.5774 0.5907 0.5774 0.7599
No log 7.9245 420 0.5937 0.6116 0.5937 0.7705
No log 7.9623 422 0.6061 0.5969 0.6061 0.7785
No log 8.0 424 0.5778 0.5878 0.5778 0.7602
No log 8.0377 426 0.5562 0.5945 0.5562 0.7458
No log 8.0755 428 0.5639 0.5602 0.5639 0.7509
No log 8.1132 430 0.5490 0.5223 0.5490 0.7410
No log 8.1509 432 0.5539 0.4687 0.5539 0.7443
No log 8.1887 434 0.6637 0.5226 0.6637 0.8147
No log 8.2264 436 0.7539 0.4587 0.7539 0.8683
No log 8.2642 438 0.7092 0.4952 0.7092 0.8422
No log 8.3019 440 0.6126 0.5316 0.6126 0.7827
No log 8.3396 442 0.5777 0.5147 0.5777 0.7601
No log 8.3774 444 0.5893 0.5266 0.5893 0.7676
No log 8.4151 446 0.5962 0.5461 0.5962 0.7721
No log 8.4528 448 0.5849 0.5543 0.5849 0.7648
No log 8.4906 450 0.6028 0.5627 0.6028 0.7764
No log 8.5283 452 0.7330 0.5020 0.7330 0.8562
No log 8.5660 454 0.8199 0.4112 0.8199 0.9055
No log 8.6038 456 0.7784 0.4760 0.7784 0.8823
No log 8.6415 458 0.6958 0.5413 0.6958 0.8341
No log 8.6792 460 0.6342 0.5522 0.6342 0.7964
No log 8.7170 462 0.5613 0.5547 0.5613 0.7492
No log 8.7547 464 0.5453 0.5459 0.5453 0.7384
No log 8.7925 466 0.5561 0.5249 0.5561 0.7457
No log 8.8302 468 0.6415 0.5384 0.6415 0.8010
No log 8.8679 470 0.6975 0.5089 0.6975 0.8352
No log 8.9057 472 0.6516 0.5268 0.6516 0.8072
No log 8.9434 474 0.5997 0.5444 0.5997 0.7744
No log 8.9811 476 0.6267 0.5226 0.6267 0.7916
No log 9.0189 478 0.6209 0.5226 0.6209 0.7880
No log 9.0566 480 0.5916 0.5506 0.5916 0.7691
No log 9.0943 482 0.6229 0.5579 0.6229 0.7892
No log 9.1321 484 0.6581 0.5379 0.6581 0.8112
No log 9.1698 486 0.6202 0.5579 0.6202 0.7876
No log 9.2075 488 0.6213 0.5782 0.6213 0.7883
No log 9.2453 490 0.5896 0.5837 0.5896 0.7678
No log 9.2830 492 0.6180 0.5777 0.6180 0.7861
No log 9.3208 494 0.6117 0.5816 0.6117 0.7821
No log 9.3585 496 0.5963 0.5816 0.5963 0.7722
No log 9.3962 498 0.5850 0.5589 0.5850 0.7649
0.3832 9.4340 500 0.5636 0.5383 0.5636 0.7508
0.3832 9.4717 502 0.5952 0.5438 0.5952 0.7715
0.3832 9.5094 504 0.5839 0.5414 0.5839 0.7642
0.3832 9.5472 506 0.5680 0.5518 0.5680 0.7537
0.3832 9.5849 508 0.5741 0.5864 0.5741 0.7577
0.3832 9.6226 510 0.5831 0.5864 0.5831 0.7636
0.3832 9.6604 512 0.6139 0.5377 0.6139 0.7835
0.3832 9.6981 514 0.6178 0.5524 0.6178 0.7860
0.3832 9.7358 516 0.6560 0.5144 0.6560 0.8100
0.3832 9.7736 518 0.6814 0.5033 0.6814 0.8254
0.3832 9.8113 520 0.7189 0.4796 0.7189 0.8479
0.3832 9.8491 522 0.6564 0.4996 0.6564 0.8102
0.3832 9.8868 524 0.5593 0.5334 0.5593 0.7479

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
4
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for MayBashendy/ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run2_AugV5_k10_task2_organization

Finetuned
(4222)
this model