ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run2_AugV5_k7_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5929
  • Qwk: 0.5234
  • Mse: 0.5929
  • Rmse: 0.7700

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0541 2 4.0802 -0.0213 4.0802 2.0199
No log 0.1081 4 2.1758 0.0463 2.1758 1.4751
No log 0.1622 6 1.3229 0.0356 1.3229 1.1502
No log 0.2162 8 1.0938 -0.0167 1.0938 1.0458
No log 0.2703 10 1.0733 -0.0487 1.0733 1.0360
No log 0.3243 12 1.0927 0.0148 1.0927 1.0453
No log 0.3784 14 1.0923 0.0940 1.0923 1.0451
No log 0.4324 16 1.0665 0.1378 1.0665 1.0327
No log 0.4865 18 1.6473 0.1147 1.6473 1.2835
No log 0.5405 20 1.3776 0.1094 1.3776 1.1737
No log 0.5946 22 0.9009 0.1805 0.9009 0.9492
No log 0.6486 24 0.9912 0.1421 0.9912 0.9956
No log 0.7027 26 1.6272 0.1440 1.6272 1.2756
No log 0.7568 28 2.1609 0.1050 2.1609 1.4700
No log 0.8108 30 1.8929 0.1410 1.8929 1.3758
No log 0.8649 32 1.2676 0.0845 1.2676 1.1259
No log 0.9189 34 0.7827 0.2624 0.7827 0.8847
No log 0.9730 36 0.6802 0.3303 0.6802 0.8248
No log 1.0270 38 0.7048 0.3599 0.7048 0.8396
No log 1.0811 40 0.9175 0.1695 0.9175 0.9579
No log 1.1351 42 1.5689 0.1452 1.5689 1.2526
No log 1.1892 44 2.2949 0.0753 2.2949 1.5149
No log 1.2432 46 2.3418 0.0897 2.3418 1.5303
No log 1.2973 48 1.9409 0.1660 1.9409 1.3932
No log 1.3514 50 1.2785 0.1326 1.2785 1.1307
No log 1.4054 52 0.7868 0.3546 0.7868 0.8870
No log 1.4595 54 0.7277 0.3533 0.7277 0.8530
No log 1.5135 56 0.8130 0.2193 0.8130 0.9017
No log 1.5676 58 0.8716 0.2250 0.8716 0.9336
No log 1.6216 60 0.8258 0.3281 0.8258 0.9087
No log 1.6757 62 0.8724 0.2995 0.8724 0.9340
No log 1.7297 64 0.7957 0.4369 0.7957 0.8920
No log 1.7838 66 0.8464 0.3494 0.8464 0.9200
No log 1.8378 68 0.9078 0.3292 0.9078 0.9528
No log 1.8919 70 0.8132 0.3714 0.8132 0.9018
No log 1.9459 72 0.6150 0.5053 0.6150 0.7842
No log 2.0 74 0.5887 0.5330 0.5887 0.7673
No log 2.0541 76 0.6126 0.5201 0.6126 0.7827
No log 2.1081 78 0.7034 0.5023 0.7034 0.8387
No log 2.1622 80 0.6659 0.5247 0.6659 0.8160
No log 2.2162 82 0.6407 0.5228 0.6407 0.8005
No log 2.2703 84 0.6723 0.5036 0.6723 0.8200
No log 2.3243 86 0.8341 0.4653 0.8341 0.9133
No log 2.3784 88 1.0930 0.3839 1.0930 1.0455
No log 2.4324 90 1.1454 0.3612 1.1454 1.0702
No log 2.4865 92 0.9582 0.4085 0.9582 0.9789
No log 2.5405 94 0.7517 0.4757 0.7517 0.8670
No log 2.5946 96 0.6244 0.4891 0.6244 0.7902
No log 2.6486 98 0.6030 0.4274 0.6030 0.7766
No log 2.7027 100 0.6245 0.4698 0.6245 0.7902
No log 2.7568 102 0.8272 0.4567 0.8272 0.9095
No log 2.8108 104 1.0874 0.3604 1.0874 1.0428
No log 2.8649 106 1.1329 0.3178 1.1329 1.0644
No log 2.9189 108 1.1371 0.2601 1.1371 1.0663
No log 2.9730 110 0.8772 0.3882 0.8772 0.9366
No log 3.0270 112 0.6164 0.4179 0.6164 0.7851
No log 3.0811 114 0.5817 0.4810 0.5817 0.7627
No log 3.1351 116 0.6638 0.5457 0.6638 0.8147
No log 3.1892 118 1.0565 0.4023 1.0565 1.0279
No log 3.2432 120 1.9199 0.2192 1.9199 1.3856
No log 3.2973 122 2.1968 0.1632 2.1968 1.4822
No log 3.3514 124 1.9959 0.2097 1.9959 1.4127
No log 3.4054 126 1.5399 0.3213 1.5399 1.2409
No log 3.4595 128 1.3521 0.3269 1.3521 1.1628
No log 3.5135 130 1.1233 0.3340 1.1233 1.0598
No log 3.5676 132 0.8622 0.4715 0.8622 0.9285
No log 3.6216 134 0.7848 0.4988 0.7848 0.8859
No log 3.6757 136 0.6522 0.4577 0.6522 0.8076
No log 3.7297 138 0.5664 0.4970 0.5664 0.7526
No log 3.7838 140 0.5723 0.5026 0.5723 0.7565
No log 3.8378 142 0.6548 0.4633 0.6548 0.8092
No log 3.8919 144 0.7812 0.4960 0.7812 0.8838
No log 3.9459 146 0.7134 0.4786 0.7134 0.8446
No log 4.0 148 0.5900 0.5728 0.5900 0.7681
No log 4.0541 150 0.5943 0.5234 0.5943 0.7709
No log 4.1081 152 0.6196 0.5881 0.6196 0.7872
No log 4.1622 154 0.7037 0.5721 0.7037 0.8388
No log 4.2162 156 0.7160 0.5706 0.7160 0.8462
No log 4.2703 158 0.8240 0.4843 0.8240 0.9077
No log 4.3243 160 0.8004 0.5026 0.8004 0.8946
No log 4.3784 162 0.6985 0.5736 0.6985 0.8358
No log 4.4324 164 0.6189 0.5969 0.6189 0.7867
No log 4.4865 166 0.6403 0.5938 0.6403 0.8002
No log 4.5405 168 0.7306 0.5097 0.7306 0.8547
No log 4.5946 170 0.8052 0.5231 0.8052 0.8974
No log 4.6486 172 0.6959 0.5064 0.6959 0.8342
No log 4.7027 174 0.6034 0.5116 0.6034 0.7768
No log 4.7568 176 0.6300 0.4999 0.6300 0.7937
No log 4.8108 178 0.6057 0.4861 0.6057 0.7782
No log 4.8649 180 0.7091 0.5424 0.7091 0.8421
No log 4.9189 182 0.9481 0.4211 0.9481 0.9737
No log 4.9730 184 0.9663 0.4068 0.9663 0.9830
No log 5.0270 186 0.7471 0.5380 0.7471 0.8644
No log 5.0811 188 0.6595 0.6047 0.6595 0.8121
No log 5.1351 190 0.6592 0.5935 0.6592 0.8119
No log 5.1892 192 0.6403 0.6137 0.6403 0.8002
No log 5.2432 194 0.6579 0.5297 0.6579 0.8111
No log 5.2973 196 0.6282 0.5901 0.6282 0.7926
No log 5.3514 198 0.6173 0.5067 0.6173 0.7857
No log 5.4054 200 0.5960 0.5145 0.5960 0.7720
No log 5.4595 202 0.6388 0.5643 0.6388 0.7993
No log 5.5135 204 0.7146 0.5046 0.7146 0.8453
No log 5.5676 206 0.7600 0.4882 0.7600 0.8718
No log 5.6216 208 0.7549 0.4860 0.7549 0.8689
No log 5.6757 210 0.6690 0.5404 0.6690 0.8179
No log 5.7297 212 0.6819 0.5781 0.6819 0.8258
No log 5.7838 214 0.7981 0.5437 0.7981 0.8934
No log 5.8378 216 0.9549 0.4420 0.9549 0.9772
No log 5.8919 218 0.8969 0.4694 0.8969 0.9470
No log 5.9459 220 0.7922 0.4826 0.7922 0.8901
No log 6.0 222 0.6850 0.5581 0.6850 0.8277
No log 6.0541 224 0.6559 0.5258 0.6559 0.8099
No log 6.1081 226 0.6132 0.4938 0.6132 0.7830
No log 6.1622 228 0.6079 0.4669 0.6079 0.7797
No log 6.2162 230 0.7024 0.4161 0.7024 0.8381
No log 6.2703 232 0.7112 0.4351 0.7112 0.8433
No log 6.3243 234 0.6227 0.4626 0.6227 0.7891
No log 6.3784 236 0.6217 0.5579 0.6217 0.7885
No log 6.4324 238 0.6615 0.4955 0.6615 0.8133
No log 6.4865 240 0.6724 0.5106 0.6724 0.8200
No log 6.5405 242 0.6231 0.5569 0.6231 0.7894
No log 6.5946 244 0.6033 0.5778 0.6033 0.7767
No log 6.6486 246 0.6230 0.5432 0.6230 0.7893
No log 6.7027 248 0.6157 0.5301 0.6157 0.7847
No log 6.7568 250 0.6399 0.5062 0.6399 0.8000
No log 6.8108 252 0.6019 0.5277 0.6019 0.7758
No log 6.8649 254 0.5653 0.5136 0.5653 0.7518
No log 6.9189 256 0.5553 0.4699 0.5553 0.7452
No log 6.9730 258 0.5556 0.5139 0.5556 0.7454
No log 7.0270 260 0.5590 0.5464 0.5590 0.7477
No log 7.0811 262 0.5909 0.5065 0.5909 0.7687
No log 7.1351 264 0.6431 0.4566 0.6431 0.8019
No log 7.1892 266 0.6427 0.4713 0.6427 0.8017
No log 7.2432 268 0.5759 0.5228 0.5759 0.7589
No log 7.2973 270 0.5598 0.5561 0.5598 0.7482
No log 7.3514 272 0.5696 0.5925 0.5696 0.7547
No log 7.4054 274 0.5768 0.5421 0.5768 0.7595
No log 7.4595 276 0.6225 0.4867 0.6225 0.7890
No log 7.5135 278 0.6686 0.4583 0.6686 0.8177
No log 7.5676 280 0.7273 0.4832 0.7273 0.8528
No log 7.6216 282 0.6372 0.4496 0.6372 0.7983
No log 7.6757 284 0.5781 0.4835 0.5781 0.7603
No log 7.7297 286 0.5638 0.5344 0.5638 0.7509
No log 7.7838 288 0.5644 0.5532 0.5644 0.7513
No log 7.8378 290 0.5842 0.5004 0.5842 0.7643
No log 7.8919 292 0.6408 0.5080 0.6408 0.8005
No log 7.9459 294 0.7009 0.5397 0.7009 0.8372
No log 8.0 296 0.7175 0.5124 0.7175 0.8471
No log 8.0541 298 0.6923 0.5038 0.6923 0.8321
No log 8.1081 300 0.6321 0.4742 0.6321 0.7950
No log 8.1622 302 0.5905 0.4920 0.5905 0.7684
No log 8.2162 304 0.6141 0.5230 0.6141 0.7837
No log 8.2703 306 0.7122 0.4902 0.7122 0.8439
No log 8.3243 308 0.6828 0.5168 0.6828 0.8263
No log 8.3784 310 0.5956 0.5799 0.5956 0.7717
No log 8.4324 312 0.6107 0.5877 0.6107 0.7815
No log 8.4865 314 0.6552 0.5291 0.6552 0.8095
No log 8.5405 316 0.6853 0.5329 0.6853 0.8279
No log 8.5946 318 0.6261 0.5373 0.6261 0.7913
No log 8.6486 320 0.5756 0.5101 0.5756 0.7587
No log 8.7027 322 0.5987 0.5320 0.5987 0.7738
No log 8.7568 324 0.5812 0.5728 0.5812 0.7624
No log 8.8108 326 0.5658 0.5737 0.5658 0.7522
No log 8.8649 328 0.5584 0.5662 0.5584 0.7473
No log 8.9189 330 0.5493 0.5231 0.5493 0.7411
No log 8.9730 332 0.5410 0.5059 0.5410 0.7355
No log 9.0270 334 0.5469 0.4867 0.5469 0.7395
No log 9.0811 336 0.5601 0.5283 0.5601 0.7484
No log 9.1351 338 0.5482 0.5310 0.5482 0.7404
No log 9.1892 340 0.5495 0.5630 0.5495 0.7413
No log 9.2432 342 0.5610 0.6228 0.5610 0.7490
No log 9.2973 344 0.6205 0.5904 0.6205 0.7877
No log 9.3514 346 0.6510 0.5439 0.6510 0.8069
No log 9.4054 348 0.6156 0.5949 0.6156 0.7846
No log 9.4595 350 0.5798 0.5708 0.5798 0.7615
No log 9.5135 352 0.5690 0.5493 0.5690 0.7543
No log 9.5676 354 0.5771 0.5598 0.5771 0.7596
No log 9.6216 356 0.5875 0.5503 0.5875 0.7665
No log 9.6757 358 0.5976 0.5375 0.5976 0.7731
No log 9.7297 360 0.5772 0.5614 0.5772 0.7597
No log 9.7838 362 0.5793 0.5670 0.5793 0.7611
No log 9.8378 364 0.5806 0.5706 0.5806 0.7620
No log 9.8919 366 0.5675 0.5105 0.5675 0.7533
No log 9.9459 368 0.5582 0.5885 0.5582 0.7471
No log 10.0 370 0.5501 0.5706 0.5501 0.7417
No log 10.0541 372 0.5541 0.5833 0.5541 0.7444
No log 10.1081 374 0.5490 0.5896 0.5490 0.7409
No log 10.1622 376 0.5437 0.5532 0.5437 0.7374
No log 10.2162 378 0.5463 0.5470 0.5463 0.7391
No log 10.2703 380 0.5501 0.5408 0.5501 0.7417
No log 10.3243 382 0.5731 0.6214 0.5731 0.7570
No log 10.3784 384 0.6497 0.5630 0.6497 0.8061
No log 10.4324 386 0.7345 0.5284 0.7345 0.8570
No log 10.4865 388 0.7420 0.5177 0.7420 0.8614
No log 10.5405 390 0.6208 0.6246 0.6208 0.7879
No log 10.5946 392 0.5632 0.5924 0.5632 0.7505
No log 10.6486 394 0.5724 0.5850 0.5724 0.7566
No log 10.7027 396 0.5695 0.5926 0.5695 0.7546
No log 10.7568 398 0.5923 0.5522 0.5923 0.7696
No log 10.8108 400 0.5712 0.5829 0.5712 0.7558
No log 10.8649 402 0.5688 0.5652 0.5688 0.7542
No log 10.9189 404 0.5673 0.5658 0.5673 0.7532
No log 10.9730 406 0.5691 0.5658 0.5691 0.7544
No log 11.0270 408 0.5742 0.5515 0.5742 0.7578
No log 11.0811 410 0.6347 0.5317 0.6347 0.7967
No log 11.1351 412 0.6902 0.5274 0.6902 0.8308
No log 11.1892 414 0.6079 0.4990 0.6079 0.7797
No log 11.2432 416 0.5491 0.4948 0.5491 0.7410
No log 11.2973 418 0.5420 0.4697 0.5420 0.7362
No log 11.3514 420 0.5349 0.4710 0.5349 0.7313
No log 11.4054 422 0.5418 0.5053 0.5418 0.7361
No log 11.4595 424 0.6050 0.5068 0.6050 0.7778
No log 11.5135 426 0.6739 0.5213 0.6739 0.8209
No log 11.5676 428 0.6285 0.5644 0.6285 0.7928
No log 11.6216 430 0.6189 0.5547 0.6189 0.7867
No log 11.6757 432 0.6093 0.5103 0.6093 0.7806
No log 11.7297 434 0.6093 0.5190 0.6093 0.7806
No log 11.7838 436 0.5700 0.4916 0.5700 0.7550
No log 11.8378 438 0.5538 0.5318 0.5538 0.7442
No log 11.8919 440 0.5502 0.5625 0.5502 0.7417
No log 11.9459 442 0.5497 0.5539 0.5497 0.7414
No log 12.0 444 0.5457 0.5306 0.5457 0.7387
No log 12.0541 446 0.5568 0.5083 0.5568 0.7462
No log 12.1081 448 0.5818 0.5921 0.5818 0.7627
No log 12.1622 450 0.6197 0.6152 0.6197 0.7872
No log 12.2162 452 0.6178 0.6088 0.6178 0.7860
No log 12.2703 454 0.6080 0.5889 0.6080 0.7798
No log 12.3243 456 0.6048 0.6040 0.6048 0.7777
No log 12.3784 458 0.6121 0.6181 0.6121 0.7824
No log 12.4324 460 0.6308 0.6504 0.6308 0.7942
No log 12.4865 462 0.6069 0.6307 0.6069 0.7790
No log 12.5405 464 0.5984 0.5864 0.5984 0.7736
No log 12.5946 466 0.5839 0.6203 0.5839 0.7642
No log 12.6486 468 0.5797 0.5989 0.5797 0.7614
No log 12.7027 470 0.5846 0.5768 0.5846 0.7646
No log 12.7568 472 0.5783 0.5593 0.5783 0.7604
No log 12.8108 474 0.5585 0.5137 0.5585 0.7473
No log 12.8649 476 0.5566 0.4731 0.5566 0.7461
No log 12.9189 478 0.5528 0.4598 0.5528 0.7435
No log 12.9730 480 0.5619 0.4721 0.5619 0.7496
No log 13.0270 482 0.5619 0.4749 0.5619 0.7496
No log 13.0811 484 0.5623 0.5095 0.5623 0.7499
No log 13.1351 486 0.5640 0.5024 0.5640 0.7510
No log 13.1892 488 0.5713 0.5103 0.5713 0.7558
No log 13.2432 490 0.5749 0.4858 0.5749 0.7582
No log 13.2973 492 0.5715 0.4749 0.5715 0.7560
No log 13.3514 494 0.5735 0.4839 0.5735 0.7573
No log 13.4054 496 0.5956 0.5170 0.5956 0.7717
No log 13.4595 498 0.5820 0.5004 0.5820 0.7629
0.3491 13.5135 500 0.5820 0.5653 0.5820 0.7629
0.3491 13.5676 502 0.5926 0.5386 0.5926 0.7698
0.3491 13.6216 504 0.6005 0.5355 0.6005 0.7749
0.3491 13.6757 506 0.6153 0.5384 0.6153 0.7844
0.3491 13.7297 508 0.6128 0.5325 0.6128 0.7828
0.3491 13.7838 510 0.5929 0.5234 0.5929 0.7700

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
4
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for MayBashendy/ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run2_AugV5_k7_task2_organization

Finetuned
(4222)
this model