ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run3_AugV5_k4_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5702
  • Qwk: 0.5276
  • Mse: 0.5702
  • Rmse: 0.7551

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0909 2 4.1289 -0.0163 4.1289 2.0320
No log 0.1818 4 2.3335 0.0928 2.3335 1.5276
No log 0.2727 6 1.3547 0.1304 1.3547 1.1639
No log 0.3636 8 0.8243 0.1612 0.8243 0.9079
No log 0.4545 10 0.8167 0.1543 0.8167 0.9037
No log 0.5455 12 0.8909 0.0600 0.8909 0.9439
No log 0.6364 14 0.8382 0.0797 0.8382 0.9155
No log 0.7273 16 0.9294 0.0710 0.9295 0.9641
No log 0.8182 18 0.9013 0.0364 0.9013 0.9494
No log 0.9091 20 0.7705 0.1758 0.7705 0.8778
No log 1.0 22 0.8491 0.1837 0.8491 0.9215
No log 1.0909 24 1.0730 -0.0064 1.0730 1.0359
No log 1.1818 26 0.9685 0.1752 0.9685 0.9841
No log 1.2727 28 0.8001 0.2758 0.8001 0.8945
No log 1.3636 30 0.6932 0.3324 0.6932 0.8326
No log 1.4545 32 0.6469 0.3085 0.6469 0.8043
No log 1.5455 34 0.7003 0.3825 0.7003 0.8369
No log 1.6364 36 0.9571 0.3413 0.9571 0.9783
No log 1.7273 38 0.8095 0.3397 0.8095 0.8997
No log 1.8182 40 0.6484 0.3199 0.6484 0.8052
No log 1.9091 42 0.6724 0.3422 0.6724 0.8200
No log 2.0 44 0.9376 0.3737 0.9376 0.9683
No log 2.0909 46 0.9724 0.2971 0.9724 0.9861
No log 2.1818 48 0.6714 0.3809 0.6714 0.8194
No log 2.2727 50 0.6992 0.4827 0.6992 0.8362
No log 2.3636 52 0.6801 0.4829 0.6801 0.8247
No log 2.4545 54 0.7007 0.3582 0.7007 0.8371
No log 2.5455 56 0.8167 0.2229 0.8167 0.9037
No log 2.6364 58 0.8054 0.3327 0.8054 0.8975
No log 2.7273 60 0.7523 0.4179 0.7523 0.8674
No log 2.8182 62 0.7476 0.3990 0.7476 0.8646
No log 2.9091 64 0.7602 0.4638 0.7602 0.8719
No log 3.0 66 0.7520 0.4320 0.7520 0.8672
No log 3.0909 68 0.7639 0.4721 0.7639 0.8740
No log 3.1818 70 0.7820 0.4843 0.7820 0.8843
No log 3.2727 72 0.7551 0.4786 0.7551 0.8689
No log 3.3636 74 0.7043 0.4932 0.7043 0.8392
No log 3.4545 76 0.6817 0.5159 0.6817 0.8257
No log 3.5455 78 0.8143 0.5015 0.8143 0.9024
No log 3.6364 80 0.7963 0.4933 0.7963 0.8924
No log 3.7273 82 0.6655 0.4843 0.6655 0.8158
No log 3.8182 84 0.6994 0.5257 0.6994 0.8363
No log 3.9091 86 0.7376 0.5171 0.7376 0.8589
No log 4.0 88 0.6865 0.5336 0.6865 0.8286
No log 4.0909 90 0.6946 0.5355 0.6946 0.8334
No log 4.1818 92 0.6506 0.5309 0.6506 0.8066
No log 4.2727 94 0.7794 0.4715 0.7794 0.8829
No log 4.3636 96 0.9749 0.4178 0.9749 0.9874
No log 4.4545 98 0.7843 0.4764 0.7843 0.8856
No log 4.5455 100 0.6275 0.5193 0.6275 0.7921
No log 4.6364 102 0.6422 0.5641 0.6422 0.8013
No log 4.7273 104 0.6895 0.5382 0.6895 0.8303
No log 4.8182 106 0.6924 0.5332 0.6924 0.8321
No log 4.9091 108 0.7141 0.5619 0.7141 0.8451
No log 5.0 110 0.7168 0.5298 0.7168 0.8466
No log 5.0909 112 0.7726 0.5584 0.7726 0.8790
No log 5.1818 114 0.6981 0.5476 0.6981 0.8355
No log 5.2727 116 0.7291 0.5198 0.7291 0.8539
No log 5.3636 118 0.6943 0.5069 0.6943 0.8332
No log 5.4545 120 0.6622 0.5123 0.6622 0.8138
No log 5.5455 122 0.6204 0.5496 0.6204 0.7877
No log 5.6364 124 0.6236 0.4926 0.6236 0.7897
No log 5.7273 126 0.6298 0.5655 0.6298 0.7936
No log 5.8182 128 0.6542 0.5416 0.6542 0.8088
No log 5.9091 130 0.6566 0.5531 0.6566 0.8103
No log 6.0 132 0.6911 0.5088 0.6911 0.8313
No log 6.0909 134 0.6520 0.5248 0.6520 0.8075
No log 6.1818 136 0.6579 0.5327 0.6579 0.8111
No log 6.2727 138 0.6381 0.5191 0.6381 0.7988
No log 6.3636 140 0.7690 0.5085 0.7690 0.8769
No log 6.4545 142 0.7417 0.5101 0.7417 0.8612
No log 6.5455 144 0.7300 0.5151 0.7300 0.8544
No log 6.6364 146 0.6336 0.5285 0.6336 0.7960
No log 6.7273 148 0.6141 0.5239 0.6141 0.7836
No log 6.8182 150 0.6567 0.4857 0.6567 0.8104
No log 6.9091 152 0.8897 0.4287 0.8897 0.9432
No log 7.0 154 0.8958 0.4398 0.8958 0.9465
No log 7.0909 156 0.6126 0.5023 0.6126 0.7827
No log 7.1818 158 0.6669 0.4960 0.6669 0.8166
No log 7.2727 160 0.7190 0.5039 0.7190 0.8479
No log 7.3636 162 0.5995 0.5081 0.5995 0.7742
No log 7.4545 164 0.6595 0.4752 0.6595 0.8121
No log 7.5455 166 0.6808 0.4536 0.6808 0.8251
No log 7.6364 168 0.6300 0.5618 0.6300 0.7937
No log 7.7273 170 0.8434 0.4557 0.8434 0.9184
No log 7.8182 172 0.9728 0.4074 0.9728 0.9863
No log 7.9091 174 0.7747 0.4881 0.7747 0.8802
No log 8.0 176 0.6230 0.5875 0.6230 0.7893
No log 8.0909 178 0.6791 0.4978 0.6791 0.8241
No log 8.1818 180 0.6281 0.4657 0.6281 0.7925
No log 8.2727 182 0.5617 0.5652 0.5617 0.7495
No log 8.3636 184 0.5606 0.5505 0.5606 0.7487
No log 8.4545 186 0.5504 0.5796 0.5504 0.7419
No log 8.5455 188 0.5510 0.5837 0.5510 0.7423
No log 8.6364 190 0.5509 0.5607 0.5509 0.7423
No log 8.7273 192 0.5612 0.6065 0.5612 0.7491
No log 8.8182 194 0.5691 0.5554 0.5691 0.7544
No log 8.9091 196 0.5771 0.5593 0.5771 0.7596
No log 9.0 198 0.5954 0.5587 0.5954 0.7716
No log 9.0909 200 0.5969 0.5646 0.5969 0.7726
No log 9.1818 202 0.6034 0.5788 0.6034 0.7768
No log 9.2727 204 0.5833 0.5182 0.5833 0.7637
No log 9.3636 206 0.5934 0.5093 0.5934 0.7703
No log 9.4545 208 0.6044 0.5188 0.6044 0.7774
No log 9.5455 210 0.6236 0.5380 0.6236 0.7897
No log 9.6364 212 0.5629 0.5505 0.5629 0.7503
No log 9.7273 214 0.6084 0.5056 0.6084 0.7800
No log 9.8182 216 0.6008 0.5297 0.6008 0.7751
No log 9.9091 218 0.5704 0.5538 0.5704 0.7552
No log 10.0 220 0.7712 0.4946 0.7712 0.8782
No log 10.0909 222 0.8232 0.4794 0.8232 0.9073
No log 10.1818 224 0.6717 0.5239 0.6717 0.8196
No log 10.2727 226 0.5981 0.5822 0.5981 0.7734
No log 10.3636 228 0.6178 0.5023 0.6178 0.7860
No log 10.4545 230 0.5906 0.5396 0.5906 0.7685
No log 10.5455 232 0.5917 0.5563 0.5917 0.7692
No log 10.6364 234 0.6051 0.5379 0.6051 0.7779
No log 10.7273 236 0.6386 0.5126 0.6386 0.7991
No log 10.8182 238 0.6946 0.5119 0.6946 0.8334
No log 10.9091 240 0.6808 0.5139 0.6808 0.8251
No log 11.0 242 0.6435 0.5568 0.6435 0.8022
No log 11.0909 244 0.6000 0.6099 0.6000 0.7746
No log 11.1818 246 0.6166 0.4752 0.6166 0.7852
No log 11.2727 248 0.6222 0.4687 0.6222 0.7888
No log 11.3636 250 0.6081 0.4735 0.6081 0.7798
No log 11.4545 252 0.5607 0.5700 0.5607 0.7488
No log 11.5455 254 0.5646 0.5525 0.5646 0.7514
No log 11.6364 256 0.6002 0.5763 0.6002 0.7747
No log 11.7273 258 0.6733 0.5753 0.6733 0.8205
No log 11.8182 260 0.6699 0.5775 0.6699 0.8185
No log 11.9091 262 0.5880 0.5019 0.5880 0.7668
No log 12.0 264 0.5843 0.5121 0.5843 0.7644
No log 12.0909 266 0.6040 0.5077 0.6040 0.7772
No log 12.1818 268 0.6133 0.4956 0.6133 0.7831
No log 12.2727 270 0.6544 0.5354 0.6544 0.8089
No log 12.3636 272 0.6838 0.5656 0.6838 0.8269
No log 12.4545 274 0.7079 0.5411 0.7079 0.8414
No log 12.5455 276 0.7326 0.5266 0.7326 0.8559
No log 12.6364 278 0.7464 0.5266 0.7464 0.8640
No log 12.7273 280 0.6757 0.5393 0.6757 0.8220
No log 12.8182 282 0.6755 0.4877 0.6755 0.8219
No log 12.9091 284 0.6310 0.5049 0.6310 0.7944
No log 13.0 286 0.5879 0.5054 0.5879 0.7667
No log 13.0909 288 0.6253 0.5079 0.6253 0.7907
No log 13.1818 290 0.6147 0.5212 0.6147 0.7840
No log 13.2727 292 0.6433 0.5331 0.6433 0.8021
No log 13.3636 294 0.6557 0.5281 0.6557 0.8098
No log 13.4545 296 0.6136 0.5360 0.6136 0.7833
No log 13.5455 298 0.6470 0.4562 0.6470 0.8044
No log 13.6364 300 0.6222 0.5155 0.6222 0.7888
No log 13.7273 302 0.6224 0.5463 0.6224 0.7889
No log 13.8182 304 0.6831 0.5348 0.6831 0.8265
No log 13.9091 306 0.7109 0.4983 0.7109 0.8432
No log 14.0 308 0.8132 0.5264 0.8132 0.9018
No log 14.0909 310 0.8106 0.5281 0.8106 0.9003
No log 14.1818 312 0.8155 0.5224 0.8155 0.9030
No log 14.2727 314 0.8112 0.5047 0.8112 0.9007
No log 14.3636 316 0.7349 0.5357 0.7349 0.8573
No log 14.4545 318 0.6635 0.5399 0.6635 0.8146
No log 14.5455 320 0.6214 0.5192 0.6214 0.7883
No log 14.6364 322 0.5802 0.4865 0.5802 0.7617
No log 14.7273 324 0.5878 0.4999 0.5878 0.7667
No log 14.8182 326 0.6374 0.4750 0.6374 0.7984
No log 14.9091 328 0.6116 0.5086 0.6116 0.7820
No log 15.0 330 0.5774 0.5774 0.5774 0.7599
No log 15.0909 332 0.5823 0.5630 0.5823 0.7631
No log 15.1818 334 0.5991 0.5491 0.5991 0.7740
No log 15.2727 336 0.6026 0.5310 0.6026 0.7763
No log 15.3636 338 0.6061 0.5100 0.6061 0.7785
No log 15.4545 340 0.6039 0.5768 0.6039 0.7771
No log 15.5455 342 0.5981 0.5768 0.5981 0.7734
No log 15.6364 344 0.5939 0.5581 0.5939 0.7707
No log 15.7273 346 0.5895 0.5097 0.5895 0.7678
No log 15.8182 348 0.6569 0.4596 0.6569 0.8105
No log 15.9091 350 0.7011 0.4467 0.7011 0.8373
No log 16.0 352 0.6249 0.4118 0.6249 0.7905
No log 16.0909 354 0.6025 0.5829 0.6025 0.7762
No log 16.1818 356 0.6402 0.5698 0.6402 0.8001
No log 16.2727 358 0.6214 0.5536 0.6214 0.7883
No log 16.3636 360 0.6119 0.5828 0.6119 0.7822
No log 16.4545 362 0.6061 0.5514 0.6061 0.7785
No log 16.5455 364 0.6042 0.5487 0.6042 0.7773
No log 16.6364 366 0.6121 0.5191 0.6121 0.7824
No log 16.7273 368 0.6307 0.4672 0.6307 0.7942
No log 16.8182 370 0.6117 0.4934 0.6117 0.7821
No log 16.9091 372 0.6233 0.5046 0.6233 0.7895
No log 17.0 374 0.6423 0.5361 0.6423 0.8014
No log 17.0909 376 0.6115 0.5180 0.6115 0.7820
No log 17.1818 378 0.5932 0.4978 0.5932 0.7702
No log 17.2727 380 0.5943 0.4733 0.5943 0.7709
No log 17.3636 382 0.5906 0.4686 0.5906 0.7685
No log 17.4545 384 0.5974 0.5132 0.5974 0.7729
No log 17.5455 386 0.6258 0.5490 0.6258 0.7911
No log 17.6364 388 0.6177 0.5638 0.6177 0.7860
No log 17.7273 390 0.5997 0.4885 0.5997 0.7744
No log 17.8182 392 0.6634 0.4593 0.6634 0.8145
No log 17.9091 394 0.6374 0.4895 0.6374 0.7983
No log 18.0 396 0.5936 0.5303 0.5936 0.7705
No log 18.0909 398 0.6190 0.5291 0.6190 0.7867
No log 18.1818 400 0.5982 0.5314 0.5982 0.7734
No log 18.2727 402 0.5839 0.5270 0.5839 0.7641
No log 18.3636 404 0.5840 0.5200 0.5840 0.7642
No log 18.4545 406 0.6128 0.5291 0.6128 0.7828
No log 18.5455 408 0.6391 0.5301 0.6391 0.7994
No log 18.6364 410 0.6248 0.5260 0.6248 0.7905
No log 18.7273 412 0.6061 0.5223 0.6061 0.7785
No log 18.8182 414 0.6178 0.5208 0.6178 0.7860
No log 18.9091 416 0.6389 0.5228 0.6389 0.7993
No log 19.0 418 0.6360 0.5523 0.6360 0.7975
No log 19.0909 420 0.6266 0.5758 0.6266 0.7916
No log 19.1818 422 0.6141 0.6061 0.6141 0.7836
No log 19.2727 424 0.6449 0.5323 0.6449 0.8031
No log 19.3636 426 0.7252 0.5203 0.7252 0.8516
No log 19.4545 428 0.6828 0.5255 0.6828 0.8263
No log 19.5455 430 0.5977 0.5641 0.5977 0.7731
No log 19.6364 432 0.5565 0.5570 0.5565 0.7460
No log 19.7273 434 0.5654 0.6175 0.5654 0.7519
No log 19.8182 436 0.5554 0.5818 0.5554 0.7453
No log 19.9091 438 0.5352 0.5562 0.5352 0.7316
No log 20.0 440 0.5657 0.4958 0.5657 0.7521
No log 20.0909 442 0.5662 0.4867 0.5662 0.7524
No log 20.1818 444 0.5347 0.5535 0.5347 0.7312
No log 20.2727 446 0.5350 0.5589 0.5350 0.7315
No log 20.3636 448 0.5352 0.5453 0.5352 0.7316
No log 20.4545 450 0.5375 0.5543 0.5375 0.7332
No log 20.5455 452 0.5444 0.5676 0.5444 0.7378
No log 20.6364 454 0.5717 0.6178 0.5717 0.7561
No log 20.7273 456 0.5784 0.5864 0.5784 0.7605
No log 20.8182 458 0.5722 0.5709 0.5722 0.7564
No log 20.9091 460 0.6167 0.5341 0.6167 0.7853
No log 21.0 462 0.5992 0.5503 0.5992 0.7741
No log 21.0909 464 0.5725 0.5087 0.5725 0.7566
No log 21.1818 466 0.5554 0.5020 0.5554 0.7453
No log 21.2727 468 0.5592 0.4928 0.5592 0.7478
No log 21.3636 470 0.5891 0.4863 0.5891 0.7675
No log 21.4545 472 0.6507 0.4865 0.6507 0.8067
No log 21.5455 474 0.6651 0.5132 0.6651 0.8155
No log 21.6364 476 0.6183 0.4811 0.6183 0.7863
No log 21.7273 478 0.5829 0.5549 0.5829 0.7635
No log 21.8182 480 0.6007 0.5189 0.6007 0.7751
No log 21.9091 482 0.5915 0.5553 0.5915 0.7691
No log 22.0 484 0.5931 0.4946 0.5931 0.7701
No log 22.0909 486 0.5943 0.5091 0.5943 0.7709
No log 22.1818 488 0.6062 0.5314 0.6062 0.7786
No log 22.2727 490 0.5855 0.5155 0.5855 0.7652
No log 22.3636 492 0.5791 0.5081 0.5791 0.7610
No log 22.4545 494 0.5685 0.5339 0.5685 0.7540
No log 22.5455 496 0.5644 0.5155 0.5644 0.7512
No log 22.6364 498 0.5690 0.5502 0.5690 0.7544
0.3089 22.7273 500 0.5745 0.5685 0.5745 0.7579
0.3089 22.8182 502 0.5815 0.5549 0.5815 0.7625
0.3089 22.9091 504 0.5910 0.5373 0.5910 0.7688
0.3089 23.0 506 0.5781 0.5402 0.5781 0.7603
0.3089 23.0909 508 0.5665 0.5440 0.5665 0.7526
0.3089 23.1818 510 0.5702 0.5276 0.5702 0.7551

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
4
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for MayBashendy/ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run3_AugV5_k4_task2_organization

Finetuned
(4222)
this model