ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run2_AugV5_k3_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6265
  • Qwk: 0.5381
  • Mse: 0.6265
  • Rmse: 0.7915

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.125 2 4.2936 -0.0141 4.2936 2.0721
No log 0.25 4 2.7849 0.0595 2.7849 1.6688
No log 0.375 6 1.3792 0.0457 1.3792 1.1744
No log 0.5 8 1.3472 -0.0430 1.3472 1.1607
No log 0.625 10 1.1281 -0.1064 1.1281 1.0621
No log 0.75 12 0.9108 0.0102 0.9108 0.9544
No log 0.875 14 0.8925 0.0443 0.8925 0.9447
No log 1.0 16 0.8855 0.1275 0.8855 0.9410
No log 1.125 18 0.8572 0.2027 0.8572 0.9258
No log 1.25 20 0.9093 0.1080 0.9093 0.9536
No log 1.375 22 1.0446 -0.0175 1.0446 1.0221
No log 1.5 24 1.0544 0.0305 1.0544 1.0268
No log 1.625 26 0.9932 0.0292 0.9932 0.9966
No log 1.75 28 0.9224 0.1311 0.9224 0.9604
No log 1.875 30 0.9192 0.1476 0.9192 0.9588
No log 2.0 32 0.8353 0.2520 0.8353 0.9139
No log 2.125 34 0.7843 0.2398 0.7843 0.8856
No log 2.25 36 0.7696 0.3063 0.7696 0.8773
No log 2.375 38 0.7714 0.3108 0.7714 0.8783
No log 2.5 40 0.8174 0.2284 0.8174 0.9041
No log 2.625 42 1.0797 0.0848 1.0797 1.0391
No log 2.75 44 1.3328 0.1108 1.3328 1.1545
No log 2.875 46 1.2924 0.1357 1.2924 1.1369
No log 3.0 48 1.0534 0.1497 1.0534 1.0263
No log 3.125 50 0.8004 0.2545 0.8004 0.8946
No log 3.25 52 0.6934 0.3903 0.6934 0.8327
No log 3.375 54 0.6609 0.3779 0.6609 0.8130
No log 3.5 56 0.6483 0.3562 0.6483 0.8052
No log 3.625 58 0.6424 0.4162 0.6424 0.8015
No log 3.75 60 0.7253 0.3408 0.7253 0.8517
No log 3.875 62 0.9169 0.1623 0.9169 0.9576
No log 4.0 64 0.8901 0.1848 0.8901 0.9434
No log 4.125 66 0.9127 0.1566 0.9127 0.9553
No log 4.25 68 0.8164 0.2670 0.8164 0.9035
No log 4.375 70 0.6892 0.3538 0.6892 0.8302
No log 4.5 72 0.6066 0.4230 0.6066 0.7789
No log 4.625 74 0.6295 0.4143 0.6295 0.7934
No log 4.75 76 0.6350 0.5214 0.6350 0.7969
No log 4.875 78 0.5877 0.4794 0.5877 0.7666
No log 5.0 80 0.6094 0.4521 0.6094 0.7806
No log 5.125 82 0.7718 0.3457 0.7718 0.8785
No log 5.25 84 0.8591 0.3615 0.8591 0.9269
No log 5.375 86 0.7594 0.3717 0.7594 0.8714
No log 5.5 88 0.6352 0.4300 0.6352 0.7970
No log 5.625 90 0.5896 0.4463 0.5896 0.7678
No log 5.75 92 0.5762 0.4375 0.5762 0.7591
No log 5.875 94 0.5787 0.4304 0.5787 0.7607
No log 6.0 96 0.5785 0.4783 0.5785 0.7606
No log 6.125 98 0.5822 0.5026 0.5822 0.7631
No log 6.25 100 0.5910 0.5651 0.5910 0.7687
No log 6.375 102 0.6013 0.5282 0.6013 0.7754
No log 6.5 104 0.6085 0.5409 0.6085 0.7801
No log 6.625 106 0.6104 0.5163 0.6104 0.7813
No log 6.75 108 0.6515 0.4895 0.6515 0.8071
No log 6.875 110 0.7217 0.4874 0.7217 0.8495
No log 7.0 112 0.7095 0.4737 0.7095 0.8423
No log 7.125 114 0.5978 0.5206 0.5978 0.7731
No log 7.25 116 0.6141 0.5179 0.6141 0.7836
No log 7.375 118 0.5890 0.5093 0.5890 0.7675
No log 7.5 120 0.5804 0.4888 0.5804 0.7619
No log 7.625 122 0.6542 0.4931 0.6542 0.8088
No log 7.75 124 0.7152 0.4573 0.7152 0.8457
No log 7.875 126 0.6887 0.4884 0.6887 0.8299
No log 8.0 128 0.6496 0.5060 0.6496 0.8060
No log 8.125 130 0.6323 0.5385 0.6323 0.7952
No log 8.25 132 0.6421 0.4815 0.6421 0.8013
No log 8.375 134 0.6343 0.5206 0.6343 0.7964
No log 8.5 136 0.6338 0.5152 0.6338 0.7961
No log 8.625 138 0.7617 0.4360 0.7617 0.8727
No log 8.75 140 0.8041 0.4191 0.8041 0.8967
No log 8.875 142 0.6731 0.5013 0.6731 0.8204
No log 9.0 144 0.6196 0.5065 0.6196 0.7871
No log 9.125 146 0.7285 0.4552 0.7285 0.8535
No log 9.25 148 0.7341 0.4628 0.7341 0.8568
No log 9.375 150 0.6162 0.5415 0.6162 0.7850
No log 9.5 152 0.7041 0.5163 0.7041 0.8391
No log 9.625 154 0.8551 0.4455 0.8551 0.9247
No log 9.75 156 0.8165 0.4966 0.8165 0.9036
No log 9.875 158 0.6295 0.5529 0.6295 0.7934
No log 10.0 160 0.6512 0.4818 0.6512 0.8070
No log 10.125 162 0.6673 0.5044 0.6673 0.8169
No log 10.25 164 0.6256 0.5234 0.6256 0.7910
No log 10.375 166 0.6133 0.5482 0.6133 0.7832
No log 10.5 168 0.5476 0.5462 0.5476 0.7400
No log 10.625 170 0.5438 0.5569 0.5438 0.7374
No log 10.75 172 0.5521 0.5321 0.5521 0.7431
No log 10.875 174 0.5484 0.5464 0.5484 0.7405
No log 11.0 176 0.5677 0.5307 0.5677 0.7535
No log 11.125 178 0.6532 0.4682 0.6532 0.8082
No log 11.25 180 0.6378 0.5500 0.6378 0.7986
No log 11.375 182 0.5649 0.5758 0.5649 0.7516
No log 11.5 184 0.6620 0.5525 0.6620 0.8136
No log 11.625 186 0.7694 0.5082 0.7694 0.8771
No log 11.75 188 0.6869 0.5692 0.6869 0.8288
No log 11.875 190 0.5907 0.5753 0.5907 0.7685
No log 12.0 192 0.7168 0.5480 0.7168 0.8467
No log 12.125 194 0.8699 0.4816 0.8699 0.9327
No log 12.25 196 0.8392 0.4879 0.8392 0.9161
No log 12.375 198 0.6605 0.5617 0.6605 0.8127
No log 12.5 200 0.5782 0.5881 0.5782 0.7604
No log 12.625 202 0.6746 0.5500 0.6746 0.8214
No log 12.75 204 0.6571 0.5500 0.6571 0.8106
No log 12.875 206 0.5823 0.5747 0.5823 0.7631
No log 13.0 208 0.5761 0.5687 0.5761 0.7590
No log 13.125 210 0.6684 0.5460 0.6684 0.8176
No log 13.25 212 0.6997 0.5424 0.6997 0.8365
No log 13.375 214 0.6041 0.5217 0.6041 0.7772
No log 13.5 216 0.5767 0.6109 0.5767 0.7594
No log 13.625 218 0.7210 0.5361 0.7210 0.8491
No log 13.75 220 0.7720 0.4613 0.7720 0.8786
No log 13.875 222 0.6857 0.5688 0.6857 0.8281
No log 14.0 224 0.5659 0.5871 0.5659 0.7522
No log 14.125 226 0.5414 0.4926 0.5414 0.7358
No log 14.25 228 0.6128 0.5233 0.6128 0.7828
No log 14.375 230 0.6590 0.4887 0.6590 0.8118
No log 14.5 232 0.5945 0.5344 0.5945 0.7711
No log 14.625 234 0.5433 0.6176 0.5433 0.7371
No log 14.75 236 0.6583 0.5328 0.6583 0.8113
No log 14.875 238 0.8825 0.4803 0.8825 0.9394
No log 15.0 240 0.9697 0.4367 0.9697 0.9847
No log 15.125 242 0.8655 0.4709 0.8655 0.9303
No log 15.25 244 0.6875 0.5305 0.6875 0.8292
No log 15.375 246 0.6205 0.5613 0.6205 0.7877
No log 15.5 248 0.6734 0.5785 0.6734 0.8206
No log 15.625 250 0.7280 0.5334 0.7280 0.8532
No log 15.75 252 0.6936 0.5206 0.6936 0.8328
No log 15.875 254 0.6140 0.5731 0.6140 0.7836
No log 16.0 256 0.5703 0.5757 0.5703 0.7552
No log 16.125 258 0.5742 0.5866 0.5742 0.7577
No log 16.25 260 0.5739 0.5666 0.5739 0.7575
No log 16.375 262 0.5585 0.5469 0.5585 0.7473
No log 16.5 264 0.5553 0.5814 0.5553 0.7452
No log 16.625 266 0.6664 0.4896 0.6664 0.8163
No log 16.75 268 0.7515 0.4178 0.7515 0.8669
No log 16.875 270 0.6855 0.4554 0.6855 0.8279
No log 17.0 272 0.5838 0.5228 0.5838 0.7641
No log 17.125 274 0.5582 0.5785 0.5582 0.7471
No log 17.25 276 0.5901 0.5374 0.5901 0.7682
No log 17.375 278 0.6448 0.5462 0.6448 0.8030
No log 17.5 280 0.6398 0.5263 0.6398 0.7999
No log 17.625 282 0.5962 0.5943 0.5962 0.7721
No log 17.75 284 0.5905 0.6138 0.5905 0.7684
No log 17.875 286 0.6006 0.5919 0.6006 0.7750
No log 18.0 288 0.5925 0.5843 0.5925 0.7697
No log 18.125 290 0.5664 0.5970 0.5664 0.7526
No log 18.25 292 0.5528 0.6086 0.5528 0.7435
No log 18.375 294 0.5477 0.5919 0.5477 0.7401
No log 18.5 296 0.5569 0.5323 0.5569 0.7462
No log 18.625 298 0.5816 0.5344 0.5816 0.7626
No log 18.75 300 0.6248 0.5510 0.6248 0.7904
No log 18.875 302 0.6372 0.4994 0.6372 0.7983
No log 19.0 304 0.5736 0.5617 0.5736 0.7574
No log 19.125 306 0.5420 0.5841 0.5420 0.7362
No log 19.25 308 0.5915 0.5304 0.5915 0.7691
No log 19.375 310 0.6455 0.5605 0.6455 0.8034
No log 19.5 312 0.6304 0.5643 0.6304 0.7940
No log 19.625 314 0.5778 0.5577 0.5778 0.7601
No log 19.75 316 0.5700 0.5934 0.5700 0.7550
No log 19.875 318 0.5638 0.6179 0.5638 0.7509
No log 20.0 320 0.5512 0.6261 0.5512 0.7424
No log 20.125 322 0.5344 0.6050 0.5344 0.7310
No log 20.25 324 0.5319 0.5784 0.5319 0.7293
No log 20.375 326 0.5426 0.5438 0.5426 0.7366
No log 20.5 328 0.5493 0.4603 0.5493 0.7411
No log 20.625 330 0.5380 0.5322 0.5380 0.7335
No log 20.75 332 0.5319 0.6091 0.5319 0.7293
No log 20.875 334 0.5481 0.6304 0.5481 0.7403
No log 21.0 336 0.5725 0.5554 0.5725 0.7567
No log 21.125 338 0.5707 0.5719 0.5707 0.7554
No log 21.25 340 0.5391 0.6018 0.5391 0.7343
No log 21.375 342 0.5207 0.5896 0.5207 0.7216
No log 21.5 344 0.5560 0.5425 0.5560 0.7456
No log 21.625 346 0.5725 0.5304 0.5725 0.7566
No log 21.75 348 0.5603 0.5425 0.5603 0.7485
No log 21.875 350 0.5333 0.5848 0.5333 0.7302
No log 22.0 352 0.5327 0.5767 0.5327 0.7298
No log 22.125 354 0.5450 0.5549 0.5450 0.7383
No log 22.25 356 0.5559 0.5426 0.5559 0.7456
No log 22.375 358 0.5867 0.5137 0.5867 0.7660
No log 22.5 360 0.5785 0.5320 0.5785 0.7606
No log 22.625 362 0.5675 0.5477 0.5675 0.7533
No log 22.75 364 0.5907 0.5807 0.5907 0.7686
No log 22.875 366 0.5941 0.6020 0.5941 0.7708
No log 23.0 368 0.6067 0.6039 0.6067 0.7789
No log 23.125 370 0.6023 0.6109 0.6023 0.7761
No log 23.25 372 0.6025 0.5782 0.6025 0.7762
No log 23.375 374 0.5938 0.5550 0.5938 0.7706
No log 23.5 376 0.5867 0.5818 0.5867 0.7659
No log 23.625 378 0.5644 0.5727 0.5644 0.7512
No log 23.75 380 0.5465 0.6090 0.5465 0.7393
No log 23.875 382 0.5477 0.6060 0.5477 0.7401
No log 24.0 384 0.5634 0.5844 0.5634 0.7506
No log 24.125 386 0.5897 0.5736 0.5897 0.7679
No log 24.25 388 0.6193 0.5324 0.6193 0.7870
No log 24.375 390 0.5950 0.5436 0.5950 0.7714
No log 24.5 392 0.5771 0.5464 0.5771 0.7597
No log 24.625 394 0.5494 0.5887 0.5494 0.7412
No log 24.75 396 0.5545 0.6044 0.5545 0.7447
No log 24.875 398 0.5996 0.5445 0.5996 0.7744
No log 25.0 400 0.6070 0.5428 0.6070 0.7791
No log 25.125 402 0.5879 0.5622 0.5879 0.7667
No log 25.25 404 0.5787 0.5273 0.5787 0.7607
No log 25.375 406 0.5914 0.5557 0.5914 0.7690
No log 25.5 408 0.5920 0.5434 0.5920 0.7694
No log 25.625 410 0.5740 0.5471 0.5740 0.7576
No log 25.75 412 0.5539 0.5919 0.5539 0.7442
No log 25.875 414 0.5499 0.6140 0.5499 0.7415
No log 26.0 416 0.5483 0.6368 0.5483 0.7405
No log 26.125 418 0.5396 0.6155 0.5396 0.7346
No log 26.25 420 0.5433 0.5998 0.5433 0.7371
No log 26.375 422 0.5369 0.6155 0.5369 0.7328
No log 26.5 424 0.5298 0.6070 0.5298 0.7279
No log 26.625 426 0.5226 0.5577 0.5226 0.7229
No log 26.75 428 0.5284 0.5905 0.5284 0.7269
No log 26.875 430 0.5333 0.5435 0.5333 0.7303
No log 27.0 432 0.5529 0.5036 0.5529 0.7436
No log 27.125 434 0.5919 0.5435 0.5919 0.7693
No log 27.25 436 0.5955 0.5435 0.5955 0.7717
No log 27.375 438 0.5857 0.5504 0.5857 0.7653
No log 27.5 440 0.5627 0.5385 0.5627 0.7501
No log 27.625 442 0.5695 0.5422 0.5695 0.7547
No log 27.75 444 0.5842 0.5403 0.5842 0.7643
No log 27.875 446 0.5679 0.5467 0.5679 0.7536
No log 28.0 448 0.5689 0.5677 0.5689 0.7543
No log 28.125 450 0.5807 0.5875 0.5807 0.7620
No log 28.25 452 0.5719 0.6104 0.5719 0.7563
No log 28.375 454 0.5654 0.5813 0.5654 0.7520
No log 28.5 456 0.5581 0.5665 0.5581 0.7471
No log 28.625 458 0.5592 0.5554 0.5592 0.7478
No log 28.75 460 0.5506 0.5554 0.5506 0.7420
No log 28.875 462 0.5413 0.5910 0.5413 0.7357
No log 29.0 464 0.5366 0.5952 0.5366 0.7325
No log 29.125 466 0.5654 0.5195 0.5654 0.7519
No log 29.25 468 0.5927 0.5437 0.5927 0.7699
No log 29.375 470 0.5784 0.5558 0.5784 0.7605
No log 29.5 472 0.5601 0.5864 0.5601 0.7484
No log 29.625 474 0.5434 0.6219 0.5434 0.7371
No log 29.75 476 0.5748 0.6085 0.5748 0.7581
No log 29.875 478 0.5876 0.6061 0.5876 0.7665
No log 30.0 480 0.5714 0.5898 0.5714 0.7559
No log 30.125 482 0.5743 0.6037 0.5743 0.7578
No log 30.25 484 0.6020 0.5696 0.6020 0.7759
No log 30.375 486 0.5831 0.5986 0.5831 0.7636
No log 30.5 488 0.5657 0.5640 0.5657 0.7521
No log 30.625 490 0.5821 0.6050 0.5821 0.7630
No log 30.75 492 0.6042 0.5434 0.6042 0.7773
No log 30.875 494 0.6049 0.5466 0.6049 0.7778
No log 31.0 496 0.5955 0.5180 0.5955 0.7717
No log 31.125 498 0.5818 0.5944 0.5818 0.7628
0.3166 31.25 500 0.5906 0.5762 0.5906 0.7685
0.3166 31.375 502 0.6130 0.5805 0.6130 0.7829
0.3166 31.5 504 0.6271 0.5565 0.6271 0.7919
0.3166 31.625 506 0.6385 0.5565 0.6385 0.7991
0.3166 31.75 508 0.6477 0.5449 0.6477 0.8048
0.3166 31.875 510 0.6373 0.5449 0.6373 0.7983
0.3166 32.0 512 0.6265 0.5381 0.6265 0.7915

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
4
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for MayBashendy/ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run2_AugV5_k3_task2_organization

Finetuned
(4222)
this model