ArabicNewSplits8_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k5_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6308
  • Qwk: 0.4938
  • Mse: 0.6308
  • Rmse: 0.7942

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0714 2 4.3508 -0.0169 4.3508 2.0859
No log 0.1429 4 3.0144 0.0376 3.0144 1.7362
No log 0.2143 6 1.4275 0.0784 1.4275 1.1948
No log 0.2857 8 1.0490 -0.0310 1.0490 1.0242
No log 0.3571 10 0.9262 0.0021 0.9262 0.9624
No log 0.4286 12 0.8643 0.1008 0.8643 0.9297
No log 0.5 14 0.8433 0.0655 0.8433 0.9183
No log 0.5714 16 0.8327 0.0726 0.8327 0.9125
No log 0.6429 18 0.8394 0.2124 0.8394 0.9162
No log 0.7143 20 1.1790 -0.0226 1.1790 1.0858
No log 0.7857 22 1.1387 -0.0014 1.1387 1.0671
No log 0.8571 24 0.9481 0.0011 0.9481 0.9737
No log 0.9286 26 0.8794 0.1898 0.8794 0.9378
No log 1.0 28 0.8852 0.1655 0.8852 0.9408
No log 1.0714 30 0.8513 0.2072 0.8513 0.9227
No log 1.1429 32 0.8184 0.2714 0.8184 0.9046
No log 1.2143 34 0.8513 0.1785 0.8513 0.9227
No log 1.2857 36 0.8873 0.1301 0.8873 0.9420
No log 1.3571 38 1.1442 0.1082 1.1442 1.0697
No log 1.4286 40 1.2651 0.1354 1.2651 1.1248
No log 1.5 42 0.9872 0.2271 0.9872 0.9936
No log 1.5714 44 0.6930 0.3452 0.6930 0.8325
No log 1.6429 46 0.6779 0.3662 0.6779 0.8233
No log 1.7143 48 0.6465 0.3865 0.6465 0.8040
No log 1.7857 50 0.6194 0.3855 0.6194 0.7870
No log 1.8571 52 0.7186 0.3840 0.7186 0.8477
No log 1.9286 54 0.6603 0.3631 0.6603 0.8126
No log 2.0 56 0.5901 0.4074 0.5901 0.7682
No log 2.0714 58 0.6159 0.4431 0.6159 0.7848
No log 2.1429 60 0.5979 0.4235 0.5979 0.7733
No log 2.2143 62 0.5915 0.4601 0.5915 0.7691
No log 2.2857 64 0.6742 0.4274 0.6742 0.8211
No log 2.3571 66 0.7193 0.4455 0.7193 0.8481
No log 2.4286 68 0.5928 0.4649 0.5928 0.7699
No log 2.5 70 0.7138 0.3915 0.7138 0.8449
No log 2.5714 72 0.6915 0.3850 0.6915 0.8316
No log 2.6429 74 0.6146 0.3818 0.6146 0.7840
No log 2.7143 76 0.7550 0.3769 0.7550 0.8689
No log 2.7857 78 0.8667 0.3638 0.8667 0.9309
No log 2.8571 80 0.7763 0.3681 0.7763 0.8811
No log 2.9286 82 0.6221 0.4650 0.6221 0.7888
No log 3.0 84 0.6019 0.5422 0.6019 0.7758
No log 3.0714 86 0.6058 0.5105 0.6058 0.7784
No log 3.1429 88 0.6264 0.4979 0.6264 0.7915
No log 3.2143 90 0.6674 0.4806 0.6674 0.8170
No log 3.2857 92 0.8318 0.4604 0.8318 0.9120
No log 3.3571 94 0.6986 0.5025 0.6986 0.8358
No log 3.4286 96 0.6057 0.5593 0.6057 0.7783
No log 3.5 98 0.6590 0.4898 0.6590 0.8118
No log 3.5714 100 0.6523 0.4353 0.6523 0.8077
No log 3.6429 102 0.6477 0.4390 0.6477 0.8048
No log 3.7143 104 0.5979 0.5079 0.5979 0.7733
No log 3.7857 106 0.6442 0.4341 0.6442 0.8026
No log 3.8571 108 0.7339 0.4488 0.7339 0.8567
No log 3.9286 110 0.9763 0.3684 0.9763 0.9881
No log 4.0 112 0.7572 0.4543 0.7572 0.8702
No log 4.0714 114 0.6089 0.5216 0.6089 0.7803
No log 4.1429 116 0.6109 0.5276 0.6109 0.7816
No log 4.2143 118 0.6321 0.5513 0.6321 0.7951
No log 4.2857 120 0.6610 0.5208 0.6610 0.8130
No log 4.3571 122 0.6937 0.5151 0.6937 0.8329
No log 4.4286 124 0.7016 0.5135 0.7016 0.8376
No log 4.5 126 0.7004 0.5186 0.7004 0.8369
No log 4.5714 128 0.6984 0.5231 0.6984 0.8357
No log 4.6429 130 0.8647 0.3797 0.8647 0.9299
No log 4.7143 132 0.7477 0.4798 0.7477 0.8647
No log 4.7857 134 0.6257 0.4930 0.6257 0.7910
No log 4.8571 136 0.6951 0.4687 0.6951 0.8337
No log 4.9286 138 0.6505 0.5199 0.6505 0.8065
No log 5.0 140 0.6424 0.5075 0.6424 0.8015
No log 5.0714 142 0.6550 0.4886 0.6550 0.8093
No log 5.1429 144 0.6461 0.4891 0.6461 0.8038
No log 5.2143 146 0.6384 0.5909 0.6384 0.7990
No log 5.2857 148 0.6332 0.5219 0.6332 0.7957
No log 5.3571 150 0.6291 0.5727 0.6291 0.7932
No log 5.4286 152 0.6333 0.5642 0.6333 0.7958
No log 5.5 154 0.6063 0.4837 0.6063 0.7787
No log 5.5714 156 0.6456 0.4610 0.6456 0.8035
No log 5.6429 158 0.6533 0.3762 0.6533 0.8083
No log 5.7143 160 0.6067 0.4434 0.6067 0.7789
No log 5.7857 162 0.5930 0.5276 0.5930 0.7701
No log 5.8571 164 0.6147 0.5880 0.6147 0.7840
No log 5.9286 166 0.6456 0.5531 0.6456 0.8035
No log 6.0 168 0.6587 0.5645 0.6587 0.8116
No log 6.0714 170 0.6764 0.5430 0.6764 0.8224
No log 6.1429 172 0.6546 0.5574 0.6546 0.8090
No log 6.2143 174 0.6438 0.5651 0.6438 0.8024
No log 6.2857 176 0.6261 0.5790 0.6261 0.7912
No log 6.3571 178 0.6107 0.4930 0.6107 0.7814
No log 6.4286 180 0.6009 0.4465 0.6009 0.7752
No log 6.5 182 0.5889 0.4888 0.5889 0.7674
No log 6.5714 184 0.5775 0.5835 0.5775 0.7599
No log 6.6429 186 0.5831 0.5902 0.5831 0.7636
No log 6.7143 188 0.6257 0.5839 0.6257 0.7910
No log 6.7857 190 0.6420 0.5823 0.6420 0.8013
No log 6.8571 192 0.7948 0.4467 0.7948 0.8915
No log 6.9286 194 0.8845 0.4052 0.8845 0.9405
No log 7.0 196 0.8641 0.3904 0.8641 0.9296
No log 7.0714 198 0.9385 0.3617 0.9385 0.9688
No log 7.1429 200 0.9563 0.3666 0.9563 0.9779
No log 7.2143 202 0.7677 0.4415 0.7677 0.8762
No log 7.2857 204 0.7215 0.4667 0.7215 0.8494
No log 7.3571 206 0.7753 0.4887 0.7753 0.8805
No log 7.4286 208 0.7221 0.4623 0.7221 0.8498
No log 7.5 210 0.7492 0.4771 0.7492 0.8655
No log 7.5714 212 0.7824 0.4012 0.7824 0.8846
No log 7.6429 214 0.7691 0.4225 0.7691 0.8770
No log 7.7143 216 0.6476 0.5386 0.6476 0.8047
No log 7.7857 218 0.6304 0.6055 0.6304 0.7940
No log 7.8571 220 0.6317 0.5730 0.6317 0.7948
No log 7.9286 222 0.6221 0.5790 0.6221 0.7887
No log 8.0 224 0.6152 0.6010 0.6152 0.7843
No log 8.0714 226 0.6132 0.5924 0.6132 0.7830
No log 8.1429 228 0.6403 0.5042 0.6403 0.8002
No log 8.2143 230 0.6364 0.5507 0.6364 0.7978
No log 8.2857 232 0.6480 0.6006 0.6480 0.8050
No log 8.3571 234 0.7961 0.4113 0.7961 0.8922
No log 8.4286 236 0.7349 0.4233 0.7349 0.8573
No log 8.5 238 0.6041 0.6086 0.6041 0.7772
No log 8.5714 240 0.6599 0.4503 0.6599 0.8123
No log 8.6429 242 0.6617 0.4717 0.6617 0.8134
No log 8.7143 244 0.6068 0.5823 0.6068 0.7790
No log 8.7857 246 0.6084 0.6274 0.6084 0.7800
No log 8.8571 248 0.6081 0.5922 0.6081 0.7798
No log 8.9286 250 0.6185 0.5979 0.6185 0.7864
No log 9.0 252 0.6531 0.5433 0.6531 0.8081
No log 9.0714 254 0.6306 0.5779 0.6306 0.7941
No log 9.1429 256 0.5789 0.6184 0.5789 0.7609
No log 9.2143 258 0.5612 0.5468 0.5612 0.7491
No log 9.2857 260 0.5579 0.5648 0.5579 0.7469
No log 9.3571 262 0.5892 0.5786 0.5892 0.7676
No log 9.4286 264 0.6987 0.4994 0.6987 0.8359
No log 9.5 266 0.6823 0.5377 0.6823 0.8260
No log 9.5714 268 0.6092 0.6101 0.6092 0.7805
No log 9.6429 270 0.5996 0.6066 0.5996 0.7743
No log 9.7143 272 0.5741 0.6065 0.5741 0.7577
No log 9.7857 274 0.5708 0.5841 0.5708 0.7555
No log 9.8571 276 0.5793 0.6308 0.5793 0.7611
No log 9.9286 278 0.6080 0.5982 0.6080 0.7797
No log 10.0 280 0.6074 0.6006 0.6074 0.7794
No log 10.0714 282 0.5988 0.6354 0.5988 0.7738
No log 10.1429 284 0.6082 0.6098 0.6082 0.7799
No log 10.2143 286 0.6394 0.5387 0.6394 0.7996
No log 10.2857 288 0.6898 0.5109 0.6898 0.8306
No log 10.3571 290 0.5975 0.5234 0.5975 0.7730
No log 10.4286 292 0.5911 0.5685 0.5911 0.7688
No log 10.5 294 0.5969 0.5602 0.5969 0.7726
No log 10.5714 296 0.5736 0.5667 0.5736 0.7573
No log 10.6429 298 0.5980 0.5579 0.5980 0.7733
No log 10.7143 300 0.6000 0.5495 0.6000 0.7746
No log 10.7857 302 0.6059 0.5725 0.6059 0.7784
No log 10.8571 304 0.6033 0.5356 0.6033 0.7767
No log 10.9286 306 0.6242 0.5035 0.6242 0.7901
No log 11.0 308 0.6215 0.5810 0.6215 0.7884
No log 11.0714 310 0.6348 0.5387 0.6348 0.7967
No log 11.1429 312 0.6691 0.4939 0.6691 0.8180
No log 11.2143 314 0.6699 0.5076 0.6699 0.8185
No log 11.2857 316 0.5751 0.6148 0.5751 0.7584
No log 11.3571 318 0.5733 0.4948 0.5733 0.7572
No log 11.4286 320 0.6005 0.5155 0.6005 0.7749
No log 11.5 322 0.5786 0.6413 0.5786 0.7606
No log 11.5714 324 0.5836 0.6588 0.5836 0.7640
No log 11.6429 326 0.5927 0.6612 0.5927 0.7699
No log 11.7143 328 0.6451 0.4959 0.6451 0.8032
No log 11.7857 330 0.7067 0.4346 0.7067 0.8407
No log 11.8571 332 0.6701 0.4564 0.6701 0.8186
No log 11.9286 334 0.5992 0.5230 0.5992 0.7741
No log 12.0 336 0.5811 0.5902 0.5811 0.7623
No log 12.0714 338 0.5982 0.5230 0.5982 0.7734
No log 12.1429 340 0.5581 0.5780 0.5581 0.7471
No log 12.2143 342 0.5685 0.5501 0.5685 0.7540
No log 12.2857 344 0.6233 0.4819 0.6233 0.7895
No log 12.3571 346 0.5989 0.5263 0.5989 0.7739
No log 12.4286 348 0.5698 0.5956 0.5698 0.7548
No log 12.5 350 0.5921 0.5785 0.5921 0.7695
No log 12.5714 352 0.5780 0.5813 0.5780 0.7603
No log 12.6429 354 0.5709 0.5779 0.5709 0.7556
No log 12.7143 356 0.5858 0.5004 0.5858 0.7654
No log 12.7857 358 0.5840 0.5266 0.5840 0.7642
No log 12.8571 360 0.5502 0.6112 0.5502 0.7417
No log 12.9286 362 0.5759 0.5680 0.5759 0.7589
No log 13.0 364 0.5683 0.6010 0.5683 0.7539
No log 13.0714 366 0.5542 0.5903 0.5542 0.7445
No log 13.1429 368 0.5696 0.5548 0.5696 0.7547
No log 13.2143 370 0.5668 0.5725 0.5668 0.7528
No log 13.2857 372 0.5591 0.6260 0.5591 0.7477
No log 13.3571 374 0.5656 0.5905 0.5656 0.7521
No log 13.4286 376 0.5640 0.5963 0.5640 0.7510
No log 13.5 378 0.5446 0.5991 0.5446 0.7379
No log 13.5714 380 0.5355 0.5914 0.5355 0.7318
No log 13.6429 382 0.5299 0.6153 0.5299 0.7279
No log 13.7143 384 0.5368 0.5851 0.5368 0.7327
No log 13.7857 386 0.5306 0.6065 0.5306 0.7284
No log 13.8571 388 0.5506 0.5880 0.5506 0.7420
No log 13.9286 390 0.5576 0.5880 0.5576 0.7467
No log 14.0 392 0.5357 0.5762 0.5357 0.7319
No log 14.0714 394 0.5360 0.5401 0.5360 0.7321
No log 14.1429 396 0.5657 0.5550 0.5657 0.7521
No log 14.2143 398 0.6543 0.5277 0.6543 0.8089
No log 14.2857 400 0.6531 0.5452 0.6531 0.8081
No log 14.3571 402 0.5617 0.6234 0.5617 0.7495
No log 14.4286 404 0.5399 0.6219 0.5399 0.7348
No log 14.5 406 0.5343 0.5895 0.5343 0.7310
No log 14.5714 408 0.5396 0.4697 0.5396 0.7346
No log 14.6429 410 0.5705 0.5039 0.5705 0.7553
No log 14.7143 412 0.5401 0.4933 0.5401 0.7349
No log 14.7857 414 0.5227 0.5155 0.5227 0.7230
No log 14.8571 416 0.5423 0.5732 0.5423 0.7364
No log 14.9286 418 0.5373 0.5982 0.5373 0.7330
No log 15.0 420 0.5375 0.5913 0.5375 0.7331
No log 15.0714 422 0.6176 0.4792 0.6176 0.7858
No log 15.1429 424 0.6670 0.4970 0.6670 0.8167
No log 15.2143 426 0.5860 0.4797 0.5860 0.7655
No log 15.2857 428 0.5372 0.5642 0.5372 0.7329
No log 15.3571 430 0.5355 0.5312 0.5355 0.7318
No log 15.4286 432 0.5419 0.4953 0.5419 0.7361
No log 15.5 434 0.5432 0.5134 0.5432 0.7370
No log 15.5714 436 0.5337 0.5218 0.5337 0.7305
No log 15.6429 438 0.5481 0.4835 0.5481 0.7403
No log 15.7143 440 0.5489 0.4640 0.5489 0.7409
No log 15.7857 442 0.5236 0.5638 0.5236 0.7236
No log 15.8571 444 0.5315 0.6376 0.5315 0.7291
No log 15.9286 446 0.6302 0.5190 0.6302 0.7939
No log 16.0 448 0.7618 0.5121 0.7618 0.8728
No log 16.0714 450 0.7692 0.5017 0.7692 0.8770
No log 16.1429 452 0.6521 0.5459 0.6521 0.8075
No log 16.2143 454 0.5934 0.5542 0.5934 0.7703
No log 16.2857 456 0.6025 0.5367 0.6025 0.7762
No log 16.3571 458 0.6399 0.5075 0.6399 0.7999
No log 16.4286 460 0.6323 0.5274 0.6323 0.7952
No log 16.5 462 0.6079 0.5698 0.6079 0.7797
No log 16.5714 464 0.5611 0.5985 0.5611 0.7491
No log 16.6429 466 0.5510 0.5905 0.5510 0.7423
No log 16.7143 468 0.5455 0.5866 0.5455 0.7386
No log 16.7857 470 0.5380 0.5891 0.5380 0.7335
No log 16.8571 472 0.5395 0.5973 0.5395 0.7345
No log 16.9286 474 0.5552 0.5880 0.5552 0.7451
No log 17.0 476 0.6037 0.5257 0.6037 0.7770
No log 17.0714 478 0.5882 0.5699 0.5882 0.7670
No log 17.1429 480 0.5384 0.5803 0.5384 0.7337
No log 17.2143 482 0.5426 0.5642 0.5426 0.7366
No log 17.2857 484 0.5432 0.6090 0.5432 0.7370
No log 17.3571 486 0.5435 0.6464 0.5435 0.7372
No log 17.4286 488 0.5433 0.5762 0.5433 0.7371
No log 17.5 490 0.5425 0.5846 0.5425 0.7366
No log 17.5714 492 0.5421 0.5762 0.5421 0.7363
No log 17.6429 494 0.5450 0.5972 0.5450 0.7382
No log 17.7143 496 0.5531 0.5871 0.5531 0.7437
No log 17.7857 498 0.5322 0.5767 0.5322 0.7295
0.3362 17.8571 500 0.5388 0.5643 0.5388 0.7340
0.3362 17.9286 502 0.5653 0.5065 0.5653 0.7519
0.3362 18.0 504 0.5536 0.5377 0.5536 0.7441
0.3362 18.0714 506 0.5311 0.5823 0.5311 0.7288
0.3362 18.1429 508 0.5650 0.6028 0.5650 0.7517
0.3362 18.2143 510 0.5749 0.6229 0.5749 0.7582
0.3362 18.2857 512 0.5719 0.5732 0.5719 0.7563
0.3362 18.3571 514 0.6005 0.5784 0.6005 0.7749
0.3362 18.4286 516 0.6296 0.5702 0.6296 0.7935
0.3362 18.5 518 0.6355 0.5624 0.6355 0.7972
0.3362 18.5714 520 0.5843 0.5632 0.5843 0.7644
0.3362 18.6429 522 0.5840 0.5398 0.5840 0.7642
0.3362 18.7143 524 0.6036 0.5578 0.6036 0.7769
0.3362 18.7857 526 0.6529 0.6036 0.6529 0.8080
0.3362 18.8571 528 0.6988 0.5546 0.6988 0.8359
0.3362 18.9286 530 0.6710 0.5546 0.6710 0.8191
0.3362 19.0 532 0.6343 0.5080 0.6343 0.7964
0.3362 19.0714 534 0.6129 0.4993 0.6129 0.7829
0.3362 19.1429 536 0.6308 0.4938 0.6308 0.7942

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
4
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for MayBashendy/ArabicNewSplits8_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k5_task2_organization

Finetuned
(4222)
this model