ArabicNewSplits8_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k1_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7056
  • Qwk: 0.5234
  • Mse: 0.7056
  • Rmse: 0.8400

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.2857 2 4.3466 -0.0057 4.3466 2.0849
No log 0.5714 4 2.5677 0.0564 2.5677 1.6024
No log 0.8571 6 1.1613 0.0405 1.1613 1.0776
No log 1.1429 8 0.8897 0.1342 0.8897 0.9432
No log 1.4286 10 0.8231 0.1417 0.8231 0.9072
No log 1.7143 12 0.8177 0.2052 0.8177 0.9043
No log 2.0 14 0.8472 0.2178 0.8472 0.9204
No log 2.2857 16 0.7913 0.1614 0.7913 0.8895
No log 2.5714 18 0.7559 0.1819 0.7559 0.8694
No log 2.8571 20 0.7356 0.1926 0.7356 0.8577
No log 3.1429 22 0.7904 0.2630 0.7904 0.8890
No log 3.4286 24 0.8721 0.2295 0.8721 0.9338
No log 3.7143 26 0.8738 0.2256 0.8738 0.9348
No log 4.0 28 0.7256 0.3020 0.7256 0.8518
No log 4.2857 30 0.7169 0.2993 0.7169 0.8467
No log 4.5714 32 0.7392 0.3697 0.7392 0.8598
No log 4.8571 34 0.7785 0.3962 0.7785 0.8824
No log 5.1429 36 0.8262 0.3560 0.8262 0.9090
No log 5.4286 38 0.7737 0.4234 0.7737 0.8796
No log 5.7143 40 0.8889 0.4587 0.8889 0.9428
No log 6.0 42 1.2169 0.3001 1.2169 1.1031
No log 6.2857 44 1.1149 0.3434 1.1149 1.0559
No log 6.5714 46 0.9928 0.4017 0.9928 0.9964
No log 6.8571 48 0.9741 0.4414 0.9741 0.9870
No log 7.1429 50 0.9602 0.4437 0.9602 0.9799
No log 7.4286 52 0.9136 0.4856 0.9136 0.9558
No log 7.7143 54 0.8061 0.4755 0.8061 0.8978
No log 8.0 56 0.7229 0.5204 0.7229 0.8502
No log 8.2857 58 0.6917 0.5058 0.6917 0.8317
No log 8.5714 60 0.6592 0.4736 0.6592 0.8119
No log 8.8571 62 0.6972 0.5118 0.6972 0.8350
No log 9.1429 64 0.6627 0.5357 0.6627 0.8141
No log 9.4286 66 0.6952 0.5271 0.6952 0.8338
No log 9.7143 68 0.7948 0.5103 0.7948 0.8915
No log 10.0 70 0.7680 0.5115 0.7680 0.8764
No log 10.2857 72 0.7331 0.5735 0.7331 0.8562
No log 10.5714 74 0.7352 0.5084 0.7352 0.8574
No log 10.8571 76 0.7068 0.5239 0.7068 0.8407
No log 11.1429 78 0.6666 0.4998 0.6666 0.8165
No log 11.4286 80 0.6589 0.4790 0.6589 0.8117
No log 11.7143 82 0.6463 0.4753 0.6463 0.8039
No log 12.0 84 0.6566 0.5044 0.6566 0.8103
No log 12.2857 86 0.6884 0.4849 0.6884 0.8297
No log 12.5714 88 0.7771 0.4510 0.7771 0.8815
No log 12.8571 90 0.8451 0.4484 0.8451 0.9193
No log 13.1429 92 0.7770 0.4853 0.7770 0.8815
No log 13.4286 94 0.8129 0.4777 0.8129 0.9016
No log 13.7143 96 0.7903 0.4872 0.7903 0.8890
No log 14.0 98 0.7943 0.4441 0.7943 0.8913
No log 14.2857 100 0.8423 0.4715 0.8423 0.9177
No log 14.5714 102 0.7908 0.4135 0.7908 0.8893
No log 14.8571 104 0.7706 0.4448 0.7706 0.8778
No log 15.1429 106 0.7240 0.4445 0.7240 0.8509
No log 15.4286 108 0.6816 0.4323 0.6816 0.8256
No log 15.7143 110 0.6941 0.4076 0.6941 0.8331
No log 16.0 112 0.6892 0.4351 0.6892 0.8302
No log 16.2857 114 0.6990 0.4390 0.6990 0.8361
No log 16.5714 116 0.6932 0.5249 0.6932 0.8326
No log 16.8571 118 0.7849 0.4942 0.7849 0.8860
No log 17.1429 120 0.7741 0.5475 0.7741 0.8798
No log 17.4286 122 0.7808 0.5036 0.7808 0.8836
No log 17.7143 124 0.7687 0.5109 0.7687 0.8767
No log 18.0 126 0.7675 0.5238 0.7675 0.8761
No log 18.2857 128 0.8603 0.4938 0.8603 0.9275
No log 18.5714 130 0.8483 0.5084 0.8483 0.9210
No log 18.8571 132 0.7854 0.5184 0.7854 0.8862
No log 19.1429 134 0.9617 0.4474 0.9617 0.9807
No log 19.4286 136 0.9312 0.4401 0.9312 0.9650
No log 19.7143 138 0.7439 0.5547 0.7439 0.8625
No log 20.0 140 0.7204 0.4928 0.7204 0.8487
No log 20.2857 142 0.7021 0.4602 0.7021 0.8379
No log 20.5714 144 0.6678 0.4547 0.6678 0.8172
No log 20.8571 146 0.7359 0.4421 0.7359 0.8579
No log 21.1429 148 0.8103 0.4381 0.8103 0.9002
No log 21.4286 150 0.8399 0.5232 0.8399 0.9165
No log 21.7143 152 0.9057 0.4556 0.9057 0.9517
No log 22.0 154 0.8716 0.4418 0.8716 0.9336
No log 22.2857 156 0.7343 0.4873 0.7343 0.8569
No log 22.5714 158 0.6281 0.4718 0.6281 0.7926
No log 22.8571 160 0.7509 0.5034 0.7509 0.8665
No log 23.1429 162 0.8809 0.4402 0.8809 0.9385
No log 23.4286 164 0.8415 0.4386 0.8415 0.9174
No log 23.7143 166 0.7327 0.5378 0.7327 0.8560
No log 24.0 168 0.8157 0.4607 0.8157 0.9031
No log 24.2857 170 0.9259 0.4604 0.9259 0.9622
No log 24.5714 172 0.9308 0.4594 0.9308 0.9648
No log 24.8571 174 0.8540 0.5192 0.8540 0.9241
No log 25.1429 176 0.8300 0.4550 0.8300 0.9110
No log 25.4286 178 0.8018 0.4803 0.8018 0.8954
No log 25.7143 180 0.7142 0.4817 0.7142 0.8451
No log 26.0 182 0.6342 0.4662 0.6342 0.7964
No log 26.2857 184 0.6734 0.4895 0.6734 0.8206
No log 26.5714 186 0.7405 0.5101 0.7405 0.8605
No log 26.8571 188 0.7636 0.5223 0.7636 0.8739
No log 27.1429 190 0.7553 0.4832 0.7553 0.8691
No log 27.4286 192 0.7653 0.4751 0.7653 0.8748
No log 27.7143 194 0.7324 0.5067 0.7324 0.8558
No log 28.0 196 0.7294 0.4682 0.7294 0.8540
No log 28.2857 198 0.7755 0.5272 0.7755 0.8807
No log 28.5714 200 0.7966 0.5272 0.7966 0.8925
No log 28.8571 202 0.7779 0.5254 0.7779 0.8820
No log 29.1429 204 0.7099 0.5400 0.7099 0.8425
No log 29.4286 206 0.6481 0.5526 0.6481 0.8050
No log 29.7143 208 0.6302 0.5124 0.6302 0.7938
No log 30.0 210 0.6192 0.5461 0.6192 0.7869
No log 30.2857 212 0.6607 0.5630 0.6607 0.8128
No log 30.5714 214 0.6919 0.5098 0.6919 0.8318
No log 30.8571 216 0.6842 0.5076 0.6842 0.8272
No log 31.1429 218 0.6488 0.5476 0.6488 0.8055
No log 31.4286 220 0.6222 0.5448 0.6222 0.7888
No log 31.7143 222 0.6494 0.5599 0.6494 0.8059
No log 32.0 224 0.6602 0.5315 0.6602 0.8125
No log 32.2857 226 0.6786 0.5384 0.6786 0.8238
No log 32.5714 228 0.6722 0.5559 0.6722 0.8199
No log 32.8571 230 0.6560 0.5315 0.6560 0.8099
No log 33.1429 232 0.6962 0.5567 0.6962 0.8344
No log 33.4286 234 0.7932 0.5001 0.7932 0.8906
No log 33.7143 236 0.8524 0.4539 0.8524 0.9232
No log 34.0 238 0.8020 0.4741 0.8020 0.8955
No log 34.2857 240 0.7301 0.52 0.7301 0.8545
No log 34.5714 242 0.6504 0.5486 0.6504 0.8065
No log 34.8571 244 0.6392 0.5022 0.6392 0.7995
No log 35.1429 246 0.6654 0.5548 0.6654 0.8157
No log 35.4286 248 0.6559 0.5543 0.6559 0.8099
No log 35.7143 250 0.6363 0.5105 0.6363 0.7977
No log 36.0 252 0.6791 0.5406 0.6791 0.8241
No log 36.2857 254 0.7985 0.5314 0.7985 0.8936
No log 36.5714 256 0.8684 0.4707 0.8684 0.9319
No log 36.8571 258 0.8587 0.5 0.8587 0.9266
No log 37.1429 260 0.7683 0.5207 0.7683 0.8765
No log 37.4286 262 0.6976 0.5339 0.6976 0.8352
No log 37.7143 264 0.6611 0.5730 0.6611 0.8131
No log 38.0 266 0.6747 0.5653 0.6747 0.8214
No log 38.2857 268 0.7086 0.5654 0.7086 0.8418
No log 38.5714 270 0.7190 0.5650 0.7190 0.8480
No log 38.8571 272 0.7372 0.5339 0.7372 0.8586
No log 39.1429 274 0.7661 0.5314 0.7661 0.8753
No log 39.4286 276 0.7496 0.5401 0.7496 0.8658
No log 39.7143 278 0.7178 0.5312 0.7178 0.8472
No log 40.0 280 0.6992 0.5390 0.6992 0.8362
No log 40.2857 282 0.7230 0.5251 0.7230 0.8503
No log 40.5714 284 0.7628 0.5309 0.7628 0.8734
No log 40.8571 286 0.7813 0.5147 0.7813 0.8839
No log 41.1429 288 0.7719 0.5355 0.7719 0.8786
No log 41.4286 290 0.7667 0.4845 0.7667 0.8756
No log 41.7143 292 0.7648 0.4982 0.7648 0.8745
No log 42.0 294 0.7478 0.5074 0.7478 0.8647
No log 42.2857 296 0.7069 0.4853 0.7069 0.8408
No log 42.5714 298 0.7041 0.5003 0.7041 0.8391
No log 42.8571 300 0.7147 0.5003 0.7147 0.8454
No log 43.1429 302 0.7001 0.4892 0.7001 0.8367
No log 43.4286 304 0.6890 0.5337 0.6890 0.8301
No log 43.7143 306 0.6949 0.5192 0.6949 0.8336
No log 44.0 308 0.7156 0.5195 0.7156 0.8459
No log 44.2857 310 0.7395 0.4903 0.7395 0.8599
No log 44.5714 312 0.7253 0.5246 0.7253 0.8516
No log 44.8571 314 0.7228 0.5405 0.7228 0.8502
No log 45.1429 316 0.7638 0.4754 0.7638 0.8740
No log 45.4286 318 0.7930 0.4966 0.7930 0.8905
No log 45.7143 320 0.7778 0.4898 0.7778 0.8819
No log 46.0 322 0.7486 0.5284 0.7486 0.8652
No log 46.2857 324 0.7450 0.5369 0.7450 0.8631
No log 46.5714 326 0.8243 0.4888 0.8243 0.9079
No log 46.8571 328 0.8827 0.4836 0.8827 0.9395
No log 47.1429 330 0.8489 0.4868 0.8489 0.9214
No log 47.4286 332 0.7786 0.5261 0.7786 0.8824
No log 47.7143 334 0.7501 0.5467 0.7501 0.8661
No log 48.0 336 0.7162 0.5097 0.7162 0.8463
No log 48.2857 338 0.7046 0.5166 0.7046 0.8394
No log 48.5714 340 0.7059 0.5421 0.7059 0.8402
No log 48.8571 342 0.7108 0.4930 0.7108 0.8431
No log 49.1429 344 0.7062 0.4988 0.7062 0.8403
No log 49.4286 346 0.7042 0.5474 0.7042 0.8391
No log 49.7143 348 0.6905 0.5474 0.6905 0.8309
No log 50.0 350 0.6657 0.5453 0.6657 0.8159
No log 50.2857 352 0.6483 0.5453 0.6483 0.8052
No log 50.5714 354 0.6242 0.5366 0.6242 0.7900
No log 50.8571 356 0.6209 0.5304 0.6209 0.7880
No log 51.1429 358 0.6394 0.5234 0.6394 0.7996
No log 51.4286 360 0.6493 0.5234 0.6493 0.8058
No log 51.7143 362 0.6620 0.5320 0.6620 0.8136
No log 52.0 364 0.6422 0.5410 0.6422 0.8014
No log 52.2857 366 0.6337 0.5448 0.6337 0.7961
No log 52.5714 368 0.6207 0.5677 0.6207 0.7878
No log 52.8571 370 0.6159 0.5768 0.6159 0.7848
No log 53.1429 372 0.6240 0.5652 0.6240 0.7899
No log 53.4286 374 0.6478 0.5503 0.6478 0.8048
No log 53.7143 376 0.6809 0.5490 0.6809 0.8252
No log 54.0 378 0.6940 0.5637 0.6940 0.8331
No log 54.2857 380 0.6900 0.5405 0.6900 0.8307
No log 54.5714 382 0.6967 0.5232 0.6967 0.8347
No log 54.8571 384 0.7132 0.4716 0.7132 0.8445
No log 55.1429 386 0.6986 0.5280 0.6986 0.8358
No log 55.4286 388 0.6940 0.5520 0.6940 0.8331
No log 55.7143 390 0.6939 0.5514 0.6939 0.8330
No log 56.0 392 0.6897 0.5474 0.6897 0.8305
No log 56.2857 394 0.6785 0.5474 0.6785 0.8237
No log 56.5714 396 0.6645 0.5599 0.6645 0.8151
No log 56.8571 398 0.6666 0.5545 0.6666 0.8164
No log 57.1429 400 0.6683 0.5644 0.6683 0.8175
No log 57.4286 402 0.6839 0.5181 0.6839 0.8270
No log 57.7143 404 0.7048 0.5489 0.7048 0.8395
No log 58.0 406 0.7159 0.5542 0.7159 0.8461
No log 58.2857 408 0.7000 0.5516 0.7000 0.8367
No log 58.5714 410 0.6652 0.5335 0.6652 0.8156
No log 58.8571 412 0.6450 0.5788 0.6450 0.8031
No log 59.1429 414 0.6435 0.5674 0.6435 0.8022
No log 59.4286 416 0.6572 0.5708 0.6572 0.8107
No log 59.7143 418 0.6793 0.5411 0.6793 0.8242
No log 60.0 420 0.6807 0.5543 0.6807 0.8251
No log 60.2857 422 0.6721 0.5584 0.6721 0.8198
No log 60.5714 424 0.6759 0.5635 0.6759 0.8221
No log 60.8571 426 0.6701 0.5783 0.6701 0.8186
No log 61.1429 428 0.6653 0.5783 0.6653 0.8156
No log 61.4286 430 0.6694 0.5624 0.6694 0.8182
No log 61.7143 432 0.6807 0.5350 0.6807 0.8250
No log 62.0 434 0.6762 0.4921 0.6762 0.8223
No log 62.2857 436 0.6667 0.5027 0.6667 0.8165
No log 62.5714 438 0.6564 0.5277 0.6564 0.8102
No log 62.8571 440 0.6417 0.5622 0.6417 0.8010
No log 63.1429 442 0.6336 0.5675 0.6336 0.7960
No log 63.4286 444 0.6189 0.5658 0.6189 0.7867
No log 63.7143 446 0.6076 0.5829 0.6076 0.7795
No log 64.0 448 0.5997 0.5566 0.5997 0.7744
No log 64.2857 450 0.6021 0.5566 0.6021 0.7760
No log 64.5714 452 0.6091 0.5743 0.6091 0.7804
No log 64.8571 454 0.6018 0.5941 0.6018 0.7758
No log 65.1429 456 0.6111 0.5827 0.6111 0.7817
No log 65.4286 458 0.6294 0.5662 0.6294 0.7933
No log 65.7143 460 0.6321 0.5662 0.6321 0.7950
No log 66.0 462 0.6277 0.5807 0.6277 0.7923
No log 66.2857 464 0.6311 0.5920 0.6311 0.7944
No log 66.5714 466 0.6384 0.6079 0.6384 0.7990
No log 66.8571 468 0.6485 0.5880 0.6485 0.8053
No log 67.1429 470 0.6618 0.5644 0.6618 0.8135
No log 67.4286 472 0.6795 0.5626 0.6795 0.8243
No log 67.7143 474 0.6887 0.5626 0.6887 0.8299
No log 68.0 476 0.6821 0.5554 0.6821 0.8259
No log 68.2857 478 0.6982 0.5446 0.6982 0.8356
No log 68.5714 480 0.7180 0.5240 0.7180 0.8473
No log 68.8571 482 0.7044 0.5334 0.7044 0.8393
No log 69.1429 484 0.6667 0.5556 0.6667 0.8165
No log 69.4286 486 0.6126 0.5687 0.6126 0.7827
No log 69.7143 488 0.5709 0.5960 0.5709 0.7556
No log 70.0 490 0.5576 0.5995 0.5576 0.7467
No log 70.2857 492 0.5600 0.5995 0.5600 0.7484
No log 70.5714 494 0.5700 0.5948 0.5700 0.7550
No log 70.8571 496 0.5908 0.6171 0.5908 0.7686
No log 71.1429 498 0.6131 0.5935 0.6131 0.7830
0.2324 71.4286 500 0.6474 0.5743 0.6474 0.8046
0.2324 71.7143 502 0.6765 0.5669 0.6765 0.8225
0.2324 72.0 504 0.6942 0.5548 0.6942 0.8332
0.2324 72.2857 506 0.6998 0.5548 0.6998 0.8366
0.2324 72.5714 508 0.6898 0.5741 0.6898 0.8306
0.2324 72.8571 510 0.6744 0.5797 0.6744 0.8212
0.2324 73.1429 512 0.6628 0.5816 0.6628 0.8141
0.2324 73.4286 514 0.6569 0.5836 0.6569 0.8105
0.2324 73.7143 516 0.6587 0.5890 0.6587 0.8116
0.2324 74.0 518 0.6717 0.5849 0.6717 0.8196
0.2324 74.2857 520 0.6736 0.5579 0.6736 0.8208
0.2324 74.5714 522 0.6920 0.5401 0.6920 0.8318
0.2324 74.8571 524 0.7067 0.5118 0.7067 0.8407
0.2324 75.1429 526 0.7023 0.5234 0.7023 0.8381
0.2324 75.4286 528 0.7056 0.5234 0.7056 0.8400

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
4
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for MayBashendy/ArabicNewSplits8_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k1_task2_organization

Finetuned
(4222)
this model