ArabicNewSplits8_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k14_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6468
  • Qwk: 0.4520
  • Mse: 0.6468
  • Rmse: 0.8042

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0267 2 4.3882 -0.0134 4.3882 2.0948
No log 0.0533 4 2.4350 0.0464 2.4350 1.5604
No log 0.08 6 1.9937 -0.0901 1.9937 1.4120
No log 0.1067 8 1.2891 -0.1077 1.2891 1.1354
No log 0.1333 10 0.8750 0.0970 0.8750 0.9354
No log 0.16 12 1.1646 0.1160 1.1646 1.0792
No log 0.1867 14 1.1533 0.1540 1.1533 1.0739
No log 0.2133 16 0.7876 0.1539 0.7876 0.8875
No log 0.24 18 0.7543 0.2376 0.7543 0.8685
No log 0.2667 20 0.8395 0.0869 0.8395 0.9162
No log 0.2933 22 0.8692 0.0147 0.8692 0.9323
No log 0.32 24 0.8141 0.0654 0.8141 0.9023
No log 0.3467 26 0.7924 0.0937 0.7924 0.8902
No log 0.3733 28 0.8073 0.1835 0.8073 0.8985
No log 0.4 30 0.8408 0.2085 0.8408 0.9170
No log 0.4267 32 0.7697 0.2662 0.7697 0.8773
No log 0.4533 34 0.7598 0.1937 0.7598 0.8717
No log 0.48 36 0.7466 0.2474 0.7466 0.8641
No log 0.5067 38 0.7960 0.1920 0.7960 0.8922
No log 0.5333 40 0.8679 0.1489 0.8679 0.9316
No log 0.56 42 0.9576 0.1248 0.9576 0.9786
No log 0.5867 44 1.0829 0.1113 1.0829 1.0406
No log 0.6133 46 0.9404 0.1805 0.9404 0.9697
No log 0.64 48 0.7629 0.2959 0.7629 0.8734
No log 0.6667 50 0.7336 0.3515 0.7336 0.8565
No log 0.6933 52 0.8819 0.2309 0.8819 0.9391
No log 0.72 54 0.8673 0.1949 0.8673 0.9313
No log 0.7467 56 0.8924 0.2048 0.8924 0.9447
No log 0.7733 58 0.9810 0.1808 0.9810 0.9904
No log 0.8 60 0.9166 0.2417 0.9166 0.9574
No log 0.8267 62 0.6991 0.3930 0.6991 0.8361
No log 0.8533 64 0.6845 0.4071 0.6845 0.8273
No log 0.88 66 0.7080 0.4139 0.7080 0.8414
No log 0.9067 68 0.7604 0.3611 0.7604 0.8720
No log 0.9333 70 0.7930 0.3946 0.7930 0.8905
No log 0.96 72 0.7994 0.3394 0.7994 0.8941
No log 0.9867 74 0.6904 0.4614 0.6904 0.8309
No log 1.0133 76 0.6694 0.4764 0.6694 0.8182
No log 1.04 78 0.6633 0.4654 0.6633 0.8144
No log 1.0667 80 0.6673 0.4625 0.6673 0.8169
No log 1.0933 82 0.6829 0.4907 0.6829 0.8264
No log 1.12 84 0.6989 0.5261 0.6989 0.8360
No log 1.1467 86 0.7629 0.4130 0.7629 0.8734
No log 1.1733 88 0.6706 0.4890 0.6706 0.8189
No log 1.2 90 0.6406 0.5167 0.6406 0.8004
No log 1.2267 92 0.6180 0.4588 0.6180 0.7862
No log 1.2533 94 0.6000 0.5093 0.6000 0.7746
No log 1.28 96 0.6071 0.4937 0.6071 0.7792
No log 1.3067 98 0.6003 0.5428 0.6003 0.7748
No log 1.3333 100 0.7624 0.4635 0.7624 0.8731
No log 1.3600 102 0.8822 0.3869 0.8822 0.9393
No log 1.3867 104 0.6705 0.4846 0.6705 0.8188
No log 1.4133 106 0.6841 0.5337 0.6841 0.8271
No log 1.44 108 0.6851 0.4861 0.6851 0.8277
No log 1.4667 110 0.7659 0.4265 0.7659 0.8752
No log 1.4933 112 0.8911 0.3925 0.8911 0.9440
No log 1.52 114 0.7744 0.4753 0.7744 0.8800
No log 1.5467 116 0.7172 0.4993 0.7172 0.8469
No log 1.5733 118 0.7616 0.5591 0.7616 0.8727
No log 1.6 120 0.6781 0.5191 0.6781 0.8235
No log 1.6267 122 0.6988 0.4878 0.6988 0.8359
No log 1.6533 124 0.6518 0.5437 0.6518 0.8073
No log 1.6800 126 0.6510 0.5479 0.6510 0.8068
No log 1.7067 128 0.7217 0.4961 0.7217 0.8495
No log 1.7333 130 0.7476 0.4333 0.7476 0.8646
No log 1.76 132 0.6585 0.5521 0.6585 0.8115
No log 1.7867 134 1.0521 0.3748 1.0521 1.0257
No log 1.8133 136 1.1245 0.3595 1.1245 1.0604
No log 1.8400 138 0.8057 0.5043 0.8056 0.8976
No log 1.8667 140 0.6864 0.4779 0.6864 0.8285
No log 1.8933 142 0.7963 0.4926 0.7963 0.8923
No log 1.92 144 0.7742 0.4455 0.7742 0.8799
No log 1.9467 146 0.6246 0.4639 0.6246 0.7903
No log 1.9733 148 0.7000 0.3946 0.7000 0.8367
No log 2.0 150 0.6720 0.4295 0.6720 0.8197
No log 2.0267 152 0.6690 0.5525 0.6690 0.8179
No log 2.0533 154 0.7405 0.5556 0.7405 0.8605
No log 2.08 156 0.7674 0.5024 0.7674 0.8760
No log 2.1067 158 0.7377 0.5353 0.7377 0.8589
No log 2.1333 160 0.7463 0.5355 0.7463 0.8639
No log 2.16 162 0.8073 0.5274 0.8073 0.8985
No log 2.1867 164 0.7908 0.5193 0.7908 0.8893
No log 2.2133 166 0.6623 0.5511 0.6623 0.8138
No log 2.24 168 0.6459 0.5075 0.6459 0.8037
No log 2.2667 170 0.6303 0.5177 0.6303 0.7939
No log 2.2933 172 0.6699 0.4605 0.6699 0.8185
No log 2.32 174 0.7298 0.3074 0.7298 0.8543
No log 2.3467 176 0.7379 0.3736 0.7379 0.8590
No log 2.3733 178 0.6722 0.4198 0.6722 0.8199
No log 2.4 180 0.6607 0.5173 0.6607 0.8128
No log 2.4267 182 0.7416 0.4372 0.7416 0.8612
No log 2.4533 184 0.9118 0.4838 0.9118 0.9549
No log 2.48 186 1.0113 0.4637 1.0113 1.0056
No log 2.5067 188 0.8220 0.4580 0.8220 0.9067
No log 2.5333 190 0.7668 0.5405 0.7668 0.8757
No log 2.56 192 0.7521 0.5736 0.7521 0.8672
No log 2.5867 194 0.7381 0.5450 0.7381 0.8591
No log 2.6133 196 0.6993 0.5352 0.6993 0.8362
No log 2.64 198 0.6796 0.5452 0.6796 0.8244
No log 2.6667 200 0.7133 0.4921 0.7133 0.8446
No log 2.6933 202 0.6641 0.4863 0.6641 0.8149
No log 2.7200 204 0.6289 0.4895 0.6289 0.7930
No log 2.7467 206 0.5907 0.4941 0.5907 0.7686
No log 2.7733 208 0.5944 0.5304 0.5944 0.7709
No log 2.8 210 0.6257 0.5731 0.6257 0.7910
No log 2.8267 212 0.7228 0.5279 0.7228 0.8502
No log 2.8533 214 0.7565 0.5345 0.7565 0.8698
No log 2.88 216 0.6859 0.5576 0.6859 0.8282
No log 2.9067 218 0.6267 0.5689 0.6267 0.7916
No log 2.9333 220 0.6523 0.5390 0.6523 0.8077
No log 2.96 222 0.6191 0.5361 0.6191 0.7869
No log 2.9867 224 0.6572 0.5623 0.6572 0.8107
No log 3.0133 226 0.6886 0.5519 0.6886 0.8298
No log 3.04 228 0.6496 0.5363 0.6496 0.8060
No log 3.0667 230 0.5894 0.5130 0.5894 0.7677
No log 3.0933 232 0.5786 0.5282 0.5786 0.7607
No log 3.12 234 0.5819 0.5045 0.5819 0.7628
No log 3.1467 236 0.6452 0.4226 0.6452 0.8032
No log 3.1733 238 0.8235 0.4119 0.8235 0.9075
No log 3.2 240 0.7444 0.4417 0.7444 0.8628
No log 3.2267 242 0.5939 0.4888 0.5939 0.7706
No log 3.2533 244 0.6340 0.5698 0.6340 0.7963
No log 3.2800 246 0.6154 0.5539 0.6154 0.7845
No log 3.3067 248 0.6259 0.5348 0.6259 0.7911
No log 3.3333 250 0.6480 0.5490 0.6480 0.8050
No log 3.36 252 0.6086 0.5359 0.6086 0.7801
No log 3.3867 254 0.6522 0.5542 0.6522 0.8076
No log 3.4133 256 0.7644 0.5105 0.7644 0.8743
No log 3.44 258 0.8134 0.4658 0.8134 0.9019
No log 3.4667 260 0.6886 0.5362 0.6886 0.8298
No log 3.4933 262 0.6858 0.5535 0.6858 0.8281
No log 3.52 264 0.8134 0.5515 0.8134 0.9019
No log 3.5467 266 0.7357 0.5220 0.7357 0.8578
No log 3.5733 268 0.7365 0.5407 0.7365 0.8582
No log 3.6 270 0.9025 0.4566 0.9025 0.9500
No log 3.6267 272 0.8965 0.4314 0.8965 0.9468
No log 3.6533 274 0.6887 0.5265 0.6887 0.8299
No log 3.68 276 0.6064 0.4920 0.6064 0.7787
No log 3.7067 278 0.6350 0.5195 0.6350 0.7968
No log 3.7333 280 0.6352 0.5150 0.6352 0.7970
No log 3.76 282 0.5863 0.4247 0.5863 0.7657
No log 3.7867 284 0.6303 0.5155 0.6303 0.7939
No log 3.8133 286 0.8605 0.4826 0.8605 0.9276
No log 3.84 288 1.0712 0.3741 1.0712 1.0350
No log 3.8667 290 0.9806 0.4228 0.9806 0.9903
No log 3.8933 292 0.7296 0.5465 0.7296 0.8542
No log 3.92 294 0.6715 0.5666 0.6715 0.8194
No log 3.9467 296 0.6805 0.5477 0.6805 0.8249
No log 3.9733 298 0.6428 0.5202 0.6428 0.8018
No log 4.0 300 0.6356 0.5223 0.6356 0.7972
No log 4.0267 302 0.6946 0.4772 0.6946 0.8334
No log 4.0533 304 0.6826 0.5116 0.6826 0.8262
No log 4.08 306 0.6041 0.4865 0.6041 0.7773
No log 4.1067 308 0.6247 0.5822 0.6247 0.7904
No log 4.1333 310 0.6491 0.5552 0.6491 0.8057
No log 4.16 312 0.6282 0.4859 0.6282 0.7926
No log 4.1867 314 0.6909 0.5050 0.6909 0.8312
No log 4.2133 316 0.7275 0.5116 0.7275 0.8529
No log 4.24 318 0.6434 0.4679 0.6434 0.8021
No log 4.2667 320 0.6357 0.5287 0.6357 0.7973
No log 4.2933 322 0.7084 0.5614 0.7084 0.8417
No log 4.32 324 0.7082 0.5855 0.7082 0.8415
No log 4.3467 326 0.6710 0.5477 0.6710 0.8192
No log 4.3733 328 0.6875 0.5730 0.6875 0.8292
No log 4.4 330 0.6828 0.6061 0.6828 0.8263
No log 4.4267 332 0.6903 0.5946 0.6903 0.8308
No log 4.4533 334 0.7444 0.5107 0.7444 0.8628
No log 4.48 336 0.8059 0.4831 0.8059 0.8977
No log 4.5067 338 0.6920 0.4774 0.6920 0.8319
No log 4.5333 340 0.6099 0.4769 0.6099 0.7810
No log 4.5600 342 0.6169 0.5028 0.6169 0.7855
No log 4.5867 344 0.6360 0.4760 0.6360 0.7975
No log 4.6133 346 0.6501 0.4760 0.6501 0.8063
No log 4.64 348 0.6644 0.5425 0.6644 0.8151
No log 4.6667 350 0.6795 0.5445 0.6795 0.8243
No log 4.6933 352 0.6581 0.4876 0.6581 0.8112
No log 4.72 354 0.7090 0.4845 0.7090 0.8420
No log 4.7467 356 0.7281 0.5088 0.7281 0.8533
No log 4.7733 358 0.7502 0.4784 0.7502 0.8661
No log 4.8 360 0.7189 0.4500 0.7189 0.8479
No log 4.8267 362 0.6713 0.4934 0.6713 0.8193
No log 4.8533 364 0.6642 0.5145 0.6642 0.8150
No log 4.88 366 0.7038 0.5392 0.7038 0.8389
No log 4.9067 368 0.7338 0.5264 0.7338 0.8566
No log 4.9333 370 0.7467 0.5534 0.7467 0.8641
No log 4.96 372 0.6577 0.5583 0.6577 0.8110
No log 4.9867 374 0.6480 0.5617 0.6480 0.8050
No log 5.0133 376 0.7165 0.5181 0.7165 0.8465
No log 5.04 378 0.6504 0.5133 0.6504 0.8065
No log 5.0667 380 0.6249 0.5609 0.6249 0.7905
No log 5.0933 382 0.6760 0.5548 0.6760 0.8222
No log 5.12 384 0.6864 0.5930 0.6864 0.8285
No log 5.1467 386 0.6735 0.5533 0.6735 0.8207
No log 5.1733 388 0.6714 0.5867 0.6714 0.8194
No log 5.2 390 0.6704 0.5609 0.6704 0.8188
No log 5.2267 392 0.6841 0.5554 0.6841 0.8271
No log 5.2533 394 0.6846 0.5712 0.6846 0.8274
No log 5.28 396 0.6642 0.5721 0.6642 0.8150
No log 5.3067 398 0.6702 0.5796 0.6702 0.8186
No log 5.3333 400 0.6707 0.5821 0.6707 0.8189
No log 5.36 402 0.6647 0.5053 0.6647 0.8153
No log 5.3867 404 0.6289 0.5121 0.6289 0.7930
No log 5.4133 406 0.5928 0.4959 0.5928 0.7699
No log 5.44 408 0.6008 0.5260 0.6008 0.7751
No log 5.4667 410 0.6098 0.5325 0.6098 0.7809
No log 5.4933 412 0.6082 0.5325 0.6082 0.7799
No log 5.52 414 0.6111 0.4954 0.6111 0.7817
No log 5.5467 416 0.5939 0.5241 0.5939 0.7706
No log 5.5733 418 0.6025 0.4832 0.6025 0.7762
No log 5.6 420 0.6194 0.5485 0.6194 0.7870
No log 5.6267 422 0.6310 0.5128 0.6310 0.7944
No log 5.6533 424 0.6648 0.5584 0.6648 0.8153
No log 5.68 426 0.6488 0.5217 0.6488 0.8055
No log 5.7067 428 0.6126 0.5245 0.6126 0.7827
No log 5.7333 430 0.6197 0.5325 0.6197 0.7872
No log 5.76 432 0.6129 0.4857 0.6129 0.7829
No log 5.7867 434 0.6114 0.4857 0.6114 0.7819
No log 5.8133 436 0.6005 0.4096 0.6005 0.7749
No log 5.84 438 0.6090 0.4557 0.6090 0.7804
No log 5.8667 440 0.6289 0.5291 0.6289 0.7930
No log 5.8933 442 0.6780 0.5095 0.6780 0.8234
No log 5.92 444 0.6742 0.5026 0.6742 0.8211
No log 5.9467 446 0.6197 0.5239 0.6197 0.7872
No log 5.9733 448 0.5876 0.4319 0.5876 0.7665
No log 6.0 450 0.6147 0.4397 0.6147 0.7840
No log 6.0267 452 0.6204 0.4072 0.6204 0.7876
No log 6.0533 454 0.6048 0.4100 0.6048 0.7777
No log 6.08 456 0.5913 0.3858 0.5913 0.7690
No log 6.1067 458 0.5979 0.5203 0.5979 0.7732
No log 6.1333 460 0.6222 0.5476 0.6222 0.7888
No log 6.16 462 0.6464 0.5142 0.6464 0.8040
No log 6.1867 464 0.6551 0.4809 0.6551 0.8094
No log 6.2133 466 0.6084 0.4877 0.6084 0.7800
No log 6.24 468 0.6084 0.4751 0.6084 0.7800
No log 6.2667 470 0.5964 0.5036 0.5964 0.7723
No log 6.2933 472 0.5780 0.4922 0.5780 0.7603
No log 6.32 474 0.5946 0.5037 0.5946 0.7711
No log 6.3467 476 0.6950 0.4399 0.6950 0.8336
No log 6.3733 478 0.8146 0.4281 0.8146 0.9026
No log 6.4 480 0.7520 0.4747 0.7520 0.8672
No log 6.4267 482 0.6644 0.5306 0.6644 0.8151
No log 6.4533 484 0.6164 0.5652 0.6164 0.7851
No log 6.48 486 0.6405 0.4678 0.6405 0.8003
No log 6.5067 488 0.6407 0.4540 0.6407 0.8004
No log 6.5333 490 0.7396 0.4718 0.7396 0.8600
No log 6.5600 492 0.7676 0.4734 0.7676 0.8762
No log 6.5867 494 0.6619 0.5073 0.6619 0.8136
No log 6.6133 496 0.6201 0.4954 0.6201 0.7874
No log 6.64 498 0.6990 0.5120 0.6990 0.8360
0.3842 6.6667 500 0.7189 0.4713 0.7189 0.8479
0.3842 6.6933 502 0.6874 0.4435 0.6874 0.8291
0.3842 6.72 504 0.6435 0.4926 0.6435 0.8022
0.3842 6.7467 506 0.6430 0.5097 0.6430 0.8019
0.3842 6.7733 508 0.6546 0.4350 0.6546 0.8091
0.3842 6.8 510 0.7201 0.4382 0.7201 0.8486
0.3842 6.8267 512 0.7396 0.4649 0.7396 0.8600
0.3842 6.8533 514 0.6930 0.4892 0.6930 0.8325
0.3842 6.88 516 0.6335 0.3964 0.6335 0.7959
0.3842 6.9067 518 0.6211 0.4274 0.6211 0.7881
0.3842 6.9333 520 0.6323 0.4307 0.6323 0.7952
0.3842 6.96 522 0.6451 0.4972 0.6451 0.8032
0.3842 6.9867 524 0.6716 0.4948 0.6716 0.8195
0.3842 7.0133 526 0.7644 0.4541 0.7644 0.8743
0.3842 7.04 528 0.8006 0.3650 0.8006 0.8947
0.3842 7.0667 530 0.7237 0.4596 0.7237 0.8507
0.3842 7.0933 532 0.6558 0.4221 0.6558 0.8098
0.3842 7.12 534 0.6468 0.4520 0.6468 0.8042

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
4
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for MayBashendy/ArabicNewSplits8_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k14_task2_organization

Finetuned
(4222)
this model