ArabicNewSplits8_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k9_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7808
  • Qwk: 0.4161
  • Mse: 0.7808
  • Rmse: 0.8837

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0417 2 4.0428 -0.0316 4.0428 2.0107
No log 0.0833 4 2.2540 -0.0160 2.2540 1.5013
No log 0.125 6 1.4314 -0.0085 1.4314 1.1964
No log 0.1667 8 1.3216 -0.1233 1.3216 1.1496
No log 0.2083 10 1.1537 -0.1118 1.1537 1.0741
No log 0.25 12 1.0800 -0.0635 1.0800 1.0392
No log 0.2917 14 0.9704 -0.0619 0.9704 0.9851
No log 0.3333 16 0.9660 -0.0896 0.9660 0.9829
No log 0.375 18 1.2409 -0.1651 1.2409 1.1140
No log 0.4167 20 1.2128 -0.2042 1.2128 1.1013
No log 0.4583 22 0.9261 0.0468 0.9261 0.9623
No log 0.5 24 0.8535 0.1451 0.8535 0.9238
No log 0.5417 26 0.8340 0.2124 0.8340 0.9132
No log 0.5833 28 0.8189 0.1535 0.8189 0.9049
No log 0.625 30 0.8479 0.1340 0.8479 0.9208
No log 0.6667 32 0.8885 0.1132 0.8885 0.9426
No log 0.7083 34 0.8979 0.1067 0.8979 0.9476
No log 0.75 36 0.9018 0.1114 0.9018 0.9496
No log 0.7917 38 1.0175 0.0509 1.0175 1.0087
No log 0.8333 40 1.2915 -0.0498 1.2915 1.1364
No log 0.875 42 1.3295 -0.0498 1.3295 1.1530
No log 0.9167 44 1.0995 0.0243 1.0995 1.0486
No log 0.9583 46 0.8975 0.1461 0.8975 0.9474
No log 1.0 48 0.8283 0.1179 0.8283 0.9101
No log 1.0417 50 0.8135 0.1245 0.8135 0.9019
No log 1.0833 52 0.8119 0.1111 0.8119 0.9011
No log 1.125 54 0.7785 0.1377 0.7785 0.8823
No log 1.1667 56 0.7700 0.2618 0.7700 0.8775
No log 1.2083 58 0.7668 0.2457 0.7668 0.8757
No log 1.25 60 0.7634 0.2194 0.7634 0.8737
No log 1.2917 62 0.7607 0.2289 0.7607 0.8722
No log 1.3333 64 0.7526 0.2484 0.7526 0.8676
No log 1.375 66 0.7412 0.2365 0.7412 0.8609
No log 1.4167 68 0.7340 0.2280 0.7340 0.8567
No log 1.4583 70 0.7212 0.2662 0.7212 0.8493
No log 1.5 72 0.7040 0.3024 0.7040 0.8390
No log 1.5417 74 0.7344 0.3330 0.7344 0.8570
No log 1.5833 76 0.7217 0.3625 0.7217 0.8495
No log 1.625 78 0.6853 0.3120 0.6853 0.8278
No log 1.6667 80 0.6834 0.3178 0.6834 0.8267
No log 1.7083 82 0.7640 0.3103 0.7640 0.8741
No log 1.75 84 0.7487 0.3942 0.7487 0.8653
No log 1.7917 86 0.7160 0.3304 0.7160 0.8462
No log 1.8333 88 0.7761 0.2694 0.7761 0.8809
No log 1.875 90 0.7079 0.3229 0.7079 0.8413
No log 1.9167 92 0.6995 0.3539 0.6995 0.8364
No log 1.9583 94 0.6905 0.3539 0.6905 0.8309
No log 2.0 96 0.6677 0.3222 0.6677 0.8171
No log 2.0417 98 0.7835 0.2988 0.7835 0.8852
No log 2.0833 100 0.9642 0.2240 0.9642 0.9819
No log 2.125 102 0.8725 0.3005 0.8725 0.9341
No log 2.1667 104 0.6664 0.3996 0.6664 0.8163
No log 2.2083 106 0.7032 0.4053 0.7032 0.8386
No log 2.25 108 0.7543 0.3736 0.7543 0.8685
No log 2.2917 110 0.7044 0.3991 0.7044 0.8393
No log 2.3333 112 0.6896 0.4009 0.6896 0.8304
No log 2.375 114 0.7576 0.4131 0.7576 0.8704
No log 2.4167 116 0.7249 0.4065 0.7249 0.8514
No log 2.4583 118 0.6707 0.4735 0.6707 0.8189
No log 2.5 120 0.7797 0.3990 0.7797 0.8830
No log 2.5417 122 0.9201 0.3302 0.9201 0.9592
No log 2.5833 124 0.9013 0.3154 0.9013 0.9494
No log 2.625 126 0.8043 0.3363 0.8043 0.8968
No log 2.6667 128 0.6579 0.5391 0.6579 0.8111
No log 2.7083 130 0.6389 0.4355 0.6389 0.7993
No log 2.75 132 0.6194 0.4122 0.6194 0.7870
No log 2.7917 134 0.6150 0.3981 0.6150 0.7842
No log 2.8333 136 0.6088 0.4204 0.6088 0.7802
No log 2.875 138 0.6114 0.3934 0.6114 0.7819
No log 2.9167 140 0.6167 0.4367 0.6167 0.7853
No log 2.9583 142 0.6064 0.4009 0.6064 0.7787
No log 3.0 144 0.6303 0.4482 0.6303 0.7939
No log 3.0417 146 0.6974 0.3937 0.6974 0.8351
No log 3.0833 148 0.6853 0.3445 0.6853 0.8278
No log 3.125 150 0.6329 0.3300 0.6329 0.7955
No log 3.1667 152 0.6399 0.4124 0.6399 0.7999
No log 3.2083 154 0.6308 0.3949 0.6308 0.7942
No log 3.25 156 0.6360 0.4237 0.6360 0.7975
No log 3.2917 158 0.6420 0.3754 0.6420 0.8012
No log 3.3333 160 0.6485 0.4389 0.6485 0.8053
No log 3.375 162 0.6669 0.4834 0.6669 0.8167
No log 3.4167 164 0.6653 0.4714 0.6653 0.8157
No log 3.4583 166 0.7854 0.4508 0.7854 0.8862
No log 3.5 168 0.7255 0.4352 0.7255 0.8517
No log 3.5417 170 0.6255 0.4680 0.6255 0.7909
No log 3.5833 172 0.7910 0.4857 0.7910 0.8894
No log 3.625 174 0.9313 0.3980 0.9313 0.9650
No log 3.6667 176 0.8486 0.4235 0.8486 0.9212
No log 3.7083 178 0.6715 0.3784 0.6715 0.8194
No log 3.75 180 0.5994 0.4505 0.5994 0.7742
No log 3.7917 182 0.6240 0.4862 0.6240 0.7899
No log 3.8333 184 0.6243 0.5233 0.6243 0.7901
No log 3.875 186 0.6360 0.4674 0.6360 0.7975
No log 3.9167 188 0.8116 0.4205 0.8116 0.9009
No log 3.9583 190 0.8092 0.4323 0.8092 0.8995
No log 4.0 192 0.6492 0.4365 0.6492 0.8057
No log 4.0417 194 0.7005 0.3914 0.7005 0.8369
No log 4.0833 196 0.8290 0.4126 0.8290 0.9105
No log 4.125 198 0.7698 0.4016 0.7698 0.8774
No log 4.1667 200 0.6368 0.4452 0.6368 0.7980
No log 4.2083 202 0.6202 0.4355 0.6202 0.7875
No log 4.25 204 0.6092 0.4252 0.6092 0.7805
No log 4.2917 206 0.6122 0.4432 0.6122 0.7824
No log 4.3333 208 0.6254 0.4677 0.6254 0.7908
No log 4.375 210 0.6450 0.4717 0.6450 0.8031
No log 4.4167 212 0.6390 0.4803 0.6390 0.7994
No log 4.4583 214 0.7000 0.4062 0.7000 0.8366
No log 4.5 216 0.6843 0.4194 0.6843 0.8272
No log 4.5417 218 0.6379 0.4482 0.6379 0.7987
No log 4.5833 220 0.6817 0.4975 0.6817 0.8256
No log 4.625 222 0.6575 0.4664 0.6575 0.8109
No log 4.6667 224 0.6471 0.4661 0.6471 0.8044
No log 4.7083 226 0.6499 0.4417 0.6499 0.8062
No log 4.75 228 0.6465 0.4328 0.6465 0.8041
No log 4.7917 230 0.6477 0.4835 0.6477 0.8048
No log 4.8333 232 0.6497 0.4643 0.6497 0.8061
No log 4.875 234 0.6504 0.4542 0.6504 0.8065
No log 4.9167 236 0.6550 0.4593 0.6550 0.8093
No log 4.9583 238 0.6606 0.4557 0.6606 0.8127
No log 5.0 240 0.6790 0.4187 0.6790 0.8240
No log 5.0417 242 0.6542 0.4253 0.6542 0.8088
No log 5.0833 244 0.6315 0.4153 0.6315 0.7947
No log 5.125 246 0.6341 0.4140 0.6341 0.7963
No log 5.1667 248 0.6288 0.4581 0.6288 0.7930
No log 5.2083 250 0.6037 0.4061 0.6037 0.7770
No log 5.25 252 0.6469 0.5092 0.6469 0.8043
No log 5.2917 254 0.6319 0.4612 0.6319 0.7949
No log 5.3333 256 0.6151 0.4634 0.6151 0.7843
No log 5.375 258 0.6344 0.4716 0.6344 0.7965
No log 5.4167 260 0.6238 0.4400 0.6238 0.7898
No log 5.4583 262 0.7026 0.4472 0.7026 0.8382
No log 5.5 264 0.6593 0.4164 0.6593 0.8120
No log 5.5417 266 0.5971 0.4230 0.5971 0.7727
No log 5.5833 268 0.6271 0.4978 0.6271 0.7919
No log 5.625 270 0.6002 0.4682 0.6002 0.7748
No log 5.6667 272 0.5873 0.4850 0.5873 0.7663
No log 5.7083 274 0.5890 0.4850 0.5890 0.7674
No log 5.75 276 0.6095 0.5344 0.6095 0.7807
No log 5.7917 278 0.6026 0.5326 0.6026 0.7763
No log 5.8333 280 0.5969 0.5326 0.5969 0.7726
No log 5.875 282 0.5851 0.5274 0.5851 0.7649
No log 5.9167 284 0.6040 0.5237 0.6040 0.7771
No log 5.9583 286 0.6583 0.5044 0.6583 0.8113
No log 6.0 288 0.6126 0.5056 0.6126 0.7827
No log 6.0417 290 0.5898 0.5789 0.5898 0.7680
No log 6.0833 292 0.5903 0.5562 0.5903 0.7683
No log 6.125 294 0.5770 0.5464 0.5770 0.7596
No log 6.1667 296 0.6314 0.5118 0.6314 0.7946
No log 6.2083 298 0.6353 0.5294 0.6353 0.7971
No log 6.25 300 0.5842 0.5396 0.5842 0.7643
No log 6.2917 302 0.5778 0.5767 0.5778 0.7601
No log 6.3333 304 0.5932 0.4857 0.5932 0.7702
No log 6.375 306 0.5940 0.4857 0.5940 0.7707
No log 6.4167 308 0.5621 0.5408 0.5621 0.7497
No log 6.4583 310 0.5901 0.5590 0.5901 0.7682
No log 6.5 312 0.6743 0.5021 0.6743 0.8211
No log 6.5417 314 0.6676 0.5007 0.6676 0.8171
No log 6.5833 316 0.5775 0.5612 0.5775 0.7599
No log 6.625 318 0.5722 0.4664 0.5722 0.7564
No log 6.6667 320 0.6249 0.5155 0.6249 0.7905
No log 6.7083 322 0.5772 0.5113 0.5772 0.7597
No log 6.75 324 0.5897 0.5854 0.5897 0.7679
No log 6.7917 326 0.8223 0.4216 0.8223 0.9068
No log 6.8333 328 0.9507 0.3969 0.9507 0.9750
No log 6.875 330 0.8498 0.4216 0.8498 0.9219
No log 6.9167 332 0.6361 0.5391 0.6361 0.7976
No log 6.9583 334 0.5727 0.5446 0.5727 0.7568
No log 7.0 336 0.6186 0.4976 0.6186 0.7865
No log 7.0417 338 0.6108 0.4976 0.6108 0.7815
No log 7.0833 340 0.5816 0.4948 0.5816 0.7627
No log 7.125 342 0.5886 0.4498 0.5886 0.7672
No log 7.1667 344 0.5999 0.4235 0.5999 0.7746
No log 7.2083 346 0.5928 0.4305 0.5928 0.7700
No log 7.25 348 0.5894 0.4465 0.5894 0.7677
No log 7.2917 350 0.5869 0.4887 0.5869 0.7661
No log 7.3333 352 0.5952 0.4976 0.5952 0.7715
No log 7.375 354 0.6014 0.5162 0.6014 0.7755
No log 7.4167 356 0.6126 0.5265 0.6126 0.7827
No log 7.4583 358 0.6180 0.5337 0.6180 0.7861
No log 7.5 360 0.6159 0.4820 0.6159 0.7848
No log 7.5417 362 0.6266 0.5011 0.6266 0.7916
No log 7.5833 364 0.6387 0.5132 0.6387 0.7992
No log 7.625 366 0.6870 0.4807 0.6870 0.8288
No log 7.6667 368 0.7992 0.4382 0.7992 0.8940
No log 7.7083 370 0.7851 0.4521 0.7851 0.8861
No log 7.75 372 0.6863 0.4492 0.6863 0.8285
No log 7.7917 374 0.6545 0.4921 0.6545 0.8090
No log 7.8333 376 0.6756 0.4491 0.6756 0.8219
No log 7.875 378 0.6657 0.4499 0.6657 0.8159
No log 7.9167 380 0.6345 0.4184 0.6345 0.7965
No log 7.9583 382 0.6350 0.4066 0.6350 0.7969
No log 8.0 384 0.6786 0.4165 0.6786 0.8238
No log 8.0417 386 0.7133 0.4247 0.7133 0.8446
No log 8.0833 388 0.6598 0.4165 0.6598 0.8123
No log 8.125 390 0.6397 0.4231 0.6397 0.7998
No log 8.1667 392 0.6108 0.4467 0.6108 0.7816
No log 8.2083 394 0.6460 0.4060 0.6460 0.8038
No log 8.25 396 0.6332 0.4157 0.6332 0.7957
No log 8.2917 398 0.5918 0.4825 0.5918 0.7693
No log 8.3333 400 0.5933 0.4828 0.5933 0.7703
No log 8.375 402 0.5910 0.4859 0.5910 0.7688
No log 8.4167 404 0.5977 0.4622 0.5977 0.7731
No log 8.4583 406 0.6700 0.4598 0.6700 0.8186
No log 8.5 408 0.6610 0.4308 0.6610 0.8130
No log 8.5417 410 0.5979 0.4523 0.5979 0.7732
No log 8.5833 412 0.5901 0.4291 0.5901 0.7682
No log 8.625 414 0.5868 0.4291 0.5868 0.7661
No log 8.6667 416 0.5985 0.4080 0.5985 0.7736
No log 8.7083 418 0.6079 0.4780 0.6079 0.7797
No log 8.75 420 0.5944 0.4555 0.5944 0.7710
No log 8.7917 422 0.5704 0.4737 0.5704 0.7552
No log 8.8333 424 0.5681 0.4844 0.5681 0.7537
No log 8.875 426 0.5747 0.4949 0.5747 0.7581
No log 8.9167 428 0.5758 0.4998 0.5758 0.7588
No log 8.9583 430 0.5765 0.4909 0.5765 0.7592
No log 9.0 432 0.5772 0.4741 0.5772 0.7597
No log 9.0417 434 0.5762 0.5011 0.5762 0.7591
No log 9.0833 436 0.5833 0.4279 0.5833 0.7637
No log 9.125 438 0.5690 0.4212 0.5690 0.7544
No log 9.1667 440 0.5675 0.4704 0.5675 0.7533
No log 9.2083 442 0.5732 0.4662 0.5732 0.7571
No log 9.25 444 0.5857 0.4565 0.5857 0.7653
No log 9.2917 446 0.6082 0.4605 0.6082 0.7799
No log 9.3333 448 0.6220 0.5064 0.6220 0.7886
No log 9.375 450 0.6478 0.5654 0.6478 0.8049
No log 9.4167 452 0.6683 0.4973 0.6683 0.8175
No log 9.4583 454 0.6384 0.4837 0.6384 0.7990
No log 9.5 456 0.5663 0.5049 0.5663 0.7526
No log 9.5417 458 0.5703 0.4633 0.5703 0.7552
No log 9.5833 460 0.6062 0.4754 0.6062 0.7786
No log 9.625 462 0.5772 0.4707 0.5772 0.7597
No log 9.6667 464 0.5749 0.4279 0.5749 0.7582
No log 9.7083 466 0.5760 0.4443 0.5760 0.7589
No log 9.75 468 0.5906 0.4577 0.5906 0.7685
No log 9.7917 470 0.6333 0.4830 0.6333 0.7958
No log 9.8333 472 0.6011 0.4448 0.6011 0.7753
No log 9.875 474 0.5855 0.4849 0.5855 0.7651
No log 9.9167 476 0.5823 0.4884 0.5823 0.7631
No log 9.9583 478 0.5738 0.4710 0.5738 0.7575
No log 10.0 480 0.5807 0.4781 0.5807 0.7621
No log 10.0417 482 0.5904 0.4828 0.5904 0.7684
No log 10.0833 484 0.5954 0.5101 0.5954 0.7716
No log 10.125 486 0.5834 0.5363 0.5834 0.7638
No log 10.1667 488 0.5647 0.5189 0.5647 0.7515
No log 10.2083 490 0.5642 0.5127 0.5642 0.7511
No log 10.25 492 0.5583 0.5189 0.5583 0.7472
No log 10.2917 494 0.5629 0.5231 0.5629 0.7502
No log 10.3333 496 0.5773 0.4869 0.5773 0.7598
No log 10.375 498 0.5478 0.5046 0.5478 0.7401
0.4042 10.4167 500 0.5506 0.5067 0.5506 0.7420
0.4042 10.4583 502 0.5751 0.5329 0.5751 0.7583
0.4042 10.5 504 0.5876 0.5622 0.5876 0.7665
0.4042 10.5417 506 0.5691 0.5075 0.5691 0.7544
0.4042 10.5833 508 0.5609 0.4650 0.5609 0.7490
0.4042 10.625 510 0.5639 0.4600 0.5639 0.7510
0.4042 10.6667 512 0.6016 0.4946 0.6016 0.7757
0.4042 10.7083 514 0.6882 0.4648 0.6882 0.8296
0.4042 10.75 516 0.7808 0.4161 0.7808 0.8837

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
3
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for MayBashendy/ArabicNewSplits8_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k9_task2_organization

Finetuned
(4222)
this model