ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run3_AugV5_k9_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6220
  • Qwk: 0.5031
  • Mse: 0.6220
  • Rmse: 0.7887

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0426 2 4.2729 -0.0091 4.2729 2.0671
No log 0.0851 4 2.4493 0.0564 2.4493 1.5650
No log 0.1277 6 1.9854 0.0984 1.9854 1.4090
No log 0.1702 8 1.4098 0.0258 1.4098 1.1874
No log 0.2128 10 0.9719 -0.0104 0.9719 0.9859
No log 0.2553 12 0.8737 0.0261 0.8737 0.9347
No log 0.2979 14 0.8067 0.1977 0.8067 0.8982
No log 0.3404 16 0.7916 0.1797 0.7916 0.8897
No log 0.3830 18 0.8956 0.1610 0.8956 0.9464
No log 0.4255 20 1.2100 0.0469 1.2100 1.1000
No log 0.4681 22 1.2783 0.0262 1.2783 1.1306
No log 0.5106 24 1.3993 0.0514 1.3993 1.1829
No log 0.5532 26 1.2659 0.0 1.2659 1.1251
No log 0.5957 28 1.2265 0.0 1.2265 1.1075
No log 0.6383 30 0.9824 0.1339 0.9824 0.9912
No log 0.6809 32 0.7797 0.2659 0.7797 0.8830
No log 0.7234 34 0.7165 0.3278 0.7165 0.8465
No log 0.7660 36 0.7520 0.2769 0.7520 0.8672
No log 0.8085 38 0.7351 0.3431 0.7351 0.8574
No log 0.8511 40 0.8688 0.2609 0.8688 0.9321
No log 0.8936 42 1.1098 0.0924 1.1098 1.0535
No log 0.9362 44 1.2935 0.1395 1.2935 1.1373
No log 0.9787 46 1.3851 0.1429 1.3851 1.1769
No log 1.0213 48 1.1903 0.1495 1.1903 1.0910
No log 1.0638 50 0.9889 0.1954 0.9889 0.9944
No log 1.1064 52 0.8813 0.2772 0.8813 0.9388
No log 1.1489 54 0.8244 0.2850 0.8244 0.9080
No log 1.1915 56 0.8478 0.2919 0.8478 0.9208
No log 1.2340 58 1.2281 0.3278 1.2281 1.1082
No log 1.2766 60 1.7780 0.1983 1.7780 1.3334
No log 1.3191 62 1.8522 0.2012 1.8522 1.3610
No log 1.3617 64 1.5459 0.2361 1.5459 1.2433
No log 1.4043 66 0.9551 0.3570 0.9551 0.9773
No log 1.4468 68 0.7083 0.4410 0.7083 0.8416
No log 1.4894 70 0.7142 0.4324 0.7142 0.8451
No log 1.5319 72 0.7339 0.4324 0.7339 0.8567
No log 1.5745 74 0.7664 0.4290 0.7664 0.8755
No log 1.6170 76 0.7347 0.4431 0.7347 0.8571
No log 1.6596 78 0.7028 0.4114 0.7028 0.8383
No log 1.7021 80 0.7740 0.3139 0.7740 0.8798
No log 1.7447 82 0.8599 0.2686 0.8599 0.9273
No log 1.7872 84 0.8692 0.2792 0.8692 0.9323
No log 1.8298 86 0.7788 0.3226 0.7788 0.8825
No log 1.8723 88 0.7680 0.3447 0.7680 0.8763
No log 1.9149 90 0.7179 0.4563 0.7179 0.8473
No log 1.9574 92 0.6748 0.4495 0.6748 0.8214
No log 2.0 94 0.6589 0.4634 0.6589 0.8117
No log 2.0426 96 0.7380 0.4222 0.7380 0.8591
No log 2.0851 98 0.7640 0.3874 0.7640 0.8741
No log 2.1277 100 0.7341 0.4633 0.7341 0.8568
No log 2.1702 102 0.8039 0.4086 0.8039 0.8966
No log 2.2128 104 0.6840 0.4708 0.6840 0.8270
No log 2.2553 106 0.6802 0.4862 0.6802 0.8247
No log 2.2979 108 0.6953 0.4883 0.6953 0.8339
No log 2.3404 110 0.7035 0.4890 0.7035 0.8387
No log 2.3830 112 0.7881 0.4242 0.7881 0.8877
No log 2.4255 114 0.8346 0.4260 0.8346 0.9136
No log 2.4681 116 0.7795 0.4997 0.7795 0.8829
No log 2.5106 118 0.7477 0.5095 0.7477 0.8647
No log 2.5532 120 0.8223 0.4572 0.8223 0.9068
No log 2.5957 122 0.9468 0.4344 0.9468 0.9730
No log 2.6383 124 1.0364 0.4241 1.0364 1.0181
No log 2.6809 126 1.1443 0.3943 1.1443 1.0697
No log 2.7234 128 1.0194 0.4053 1.0194 1.0097
No log 2.7660 130 0.8134 0.4502 0.8134 0.9019
No log 2.8085 132 0.7468 0.4689 0.7468 0.8642
No log 2.8511 134 0.6996 0.5201 0.6996 0.8364
No log 2.8936 136 0.6836 0.4790 0.6836 0.8268
No log 2.9362 138 0.6940 0.4532 0.6940 0.8330
No log 2.9787 140 0.6894 0.4587 0.6894 0.8303
No log 3.0213 142 0.7001 0.4587 0.7001 0.8367
No log 3.0638 144 0.7240 0.4557 0.7240 0.8509
No log 3.1064 146 0.7773 0.4173 0.7773 0.8816
No log 3.1489 148 0.8353 0.4327 0.8353 0.9139
No log 3.1915 150 0.8355 0.4228 0.8355 0.9140
No log 3.2340 152 0.7437 0.4602 0.7437 0.8624
No log 3.2766 154 0.7600 0.4315 0.7600 0.8718
No log 3.3191 156 0.8410 0.3950 0.8410 0.9171
No log 3.3617 158 0.7785 0.3860 0.7785 0.8823
No log 3.4043 160 0.6663 0.4240 0.6663 0.8162
No log 3.4468 162 0.6523 0.4531 0.6523 0.8077
No log 3.4894 164 0.7148 0.4755 0.7148 0.8455
No log 3.5319 166 0.8954 0.4714 0.8954 0.9462
No log 3.5745 168 1.0368 0.4060 1.0368 1.0182
No log 3.6170 170 1.0616 0.4124 1.0616 1.0304
No log 3.6596 172 1.0842 0.4062 1.0842 1.0413
No log 3.7021 174 0.8773 0.4411 0.8773 0.9366
No log 3.7447 176 0.7011 0.5299 0.7011 0.8373
No log 3.7872 178 0.6652 0.5082 0.6652 0.8156
No log 3.8298 180 0.6894 0.5482 0.6894 0.8303
No log 3.8723 182 0.6861 0.5405 0.6861 0.8283
No log 3.9149 184 0.7020 0.5556 0.7020 0.8379
No log 3.9574 186 0.7290 0.4896 0.7290 0.8538
No log 4.0 188 0.7216 0.4674 0.7216 0.8495
No log 4.0426 190 0.6787 0.5078 0.6787 0.8238
No log 4.0851 192 0.6666 0.5141 0.6666 0.8165
No log 4.1277 194 0.7076 0.4789 0.7076 0.8412
No log 4.1702 196 0.6865 0.5196 0.6865 0.8286
No log 4.2128 198 0.6570 0.5311 0.6570 0.8106
No log 4.2553 200 0.7025 0.4737 0.7025 0.8382
No log 4.2979 202 0.8629 0.4276 0.8629 0.9289
No log 4.3404 204 0.9039 0.4152 0.9039 0.9507
No log 4.3830 206 0.8206 0.4459 0.8206 0.9059
No log 4.4255 208 0.7428 0.5196 0.7428 0.8619
No log 4.4681 210 0.6923 0.4958 0.6923 0.8321
No log 4.5106 212 0.6877 0.4834 0.6877 0.8293
No log 4.5532 214 0.7000 0.4949 0.7000 0.8367
No log 4.5957 216 0.8042 0.4049 0.8042 0.8968
No log 4.6383 218 0.9265 0.3856 0.9265 0.9625
No log 4.6809 220 0.9679 0.4033 0.9679 0.9838
No log 4.7234 222 0.9097 0.4250 0.9097 0.9538
No log 4.7660 224 0.8253 0.4720 0.8253 0.9085
No log 4.8085 226 0.8147 0.4735 0.8147 0.9026
No log 4.8511 228 0.8962 0.4576 0.8962 0.9467
No log 4.8936 230 1.0199 0.4479 1.0199 1.0099
No log 4.9362 232 1.0190 0.4664 1.0190 1.0095
No log 4.9787 234 0.8215 0.5202 0.8215 0.9063
No log 5.0213 236 0.6911 0.5555 0.6911 0.8313
No log 5.0638 238 0.6523 0.4927 0.6523 0.8077
No log 5.1064 240 0.6834 0.4845 0.6834 0.8267
No log 5.1489 242 0.7690 0.5384 0.7690 0.8769
No log 5.1915 244 0.9164 0.4393 0.9164 0.9573
No log 5.2340 246 0.8709 0.4233 0.8709 0.9332
No log 5.2766 248 0.7602 0.4879 0.7602 0.8719
No log 5.3191 250 0.6960 0.5324 0.6960 0.8342
No log 5.3617 252 0.6561 0.5490 0.6561 0.8100
No log 5.4043 254 0.6128 0.5193 0.6128 0.7828
No log 5.4468 256 0.6010 0.5138 0.6010 0.7753
No log 5.4894 258 0.5972 0.5433 0.5972 0.7728
No log 5.5319 260 0.5778 0.5451 0.5778 0.7601
No log 5.5745 262 0.6018 0.5438 0.6018 0.7758
No log 5.6170 264 0.6364 0.5357 0.6364 0.7977
No log 5.6596 266 0.6502 0.5368 0.6502 0.8064
No log 5.7021 268 0.6630 0.5445 0.6630 0.8142
No log 5.7447 270 0.6989 0.5069 0.6989 0.8360
No log 5.7872 272 0.6985 0.5302 0.6985 0.8358
No log 5.8298 274 0.6813 0.5508 0.6813 0.8254
No log 5.8723 276 0.6803 0.5675 0.6803 0.8248
No log 5.9149 278 0.6697 0.5297 0.6697 0.8184
No log 5.9574 280 0.6411 0.5032 0.6411 0.8007
No log 6.0 282 0.6157 0.5041 0.6157 0.7847
No log 6.0426 284 0.5912 0.5641 0.5912 0.7689
No log 6.0851 286 0.6137 0.5075 0.6137 0.7834
No log 6.1277 288 0.6690 0.5151 0.6690 0.8179
No log 6.1702 290 0.6608 0.5164 0.6608 0.8129
No log 6.2128 292 0.6207 0.5266 0.6207 0.7878
No log 6.2553 294 0.6788 0.4859 0.6788 0.8239
No log 6.2979 296 0.7775 0.4352 0.7775 0.8818
No log 6.3404 298 0.7445 0.4461 0.7445 0.8628
No log 6.3830 300 0.6474 0.5155 0.6474 0.8046
No log 6.4255 302 0.6260 0.5200 0.6260 0.7912
No log 6.4681 304 0.6910 0.5018 0.6910 0.8313
No log 6.5106 306 0.6950 0.5006 0.6950 0.8337
No log 6.5532 308 0.6496 0.5789 0.6496 0.8060
No log 6.5957 310 0.6443 0.5073 0.6443 0.8027
No log 6.6383 312 0.7282 0.4467 0.7282 0.8534
No log 6.6809 314 0.7840 0.4139 0.7840 0.8854
No log 6.7234 316 0.7525 0.4079 0.7525 0.8675
No log 6.7660 318 0.6904 0.5042 0.6904 0.8309
No log 6.8085 320 0.6610 0.5120 0.6610 0.8130
No log 6.8511 322 0.6564 0.5233 0.6564 0.8102
No log 6.8936 324 0.6610 0.5233 0.6610 0.8130
No log 6.9362 326 0.6606 0.5067 0.6606 0.8128
No log 6.9787 328 0.7031 0.4925 0.7031 0.8385
No log 7.0213 330 0.7821 0.4390 0.7821 0.8844
No log 7.0638 332 0.7948 0.4273 0.7948 0.8915
No log 7.1064 334 0.7614 0.4515 0.7614 0.8726
No log 7.1489 336 0.6897 0.5598 0.6897 0.8305
No log 7.1915 338 0.6732 0.5535 0.6732 0.8205
No log 7.2340 340 0.6605 0.5303 0.6605 0.8127
No log 7.2766 342 0.6523 0.5087 0.6523 0.8077
No log 7.3191 344 0.6915 0.4971 0.6915 0.8316
No log 7.3617 346 0.7260 0.4728 0.7260 0.8520
No log 7.4043 348 0.7592 0.4646 0.7592 0.8713
No log 7.4468 350 0.7441 0.4980 0.7441 0.8626
No log 7.4894 352 0.6598 0.5743 0.6598 0.8123
No log 7.5319 354 0.6532 0.6019 0.6532 0.8082
No log 7.5745 356 0.6517 0.6019 0.6517 0.8073
No log 7.6170 358 0.6458 0.5392 0.6458 0.8036
No log 7.6596 360 0.6432 0.5193 0.6432 0.8020
No log 7.7021 362 0.6289 0.5063 0.6289 0.7930
No log 7.7447 364 0.6182 0.5199 0.6182 0.7863
No log 7.7872 366 0.6138 0.5364 0.6138 0.7835
No log 7.8298 368 0.6132 0.5153 0.6132 0.7831
No log 7.8723 370 0.6074 0.5138 0.6074 0.7793
No log 7.9149 372 0.6163 0.5462 0.6163 0.7851
No log 7.9574 374 0.5917 0.4914 0.5917 0.7692
No log 8.0 376 0.5865 0.5314 0.5865 0.7658
No log 8.0426 378 0.5911 0.5525 0.5911 0.7688
No log 8.0851 380 0.5998 0.5668 0.5998 0.7745
No log 8.1277 382 0.6350 0.5376 0.6350 0.7969
No log 8.1702 384 0.7261 0.5096 0.7261 0.8521
No log 8.2128 386 0.7495 0.4927 0.7495 0.8657
No log 8.2553 388 0.6849 0.53 0.6849 0.8276
No log 8.2979 390 0.6284 0.5282 0.6284 0.7927
No log 8.3404 392 0.6234 0.5450 0.6234 0.7896
No log 8.3830 394 0.6281 0.5024 0.6281 0.7925
No log 8.4255 396 0.6481 0.4707 0.6481 0.8050
No log 8.4681 398 0.6900 0.4341 0.6900 0.8307
No log 8.5106 400 0.7158 0.4233 0.7158 0.8461
No log 8.5532 402 0.6765 0.4454 0.6765 0.8225
No log 8.5957 404 0.6934 0.4654 0.6934 0.8327
No log 8.6383 406 0.6858 0.4483 0.6858 0.8282
No log 8.6809 408 0.6912 0.4736 0.6912 0.8314
No log 8.7234 410 0.7168 0.4283 0.7168 0.8467
No log 8.7660 412 0.7677 0.4716 0.7677 0.8762
No log 8.8085 414 0.7809 0.4608 0.7809 0.8837
No log 8.8511 416 0.7194 0.4349 0.7194 0.8482
No log 8.8936 418 0.6579 0.4729 0.6579 0.8111
No log 8.9362 420 0.6356 0.4705 0.6356 0.7973
No log 8.9787 422 0.6317 0.4945 0.6317 0.7948
No log 9.0213 424 0.6511 0.4880 0.6511 0.8069
No log 9.0638 426 0.6999 0.4394 0.6999 0.8366
No log 9.1064 428 0.7213 0.4333 0.7213 0.8493
No log 9.1489 430 0.7454 0.4515 0.7454 0.8634
No log 9.1915 432 0.7435 0.4696 0.7435 0.8622
No log 9.2340 434 0.7584 0.4566 0.7584 0.8708
No log 9.2766 436 0.7714 0.4628 0.7714 0.8783
No log 9.3191 438 0.7728 0.4663 0.7728 0.8791
No log 9.3617 440 0.7354 0.5214 0.7354 0.8575
No log 9.4043 442 0.7196 0.5610 0.7196 0.8483
No log 9.4468 444 0.7370 0.5595 0.7370 0.8585
No log 9.4894 446 0.8053 0.4778 0.8053 0.8974
No log 9.5319 448 0.9010 0.4228 0.9010 0.9492
No log 9.5745 450 0.9405 0.3814 0.9405 0.9698
No log 9.6170 452 0.9676 0.3869 0.9676 0.9837
No log 9.6596 454 0.8391 0.4495 0.8391 0.9160
No log 9.7021 456 0.6732 0.4474 0.6732 0.8205
No log 9.7447 458 0.6078 0.4992 0.6078 0.7796
No log 9.7872 460 0.6010 0.4952 0.6010 0.7752
No log 9.8298 462 0.6021 0.4952 0.6021 0.7760
No log 9.8723 464 0.6207 0.4512 0.6207 0.7878
No log 9.9149 466 0.6326 0.4577 0.6326 0.7954
No log 9.9574 468 0.6256 0.4716 0.6256 0.7910
No log 10.0 470 0.6330 0.4477 0.6330 0.7956
No log 10.0426 472 0.6544 0.4820 0.6544 0.8090
No log 10.0851 474 0.7100 0.5237 0.7100 0.8426
No log 10.1277 476 0.7719 0.5585 0.7719 0.8786
No log 10.1702 478 0.8097 0.5575 0.8097 0.8998
No log 10.2128 480 0.8143 0.5575 0.8143 0.9024
No log 10.2553 482 0.7374 0.5638 0.7374 0.8587
No log 10.2979 484 0.7128 0.5414 0.7128 0.8443
No log 10.3404 486 0.6929 0.5324 0.6929 0.8324
No log 10.3830 488 0.6502 0.5069 0.6502 0.8063
No log 10.4255 490 0.6540 0.53 0.6540 0.8087
No log 10.4681 492 0.6529 0.5447 0.6529 0.8081
No log 10.5106 494 0.6134 0.5454 0.6134 0.7832
No log 10.5532 496 0.5886 0.5665 0.5886 0.7672
No log 10.5957 498 0.5886 0.5368 0.5886 0.7672
0.3691 10.6383 500 0.6029 0.5386 0.6029 0.7765
0.3691 10.6809 502 0.5783 0.5191 0.5783 0.7605
0.3691 10.7234 504 0.5645 0.5147 0.5645 0.7513
0.3691 10.7660 506 0.5786 0.5433 0.5786 0.7607
0.3691 10.8085 508 0.5745 0.5251 0.5745 0.7580
0.3691 10.8511 510 0.5761 0.4971 0.5761 0.7590
0.3691 10.8936 512 0.5901 0.4862 0.5901 0.7682
0.3691 10.9362 514 0.6031 0.4824 0.6031 0.7766
0.3691 10.9787 516 0.6220 0.5031 0.6220 0.7887

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
3
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for MayBashendy/ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run3_AugV5_k9_task2_organization

Finetuned
(4222)
this model