ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run3_AugV5_k6_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5739
  • Qwk: 0.4276
  • Mse: 0.5739
  • Rmse: 0.7575

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0625 2 4.2977 -0.0038 4.2977 2.0731
No log 0.125 4 2.1232 0.0822 2.1232 1.4571
No log 0.1875 6 1.1544 0.0154 1.1544 1.0744
No log 0.25 8 0.8479 0.1616 0.8479 0.9208
No log 0.3125 10 0.8473 0.0868 0.8473 0.9205
No log 0.375 12 0.8654 0.1265 0.8654 0.9303
No log 0.4375 14 0.8437 0.2643 0.8437 0.9185
No log 0.5 16 0.9393 0.1416 0.9393 0.9692
No log 0.5625 18 1.3947 0.0811 1.3947 1.1810
No log 0.625 20 1.3401 0.1007 1.3401 1.1576
No log 0.6875 22 0.9484 0.1741 0.9484 0.9739
No log 0.75 24 0.6702 0.3947 0.6702 0.8187
No log 0.8125 26 0.6545 0.4124 0.6545 0.8090
No log 0.875 28 0.6464 0.3915 0.6464 0.8040
No log 0.9375 30 0.7866 0.3203 0.7866 0.8869
No log 1.0 32 1.0283 0.3178 1.0283 1.0140
No log 1.0625 34 1.3757 0.2071 1.3757 1.1729
No log 1.125 36 1.4568 0.2133 1.4568 1.2070
No log 1.1875 38 1.0331 0.2617 1.0331 1.0164
No log 1.25 40 0.6794 0.3591 0.6794 0.8243
No log 1.3125 42 0.6357 0.4112 0.6357 0.7973
No log 1.375 44 0.6701 0.3523 0.6701 0.8186
No log 1.4375 46 0.6743 0.3631 0.6743 0.8212
No log 1.5 48 0.7662 0.3254 0.7662 0.8753
No log 1.5625 50 0.7815 0.3219 0.7815 0.8840
No log 1.625 52 0.7622 0.3469 0.7622 0.8730
No log 1.6875 54 0.6263 0.4112 0.6263 0.7914
No log 1.75 56 0.5540 0.4822 0.5540 0.7443
No log 1.8125 58 0.5557 0.4852 0.5557 0.7454
No log 1.875 60 0.5691 0.4821 0.5691 0.7544
No log 1.9375 62 0.7689 0.4786 0.7689 0.8769
No log 2.0 64 0.9560 0.3972 0.9560 0.9778
No log 2.0625 66 0.6650 0.4814 0.6650 0.8155
No log 2.125 68 0.5569 0.4675 0.5569 0.7463
No log 2.1875 70 0.5729 0.4842 0.5729 0.7569
No log 2.25 72 0.5574 0.4873 0.5574 0.7466
No log 2.3125 74 0.5483 0.4897 0.5483 0.7405
No log 2.375 76 0.6252 0.4746 0.6252 0.7907
No log 2.4375 78 0.5725 0.4645 0.5725 0.7567
No log 2.5 80 0.5522 0.5339 0.5522 0.7431
No log 2.5625 82 0.5593 0.5539 0.5593 0.7479
No log 2.625 84 0.5673 0.5497 0.5673 0.7532
No log 2.6875 86 0.5748 0.5870 0.5748 0.7582
No log 2.75 88 0.5826 0.5541 0.5826 0.7633
No log 2.8125 90 0.6101 0.5777 0.6101 0.7811
No log 2.875 92 0.7987 0.5521 0.7987 0.8937
No log 2.9375 94 0.9902 0.4420 0.9902 0.9951
No log 3.0 96 0.8812 0.5088 0.8812 0.9387
No log 3.0625 98 0.6936 0.5152 0.6936 0.8328
No log 3.125 100 0.6367 0.5309 0.6367 0.7979
No log 3.1875 102 0.6148 0.5649 0.6148 0.7841
No log 3.25 104 0.6116 0.5510 0.6116 0.7820
No log 3.3125 106 0.7515 0.5048 0.7515 0.8669
No log 3.375 108 0.8913 0.3364 0.8913 0.9441
No log 3.4375 110 0.7429 0.4707 0.7429 0.8619
No log 3.5 112 0.6454 0.5011 0.6454 0.8034
No log 3.5625 114 0.6354 0.4985 0.6354 0.7971
No log 3.625 116 0.6257 0.4936 0.6257 0.7910
No log 3.6875 118 0.6786 0.5358 0.6786 0.8238
No log 3.75 120 0.6558 0.5139 0.6558 0.8098
No log 3.8125 122 0.6004 0.5040 0.6004 0.7749
No log 3.875 124 0.6033 0.4688 0.6033 0.7767
No log 3.9375 126 0.5999 0.4579 0.5999 0.7745
No log 4.0 128 0.6846 0.5074 0.6846 0.8274
No log 4.0625 130 0.6587 0.5201 0.6587 0.8116
No log 4.125 132 0.6016 0.4833 0.6016 0.7756
No log 4.1875 134 0.7315 0.5163 0.7315 0.8553
No log 4.25 136 0.7535 0.5163 0.7535 0.8681
No log 4.3125 138 0.6418 0.5642 0.6418 0.8011
No log 4.375 140 0.6989 0.5544 0.6989 0.8360
No log 4.4375 142 0.6844 0.5309 0.6844 0.8273
No log 4.5 144 0.6826 0.5436 0.6826 0.8262
No log 4.5625 146 0.8225 0.5018 0.8225 0.9069
No log 4.625 148 0.6747 0.5265 0.6747 0.8214
No log 4.6875 150 0.6026 0.4920 0.6026 0.7762
No log 4.75 152 0.6082 0.4908 0.6082 0.7799
No log 4.8125 154 0.6400 0.4789 0.6400 0.8000
No log 4.875 156 0.6100 0.4992 0.6100 0.7810
No log 4.9375 158 0.6104 0.5057 0.6104 0.7813
No log 5.0 160 0.6492 0.5019 0.6492 0.8057
No log 5.0625 162 0.6546 0.5018 0.6546 0.8091
No log 5.125 164 0.6223 0.4733 0.6223 0.7889
No log 5.1875 166 0.6171 0.4608 0.6171 0.7855
No log 5.25 168 0.6122 0.4971 0.6122 0.7825
No log 5.3125 170 0.6240 0.5384 0.6240 0.7900
No log 5.375 172 0.6565 0.5135 0.6565 0.8102
No log 5.4375 174 0.6365 0.5247 0.6365 0.7978
No log 5.5 176 0.6053 0.4984 0.6053 0.7780
No log 5.5625 178 0.5980 0.5567 0.5980 0.7733
No log 5.625 180 0.6076 0.5065 0.6076 0.7795
No log 5.6875 182 0.6298 0.4916 0.6298 0.7936
No log 5.75 184 0.8423 0.4590 0.8423 0.9178
No log 5.8125 186 0.7731 0.5019 0.7731 0.8793
No log 5.875 188 0.6221 0.5050 0.6221 0.7887
No log 5.9375 190 0.6341 0.4781 0.6341 0.7963
No log 6.0 192 0.6293 0.4695 0.6293 0.7933
No log 6.0625 194 0.6659 0.5438 0.6659 0.8160
No log 6.125 196 0.6849 0.5420 0.6849 0.8276
No log 6.1875 198 0.6442 0.4757 0.6442 0.8026
No log 6.25 200 0.6954 0.4959 0.6954 0.8339
No log 6.3125 202 0.6656 0.4529 0.6656 0.8158
No log 6.375 204 0.6586 0.5114 0.6586 0.8115
No log 6.4375 206 0.6795 0.5111 0.6795 0.8243
No log 6.5 208 0.8455 0.4182 0.8455 0.9195
No log 6.5625 210 0.8069 0.4212 0.8069 0.8983
No log 6.625 212 0.6412 0.5011 0.6412 0.8007
No log 6.6875 214 0.7313 0.4422 0.7313 0.8552
No log 6.75 216 0.8508 0.3369 0.8508 0.9224
No log 6.8125 218 0.7695 0.3756 0.7695 0.8772
No log 6.875 220 0.6261 0.5006 0.6261 0.7912
No log 6.9375 222 0.6978 0.4813 0.6978 0.8353
No log 7.0 224 0.7723 0.4925 0.7723 0.8788
No log 7.0625 226 0.6822 0.4973 0.6822 0.8260
No log 7.125 228 0.6743 0.4706 0.6743 0.8211
No log 7.1875 230 0.6799 0.4706 0.6799 0.8246
No log 7.25 232 0.7260 0.4914 0.7260 0.8521
No log 7.3125 234 0.7906 0.5322 0.7906 0.8892
No log 7.375 236 0.6844 0.4540 0.6844 0.8273
No log 7.4375 238 0.6783 0.4486 0.6783 0.8236
No log 7.5 240 0.8836 0.4428 0.8836 0.9400
No log 7.5625 242 0.8793 0.4264 0.8793 0.9377
No log 7.625 244 0.7289 0.4570 0.7289 0.8537
No log 7.6875 246 0.5948 0.4913 0.5948 0.7712
No log 7.75 248 0.6994 0.4635 0.6994 0.8363
No log 7.8125 250 0.7958 0.3927 0.7958 0.8921
No log 7.875 252 0.7192 0.4324 0.7192 0.8481
No log 7.9375 254 0.6122 0.4774 0.6122 0.7825
No log 8.0 256 0.6433 0.5115 0.6433 0.8021
No log 8.0625 258 0.6220 0.4446 0.6220 0.7887
No log 8.125 260 0.6339 0.4626 0.6339 0.7962
No log 8.1875 262 0.7102 0.4814 0.7102 0.8427
No log 8.25 264 0.6479 0.4664 0.6479 0.8050
No log 8.3125 266 0.6031 0.4978 0.6031 0.7766
No log 8.375 268 0.6822 0.5423 0.6822 0.8259
No log 8.4375 270 0.6713 0.5163 0.6713 0.8193
No log 8.5 272 0.5960 0.5043 0.5960 0.7720
No log 8.5625 274 0.6087 0.4091 0.6087 0.7802
No log 8.625 276 0.6997 0.4538 0.6997 0.8365
No log 8.6875 278 0.7370 0.4598 0.7370 0.8585
No log 8.75 280 0.6487 0.5214 0.6487 0.8054
No log 8.8125 282 0.6632 0.4901 0.6632 0.8144
No log 8.875 284 0.8019 0.4881 0.8019 0.8955
No log 8.9375 286 0.7709 0.4838 0.7709 0.8780
No log 9.0 288 0.6375 0.4669 0.6375 0.7984
No log 9.0625 290 0.6182 0.5054 0.6182 0.7863
No log 9.125 292 0.6193 0.4763 0.6193 0.7870
No log 9.1875 294 0.5990 0.4897 0.5990 0.7740
No log 9.25 296 0.6192 0.5014 0.6192 0.7869
No log 9.3125 298 0.6062 0.5058 0.6062 0.7786
No log 9.375 300 0.5980 0.4726 0.5980 0.7733
No log 9.4375 302 0.6114 0.4872 0.6114 0.7819
No log 9.5 304 0.6109 0.5383 0.6109 0.7816
No log 9.5625 306 0.6128 0.5106 0.6128 0.7828
No log 9.625 308 0.6171 0.5152 0.6171 0.7856
No log 9.6875 310 0.6129 0.5012 0.6129 0.7829
No log 9.75 312 0.6076 0.5203 0.6076 0.7795
No log 9.8125 314 0.6129 0.5476 0.6129 0.7829
No log 9.875 316 0.6416 0.5185 0.6416 0.8010
No log 9.9375 318 0.6311 0.5308 0.6311 0.7944
No log 10.0 320 0.6113 0.4915 0.6113 0.7819
No log 10.0625 322 0.6329 0.4721 0.6329 0.7956
No log 10.125 324 0.6410 0.4721 0.6410 0.8006
No log 10.1875 326 0.6201 0.4740 0.6201 0.7875
No log 10.25 328 0.5851 0.4783 0.5851 0.7649
No log 10.3125 330 0.5741 0.4603 0.5741 0.7577
No log 10.375 332 0.5851 0.5108 0.5851 0.7649
No log 10.4375 334 0.6111 0.4811 0.6111 0.7817
No log 10.5 336 0.6018 0.4827 0.6018 0.7758
No log 10.5625 338 0.6000 0.4709 0.6000 0.7746
No log 10.625 340 0.6022 0.4901 0.6022 0.7760
No log 10.6875 342 0.5767 0.4930 0.5767 0.7594
No log 10.75 344 0.5871 0.5494 0.5871 0.7662
No log 10.8125 346 0.5970 0.5916 0.5970 0.7727
No log 10.875 348 0.5732 0.5333 0.5732 0.7571
No log 10.9375 350 0.6124 0.5088 0.6124 0.7826
No log 11.0 352 0.6176 0.5074 0.6176 0.7859
No log 11.0625 354 0.5780 0.5292 0.5780 0.7603
No log 11.125 356 0.5748 0.5271 0.5748 0.7582
No log 11.1875 358 0.5636 0.5263 0.5636 0.7508
No log 11.25 360 0.5501 0.5371 0.5501 0.7417
No log 11.3125 362 0.5499 0.5492 0.5499 0.7415
No log 11.375 364 0.5868 0.5416 0.5868 0.7660
No log 11.4375 366 0.5671 0.5659 0.5671 0.7530
No log 11.5 368 0.5651 0.5159 0.5651 0.7517
No log 11.5625 370 0.5596 0.5005 0.5596 0.7481
No log 11.625 372 0.5736 0.5051 0.5736 0.7574
No log 11.6875 374 0.6130 0.4755 0.6130 0.7829
No log 11.75 376 0.6847 0.5433 0.6847 0.8275
No log 11.8125 378 0.6846 0.5433 0.6846 0.8274
No log 11.875 380 0.6122 0.4822 0.6122 0.7825
No log 11.9375 382 0.5870 0.5269 0.5870 0.7662
No log 12.0 384 0.5978 0.5708 0.5978 0.7732
No log 12.0625 386 0.5928 0.5385 0.5928 0.7699
No log 12.125 388 0.5881 0.5479 0.5881 0.7669
No log 12.1875 390 0.5825 0.5282 0.5825 0.7632
No log 12.25 392 0.5960 0.5026 0.5960 0.7720
No log 12.3125 394 0.6686 0.4772 0.6686 0.8177
No log 12.375 396 0.6472 0.5027 0.6472 0.8045
No log 12.4375 398 0.6027 0.4845 0.6027 0.7764
No log 12.5 400 0.5724 0.5124 0.5724 0.7566
No log 12.5625 402 0.6213 0.5347 0.6213 0.7882
No log 12.625 404 0.6475 0.5605 0.6475 0.8047
No log 12.6875 406 0.6065 0.5868 0.6065 0.7788
No log 12.75 408 0.6323 0.5115 0.6323 0.7952
No log 12.8125 410 0.7075 0.4844 0.7075 0.8411
No log 12.875 412 0.6520 0.5076 0.6520 0.8075
No log 12.9375 414 0.5845 0.5139 0.5845 0.7645
No log 13.0 416 0.5615 0.5019 0.5615 0.7494
No log 13.0625 418 0.5559 0.5026 0.5559 0.7456
No log 13.125 420 0.5641 0.5084 0.5641 0.7510
No log 13.1875 422 0.5687 0.5148 0.5687 0.7541
No log 13.25 424 0.5639 0.4963 0.5639 0.7510
No log 13.3125 426 0.5698 0.4827 0.5698 0.7549
No log 13.375 428 0.5715 0.5012 0.5715 0.7559
No log 13.4375 430 0.6324 0.5019 0.6324 0.7952
No log 13.5 432 0.6303 0.4960 0.6303 0.7939
No log 13.5625 434 0.5699 0.4738 0.5699 0.7549
No log 13.625 436 0.5681 0.4418 0.5681 0.7537
No log 13.6875 438 0.5694 0.4830 0.5694 0.7546
No log 13.75 440 0.5724 0.4808 0.5724 0.7566
No log 13.8125 442 0.5834 0.5005 0.5834 0.7638
No log 13.875 444 0.5817 0.5079 0.5817 0.7627
No log 13.9375 446 0.5650 0.5145 0.5650 0.7517
No log 14.0 448 0.5771 0.4560 0.5771 0.7597
No log 14.0625 450 0.6203 0.4385 0.6203 0.7876
No log 14.125 452 0.5898 0.4758 0.5898 0.7680
No log 14.1875 454 0.5727 0.4932 0.5727 0.7568
No log 14.25 456 0.5877 0.4928 0.5877 0.7666
No log 14.3125 458 0.5847 0.5081 0.5847 0.7647
No log 14.375 460 0.5791 0.4851 0.5791 0.7610
No log 14.4375 462 0.5804 0.4924 0.5804 0.7619
No log 14.5 464 0.5729 0.5121 0.5729 0.7569
No log 14.5625 466 0.5836 0.5100 0.5836 0.7639
No log 14.625 468 0.5739 0.4827 0.5739 0.7576
No log 14.6875 470 0.5708 0.5040 0.5708 0.7555
No log 14.75 472 0.5752 0.4609 0.5752 0.7584
No log 14.8125 474 0.5857 0.4827 0.5857 0.7653
No log 14.875 476 0.5889 0.4733 0.5889 0.7674
No log 14.9375 478 0.5905 0.5290 0.5905 0.7684
No log 15.0 480 0.5941 0.4941 0.5941 0.7708
No log 15.0625 482 0.5952 0.4849 0.5952 0.7715
No log 15.125 484 0.6003 0.5293 0.6003 0.7748
No log 15.1875 486 0.6253 0.4917 0.6253 0.7907
No log 15.25 488 0.6366 0.4806 0.6366 0.7979
No log 15.3125 490 0.6165 0.4917 0.6165 0.7852
No log 15.375 492 0.6041 0.5123 0.6041 0.7772
No log 15.4375 494 0.6048 0.4457 0.6048 0.7777
No log 15.5 496 0.6090 0.4921 0.6090 0.7804
No log 15.5625 498 0.6381 0.4965 0.6381 0.7988
0.326 15.625 500 0.6474 0.4785 0.6474 0.8046
0.326 15.6875 502 0.6194 0.4714 0.6194 0.7870
0.326 15.75 504 0.6188 0.5296 0.6188 0.7867
0.326 15.8125 506 0.6182 0.4573 0.6182 0.7862
0.326 15.875 508 0.6071 0.4434 0.6071 0.7792
0.326 15.9375 510 0.5855 0.4368 0.5855 0.7652
0.326 16.0 512 0.5856 0.4930 0.5856 0.7652
0.326 16.0625 514 0.6062 0.5077 0.6062 0.7786
0.326 16.125 516 0.5856 0.4934 0.5856 0.7653
0.326 16.1875 518 0.5770 0.4745 0.5770 0.7596
0.326 16.25 520 0.5696 0.4676 0.5696 0.7547
0.326 16.3125 522 0.5746 0.5173 0.5746 0.7580
0.326 16.375 524 0.5944 0.5011 0.5944 0.7709
0.326 16.4375 526 0.6193 0.5003 0.6193 0.7870
0.326 16.5 528 0.6344 0.4872 0.6344 0.7965
0.326 16.5625 530 0.6797 0.4824 0.6797 0.8244
0.326 16.625 532 0.6438 0.4659 0.6438 0.8024
0.326 16.6875 534 0.5807 0.4725 0.5807 0.7620
0.326 16.75 536 0.5788 0.4410 0.5788 0.7608
0.326 16.8125 538 0.5878 0.4530 0.5878 0.7667
0.326 16.875 540 0.5739 0.4276 0.5739 0.7575

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
4
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for MayBashendy/ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run3_AugV5_k6_task2_organization

Finetuned
(4222)
this model