ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run2_AugV5_k13_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5973
  • Qwk: 0.5510
  • Mse: 0.5973
  • Rmse: 0.7728

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0303 2 4.2437 -0.0200 4.2437 2.0600
No log 0.0606 4 2.3060 0.0518 2.3060 1.5185
No log 0.0909 6 1.4451 0.0251 1.4451 1.2021
No log 0.1212 8 1.0961 -0.0377 1.0961 1.0469
No log 0.1515 10 0.9391 0.0519 0.9391 0.9691
No log 0.1818 12 0.8076 0.2537 0.8076 0.8986
No log 0.2121 14 0.7883 0.2733 0.7883 0.8879
No log 0.2424 16 0.7539 0.3051 0.7539 0.8683
No log 0.2727 18 0.7718 0.3076 0.7718 0.8785
No log 0.3030 20 0.8926 0.3108 0.8926 0.9448
No log 0.3333 22 0.9831 0.2776 0.9831 0.9915
No log 0.3636 24 0.8266 0.3425 0.8266 0.9092
No log 0.3939 26 0.7172 0.3307 0.7172 0.8469
No log 0.4242 28 0.6883 0.3499 0.6883 0.8296
No log 0.4545 30 0.7305 0.3528 0.7305 0.8547
No log 0.4848 32 1.0912 0.2352 1.0912 1.0446
No log 0.5152 34 1.6630 0.1878 1.6630 1.2896
No log 0.5455 36 2.2481 0.0608 2.2481 1.4994
No log 0.5758 38 2.1379 0.1290 2.1379 1.4622
No log 0.6061 40 1.5969 0.2011 1.5969 1.2637
No log 0.6364 42 0.9295 0.3576 0.9295 0.9641
No log 0.6667 44 0.7512 0.3579 0.7512 0.8667
No log 0.6970 46 0.7392 0.2971 0.7392 0.8598
No log 0.7273 48 0.8486 0.3195 0.8486 0.9212
No log 0.7576 50 1.1241 0.2765 1.1241 1.0602
No log 0.7879 52 1.7159 0.1552 1.7159 1.3099
No log 0.8182 54 1.7376 0.1636 1.7376 1.3182
No log 0.8485 56 1.5869 0.1699 1.5869 1.2597
No log 0.8788 58 1.2745 0.1820 1.2745 1.1289
No log 0.9091 60 1.1529 0.2588 1.1529 1.0737
No log 0.9394 62 1.1826 0.1976 1.1826 1.0875
No log 0.9697 64 1.2701 0.1517 1.2701 1.1270
No log 1.0 66 1.2298 0.1696 1.2298 1.1090
No log 1.0303 68 0.9636 0.3117 0.9636 0.9816
No log 1.0606 70 0.7629 0.3801 0.7629 0.8734
No log 1.0909 72 0.6990 0.3818 0.6990 0.8361
No log 1.1212 74 0.6661 0.4560 0.6661 0.8161
No log 1.1515 76 0.6892 0.3984 0.6892 0.8302
No log 1.1818 78 0.8702 0.3295 0.8702 0.9329
No log 1.2121 80 1.1036 0.2471 1.1036 1.0505
No log 1.2424 82 1.2145 0.2291 1.2145 1.1021
No log 1.2727 84 1.3118 0.1965 1.3118 1.1453
No log 1.3030 86 1.2422 0.2240 1.2422 1.1145
No log 1.3333 88 1.0986 0.2681 1.0986 1.0481
No log 1.3636 90 0.8130 0.3839 0.8130 0.9017
No log 1.3939 92 0.6385 0.4806 0.6385 0.7990
No log 1.4242 94 0.5971 0.5166 0.5971 0.7727
No log 1.4545 96 0.5944 0.5192 0.5944 0.7710
No log 1.4848 98 0.6175 0.4970 0.6175 0.7858
No log 1.5152 100 0.7169 0.4726 0.7169 0.8467
No log 1.5455 102 0.7600 0.4255 0.7600 0.8718
No log 1.5758 104 0.8903 0.3457 0.8903 0.9435
No log 1.6061 106 0.9261 0.3452 0.9261 0.9624
No log 1.6364 108 0.9773 0.3384 0.9773 0.9886
No log 1.6667 110 0.9426 0.3556 0.9426 0.9709
No log 1.6970 112 0.8051 0.3714 0.8051 0.8973
No log 1.7273 114 0.7439 0.3883 0.7439 0.8625
No log 1.7576 116 0.7858 0.4618 0.7858 0.8864
No log 1.7879 118 0.6886 0.4929 0.6886 0.8298
No log 1.8182 120 0.6743 0.5606 0.6743 0.8212
No log 1.8485 122 0.6606 0.5106 0.6606 0.8128
No log 1.8788 124 0.6441 0.5098 0.6441 0.8026
No log 1.9091 126 0.7574 0.4554 0.7574 0.8703
No log 1.9394 128 0.9085 0.3741 0.9085 0.9532
No log 1.9697 130 0.9922 0.3516 0.9922 0.9961
No log 2.0 132 0.9563 0.3737 0.9563 0.9779
No log 2.0303 134 0.7497 0.3943 0.7497 0.8658
No log 2.0606 136 0.6961 0.4517 0.6961 0.8344
No log 2.0909 138 0.6581 0.4428 0.6581 0.8112
No log 2.1212 140 0.6761 0.4540 0.6761 0.8222
No log 2.1515 142 0.7549 0.4049 0.7549 0.8689
No log 2.1818 144 0.9134 0.3728 0.9134 0.9557
No log 2.2121 146 0.9871 0.3486 0.9871 0.9936
No log 2.2424 148 0.9842 0.3842 0.9842 0.9921
No log 2.2727 150 0.9552 0.4106 0.9552 0.9774
No log 2.3030 152 0.8266 0.4092 0.8266 0.9092
No log 2.3333 154 0.7703 0.4085 0.7703 0.8776
No log 2.3636 156 0.7158 0.4052 0.7158 0.8461
No log 2.3939 158 0.6616 0.4307 0.6616 0.8134
No log 2.4242 160 0.6677 0.4802 0.6677 0.8171
No log 2.4545 162 0.6730 0.3742 0.6730 0.8204
No log 2.4848 164 0.6434 0.4356 0.6434 0.8022
No log 2.5152 166 0.6852 0.4214 0.6852 0.8278
No log 2.5455 168 0.8728 0.4172 0.8728 0.9342
No log 2.5758 170 0.9375 0.3845 0.9375 0.9683
No log 2.6061 172 0.8500 0.4162 0.8500 0.9220
No log 2.6364 174 0.8063 0.4556 0.8063 0.8979
No log 2.6667 176 0.7207 0.4801 0.7207 0.8489
No log 2.6970 178 0.7056 0.5344 0.7056 0.8400
No log 2.7273 180 0.6624 0.5813 0.6624 0.8139
No log 2.7576 182 0.6776 0.4697 0.6776 0.8231
No log 2.7879 184 0.8653 0.4475 0.8653 0.9302
No log 2.8182 186 0.9436 0.4232 0.9436 0.9714
No log 2.8485 188 0.9605 0.4177 0.9605 0.9800
No log 2.8788 190 0.7743 0.4732 0.7743 0.8799
No log 2.9091 192 0.6480 0.4658 0.6480 0.8050
No log 2.9394 194 0.6186 0.5170 0.6186 0.7865
No log 2.9697 196 0.6848 0.4594 0.6848 0.8275
No log 3.0 198 0.7270 0.4573 0.7270 0.8526
No log 3.0303 200 0.6605 0.4795 0.6605 0.8127
No log 3.0606 202 0.6163 0.4545 0.6163 0.7851
No log 3.0909 204 0.6271 0.4861 0.6271 0.7919
No log 3.1212 206 0.7024 0.4474 0.7024 0.8381
No log 3.1515 208 0.6934 0.4734 0.6934 0.8327
No log 3.1818 210 0.6230 0.4437 0.6230 0.7893
No log 3.2121 212 0.6476 0.3924 0.6476 0.8047
No log 3.2424 214 0.7016 0.4152 0.7016 0.8376
No log 3.2727 216 0.6796 0.3954 0.6796 0.8244
No log 3.3030 218 0.6668 0.5150 0.6668 0.8166
No log 3.3333 220 0.6829 0.4909 0.6829 0.8264
No log 3.3636 222 0.7703 0.4819 0.7703 0.8777
No log 3.3939 224 0.8916 0.4026 0.8916 0.9442
No log 3.4242 226 0.8534 0.4305 0.8534 0.9238
No log 3.4545 228 0.7043 0.4996 0.7043 0.8392
No log 3.4848 230 0.6225 0.4649 0.6225 0.7890
No log 3.5152 232 0.6109 0.4398 0.6109 0.7816
No log 3.5455 234 0.6014 0.4192 0.6014 0.7755
No log 3.5758 236 0.5981 0.4860 0.5981 0.7733
No log 3.6061 238 0.6038 0.5377 0.6038 0.7770
No log 3.6364 240 0.6700 0.5208 0.6700 0.8186
No log 3.6667 242 0.6899 0.4972 0.6899 0.8306
No log 3.6970 244 0.6541 0.5190 0.6541 0.8088
No log 3.7273 246 0.6519 0.5150 0.6519 0.8074
No log 3.7576 248 0.6108 0.5384 0.6108 0.7815
No log 3.7879 250 0.6204 0.5644 0.6204 0.7876
No log 3.8182 252 0.6264 0.5963 0.6264 0.7914
No log 3.8485 254 0.6162 0.5929 0.6162 0.7850
No log 3.8788 256 0.6305 0.5220 0.6305 0.7941
No log 3.9091 258 0.6487 0.4975 0.6487 0.8054
No log 3.9394 260 0.7017 0.4724 0.7017 0.8377
No log 3.9697 262 0.7196 0.4786 0.7196 0.8483
No log 4.0 264 0.6901 0.4734 0.6901 0.8307
No log 4.0303 266 0.7129 0.4439 0.7129 0.8443
No log 4.0606 268 0.7643 0.5246 0.7643 0.8743
No log 4.0909 270 0.7123 0.5307 0.7123 0.8440
No log 4.1212 272 0.7101 0.4564 0.7101 0.8427
No log 4.1515 274 0.7462 0.4426 0.7462 0.8638
No log 4.1818 276 0.7256 0.4336 0.7256 0.8518
No log 4.2121 278 0.7002 0.4814 0.7002 0.8368
No log 4.2424 280 0.6546 0.4824 0.6546 0.8091
No log 4.2727 282 0.6135 0.4467 0.6135 0.7832
No log 4.3030 284 0.6158 0.5048 0.6158 0.7847
No log 4.3333 286 0.6384 0.5016 0.6384 0.7990
No log 4.3636 288 0.6290 0.5206 0.6290 0.7931
No log 4.3939 290 0.6151 0.5206 0.6151 0.7843
No log 4.4242 292 0.6281 0.5040 0.6281 0.7925
No log 4.4545 294 0.6288 0.5184 0.6288 0.7930
No log 4.4848 296 0.6033 0.5431 0.6033 0.7767
No log 4.5152 298 0.6187 0.5315 0.6187 0.7866
No log 4.5455 300 0.6073 0.5079 0.6073 0.7793
No log 4.5758 302 0.5914 0.5444 0.5914 0.7690
No log 4.6061 304 0.5845 0.5479 0.5845 0.7645
No log 4.6364 306 0.5790 0.5255 0.5790 0.7609
No log 4.6667 308 0.5881 0.4822 0.5881 0.7669
No log 4.6970 310 0.6147 0.4835 0.6147 0.7840
No log 4.7273 312 0.5955 0.4310 0.5955 0.7717
No log 4.7576 314 0.5911 0.4133 0.5911 0.7688
No log 4.7879 316 0.5873 0.4559 0.5873 0.7664
No log 4.8182 318 0.5877 0.4943 0.5877 0.7666
No log 4.8485 320 0.6782 0.5292 0.6782 0.8235
No log 4.8788 322 0.8162 0.4397 0.8162 0.9035
No log 4.9091 324 0.7810 0.4346 0.7810 0.8837
No log 4.9394 326 0.6441 0.5114 0.6441 0.8026
No log 4.9697 328 0.5887 0.4703 0.5887 0.7673
No log 5.0 330 0.6041 0.4353 0.6041 0.7772
No log 5.0303 332 0.6086 0.3868 0.6086 0.7801
No log 5.0606 334 0.6157 0.3371 0.6157 0.7847
No log 5.0909 336 0.6276 0.3337 0.6276 0.7922
No log 5.1212 338 0.6587 0.4133 0.6587 0.8116
No log 5.1515 340 0.6566 0.4265 0.6566 0.8103
No log 5.1818 342 0.6077 0.4757 0.6077 0.7796
No log 5.2121 344 0.6120 0.4599 0.6120 0.7823
No log 5.2424 346 0.6402 0.5211 0.6402 0.8001
No log 5.2727 348 0.6258 0.5524 0.6258 0.7911
No log 5.3030 350 0.6010 0.4671 0.6010 0.7752
No log 5.3333 352 0.5899 0.4858 0.5899 0.7681
No log 5.3636 354 0.5875 0.4814 0.5875 0.7665
No log 5.3939 356 0.5882 0.4978 0.5882 0.7669
No log 5.4242 358 0.5833 0.5234 0.5833 0.7637
No log 5.4545 360 0.6310 0.5351 0.6310 0.7944
No log 5.4848 362 0.6482 0.5141 0.6482 0.8051
No log 5.5152 364 0.6195 0.5732 0.6195 0.7871
No log 5.5455 366 0.6264 0.5354 0.6264 0.7914
No log 5.5758 368 0.6481 0.5309 0.6481 0.8050
No log 5.6061 370 0.6780 0.5176 0.6780 0.8234
No log 5.6364 372 0.6843 0.4896 0.6843 0.8272
No log 5.6667 374 0.7078 0.5058 0.7078 0.8413
No log 5.6970 376 0.6775 0.4939 0.6775 0.8231
No log 5.7273 378 0.6263 0.5571 0.6263 0.7914
No log 5.7576 380 0.5952 0.5310 0.5952 0.7715
No log 5.7879 382 0.5863 0.5111 0.5863 0.7657
No log 5.8182 384 0.5792 0.4965 0.5792 0.7611
No log 5.8485 386 0.5600 0.5489 0.5600 0.7484
No log 5.8788 388 0.5837 0.5457 0.5837 0.7640
No log 5.9091 390 0.5991 0.5542 0.5991 0.7740
No log 5.9394 392 0.5875 0.5705 0.5875 0.7665
No log 5.9697 394 0.5698 0.5331 0.5698 0.7548
No log 6.0 396 0.5525 0.5124 0.5525 0.7433
No log 6.0303 398 0.5528 0.5054 0.5528 0.7435
No log 6.0606 400 0.5549 0.5118 0.5549 0.7449
No log 6.0909 402 0.5566 0.5155 0.5566 0.7461
No log 6.1212 404 0.5606 0.5230 0.5606 0.7488
No log 6.1515 406 0.5491 0.4876 0.5491 0.7410
No log 6.1818 408 0.5664 0.4924 0.5664 0.7526
No log 6.2121 410 0.6042 0.5070 0.6042 0.7773
No log 6.2424 412 0.5943 0.5214 0.5943 0.7709
No log 6.2727 414 0.5846 0.4955 0.5846 0.7646
No log 6.3030 416 0.5858 0.5174 0.5858 0.7654
No log 6.3333 418 0.5997 0.5110 0.5997 0.7744
No log 6.3636 420 0.6063 0.5082 0.6063 0.7786
No log 6.3939 422 0.6144 0.5004 0.6144 0.7838
No log 6.4242 424 0.6124 0.5131 0.6124 0.7825
No log 6.4545 426 0.6110 0.5350 0.6110 0.7817
No log 6.4848 428 0.6014 0.5124 0.6014 0.7755
No log 6.5152 430 0.6131 0.5820 0.6131 0.7830
No log 6.5455 432 0.6265 0.5610 0.6265 0.7915
No log 6.5758 434 0.6247 0.5451 0.6247 0.7904
No log 6.6061 436 0.6275 0.4755 0.6275 0.7922
No log 6.6364 438 0.6141 0.4766 0.6141 0.7836
No log 6.6667 440 0.6072 0.5101 0.6072 0.7792
No log 6.6970 442 0.5886 0.4542 0.5886 0.7672
No log 6.7273 444 0.5877 0.4603 0.5877 0.7666
No log 6.7576 446 0.5861 0.5005 0.5861 0.7656
No log 6.7879 448 0.5928 0.5153 0.5928 0.7699
No log 6.8182 450 0.5979 0.5546 0.5979 0.7732
No log 6.8485 452 0.6107 0.5338 0.6107 0.7815
No log 6.8788 454 0.6317 0.5338 0.6317 0.7948
No log 6.9091 456 0.6215 0.5214 0.6215 0.7884
No log 6.9394 458 0.6050 0.5624 0.6050 0.7778
No log 6.9697 460 0.6114 0.5216 0.6114 0.7819
No log 7.0 462 0.6390 0.5461 0.6390 0.7993
No log 7.0303 464 0.7177 0.5438 0.7177 0.8472
No log 7.0606 466 0.7083 0.5217 0.7083 0.8416
No log 7.0909 468 0.7185 0.4620 0.7185 0.8476
No log 7.1212 470 0.7301 0.4508 0.7301 0.8545
No log 7.1515 472 0.7060 0.5145 0.7060 0.8403
No log 7.1818 474 0.6594 0.4708 0.6594 0.8121
No log 7.2121 476 0.6332 0.4750 0.6332 0.7958
No log 7.2424 478 0.6464 0.5372 0.6464 0.8040
No log 7.2727 480 0.6444 0.5177 0.6444 0.8027
No log 7.3030 482 0.6237 0.5458 0.6237 0.7897
No log 7.3333 484 0.5972 0.5163 0.5972 0.7728
No log 7.3636 486 0.5852 0.5338 0.5852 0.7650
No log 7.3939 488 0.5747 0.4985 0.5747 0.7581
No log 7.4242 490 0.5874 0.4986 0.5874 0.7664
No log 7.4545 492 0.5865 0.4562 0.5865 0.7658
No log 7.4848 494 0.5435 0.4896 0.5435 0.7372
No log 7.5152 496 0.5280 0.5184 0.5280 0.7266
No log 7.5455 498 0.5488 0.5159 0.5488 0.7408
0.3849 7.5758 500 0.5593 0.5674 0.5593 0.7479
0.3849 7.6061 502 0.5911 0.5368 0.5911 0.7689
0.3849 7.6364 504 0.6278 0.5624 0.6278 0.7924
0.3849 7.6667 506 0.6610 0.5441 0.6610 0.8130
0.3849 7.6970 508 0.6429 0.5192 0.6429 0.8018
0.3849 7.7273 510 0.5973 0.5510 0.5973 0.7728

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
4
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for MayBashendy/ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run2_AugV5_k13_task2_organization

Finetuned
(4222)
this model