ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k5_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.1049
  • Qwk: 0.5882
  • Mse: 1.1049
  • Rmse: 1.0511

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0526 2 6.9746 0.0167 6.9746 2.6409
No log 0.1053 4 4.0421 0.0800 4.0421 2.0105
No log 0.1579 6 2.6353 0.0633 2.6353 1.6234
No log 0.2105 8 1.8905 0.1789 1.8905 1.3750
No log 0.2632 10 1.6413 0.2342 1.6413 1.2811
No log 0.3158 12 1.5711 0.1835 1.5711 1.2534
No log 0.3684 14 1.4404 0.3898 1.4404 1.2002
No log 0.4211 16 1.3865 0.4138 1.3865 1.1775
No log 0.4737 18 1.3875 0.2778 1.3875 1.1779
No log 0.5263 20 1.6900 0.2542 1.6900 1.3000
No log 0.5789 22 1.6897 0.2975 1.6897 1.2999
No log 0.6316 24 1.6222 0.3390 1.6222 1.2737
No log 0.6842 26 1.5033 0.3333 1.5033 1.2261
No log 0.7368 28 1.3790 0.2243 1.3790 1.1743
No log 0.7895 30 1.3298 0.3784 1.3298 1.1532
No log 0.8421 32 1.2840 0.4348 1.2840 1.1331
No log 0.8947 34 1.3097 0.3571 1.3097 1.1444
No log 0.9474 36 1.5740 0.2301 1.5740 1.2546
No log 1.0 38 1.7084 0.2241 1.7084 1.3071
No log 1.0526 40 1.4326 0.3193 1.4326 1.1969
No log 1.1053 42 1.1975 0.4576 1.1975 1.0943
No log 1.1579 44 1.0707 0.4833 1.0707 1.0347
No log 1.2105 46 1.0167 0.5354 1.0167 1.0083
No log 1.2632 48 1.0475 0.608 1.0475 1.0235
No log 1.3158 50 1.2131 0.5312 1.2131 1.1014
No log 1.3684 52 1.1775 0.5692 1.1775 1.0851
No log 1.4211 54 1.2171 0.5077 1.2171 1.1032
No log 1.4737 56 1.3180 0.4328 1.3180 1.1480
No log 1.5263 58 1.4516 0.3731 1.4516 1.2048
No log 1.5789 60 1.5976 0.3582 1.5976 1.2640
No log 1.6316 62 1.6532 0.4138 1.6532 1.2858
No log 1.6842 64 1.4937 0.3636 1.4937 1.2222
No log 1.7368 66 1.4875 0.4203 1.4875 1.2196
No log 1.7895 68 1.5525 0.4416 1.5525 1.2460
No log 1.8421 70 1.4790 0.4384 1.4790 1.2161
No log 1.8947 72 1.5991 0.3597 1.5991 1.2646
No log 1.9474 74 1.5824 0.3852 1.5824 1.2580
No log 2.0 76 1.1428 0.5649 1.1428 1.0690
No log 2.0526 78 0.8653 0.6395 0.8653 0.9302
No log 2.1053 80 0.9891 0.7093 0.9891 0.9945
No log 2.1579 82 0.9977 0.7232 0.9977 0.9989
No log 2.2105 84 0.8806 0.6667 0.8806 0.9384
No log 2.2632 86 1.4881 0.3279 1.4881 1.2199
No log 2.3158 88 2.3812 0.0458 2.3812 1.5431
No log 2.3684 90 2.6567 0.0432 2.6567 1.6299
No log 2.4211 92 2.3915 0.1471 2.3915 1.5465
No log 2.4737 94 1.8075 0.2636 1.8075 1.3444
No log 2.5263 96 1.3109 0.4848 1.3109 1.1450
No log 2.5789 98 1.1344 0.512 1.1344 1.0651
No log 2.6316 100 1.3456 0.5 1.3456 1.1600
No log 2.6842 102 1.3684 0.3833 1.3684 1.1698
No log 2.7368 104 1.2126 0.4918 1.2126 1.1012
No log 2.7895 106 1.1143 0.5802 1.1143 1.0556
No log 2.8421 108 1.1552 0.5496 1.1552 1.0748
No log 2.8947 110 1.2114 0.5588 1.2114 1.1006
No log 2.9474 112 1.2116 0.4748 1.2116 1.1007
No log 3.0 114 1.2360 0.5241 1.2360 1.1117
No log 3.0526 116 1.3320 0.5 1.3320 1.1541
No log 3.1053 118 1.2206 0.6056 1.2206 1.1048
No log 3.1579 120 1.0082 0.6324 1.0082 1.0041
No log 3.2105 122 0.8508 0.6190 0.8508 0.9224
No log 3.2632 124 0.8296 0.6406 0.8296 0.9108
No log 3.3158 126 0.8050 0.6667 0.8050 0.8972
No log 3.3684 128 0.7426 0.6857 0.7426 0.8618
No log 3.4211 130 0.8318 0.6883 0.8318 0.9120
No log 3.4737 132 1.0191 0.6667 1.0191 1.0095
No log 3.5263 134 1.2713 0.6420 1.2713 1.1275
No log 3.5789 136 1.3426 0.4966 1.3426 1.1587
No log 3.6316 138 1.3311 0.5035 1.3311 1.1537
No log 3.6842 140 1.3922 0.4722 1.3922 1.1799
No log 3.7368 142 1.4159 0.4476 1.4159 1.1899
No log 3.7895 144 1.4553 0.4110 1.4553 1.2064
No log 3.8421 146 1.4288 0.3741 1.4288 1.1953
No log 3.8947 148 1.3819 0.4296 1.3819 1.1755
No log 3.9474 150 1.2125 0.5414 1.2125 1.1012
No log 4.0 152 1.0672 0.6269 1.0672 1.0330
No log 4.0526 154 0.9690 0.6364 0.9690 0.9844
No log 4.1053 156 0.9293 0.6466 0.9293 0.9640
No log 4.1579 158 0.9038 0.6131 0.9038 0.9507
No log 4.2105 160 0.8848 0.6241 0.8848 0.9406
No log 4.2632 162 0.7979 0.7105 0.7979 0.8932
No log 4.3158 164 0.7598 0.6986 0.7598 0.8716
No log 4.3684 166 0.8346 0.6761 0.8346 0.9135
No log 4.4211 168 0.9083 0.6087 0.9083 0.9531
No log 4.4737 170 0.8673 0.6575 0.8673 0.9313
No log 4.5263 172 0.7607 0.6944 0.7607 0.8722
No log 4.5789 174 0.7268 0.6901 0.7268 0.8525
No log 4.6316 176 0.7154 0.7206 0.7154 0.8458
No log 4.6842 178 0.7203 0.7391 0.7203 0.8487
No log 4.7368 180 0.7318 0.7042 0.7318 0.8555
No log 4.7895 182 0.7482 0.7092 0.7482 0.8650
No log 4.8421 184 0.9835 0.6708 0.9835 0.9917
No log 4.8947 186 1.2783 0.5732 1.2783 1.1306
No log 4.9474 188 1.3347 0.4722 1.3347 1.1553
No log 5.0 190 1.3543 0.4493 1.3543 1.1638
No log 5.0526 192 1.3571 0.4539 1.3571 1.1649
No log 5.1053 194 1.2712 0.5324 1.2712 1.1275
No log 5.1579 196 1.1567 0.5649 1.1567 1.0755
No log 5.2105 198 1.0292 0.5846 1.0292 1.0145
No log 5.2632 200 0.9073 0.6074 0.9073 0.9525
No log 5.3158 202 0.8090 0.6667 0.8090 0.8994
No log 5.3684 204 0.7895 0.7034 0.7895 0.8885
No log 5.4211 206 0.8241 0.7179 0.8241 0.9078
No log 5.4737 208 0.7910 0.7067 0.7910 0.8894
No log 5.5263 210 0.7442 0.7183 0.7442 0.8627
No log 5.5789 212 0.7155 0.7246 0.7155 0.8459
No log 5.6316 214 0.7492 0.6870 0.7492 0.8656
No log 5.6842 216 0.7589 0.7176 0.7589 0.8711
No log 5.7368 218 0.8124 0.6667 0.8124 0.9013
No log 5.7895 220 0.8704 0.64 0.8704 0.9329
No log 5.8421 222 0.8785 0.6016 0.8785 0.9373
No log 5.8947 224 0.8310 0.6290 0.8310 0.9116
No log 5.9474 226 0.7858 0.6715 0.7858 0.8864
No log 6.0 228 0.8168 0.6892 0.8168 0.9037
No log 6.0526 230 0.8112 0.7160 0.8112 0.9007
No log 6.1053 232 0.7355 0.7034 0.7355 0.8576
No log 6.1579 234 0.7071 0.6901 0.7071 0.8409
No log 6.2105 236 0.6810 0.7092 0.6810 0.8253
No log 6.2632 238 0.7015 0.6906 0.7015 0.8376
No log 6.3158 240 0.7118 0.6901 0.7118 0.8437
No log 6.3684 242 0.7816 0.7051 0.7816 0.8841
No log 6.4211 244 0.8314 0.6483 0.8314 0.9118
No log 6.4737 246 0.8358 0.6338 0.8358 0.9142
No log 6.5263 248 0.8464 0.6475 0.8464 0.9200
No log 6.5789 250 0.8245 0.6667 0.8245 0.9080
No log 6.6316 252 0.8242 0.6619 0.8242 0.9078
No log 6.6842 254 0.8147 0.6479 0.8147 0.9026
No log 6.7368 256 0.8984 0.6438 0.8984 0.9479
No log 6.7895 258 0.9666 0.6667 0.9666 0.9831
No log 6.8421 260 0.8535 0.6438 0.8535 0.9239
No log 6.8947 262 0.7136 0.75 0.7136 0.8447
No log 6.9474 264 0.7150 0.7273 0.7150 0.8456
No log 7.0 266 0.7360 0.7273 0.7360 0.8579
No log 7.0526 268 0.8623 0.6143 0.8623 0.9286
No log 7.1053 270 1.0179 0.6154 1.0179 1.0089
No log 7.1579 272 1.0912 0.6111 1.0912 1.0446
No log 7.2105 274 1.1054 0.6667 1.1054 1.0514
No log 7.2632 276 1.0837 0.6667 1.0837 1.0410
No log 7.3158 278 0.8642 0.6122 0.8642 0.9296
No log 7.3684 280 0.7976 0.6573 0.7976 0.8931
No log 7.4211 282 0.7846 0.6713 0.7846 0.8858
No log 7.4737 284 0.8065 0.6571 0.8065 0.8980
No log 7.5263 286 0.9399 0.6383 0.9399 0.9695
No log 7.5789 288 1.0732 0.6309 1.0732 1.0360
No log 7.6316 290 1.0167 0.6197 1.0167 1.0083
No log 7.6842 292 0.9214 0.6418 0.9214 0.9599
No log 7.7368 294 0.9040 0.6154 0.9040 0.9508
No log 7.7895 296 0.9050 0.6715 0.9050 0.9513
No log 7.8421 298 0.9489 0.6475 0.9489 0.9741
No log 7.8947 300 1.0621 0.6345 1.0621 1.0306
No log 7.9474 302 1.1712 0.6087 1.1712 1.0822
No log 8.0 304 1.1365 0.6395 1.1365 1.0661
No log 8.0526 306 1.0163 0.6165 1.0163 1.0081
No log 8.1053 308 0.9613 0.6154 0.9613 0.9804
No log 8.1579 310 0.9813 0.6154 0.9813 0.9906
No log 8.2105 312 0.9958 0.6154 0.9958 0.9979
No log 8.2632 314 1.0176 0.6029 1.0176 1.0087
No log 8.3158 316 1.0025 0.6029 1.0025 1.0012
No log 8.3684 318 0.9628 0.6187 0.9628 0.9812
No log 8.4211 320 0.9296 0.6331 0.9296 0.9642
No log 8.4737 322 0.9337 0.6434 0.9337 0.9663
No log 8.5263 324 1.0286 0.6541 1.0286 1.0142
No log 8.5789 326 0.9525 0.6625 0.9525 0.9759
No log 8.6316 328 0.8520 0.6667 0.8520 0.9230
No log 8.6842 330 0.7534 0.6434 0.7534 0.8680
No log 8.7368 332 0.7706 0.6571 0.7706 0.8778
No log 8.7895 334 0.8344 0.6338 0.8344 0.9134
No log 8.8421 336 1.0324 0.6122 1.0324 1.0161
No log 8.8947 338 1.1776 0.5625 1.1776 1.0852
No log 8.9474 340 1.2065 0.55 1.2065 1.0984
No log 9.0 342 1.1013 0.5442 1.1013 1.0494
No log 9.0526 344 0.9965 0.6087 0.9965 0.9983
No log 9.1053 346 0.9069 0.6515 0.9069 0.9523
No log 9.1579 348 0.8589 0.6316 0.8589 0.9268
No log 9.2105 350 0.8387 0.6412 0.8387 0.9158
No log 9.2632 352 0.8336 0.6383 0.8336 0.9130
No log 9.3158 354 0.8491 0.6241 0.8491 0.9215
No log 9.3684 356 0.7910 0.6950 0.7910 0.8894
No log 9.4211 358 0.7147 0.7286 0.7147 0.8454
No log 9.4737 360 0.7023 0.7194 0.7023 0.8380
No log 9.5263 362 0.7073 0.75 0.7073 0.8410
No log 9.5789 364 0.7662 0.7333 0.7662 0.8754
No log 9.6316 366 0.8309 0.6800 0.8309 0.9115
No log 9.6842 368 0.8803 0.6623 0.8803 0.9383
No log 9.7368 370 0.8414 0.6757 0.8414 0.9173
No log 9.7895 372 0.7624 0.6950 0.7624 0.8732
No log 9.8421 374 0.7290 0.7218 0.7290 0.8538
No log 9.8947 376 0.7204 0.7407 0.7204 0.8488
No log 9.9474 378 0.7181 0.7407 0.7181 0.8474
No log 10.0 380 0.7476 0.7222 0.7476 0.8647
No log 10.0526 382 0.9072 0.6951 0.9072 0.9525
No log 10.1053 384 1.1234 0.6437 1.1234 1.0599
No log 10.1579 386 1.1215 0.6471 1.1215 1.0590
No log 10.2105 388 0.9986 0.6443 0.9986 0.9993
No log 10.2632 390 0.8506 0.6418 0.8506 0.9223
No log 10.3158 392 0.7912 0.6923 0.7912 0.8895
No log 10.3684 394 0.7855 0.6923 0.7855 0.8863
No log 10.4211 396 0.8103 0.6571 0.8103 0.9002
No log 10.4737 398 0.9447 0.6711 0.9447 0.9719
No log 10.5263 400 1.0041 0.6497 1.0041 1.0020
No log 10.5789 402 0.9867 0.6358 0.9867 0.9933
No log 10.6316 404 0.9519 0.6447 0.9519 0.9757
No log 10.6842 406 0.9617 0.6579 0.9617 0.9807
No log 10.7368 408 0.9044 0.6528 0.9044 0.9510
No log 10.7895 410 0.9059 0.6377 0.9059 0.9518
No log 10.8421 412 0.8774 0.6567 0.8774 0.9367
No log 10.8947 414 0.8581 0.6667 0.8581 0.9263
No log 10.9474 416 0.8466 0.6667 0.8466 0.9201
No log 11.0 418 0.8665 0.6617 0.8665 0.9308
No log 11.0526 420 0.9394 0.6620 0.9394 0.9692
No log 11.1053 422 0.9376 0.6475 0.9376 0.9683
No log 11.1579 424 0.8701 0.6515 0.8701 0.9328
No log 11.2105 426 0.8269 0.6667 0.8269 0.9094
No log 11.2632 428 0.8166 0.6849 0.8166 0.9036
No log 11.3158 430 0.8056 0.6621 0.8056 0.8975
No log 11.3684 432 0.7838 0.6761 0.7838 0.8853
No log 11.4211 434 0.7975 0.6765 0.7975 0.8930
No log 11.4737 436 0.8347 0.6515 0.8347 0.9136
No log 11.5263 438 0.8876 0.6047 0.8876 0.9421
No log 11.5789 440 0.9808 0.6434 0.9808 0.9904
No log 11.6316 442 1.1094 0.6275 1.1094 1.0533
No log 11.6842 444 1.1477 0.6125 1.1477 1.0713
No log 11.7368 446 1.0895 0.6234 1.0895 1.0438
No log 11.7895 448 1.0017 0.6099 1.0017 1.0009
No log 11.8421 450 0.9859 0.6143 0.9859 0.9929
No log 11.8947 452 0.9886 0.6099 0.9886 0.9943
No log 11.9474 454 0.9025 0.6522 0.9025 0.9500
No log 12.0 456 0.8577 0.6522 0.8577 0.9261
No log 12.0526 458 0.7998 0.6765 0.7998 0.8943
No log 12.1053 460 0.7373 0.6812 0.7373 0.8587
No log 12.1579 462 0.7166 0.6812 0.7166 0.8465
No log 12.2105 464 0.8053 0.6849 0.8053 0.8974
No log 12.2632 466 0.9128 0.6624 0.9128 0.9554
No log 12.3158 468 0.8908 0.6667 0.8908 0.9438
No log 12.3684 470 0.8094 0.6667 0.8094 0.8996
No log 12.4211 472 0.7073 0.7164 0.7073 0.8410
No log 12.4737 474 0.7016 0.7368 0.7016 0.8376
No log 12.5263 476 0.7198 0.6818 0.7198 0.8484
No log 12.5789 478 0.7994 0.6767 0.7994 0.8941
No log 12.6316 480 0.9242 0.6622 0.9242 0.9613
No log 12.6842 482 1.0968 0.6316 1.0968 1.0473
No log 12.7368 484 1.0812 0.6081 1.0812 1.0398
No log 12.7895 486 1.0790 0.6277 1.0790 1.0387
No log 12.8421 488 1.0081 0.6119 1.0081 1.0040
No log 12.8947 490 0.8668 0.6212 0.8668 0.9310
No log 12.9474 492 0.7777 0.6923 0.7777 0.8819
No log 13.0 494 0.7714 0.6667 0.7714 0.8783
No log 13.0526 496 0.8431 0.6622 0.8431 0.9182
No log 13.1053 498 0.8979 0.6797 0.8979 0.9476
0.4391 13.1579 500 0.8877 0.6294 0.8877 0.9422
0.4391 13.2105 502 0.8970 0.6423 0.8970 0.9471
0.4391 13.2632 504 0.9060 0.6212 0.9060 0.9519
0.4391 13.3158 506 0.9353 0.6061 0.9353 0.9671
0.4391 13.3684 508 0.9942 0.6061 0.9942 0.9971
0.4391 13.4211 510 1.0491 0.6061 1.0491 1.0242
0.4391 13.4737 512 1.1049 0.5882 1.1049 1.0511

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
4
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k5_task1_organization

Finetuned
(4222)
this model