ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run2_AugV5_k17_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6288
  • Qwk: 0.4441
  • Mse: 0.6288
  • Rmse: 0.7930

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0227 2 4.3590 -0.0291 4.3590 2.0878
No log 0.0455 4 2.4651 0.0405 2.4651 1.5701
No log 0.0682 6 1.4900 0.0006 1.4900 1.2207
No log 0.0909 8 1.3120 -0.0172 1.3120 1.1454
No log 0.1136 10 0.9639 0.0051 0.9639 0.9818
No log 0.1364 12 1.0618 0.0243 1.0618 1.0305
No log 0.1591 14 0.9415 0.1075 0.9415 0.9703
No log 0.1818 16 0.8496 0.2534 0.8496 0.9217
No log 0.2045 18 0.8673 0.2339 0.8673 0.9313
No log 0.2273 20 0.7595 0.2292 0.7595 0.8715
No log 0.25 22 0.7790 0.2374 0.7790 0.8826
No log 0.2727 24 1.0773 0.1499 1.0773 1.0379
No log 0.2955 26 1.2080 0.0994 1.2080 1.0991
No log 0.3182 28 0.9261 0.1483 0.9261 0.9623
No log 0.3409 30 0.7897 0.2325 0.7897 0.8887
No log 0.3636 32 0.7538 0.2988 0.7538 0.8682
No log 0.3864 34 0.7845 0.2174 0.7845 0.8857
No log 0.4091 36 1.1452 0.1144 1.1452 1.0702
No log 0.4318 38 1.4474 0.1299 1.4474 1.2031
No log 0.4545 40 1.2197 0.1612 1.2197 1.1044
No log 0.4773 42 1.1357 0.2337 1.1357 1.0657
No log 0.5 44 0.9862 0.2397 0.9862 0.9931
No log 0.5227 46 1.0333 0.2123 1.0333 1.0165
No log 0.5455 48 0.9638 0.2468 0.9638 0.9817
No log 0.5682 50 1.0153 0.2229 1.0153 1.0076
No log 0.5909 52 1.0610 0.1163 1.0610 1.0300
No log 0.6136 54 1.0939 0.1546 1.0939 1.0459
No log 0.6364 56 0.8650 0.3151 0.8650 0.9300
No log 0.6591 58 0.6742 0.3564 0.6742 0.8211
No log 0.6818 60 0.6767 0.3438 0.6767 0.8226
No log 0.7045 62 0.6591 0.3569 0.6591 0.8118
No log 0.7273 64 0.7669 0.3618 0.7669 0.8758
No log 0.75 66 0.8930 0.3148 0.8930 0.9450
No log 0.7727 68 0.9064 0.2518 0.9064 0.9520
No log 0.7955 70 0.9105 0.2457 0.9105 0.9542
No log 0.8182 72 0.7745 0.3086 0.7745 0.8801
No log 0.8409 74 0.6774 0.3646 0.6774 0.8231
No log 0.8636 76 0.6672 0.3552 0.6672 0.8168
No log 0.8864 78 0.7002 0.4097 0.7002 0.8368
No log 0.9091 80 1.0726 0.2930 1.0726 1.0356
No log 0.9318 82 1.5337 0.1836 1.5337 1.2384
No log 0.9545 84 1.4684 0.2000 1.4684 1.2118
No log 0.9773 86 1.0910 0.3167 1.0910 1.0445
No log 1.0 88 0.7624 0.3892 0.7624 0.8731
No log 1.0227 90 0.6765 0.3953 0.6765 0.8225
No log 1.0455 92 0.7397 0.4137 0.7397 0.8601
No log 1.0682 94 0.8852 0.3421 0.8852 0.9409
No log 1.0909 96 0.8778 0.3459 0.8778 0.9369
No log 1.1136 98 0.8284 0.3490 0.8284 0.9102
No log 1.1364 100 0.8031 0.3818 0.8031 0.8962
No log 1.1591 102 0.7331 0.4417 0.7331 0.8562
No log 1.1818 104 0.7158 0.4035 0.7158 0.8461
No log 1.2045 106 0.6719 0.4450 0.6719 0.8197
No log 1.2273 108 0.6861 0.4544 0.6861 0.8283
No log 1.25 110 0.7422 0.4211 0.7422 0.8615
No log 1.2727 112 0.8463 0.4021 0.8463 0.9199
No log 1.2955 114 0.9181 0.3527 0.9181 0.9582
No log 1.3182 116 0.9964 0.3046 0.9964 0.9982
No log 1.3409 118 0.7946 0.4495 0.7946 0.8914
No log 1.3636 120 0.6871 0.4497 0.6871 0.8289
No log 1.3864 122 0.6698 0.4498 0.6698 0.8184
No log 1.4091 124 0.6904 0.3832 0.6904 0.8309
No log 1.4318 126 0.8577 0.3282 0.8577 0.9261
No log 1.4545 128 0.9422 0.3570 0.9422 0.9707
No log 1.4773 130 1.1305 0.3033 1.1305 1.0633
No log 1.5 132 0.9240 0.3645 0.9240 0.9612
No log 1.5227 134 0.6485 0.3667 0.6485 0.8053
No log 1.5455 136 0.6548 0.4157 0.6548 0.8092
No log 1.5682 138 0.6476 0.4042 0.6476 0.8047
No log 1.5909 140 0.8107 0.3660 0.8107 0.9004
No log 1.6136 142 0.8493 0.3499 0.8493 0.9216
No log 1.6364 144 0.6984 0.4674 0.6984 0.8357
No log 1.6591 146 0.6568 0.4608 0.6568 0.8104
No log 1.6818 148 0.6478 0.4792 0.6478 0.8049
No log 1.7045 150 0.6844 0.4626 0.6844 0.8273
No log 1.7273 152 0.8699 0.4056 0.8699 0.9327
No log 1.75 154 0.9012 0.3776 0.9012 0.9493
No log 1.7727 156 0.7388 0.4693 0.7388 0.8595
No log 1.7955 158 0.6299 0.5091 0.6299 0.7936
No log 1.8182 160 0.6299 0.4797 0.6299 0.7937
No log 1.8409 162 0.6875 0.4860 0.6875 0.8291
No log 1.8636 164 0.8978 0.3569 0.8978 0.9475
No log 1.8864 166 0.9651 0.3229 0.9651 0.9824
No log 1.9091 168 0.8324 0.3817 0.8324 0.9124
No log 1.9318 170 0.6658 0.5175 0.6658 0.8160
No log 1.9545 172 0.6431 0.4395 0.6431 0.8019
No log 1.9773 174 0.7632 0.4989 0.7632 0.8736
No log 2.0 176 1.0390 0.2991 1.0390 1.0193
No log 2.0227 178 1.1197 0.3118 1.1197 1.0582
No log 2.0455 180 0.8435 0.3748 0.8435 0.9184
No log 2.0682 182 0.6218 0.4375 0.6218 0.7886
No log 2.0909 184 0.6221 0.4180 0.6221 0.7887
No log 2.1136 186 0.6223 0.4238 0.6223 0.7888
No log 2.1364 188 0.6342 0.3886 0.6342 0.7964
No log 2.1591 190 0.6465 0.3753 0.6465 0.8041
No log 2.1818 192 0.6569 0.3568 0.6569 0.8105
No log 2.2045 194 0.6750 0.3773 0.6750 0.8216
No log 2.2273 196 0.7375 0.3723 0.7375 0.8588
No log 2.25 198 0.7176 0.3691 0.7176 0.8471
No log 2.2727 200 0.6940 0.3288 0.6940 0.8331
No log 2.2955 202 0.7041 0.4492 0.7041 0.8391
No log 2.3182 204 0.6998 0.3376 0.6998 0.8365
No log 2.3409 206 0.8453 0.4372 0.8453 0.9194
No log 2.3636 208 1.0130 0.3379 1.0130 1.0065
No log 2.3864 210 0.9629 0.3440 0.9629 0.9813
No log 2.4091 212 0.8417 0.3606 0.8417 0.9174
No log 2.4318 214 0.8051 0.3594 0.8051 0.8973
No log 2.4545 216 0.7132 0.3894 0.7132 0.8445
No log 2.4773 218 0.6706 0.3896 0.6706 0.8189
No log 2.5 220 0.6681 0.4068 0.6681 0.8174
No log 2.5227 222 0.7011 0.4172 0.7011 0.8373
No log 2.5455 224 0.8655 0.4098 0.8655 0.9303
No log 2.5682 226 1.1171 0.3042 1.1171 1.0569
No log 2.5909 228 1.0186 0.3609 1.0186 1.0093
No log 2.6136 230 0.7910 0.4537 0.7910 0.8894
No log 2.6364 232 0.6868 0.4603 0.6868 0.8287
No log 2.6591 234 0.6593 0.4610 0.6593 0.8120
No log 2.6818 236 0.6878 0.4182 0.6878 0.8293
No log 2.7045 238 0.7607 0.3830 0.7607 0.8722
No log 2.7273 240 0.8128 0.3900 0.8128 0.9015
No log 2.75 242 0.7916 0.4465 0.7916 0.8897
No log 2.7727 244 0.7011 0.4451 0.7011 0.8373
No log 2.7955 246 0.6810 0.4511 0.6810 0.8252
No log 2.8182 248 0.7703 0.4771 0.7703 0.8777
No log 2.8409 250 0.7607 0.4846 0.7607 0.8722
No log 2.8636 252 0.6950 0.5070 0.6950 0.8337
No log 2.8864 254 0.6671 0.5092 0.6671 0.8167
No log 2.9091 256 0.6316 0.4686 0.6316 0.7947
No log 2.9318 258 0.6472 0.4744 0.6472 0.8045
No log 2.9545 260 0.7391 0.5260 0.7391 0.8597
No log 2.9773 262 0.7257 0.4996 0.7257 0.8519
No log 3.0 264 0.7197 0.4910 0.7197 0.8483
No log 3.0227 266 0.6525 0.4226 0.6525 0.8078
No log 3.0455 268 0.6749 0.4444 0.6749 0.8215
No log 3.0682 270 0.6514 0.4483 0.6514 0.8071
No log 3.0909 272 0.6497 0.4219 0.6497 0.8060
No log 3.1136 274 0.6670 0.4172 0.6670 0.8167
No log 3.1364 276 0.7256 0.5079 0.7256 0.8518
No log 3.1591 278 0.7924 0.4929 0.7924 0.8902
No log 3.1818 280 0.8168 0.4816 0.8168 0.9038
No log 3.2045 282 0.9155 0.4615 0.9155 0.9568
No log 3.2273 284 0.9212 0.4615 0.9212 0.9598
No log 3.25 286 0.7641 0.4861 0.7641 0.8741
No log 3.2727 288 0.6829 0.4490 0.6829 0.8264
No log 3.2955 290 0.6909 0.4431 0.6909 0.8312
No log 3.3182 292 0.6752 0.4296 0.6752 0.8217
No log 3.3409 294 0.7196 0.4431 0.7196 0.8483
No log 3.3636 296 0.7286 0.4234 0.7286 0.8536
No log 3.3864 298 0.7410 0.4357 0.7410 0.8608
No log 3.4091 300 0.6729 0.4161 0.6729 0.8203
No log 3.4318 302 0.6712 0.4418 0.6712 0.8193
No log 3.4545 304 0.7143 0.4438 0.7143 0.8452
No log 3.4773 306 0.7567 0.4734 0.7567 0.8699
No log 3.5 308 0.7473 0.4439 0.7473 0.8645
No log 3.5227 310 0.6817 0.4948 0.6817 0.8257
No log 3.5455 312 0.6894 0.4973 0.6894 0.8303
No log 3.5682 314 0.7349 0.4844 0.7349 0.8573
No log 3.5909 316 0.6612 0.5100 0.6612 0.8132
No log 3.6136 318 0.6437 0.4883 0.6437 0.8023
No log 3.6364 320 0.6277 0.4830 0.6277 0.7923
No log 3.6591 322 0.6418 0.5239 0.6418 0.8011
No log 3.6818 324 0.6435 0.5239 0.6435 0.8022
No log 3.7045 326 0.6364 0.4934 0.6364 0.7978
No log 3.7273 328 0.6206 0.4800 0.6206 0.7878
No log 3.75 330 0.6259 0.4647 0.6259 0.7911
No log 3.7727 332 0.6265 0.4677 0.6265 0.7915
No log 3.7955 334 0.6628 0.4826 0.6628 0.8141
No log 3.8182 336 0.7733 0.5066 0.7733 0.8794
No log 3.8409 338 0.8065 0.5054 0.8065 0.8981
No log 3.8636 340 0.7363 0.5015 0.7363 0.8581
No log 3.8864 342 0.7323 0.5220 0.7323 0.8557
No log 3.9091 344 0.7863 0.4989 0.7863 0.8867
No log 3.9318 346 0.6754 0.5081 0.6754 0.8218
No log 3.9545 348 0.6630 0.5085 0.6630 0.8143
No log 3.9773 350 0.8836 0.4419 0.8836 0.9400
No log 4.0 352 0.8953 0.4328 0.8953 0.9462
No log 4.0227 354 0.6998 0.4876 0.6998 0.8365
No log 4.0455 356 0.6139 0.5374 0.6139 0.7835
No log 4.0682 358 0.6154 0.5071 0.6154 0.7845
No log 4.0909 360 0.6292 0.5343 0.6292 0.7932
No log 4.1136 362 0.7183 0.5327 0.7183 0.8475
No log 4.1364 364 0.7663 0.4973 0.7663 0.8754
No log 4.1591 366 0.6732 0.5420 0.6732 0.8205
No log 4.1818 368 0.6780 0.4999 0.6780 0.8234
No log 4.2045 370 0.6721 0.5082 0.6721 0.8198
No log 4.2273 372 0.6484 0.5016 0.6484 0.8052
No log 4.25 374 0.7193 0.5133 0.7193 0.8481
No log 4.2727 376 0.7053 0.5186 0.7053 0.8398
No log 4.2955 378 0.6329 0.5010 0.6329 0.7956
No log 4.3182 380 0.6345 0.4823 0.6345 0.7965
No log 4.3409 382 0.6416 0.5092 0.6416 0.8010
No log 4.3636 384 0.6024 0.4644 0.6024 0.7761
No log 4.3864 386 0.6110 0.4684 0.6110 0.7817
No log 4.4091 388 0.6475 0.5124 0.6475 0.8047
No log 4.4318 390 0.6886 0.4474 0.6886 0.8298
No log 4.4545 392 0.7271 0.4761 0.7271 0.8527
No log 4.4773 394 0.6679 0.4662 0.6679 0.8172
No log 4.5 396 0.6359 0.4721 0.6359 0.7974
No log 4.5227 398 0.6346 0.4683 0.6346 0.7966
No log 4.5455 400 0.6274 0.4506 0.6274 0.7921
No log 4.5682 402 0.6155 0.4393 0.6155 0.7845
No log 4.5909 404 0.6310 0.4788 0.6310 0.7944
No log 4.6136 406 0.6230 0.4726 0.6230 0.7893
No log 4.6364 408 0.6033 0.4313 0.6033 0.7767
No log 4.6591 410 0.6066 0.4423 0.6066 0.7789
No log 4.6818 412 0.6264 0.5056 0.6264 0.7915
No log 4.7045 414 0.6294 0.4852 0.6294 0.7933
No log 4.7273 416 0.6172 0.5030 0.6172 0.7856
No log 4.75 418 0.6444 0.5324 0.6444 0.8027
No log 4.7727 420 0.6754 0.5329 0.6754 0.8218
No log 4.7955 422 0.6430 0.5205 0.6430 0.8019
No log 4.8182 424 0.6015 0.4364 0.6015 0.7755
No log 4.8409 426 0.6235 0.4522 0.6235 0.7896
No log 4.8636 428 0.6315 0.4574 0.6315 0.7947
No log 4.8864 430 0.6175 0.4467 0.6175 0.7858
No log 4.9091 432 0.6977 0.4727 0.6977 0.8353
No log 4.9318 434 0.8334 0.4607 0.8334 0.9129
No log 4.9545 436 0.8154 0.4454 0.8154 0.9030
No log 4.9773 438 0.7531 0.4722 0.7531 0.8678
No log 5.0 440 0.6996 0.4719 0.6996 0.8364
No log 5.0227 442 0.6587 0.5022 0.6587 0.8116
No log 5.0455 444 0.6593 0.5337 0.6593 0.8120
No log 5.0682 446 0.6345 0.5208 0.6345 0.7966
No log 5.0909 448 0.6734 0.4765 0.6734 0.8206
No log 5.1136 450 0.8865 0.4502 0.8865 0.9415
No log 5.1364 452 0.9620 0.4381 0.9620 0.9808
No log 5.1591 454 0.8208 0.4614 0.8208 0.9060
No log 5.1818 456 0.6467 0.4526 0.6467 0.8042
No log 5.2045 458 0.6190 0.4453 0.6190 0.7868
No log 5.2273 460 0.6299 0.4763 0.6299 0.7937
No log 5.25 462 0.6568 0.4827 0.6568 0.8105
No log 5.2727 464 0.6531 0.4548 0.6531 0.8081
No log 5.2955 466 0.6289 0.4555 0.6289 0.7930
No log 5.3182 468 0.6376 0.4541 0.6376 0.7985
No log 5.3409 470 0.6667 0.4819 0.6667 0.8165
No log 5.3636 472 0.6829 0.4919 0.6829 0.8264
No log 5.3864 474 0.6338 0.5307 0.6338 0.7961
No log 5.4091 476 0.6090 0.6099 0.6090 0.7804
No log 5.4318 478 0.6079 0.5885 0.6079 0.7797
No log 5.4545 480 0.6135 0.5091 0.6135 0.7833
No log 5.4773 482 0.6466 0.4669 0.6466 0.8041
No log 5.5 484 0.6134 0.4662 0.6134 0.7832
No log 5.5227 486 0.5922 0.5065 0.5922 0.7695
No log 5.5455 488 0.5999 0.5011 0.5999 0.7745
No log 5.5682 490 0.6152 0.4743 0.6152 0.7843
No log 5.5909 492 0.6019 0.4512 0.6019 0.7758
No log 5.6136 494 0.6167 0.4681 0.6167 0.7853
No log 5.6364 496 0.6144 0.4976 0.6144 0.7838
No log 5.6591 498 0.6036 0.4474 0.6036 0.7769
0.3722 5.6818 500 0.6761 0.4507 0.6761 0.8222
0.3722 5.7045 502 0.7702 0.4279 0.7702 0.8776
0.3722 5.7273 504 0.7338 0.4339 0.7338 0.8566
0.3722 5.75 506 0.6611 0.4804 0.6611 0.8131
0.3722 5.7727 508 0.7351 0.4371 0.7351 0.8574
0.3722 5.7955 510 0.7624 0.4637 0.7624 0.8731
0.3722 5.8182 512 0.6856 0.4824 0.6856 0.8280
0.3722 5.8409 514 0.6184 0.4610 0.6184 0.7864
0.3722 5.8636 516 0.6076 0.4289 0.6076 0.7795
0.3722 5.8864 518 0.6032 0.4181 0.6032 0.7767
0.3722 5.9091 520 0.6080 0.4116 0.6080 0.7797
0.3722 5.9318 522 0.6315 0.4257 0.6315 0.7947
0.3722 5.9545 524 0.6288 0.4441 0.6288 0.7930

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
4
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for MayBashendy/ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run2_AugV5_k17_task2_organization

Finetuned
(4222)
this model