ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k20_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5745
  • Qwk: 0.4315
  • Mse: 0.5745
  • Rmse: 0.7580

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0198 2 2.5746 -0.0407 2.5746 1.6045
No log 0.0396 4 1.3859 0.0503 1.3859 1.1772
No log 0.0594 6 1.2091 -0.1569 1.2091 1.0996
No log 0.0792 8 1.0766 -0.1238 1.0766 1.0376
No log 0.0990 10 0.9276 0.2164 0.9276 0.9631
No log 0.1188 12 0.6635 0.2280 0.6635 0.8146
No log 0.1386 14 0.7953 0.3907 0.7953 0.8918
No log 0.1584 16 0.8953 0.3027 0.8953 0.9462
No log 0.1782 18 0.7839 0.3620 0.7839 0.8854
No log 0.1980 20 0.7637 0.2738 0.7637 0.8739
No log 0.2178 22 1.0273 0.2496 1.0273 1.0136
No log 0.2376 24 1.4286 0.1444 1.4286 1.1952
No log 0.2574 26 1.4600 0.0381 1.4600 1.2083
No log 0.2772 28 1.1129 0.2421 1.1129 1.0549
No log 0.2970 30 0.7142 0.1272 0.7142 0.8451
No log 0.3168 32 0.6670 0.2374 0.6670 0.8167
No log 0.3366 34 0.6820 0.0846 0.6820 0.8258
No log 0.3564 36 0.6726 0.0393 0.6726 0.8201
No log 0.3762 38 0.6496 0.2405 0.6496 0.8060
No log 0.3960 40 0.6508 0.1327 0.6508 0.8067
No log 0.4158 42 0.7201 0.2142 0.7201 0.8486
No log 0.4356 44 0.8089 0.2408 0.8089 0.8994
No log 0.4554 46 0.8013 0.2087 0.8013 0.8952
No log 0.4752 48 0.6753 0.0889 0.6753 0.8218
No log 0.4950 50 0.6596 0.2930 0.6596 0.8121
No log 0.5149 52 0.6995 0.2930 0.6995 0.8363
No log 0.5347 54 0.6690 0.0428 0.6690 0.8179
No log 0.5545 56 0.6958 0.0 0.6958 0.8342
No log 0.5743 58 0.7065 0.0 0.7065 0.8405
No log 0.5941 60 0.6811 0.0428 0.6811 0.8253
No log 0.6139 62 0.6740 0.0393 0.6740 0.8210
No log 0.6337 64 0.6885 0.0410 0.6885 0.8297
No log 0.6535 66 0.7576 0.0495 0.7576 0.8704
No log 0.6733 68 0.9030 0.1328 0.9030 0.9503
No log 0.6931 70 1.0777 0.3019 1.0777 1.0381
No log 0.7129 72 1.1619 0.1765 1.1619 1.0779
No log 0.7327 74 0.9549 0.1650 0.9549 0.9772
No log 0.7525 76 0.8166 0.0905 0.8166 0.9037
No log 0.7723 78 0.6950 0.2002 0.6950 0.8336
No log 0.7921 80 0.6688 0.3084 0.6688 0.8178
No log 0.8119 82 0.6690 0.2751 0.6690 0.8179
No log 0.8317 84 0.7580 0.0460 0.7580 0.8706
No log 0.8515 86 0.9174 0.1318 0.9174 0.9578
No log 0.8713 88 0.9675 0.2844 0.9675 0.9836
No log 0.8911 90 0.8553 0.1323 0.8553 0.9248
No log 0.9109 92 0.6535 0.2374 0.6535 0.8084
No log 0.9307 94 0.6388 0.3020 0.6388 0.7993
No log 0.9505 96 0.6276 0.3020 0.6276 0.7922
No log 0.9703 98 0.6279 0.3437 0.6279 0.7924
No log 0.9901 100 0.7846 0.1993 0.7846 0.8858
No log 1.0099 102 0.8733 0.3085 0.8733 0.9345
No log 1.0297 104 0.9202 0.2928 0.9202 0.9593
No log 1.0495 106 0.7719 0.2948 0.7719 0.8786
No log 1.0693 108 0.7263 0.2454 0.7263 0.8522
No log 1.0891 110 0.6365 0.2713 0.6365 0.7978
No log 1.1089 112 0.6248 0.2713 0.6248 0.7905
No log 1.1287 114 0.6785 0.3029 0.6785 0.8237
No log 1.1485 116 0.6775 0.2947 0.6775 0.8231
No log 1.1683 118 0.6894 0.2909 0.6894 0.8303
No log 1.1881 120 0.6539 0.3183 0.6539 0.8086
No log 1.2079 122 0.6959 0.4134 0.6959 0.8342
No log 1.2277 124 0.6813 0.4206 0.6813 0.8254
No log 1.2475 126 0.6676 0.2513 0.6676 0.8171
No log 1.2673 128 0.7052 0.3854 0.7052 0.8397
No log 1.2871 130 0.8401 0.3807 0.8401 0.9166
No log 1.3069 132 0.7525 0.4219 0.7525 0.8675
No log 1.3267 134 0.7924 0.4076 0.7924 0.8902
No log 1.3465 136 0.7309 0.4783 0.7309 0.8549
No log 1.3663 138 0.6363 0.4215 0.6363 0.7977
No log 1.3861 140 0.6250 0.3681 0.6250 0.7906
No log 1.4059 142 0.7059 0.3876 0.7059 0.8402
No log 1.4257 144 0.6969 0.3902 0.6969 0.8348
No log 1.4455 146 0.6326 0.3399 0.6326 0.7954
No log 1.4653 148 0.5644 0.5362 0.5644 0.7513
No log 1.4851 150 0.5916 0.4542 0.5916 0.7691
No log 1.5050 152 0.5783 0.5016 0.5783 0.7605
No log 1.5248 154 0.5660 0.5306 0.5660 0.7523
No log 1.5446 156 0.5885 0.3541 0.5885 0.7671
No log 1.5644 158 0.6761 0.4463 0.6761 0.8223
No log 1.5842 160 0.6325 0.4037 0.6325 0.7953
No log 1.6040 162 0.5630 0.4747 0.5630 0.7503
No log 1.6238 164 0.6376 0.4905 0.6376 0.7985
No log 1.6436 166 0.6315 0.4905 0.6315 0.7947
No log 1.6634 168 0.5703 0.4402 0.5703 0.7552
No log 1.6832 170 0.6948 0.3017 0.6948 0.8335
No log 1.7030 172 0.7020 0.3551 0.7020 0.8378
No log 1.7228 174 0.6919 0.3805 0.6919 0.8318
No log 1.7426 176 0.6422 0.4058 0.6422 0.8014
No log 1.7624 178 0.5703 0.3738 0.5703 0.7552
No log 1.7822 180 0.5666 0.4105 0.5666 0.7527
No log 1.8020 182 0.6049 0.4335 0.6049 0.7778
No log 1.8218 184 0.6020 0.4858 0.6020 0.7759
No log 1.8416 186 0.5675 0.4692 0.5675 0.7534
No log 1.8614 188 0.5779 0.4906 0.5779 0.7602
No log 1.8812 190 0.6783 0.4768 0.6783 0.8236
No log 1.9010 192 0.8584 0.3845 0.8584 0.9265
No log 1.9208 194 0.8794 0.3845 0.8794 0.9378
No log 1.9406 196 0.7958 0.3924 0.7958 0.8921
No log 1.9604 198 0.6304 0.4598 0.6304 0.7940
No log 1.9802 200 0.5410 0.5846 0.5410 0.7356
No log 2.0 202 0.5363 0.5625 0.5363 0.7323
No log 2.0198 204 0.5819 0.4747 0.5819 0.7629
No log 2.0396 206 0.6377 0.4057 0.6377 0.7986
No log 2.0594 208 0.5969 0.4150 0.5969 0.7726
No log 2.0792 210 0.5640 0.5603 0.5640 0.7510
No log 2.0990 212 0.5959 0.4377 0.5959 0.7719
No log 2.1188 214 0.7111 0.4114 0.7111 0.8433
No log 2.1386 216 0.7123 0.3889 0.7123 0.8440
No log 2.1584 218 0.5385 0.5300 0.5385 0.7338
No log 2.1782 220 0.4921 0.5151 0.4921 0.7015
No log 2.1980 222 0.4799 0.5488 0.4799 0.6927
No log 2.2178 224 0.5628 0.5168 0.5628 0.7502
No log 2.2376 226 0.7538 0.4617 0.7538 0.8682
No log 2.2574 228 0.7127 0.4906 0.7127 0.8442
No log 2.2772 230 0.5659 0.5421 0.5659 0.7522
No log 2.2970 232 0.5803 0.5236 0.5803 0.7618
No log 2.3168 234 0.7012 0.4462 0.7012 0.8374
No log 2.3366 236 0.6979 0.4462 0.6979 0.8354
No log 2.3564 238 0.6186 0.4170 0.6186 0.7865
No log 2.3762 240 0.6538 0.4462 0.6538 0.8086
No log 2.3960 242 0.6136 0.5175 0.6136 0.7833
No log 2.4158 244 0.5167 0.6034 0.5167 0.7188
No log 2.4356 246 0.5145 0.5826 0.5145 0.7173
No log 2.4554 248 0.6288 0.4494 0.6288 0.7930
No log 2.4752 250 0.6736 0.4511 0.6736 0.8207
No log 2.4950 252 0.6249 0.4477 0.6249 0.7905
No log 2.5149 254 0.5694 0.5200 0.5694 0.7546
No log 2.5347 256 0.6386 0.4261 0.6386 0.7991
No log 2.5545 258 0.6118 0.5200 0.6118 0.7822
No log 2.5743 260 0.5558 0.4517 0.5558 0.7455
No log 2.5941 262 0.5377 0.4752 0.5377 0.7333
No log 2.6139 264 0.5298 0.5868 0.5298 0.7279
No log 2.6337 266 0.5304 0.5991 0.5304 0.7283
No log 2.6535 268 0.6379 0.4670 0.6379 0.7987
No log 2.6733 270 0.9407 0.4257 0.9407 0.9699
No log 2.6931 272 0.9357 0.4257 0.9357 0.9673
No log 2.7129 274 0.6632 0.4874 0.6632 0.8144
No log 2.7327 276 0.5016 0.6245 0.5016 0.7083
No log 2.7525 278 0.4953 0.6158 0.4953 0.7038
No log 2.7723 280 0.5094 0.6491 0.5094 0.7138
No log 2.7921 282 0.6173 0.5003 0.6173 0.7857
No log 2.8119 284 0.6780 0.4719 0.6780 0.8234
No log 2.8317 286 0.5924 0.4987 0.5924 0.7696
No log 2.8515 288 0.5210 0.5929 0.5210 0.7218
No log 2.8713 290 0.5156 0.6223 0.5156 0.7181
No log 2.8911 292 0.5333 0.5348 0.5333 0.7303
No log 2.9109 294 0.5964 0.5387 0.5964 0.7722
No log 2.9307 296 0.7525 0.4177 0.7525 0.8675
No log 2.9505 298 0.7760 0.4177 0.7760 0.8809
No log 2.9703 300 0.6457 0.5267 0.6457 0.8036
No log 2.9901 302 0.5364 0.5738 0.5364 0.7324
No log 3.0099 304 0.5130 0.5614 0.5130 0.7162
No log 3.0297 306 0.5225 0.5797 0.5225 0.7228
No log 3.0495 308 0.5063 0.5683 0.5063 0.7116
No log 3.0693 310 0.5010 0.5494 0.5010 0.7078
No log 3.0891 312 0.5402 0.4614 0.5402 0.7350
No log 3.1089 314 0.5217 0.4345 0.5217 0.7223
No log 3.1287 316 0.5173 0.5393 0.5173 0.7192
No log 3.1485 318 0.5754 0.4642 0.5754 0.7585
No log 3.1683 320 0.5423 0.4414 0.5423 0.7364
No log 3.1881 322 0.4989 0.5657 0.4989 0.7063
No log 3.2079 324 0.5064 0.5143 0.5064 0.7116
No log 3.2277 326 0.5537 0.5200 0.5537 0.7441
No log 3.2475 328 0.6676 0.4756 0.6676 0.8170
No log 3.2673 330 0.6795 0.4946 0.6795 0.8243
No log 3.2871 332 0.5815 0.4711 0.5815 0.7626
No log 3.3069 334 0.5451 0.4719 0.5451 0.7383
No log 3.3267 336 0.5548 0.4788 0.5548 0.7448
No log 3.3465 338 0.5429 0.5248 0.5429 0.7368
No log 3.3663 340 0.5974 0.3590 0.5974 0.7729
No log 3.3861 342 0.6656 0.3471 0.6656 0.8158
No log 3.4059 344 0.6704 0.3718 0.6704 0.8188
No log 3.4257 346 0.6122 0.4009 0.6122 0.7824
No log 3.4455 348 0.5254 0.5213 0.5254 0.7248
No log 3.4653 350 0.5146 0.5734 0.5146 0.7174
No log 3.4851 352 0.5619 0.4834 0.5619 0.7496
No log 3.5050 354 0.6331 0.4527 0.6331 0.7957
No log 3.5248 356 0.5798 0.5046 0.5798 0.7615
No log 3.5446 358 0.5305 0.5956 0.5305 0.7284
No log 3.5644 360 0.5254 0.5687 0.5254 0.7248
No log 3.5842 362 0.5372 0.5611 0.5372 0.7329
No log 3.6040 364 0.5419 0.5611 0.5419 0.7361
No log 3.6238 366 0.5295 0.5505 0.5295 0.7277
No log 3.6436 368 0.5353 0.5268 0.5353 0.7316
No log 3.6634 370 0.5535 0.5373 0.5535 0.7440
No log 3.6832 372 0.5677 0.5143 0.5677 0.7535
No log 3.7030 374 0.5762 0.4972 0.5762 0.7591
No log 3.7228 376 0.5963 0.4742 0.5963 0.7722
No log 3.7426 378 0.6153 0.3443 0.6153 0.7844
No log 3.7624 380 0.5972 0.4361 0.5972 0.7728
No log 3.7822 382 0.6033 0.3894 0.6033 0.7767
No log 3.8020 384 0.5970 0.4182 0.5970 0.7727
No log 3.8218 386 0.6131 0.3916 0.6131 0.7830
No log 3.8416 388 0.7017 0.3202 0.7017 0.8377
No log 3.8614 390 0.7317 0.3633 0.7317 0.8554
No log 3.8812 392 0.6508 0.4334 0.6508 0.8067
No log 3.9010 394 0.6183 0.4681 0.6183 0.7864
No log 3.9208 396 0.6394 0.5246 0.6394 0.7996
No log 3.9406 398 0.5950 0.5457 0.5950 0.7714
No log 3.9604 400 0.5778 0.4697 0.5778 0.7601
No log 3.9802 402 0.6643 0.5363 0.6643 0.8151
No log 4.0 404 0.7028 0.4634 0.7028 0.8383
No log 4.0198 406 0.7286 0.4686 0.7286 0.8536
No log 4.0396 408 0.6401 0.4892 0.6401 0.8001
No log 4.0594 410 0.5863 0.5406 0.5863 0.7657
No log 4.0792 412 0.5617 0.4524 0.5617 0.7495
No log 4.0990 414 0.5749 0.4601 0.5749 0.7582
No log 4.1188 416 0.5952 0.4864 0.5952 0.7715
No log 4.1386 418 0.5879 0.4268 0.5879 0.7667
No log 4.1584 420 0.5999 0.3948 0.5999 0.7746
No log 4.1782 422 0.6221 0.5237 0.6221 0.7887
No log 4.1980 424 0.6152 0.4308 0.6152 0.7844
No log 4.2178 426 0.6107 0.4094 0.6107 0.7815
No log 4.2376 428 0.6205 0.4801 0.6205 0.7877
No log 4.2574 430 0.6194 0.4782 0.6194 0.7870
No log 4.2772 432 0.5970 0.4717 0.5970 0.7726
No log 4.2970 434 0.6351 0.3739 0.6351 0.7970
No log 4.3168 436 0.6881 0.3892 0.6881 0.8295
No log 4.3366 438 0.6626 0.3937 0.6626 0.8140
No log 4.3564 440 0.5846 0.4171 0.5846 0.7646
No log 4.3762 442 0.5704 0.4704 0.5704 0.7552
No log 4.3960 444 0.5914 0.4190 0.5914 0.7690
No log 4.4158 446 0.5713 0.4451 0.5713 0.7558
No log 4.4356 448 0.5706 0.4217 0.5706 0.7554
No log 4.4554 450 0.5881 0.4022 0.5881 0.7669
No log 4.4752 452 0.5873 0.3762 0.5873 0.7663
No log 4.4950 454 0.5825 0.4514 0.5825 0.7632
No log 4.5149 456 0.6210 0.4239 0.6210 0.7880
No log 4.5347 458 0.6497 0.4531 0.6497 0.8061
No log 4.5545 460 0.6440 0.4444 0.6440 0.8025
No log 4.5743 462 0.6284 0.4576 0.6284 0.7927
No log 4.5941 464 0.5730 0.4663 0.5730 0.7569
No log 4.6139 466 0.5442 0.5271 0.5442 0.7377
No log 4.6337 468 0.5339 0.4526 0.5339 0.7307
No log 4.6535 470 0.5462 0.5479 0.5462 0.7391
No log 4.6733 472 0.5890 0.5268 0.5890 0.7675
No log 4.6931 474 0.6705 0.4859 0.6705 0.8188
No log 4.7129 476 0.6855 0.4805 0.6855 0.8279
No log 4.7327 478 0.6485 0.4914 0.6485 0.8053
No log 4.7525 480 0.6016 0.4817 0.6016 0.7757
No log 4.7723 482 0.6044 0.4706 0.6044 0.7774
No log 4.7921 484 0.6113 0.4634 0.6113 0.7818
No log 4.8119 486 0.6252 0.4634 0.6252 0.7907
No log 4.8317 488 0.5725 0.4282 0.5725 0.7566
No log 4.8515 490 0.5482 0.4857 0.5482 0.7404
No log 4.8713 492 0.5582 0.4375 0.5582 0.7471
No log 4.8911 494 0.5576 0.4292 0.5576 0.7468
No log 4.9109 496 0.5467 0.5003 0.5467 0.7394
No log 4.9307 498 0.5477 0.5079 0.5477 0.7401
0.3339 4.9505 500 0.5507 0.5178 0.5507 0.7421
0.3339 4.9703 502 0.5585 0.4427 0.5585 0.7473
0.3339 4.9901 504 0.5799 0.4087 0.5799 0.7615
0.3339 5.0099 506 0.5816 0.4087 0.5816 0.7626
0.3339 5.0297 508 0.5644 0.4066 0.5644 0.7513
0.3339 5.0495 510 0.5745 0.4315 0.5745 0.7580

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
9
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k20_task7_organization

Finetuned
(4222)
this model