ArabicNewSplits7_FineTuningAraBERT_run3_AugV5_k15_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.1783
  • Qwk: -0.0028
  • Mse: 1.1783
  • Rmse: 1.0855

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0488 2 3.7846 -0.0068 3.7846 1.9454
No log 0.0976 4 1.9715 0.0531 1.9715 1.4041
No log 0.1463 6 2.3801 0.0038 2.3801 1.5428
No log 0.1951 8 2.0195 -0.0060 2.0195 1.4211
No log 0.2439 10 0.9274 -0.0056 0.9274 0.9630
No log 0.2927 12 0.7857 -0.0287 0.7857 0.8864
No log 0.3415 14 0.7818 -0.1244 0.7818 0.8842
No log 0.3902 16 1.3540 -0.0479 1.3540 1.1636
No log 0.4390 18 1.3660 -0.0500 1.3660 1.1688
No log 0.4878 20 1.0813 -0.0695 1.0813 1.0399
No log 0.5366 22 0.8799 0.0609 0.8799 0.9380
No log 0.5854 24 0.8309 0.0377 0.8309 0.9115
No log 0.6341 26 0.7573 -0.0101 0.7573 0.8702
No log 0.6829 28 0.7208 -0.0069 0.7208 0.8490
No log 0.7317 30 0.7683 0.0416 0.7683 0.8765
No log 0.7805 32 0.9391 -0.0101 0.9391 0.9691
No log 0.8293 34 0.9606 0.0176 0.9606 0.9801
No log 0.8780 36 1.0409 -0.0468 1.0409 1.0202
No log 0.9268 38 0.9998 -0.0423 0.9998 0.9999
No log 0.9756 40 0.8364 0.0588 0.8364 0.9145
No log 1.0244 42 0.7584 -0.0131 0.7584 0.8709
No log 1.0732 44 0.6898 0.0 0.6898 0.8305
No log 1.1220 46 0.6922 0.0 0.6922 0.8320
No log 1.1707 48 0.7318 0.0506 0.7318 0.8554
No log 1.2195 50 0.9772 -0.0982 0.9772 0.9885
No log 1.2683 52 1.0739 -0.0468 1.0739 1.0363
No log 1.3171 54 0.8417 0.0017 0.8417 0.9174
No log 1.3659 56 0.7287 0.0506 0.7287 0.8537
No log 1.4146 58 0.7593 0.1202 0.7593 0.8714
No log 1.4634 60 1.1596 0.0518 1.1596 1.0769
No log 1.5122 62 1.0117 -0.0943 1.0117 1.0058
No log 1.5610 64 0.7548 0.1259 0.7548 0.8688
No log 1.6098 66 0.8317 -0.0371 0.8317 0.9120
No log 1.6585 68 0.7260 0.0416 0.7260 0.8521
No log 1.7073 70 0.7197 0.0555 0.7197 0.8483
No log 1.7561 72 0.7025 0.0506 0.7025 0.8381
No log 1.8049 74 1.2826 0.0133 1.2826 1.1325
No log 1.8537 76 1.8716 -0.0000 1.8716 1.3681
No log 1.9024 78 1.5932 -0.0475 1.5932 1.2622
No log 1.9512 80 0.9978 -0.1605 0.9978 0.9989
No log 2.0 82 0.8603 0.0953 0.8603 0.9275
No log 2.0488 84 0.7749 0.1675 0.7749 0.8803
No log 2.0976 86 0.7547 -0.0091 0.7547 0.8687
No log 2.1463 88 0.9294 -0.0007 0.9294 0.9641
No log 2.1951 90 1.1780 -0.0842 1.1780 1.0854
No log 2.2439 92 0.9456 -0.0391 0.9456 0.9724
No log 2.2927 94 1.0249 -0.0837 1.0249 1.0124
No log 2.3415 96 0.8823 -0.0630 0.8823 0.9393
No log 2.3902 98 0.8372 0.0209 0.8372 0.9150
No log 2.4390 100 0.8175 0.1030 0.8175 0.9041
No log 2.4878 102 0.8407 0.0245 0.8407 0.9169
No log 2.5366 104 0.9844 -0.0076 0.9844 0.9922
No log 2.5854 106 0.9446 -0.0030 0.9446 0.9719
No log 2.6341 108 0.8405 -0.0566 0.8405 0.9168
No log 2.6829 110 0.9071 -0.0408 0.9071 0.9524
No log 2.7317 112 0.7937 0.1080 0.7937 0.8909
No log 2.7805 114 1.1579 -0.0648 1.1579 1.0761
No log 2.8293 116 1.5235 -0.0221 1.5235 1.2343
No log 2.8780 118 1.2502 -0.0468 1.2502 1.1181
No log 2.9268 120 0.8799 -0.0122 0.8799 0.9380
No log 2.9756 122 0.7485 0.1828 0.7485 0.8652
No log 3.0244 124 0.7385 0.0662 0.7385 0.8593
No log 3.0732 126 0.7350 0.0436 0.7350 0.8573
No log 3.1220 128 1.1039 -0.0200 1.1039 1.0507
No log 3.1707 130 1.3297 -0.0084 1.3297 1.1531
No log 3.2195 132 0.9381 -0.0163 0.9381 0.9685
No log 3.2683 134 0.7822 -0.0271 0.7822 0.8844
No log 3.3171 136 0.8146 -0.0648 0.8146 0.9026
No log 3.3659 138 0.8647 0.0964 0.8647 0.9299
No log 3.4146 140 0.8940 0.0497 0.8940 0.9455
No log 3.4634 142 0.8137 0.0129 0.8137 0.9020
No log 3.5122 144 0.9735 -0.0099 0.9735 0.9867
No log 3.5610 146 0.9487 0.0180 0.9487 0.9740
No log 3.6098 148 0.7929 0.0856 0.7929 0.8905
No log 3.6585 150 0.9222 0.0424 0.9222 0.9603
No log 3.7073 152 0.8354 0.0913 0.8354 0.9140
No log 3.7561 154 0.7320 0.0513 0.7320 0.8556
No log 3.8049 156 0.7537 -0.0939 0.7537 0.8682
No log 3.8537 158 0.7834 -0.1266 0.7834 0.8851
No log 3.9024 160 0.8387 0.0289 0.8387 0.9158
No log 3.9512 162 1.1367 -0.0870 1.1367 1.0662
No log 4.0 164 1.1121 -0.0513 1.1121 1.0546
No log 4.0488 166 0.9390 -0.0694 0.9390 0.9690
No log 4.0976 168 0.9010 0.0580 0.9010 0.9492
No log 4.1463 170 0.9162 0.0728 0.9162 0.9572
No log 4.1951 172 0.8524 0.0851 0.8524 0.9233
No log 4.2439 174 0.9676 -0.0393 0.9676 0.9837
No log 4.2927 176 0.9378 -0.0391 0.9378 0.9684
No log 4.3415 178 0.8493 0.0600 0.8493 0.9216
No log 4.3902 180 0.8014 0.0741 0.8014 0.8952
No log 4.4390 182 0.7980 -0.0717 0.7980 0.8933
No log 4.4878 184 0.8580 0.0364 0.8580 0.9263
No log 4.5366 186 0.7979 -0.1597 0.7979 0.8932
No log 4.5854 188 0.8027 0.0732 0.8027 0.8959
No log 4.6341 190 1.0961 -0.0533 1.0961 1.0470
No log 4.6829 192 1.0096 -0.0097 1.0096 1.0048
No log 4.7317 194 0.7708 0.0869 0.7708 0.8780
No log 4.7805 196 0.7566 0.0 0.7566 0.8698
No log 4.8293 198 0.7695 -0.0032 0.7695 0.8772
No log 4.8780 200 0.8172 0.1249 0.8172 0.9040
No log 4.9268 202 0.8275 0.0828 0.8275 0.9097
No log 4.9756 204 0.8772 0.0700 0.8772 0.9366
No log 5.0244 206 0.9327 -0.0711 0.9327 0.9658
No log 5.0732 208 0.8396 -0.0118 0.8396 0.9163
No log 5.1220 210 0.8459 -0.0705 0.8459 0.9197
No log 5.1707 212 0.8516 -0.1151 0.8516 0.9228
No log 5.2195 214 0.8804 -0.0444 0.8804 0.9383
No log 5.2683 216 0.9223 0.0741 0.9223 0.9604
No log 5.3171 218 0.9496 -0.0123 0.9496 0.9745
No log 5.3659 220 1.0287 -0.0214 1.0287 1.0143
No log 5.4146 222 0.9515 -0.0295 0.9515 0.9754
No log 5.4634 224 0.9331 -0.0095 0.9331 0.9660
No log 5.5122 226 0.8992 0.0361 0.8992 0.9482
No log 5.5610 228 0.9309 0.0525 0.9309 0.9648
No log 5.6098 230 0.8755 0.1485 0.8755 0.9357
No log 5.6585 232 0.8641 -0.0200 0.8641 0.9296
No log 5.7073 234 0.8657 0.0239 0.8657 0.9304
No log 5.7561 236 0.8925 0.0574 0.8925 0.9447
No log 5.8049 238 0.8940 0.0574 0.8940 0.9455
No log 5.8537 240 0.7800 0.0798 0.7800 0.8832
No log 5.9024 242 0.7878 0.0592 0.7878 0.8876
No log 5.9512 244 0.7967 0.0522 0.7967 0.8926
No log 6.0 246 0.8612 0.1986 0.8612 0.9280
No log 6.0488 248 0.8527 0.2005 0.8527 0.9234
No log 6.0976 250 0.8856 0.2005 0.8856 0.9410
No log 6.1463 252 1.0245 -0.1224 1.0245 1.0122
No log 6.1951 254 1.0861 -0.1228 1.0861 1.0422
No log 6.2439 256 0.9359 0.0377 0.9359 0.9674
No log 6.2927 258 0.8866 0.1407 0.8866 0.9416
No log 6.3415 260 0.8256 0.0451 0.8256 0.9086
No log 6.3902 262 0.8108 0.1475 0.8108 0.9004
No log 6.4390 264 0.9721 -0.0854 0.9721 0.9859
No log 6.4878 266 1.0590 -0.1228 1.0590 1.0291
No log 6.5366 268 1.0097 -0.0854 1.0097 1.0049
No log 6.5854 270 0.8872 0.1475 0.8872 0.9419
No log 6.6341 272 0.8539 0.0301 0.8539 0.9241
No log 6.6829 274 0.8102 0.1722 0.8102 0.9001
No log 6.7317 276 0.8412 0.1379 0.8412 0.9171
No log 6.7805 278 0.9610 -0.0456 0.9610 0.9803
No log 6.8293 280 0.9475 0.0438 0.9475 0.9734
No log 6.8780 282 0.8743 0.0827 0.8743 0.9350
No log 6.9268 284 0.8837 -0.0056 0.8837 0.9400
No log 6.9756 286 0.8496 0.0093 0.8496 0.9217
No log 7.0244 288 0.8243 0.0732 0.8243 0.9079
No log 7.0732 290 0.8377 0.0562 0.8377 0.9153
No log 7.1220 292 0.9053 0.0362 0.9053 0.9515
No log 7.1707 294 0.9462 -0.0163 0.9462 0.9727
No log 7.2195 296 0.8664 0.1286 0.8664 0.9308
No log 7.2683 298 0.8253 0.0639 0.8253 0.9085
No log 7.3171 300 0.8518 0.0851 0.8518 0.9229
No log 7.3659 302 0.9425 0.0301 0.9425 0.9708
No log 7.4146 304 1.1073 -0.0409 1.1073 1.0523
No log 7.4634 306 1.3191 -0.0238 1.3191 1.1485
No log 7.5122 308 1.1636 -0.0855 1.1636 1.0787
No log 7.5610 310 0.9439 0.0831 0.9439 0.9715
No log 7.6098 312 0.8789 -0.0377 0.8789 0.9375
No log 7.6585 314 0.8676 -0.1176 0.8676 0.9314
No log 7.7073 316 0.8728 0.0538 0.8728 0.9343
No log 7.7561 318 0.9508 -0.0471 0.9508 0.9751
No log 7.8049 320 1.0652 -0.0854 1.0652 1.0321
No log 7.8537 322 0.9572 0.0793 0.9572 0.9784
No log 7.9024 324 0.8902 -0.0806 0.8902 0.9435
No log 7.9512 326 0.8676 -0.1278 0.8676 0.9314
No log 8.0 328 0.8623 0.1001 0.8623 0.9286
No log 8.0488 330 0.8691 0.0913 0.8691 0.9323
No log 8.0976 332 0.8427 -0.0054 0.8427 0.9180
No log 8.1463 334 0.8229 0.0776 0.8229 0.9071
No log 8.1951 336 0.7904 0.1311 0.7904 0.8890
No log 8.2439 338 0.7584 0.1311 0.7584 0.8709
No log 8.2927 340 0.7510 0.1311 0.7510 0.8666
No log 8.3415 342 0.7737 0.1254 0.7737 0.8796
No log 8.3902 344 0.8217 0.1199 0.8217 0.9065
No log 8.4390 346 0.8739 -0.1392 0.8739 0.9348
No log 8.4878 348 0.9747 -0.0076 0.9747 0.9873
No log 8.5366 350 1.0127 0.0267 1.0127 1.0063
No log 8.5854 352 0.9357 0.0062 0.9357 0.9673
No log 8.6341 354 0.8861 0.0049 0.8861 0.9413
No log 8.6829 356 0.8837 0.0364 0.8837 0.9401
No log 8.7317 358 0.8191 -0.0259 0.8191 0.9051
No log 8.7805 360 0.7725 0.1254 0.7725 0.8789
No log 8.8293 362 1.0963 0.0104 1.0963 1.0471
No log 8.8780 364 1.3258 -0.0116 1.3258 1.1514
No log 8.9268 366 1.0832 -0.0557 1.0832 1.0408
No log 8.9756 368 0.8019 0.1565 0.8019 0.8955
No log 9.0244 370 0.7098 0.0914 0.7098 0.8425
No log 9.0732 372 0.7306 0.0914 0.7306 0.8547
No log 9.1220 374 0.7591 0.1371 0.7591 0.8713
No log 9.1707 376 0.9385 -0.0031 0.9385 0.9688
No log 9.2195 378 1.0827 0.0125 1.0827 1.0405
No log 9.2683 380 0.9839 -0.0471 0.9839 0.9919
No log 9.3171 382 0.8826 0.0426 0.8826 0.9395
No log 9.3659 384 0.8767 -0.0598 0.8767 0.9363
No log 9.4146 386 0.8794 0.0778 0.8794 0.9378
No log 9.4634 388 0.9363 0.1286 0.9363 0.9676
No log 9.5122 390 0.9824 -0.0471 0.9824 0.9912
No log 9.5610 392 0.9337 0.0793 0.9337 0.9663
No log 9.6098 394 0.8494 0.0828 0.8494 0.9216
No log 9.6585 396 0.8914 0.0913 0.8914 0.9441
No log 9.7073 398 0.8855 0.0600 0.8855 0.9410
No log 9.7561 400 0.9164 0.0456 0.9164 0.9573
No log 9.8049 402 0.9071 0.0913 0.9071 0.9524
No log 9.8537 404 0.8469 0.0282 0.8469 0.9203
No log 9.9024 406 0.8337 0.0282 0.8337 0.9130
No log 9.9512 408 0.8680 0.1001 0.8680 0.9317
No log 10.0 410 0.8810 0.0525 0.8810 0.9386
No log 10.0488 412 0.8712 0.0247 0.8712 0.9334
No log 10.0976 414 0.8750 -0.0444 0.8750 0.9354
No log 10.1463 416 0.8591 -0.0320 0.8591 0.9269
No log 10.1951 418 0.8114 -0.0086 0.8114 0.9008
No log 10.2439 420 0.8877 0.0346 0.8877 0.9422
No log 10.2927 422 0.9568 -0.0101 0.9568 0.9782
No log 10.3415 424 0.9854 -0.0532 0.9854 0.9927
No log 10.3902 426 0.9411 -0.0031 0.9411 0.9701
No log 10.4390 428 0.9164 0.0205 0.9164 0.9573
No log 10.4878 430 0.9552 -0.0441 0.9552 0.9773
No log 10.5366 432 0.9781 -0.0991 0.9781 0.9890
No log 10.5854 434 1.0110 -0.0441 1.0110 1.0055
No log 10.6341 436 1.0048 -0.0456 1.0048 1.0024
No log 10.6829 438 0.9260 -0.0694 0.9260 0.9623
No log 10.7317 440 0.8900 -0.0878 0.8900 0.9434
No log 10.7805 442 0.8921 -0.1628 0.8921 0.9445
No log 10.8293 444 0.9142 -0.0985 0.9142 0.9561
No log 10.8780 446 0.9679 -0.0441 0.9679 0.9838
No log 10.9268 448 0.9450 -0.1102 0.9450 0.9721
No log 10.9756 450 0.8976 -0.0549 0.8976 0.9474
No log 11.0244 452 0.9045 -0.0357 0.9045 0.9511
No log 11.0732 454 0.9187 -0.1280 0.9187 0.9585
No log 11.1220 456 0.9148 -0.0923 0.9148 0.9565
No log 11.1707 458 0.8986 -0.1335 0.8986 0.9479
No log 11.2195 460 0.9011 0.0152 0.9011 0.9492
No log 11.2683 462 0.8638 0.0600 0.8638 0.9294
No log 11.3171 464 0.7834 0.1675 0.7834 0.8851
No log 11.3659 466 0.7743 0.2078 0.7743 0.8799
No log 11.4146 468 0.8275 0.1775 0.8275 0.9097
No log 11.4634 470 0.9014 -0.0008 0.9014 0.9494
No log 11.5122 472 0.8894 0.0871 0.8894 0.9431
No log 11.5610 474 0.8095 0.1740 0.8095 0.8997
No log 11.6098 476 0.8309 0.0776 0.8309 0.9116
No log 11.6585 478 0.8381 0.0821 0.8381 0.9155
No log 11.7073 480 0.8492 0.2431 0.8492 0.9215
No log 11.7561 482 0.8349 0.2431 0.8349 0.9137
No log 11.8049 484 0.8290 0.1807 0.8290 0.9105
No log 11.8537 486 0.8490 -0.0152 0.8490 0.9214
No log 11.9024 488 0.8503 -0.0059 0.8503 0.9221
No log 11.9512 490 0.8591 0.1311 0.8591 0.9269
No log 12.0 492 0.9357 -0.0441 0.9357 0.9673
No log 12.0488 494 0.9457 -0.0456 0.9457 0.9725
No log 12.0976 496 0.9108 0.0490 0.9108 0.9544
No log 12.1463 498 0.7997 0.0357 0.7997 0.8942
0.3007 12.1951 500 0.7665 0.0412 0.7665 0.8755
0.3007 12.2439 502 0.8110 0.0357 0.8110 0.9006
0.3007 12.2927 504 0.8912 -0.0283 0.8912 0.9440
0.3007 12.3415 506 0.9449 -0.0341 0.9449 0.9721
0.3007 12.3902 508 1.0440 -0.0496 1.0440 1.0218
0.3007 12.4390 510 1.1783 -0.0028 1.1783 1.0855

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
7
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run3_AugV5_k15_task3_organization

Finetuned
(4205)
this model