|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 2.997632202052092, |
|
"eval_steps": 100, |
|
"global_step": 1899, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.01578531965272297, |
|
"grad_norm": 0.7275460362434387, |
|
"learning_rate": 4.0000000000000003e-07, |
|
"loss": 0.6962, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.03157063930544594, |
|
"grad_norm": 0.5963730216026306, |
|
"learning_rate": 8.000000000000001e-07, |
|
"loss": 0.6989, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.0473559589581689, |
|
"grad_norm": 1.253847360610962, |
|
"learning_rate": 1.2000000000000002e-06, |
|
"loss": 0.6935, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.06314127861089187, |
|
"grad_norm": 1.5691624879837036, |
|
"learning_rate": 1.6000000000000001e-06, |
|
"loss": 0.6947, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.07892659826361484, |
|
"grad_norm": 1.1631684303283691, |
|
"learning_rate": 2.0000000000000003e-06, |
|
"loss": 0.6889, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.0947119179163378, |
|
"grad_norm": 1.152625322341919, |
|
"learning_rate": 2.4000000000000003e-06, |
|
"loss": 0.6776, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.11049723756906077, |
|
"grad_norm": 1.720926284790039, |
|
"learning_rate": 2.8000000000000003e-06, |
|
"loss": 0.6468, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.12628255722178375, |
|
"grad_norm": 3.781614065170288, |
|
"learning_rate": 3.2000000000000003e-06, |
|
"loss": 0.6133, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.1420678768745067, |
|
"grad_norm": 2.526336669921875, |
|
"learning_rate": 3.6000000000000003e-06, |
|
"loss": 0.5555, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.15785319652722968, |
|
"grad_norm": 2.09712290763855, |
|
"learning_rate": 4.000000000000001e-06, |
|
"loss": 0.5515, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.15785319652722968, |
|
"eval_accuracy": 0.7578864353312302, |
|
"eval_accuracy_label_test": 0.5070365902694008, |
|
"eval_accuracy_label_train": 0.9992263056092844, |
|
"eval_f1": 0.7415473366691369, |
|
"eval_loss": 0.5193965435028076, |
|
"eval_precision": 0.8351793628382634, |
|
"eval_recall": 0.7578864353312302, |
|
"eval_runtime": 14.5015, |
|
"eval_samples_per_second": 349.758, |
|
"eval_steps_per_second": 21.86, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.17363851617995266, |
|
"grad_norm": 1.363462209701538, |
|
"learning_rate": 4.4e-06, |
|
"loss": 0.5304, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.1894238358326756, |
|
"grad_norm": 1.779582142829895, |
|
"learning_rate": 4.800000000000001e-06, |
|
"loss": 0.5147, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.2052091554853986, |
|
"grad_norm": 3.7067956924438477, |
|
"learning_rate": 5.2e-06, |
|
"loss": 0.4892, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.22099447513812154, |
|
"grad_norm": 19.64886474609375, |
|
"learning_rate": 5.600000000000001e-06, |
|
"loss": 0.4963, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.23677979479084452, |
|
"grad_norm": 12.231091499328613, |
|
"learning_rate": 6e-06, |
|
"loss": 0.451, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.2525651144435675, |
|
"grad_norm": 5.452483654022217, |
|
"learning_rate": 6.4000000000000006e-06, |
|
"loss": 0.3406, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.26835043409629045, |
|
"grad_norm": 4.669571876525879, |
|
"learning_rate": 6.800000000000001e-06, |
|
"loss": 0.3745, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.2841357537490134, |
|
"grad_norm": 2.206510305404663, |
|
"learning_rate": 7.2000000000000005e-06, |
|
"loss": 0.3036, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.2999210734017364, |
|
"grad_norm": 6.932628631591797, |
|
"learning_rate": 7.600000000000001e-06, |
|
"loss": 0.2632, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.31570639305445936, |
|
"grad_norm": 3.4731557369232178, |
|
"learning_rate": 8.000000000000001e-06, |
|
"loss": 0.2205, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.31570639305445936, |
|
"eval_accuracy": 0.9300078864353313, |
|
"eval_accuracy_label_test": 0.9883393646964214, |
|
"eval_accuracy_label_train": 0.8738878143133463, |
|
"eval_f1": 0.9298468332917054, |
|
"eval_loss": 0.25367653369903564, |
|
"eval_precision": 0.9361225540545063, |
|
"eval_recall": 0.9300078864353313, |
|
"eval_runtime": 14.577, |
|
"eval_samples_per_second": 347.944, |
|
"eval_steps_per_second": 21.747, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.3314917127071823, |
|
"grad_norm": 3.005352735519409, |
|
"learning_rate": 8.400000000000001e-06, |
|
"loss": 0.217, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.3472770323599053, |
|
"grad_norm": 5.314383506774902, |
|
"learning_rate": 8.8e-06, |
|
"loss": 0.1349, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.36306235201262826, |
|
"grad_norm": 1.332632303237915, |
|
"learning_rate": 9.200000000000002e-06, |
|
"loss": 0.092, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.3788476716653512, |
|
"grad_norm": 14.597169876098633, |
|
"learning_rate": 9.600000000000001e-06, |
|
"loss": 0.0767, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.39463299131807417, |
|
"grad_norm": 0.3883499503135681, |
|
"learning_rate": 1e-05, |
|
"loss": 0.1029, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.4104183109707972, |
|
"grad_norm": 18.310523986816406, |
|
"learning_rate": 1.04e-05, |
|
"loss": 0.0856, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.4262036306235201, |
|
"grad_norm": 6.817958354949951, |
|
"learning_rate": 1.0800000000000002e-05, |
|
"loss": 0.0514, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.4419889502762431, |
|
"grad_norm": 10.453371047973633, |
|
"learning_rate": 1.1200000000000001e-05, |
|
"loss": 0.0925, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.4577742699289661, |
|
"grad_norm": 0.29950886964797974, |
|
"learning_rate": 1.16e-05, |
|
"loss": 0.0727, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.47355958958168903, |
|
"grad_norm": 3.3916432857513428, |
|
"learning_rate": 1.2e-05, |
|
"loss": 0.1106, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.47355958958168903, |
|
"eval_accuracy": 0.9128548895899053, |
|
"eval_accuracy_label_test": 0.995979091274628, |
|
"eval_accuracy_label_train": 0.8328820116054159, |
|
"eval_f1": 0.9123898201159903, |
|
"eval_loss": 0.3449595868587494, |
|
"eval_precision": 0.9248260970850004, |
|
"eval_recall": 0.9128548895899053, |
|
"eval_runtime": 14.6481, |
|
"eval_samples_per_second": 346.258, |
|
"eval_steps_per_second": 21.641, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.489344909234412, |
|
"grad_norm": 12.914605140686035, |
|
"learning_rate": 1.2400000000000002e-05, |
|
"loss": 0.061, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.505130228887135, |
|
"grad_norm": 0.24762243032455444, |
|
"learning_rate": 1.2800000000000001e-05, |
|
"loss": 0.0929, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.5209155485398579, |
|
"grad_norm": 0.3279378116130829, |
|
"learning_rate": 1.3200000000000002e-05, |
|
"loss": 0.0511, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.5367008681925809, |
|
"grad_norm": 7.465598106384277, |
|
"learning_rate": 1.3600000000000002e-05, |
|
"loss": 0.0727, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.5524861878453039, |
|
"grad_norm": 0.9897856712341309, |
|
"learning_rate": 1.4e-05, |
|
"loss": 0.095, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.5682715074980268, |
|
"grad_norm": 19.12044334411621, |
|
"learning_rate": 1.4400000000000001e-05, |
|
"loss": 0.0386, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.5840568271507498, |
|
"grad_norm": 3.458350419998169, |
|
"learning_rate": 1.48e-05, |
|
"loss": 0.0467, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.5998421468034728, |
|
"grad_norm": 0.10713621973991394, |
|
"learning_rate": 1.5200000000000002e-05, |
|
"loss": 0.0316, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.6156274664561957, |
|
"grad_norm": 0.3854784071445465, |
|
"learning_rate": 1.5600000000000003e-05, |
|
"loss": 0.0396, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.6314127861089187, |
|
"grad_norm": 6.402571201324463, |
|
"learning_rate": 1.6000000000000003e-05, |
|
"loss": 0.0384, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.6314127861089187, |
|
"eval_accuracy": 0.9682570977917981, |
|
"eval_accuracy_label_test": 0.983514274225975, |
|
"eval_accuracy_label_train": 0.9535783365570599, |
|
"eval_f1": 0.9682589499365617, |
|
"eval_loss": 0.14084377884864807, |
|
"eval_precision": 0.968730671942094, |
|
"eval_recall": 0.9682570977917981, |
|
"eval_runtime": 14.7066, |
|
"eval_samples_per_second": 344.878, |
|
"eval_steps_per_second": 21.555, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.6471981057616417, |
|
"grad_norm": 7.416443347930908, |
|
"learning_rate": 1.64e-05, |
|
"loss": 0.0354, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.6629834254143646, |
|
"grad_norm": 25.271848678588867, |
|
"learning_rate": 1.6800000000000002e-05, |
|
"loss": 0.1069, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.6787687450670876, |
|
"grad_norm": 0.10532895475625992, |
|
"learning_rate": 1.72e-05, |
|
"loss": 0.0411, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.6945540647198106, |
|
"grad_norm": 0.05838713422417641, |
|
"learning_rate": 1.76e-05, |
|
"loss": 0.0593, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.7103393843725335, |
|
"grad_norm": 0.06444909423589706, |
|
"learning_rate": 1.8e-05, |
|
"loss": 0.0073, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.7261247040252565, |
|
"grad_norm": 6.492961883544922, |
|
"learning_rate": 1.8400000000000003e-05, |
|
"loss": 0.0539, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.7419100236779794, |
|
"grad_norm": 0.08374536037445068, |
|
"learning_rate": 1.88e-05, |
|
"loss": 0.0757, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.7576953433307024, |
|
"grad_norm": 0.05994931235909462, |
|
"learning_rate": 1.9200000000000003e-05, |
|
"loss": 0.0201, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.7734806629834254, |
|
"grad_norm": 17.886022567749023, |
|
"learning_rate": 1.9600000000000002e-05, |
|
"loss": 0.0536, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.7892659826361483, |
|
"grad_norm": 2.434602737426758, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0631, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.7892659826361483, |
|
"eval_accuracy": 0.9631309148264984, |
|
"eval_accuracy_label_test": 0.989545637314033, |
|
"eval_accuracy_label_train": 0.9377176015473888, |
|
"eval_f1": 0.9631237556597715, |
|
"eval_loss": 0.15165044367313385, |
|
"eval_precision": 0.964482811637024, |
|
"eval_recall": 0.9631309148264984, |
|
"eval_runtime": 14.5963, |
|
"eval_samples_per_second": 347.485, |
|
"eval_steps_per_second": 21.718, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.8050513022888713, |
|
"grad_norm": 2.8019778728485107, |
|
"learning_rate": 1.9857040743388135e-05, |
|
"loss": 0.0191, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.8208366219415943, |
|
"grad_norm": 2.233394145965576, |
|
"learning_rate": 1.971408148677627e-05, |
|
"loss": 0.0317, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.8366219415943172, |
|
"grad_norm": 0.1284877210855484, |
|
"learning_rate": 1.9571122230164405e-05, |
|
"loss": 0.0467, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.8524072612470402, |
|
"grad_norm": 0.0791168138384819, |
|
"learning_rate": 1.942816297355254e-05, |
|
"loss": 0.0455, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.8681925808997633, |
|
"grad_norm": 0.35995563864707947, |
|
"learning_rate": 1.928520371694067e-05, |
|
"loss": 0.0253, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.8839779005524862, |
|
"grad_norm": 2.419013738632202, |
|
"learning_rate": 1.9142244460328808e-05, |
|
"loss": 0.0551, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.8997632202052092, |
|
"grad_norm": 12.438931465148926, |
|
"learning_rate": 1.899928520371694e-05, |
|
"loss": 0.0563, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.9155485398579322, |
|
"grad_norm": 0.049261171370744705, |
|
"learning_rate": 1.8856325947105075e-05, |
|
"loss": 0.0291, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.9313338595106551, |
|
"grad_norm": 0.04032573848962784, |
|
"learning_rate": 1.871336669049321e-05, |
|
"loss": 0.0026, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.9471191791633781, |
|
"grad_norm": 1.215957522392273, |
|
"learning_rate": 1.8570407433881345e-05, |
|
"loss": 0.0276, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.9471191791633781, |
|
"eval_accuracy": 0.9386829652996845, |
|
"eval_accuracy_label_test": 0.9947728186570165, |
|
"eval_accuracy_label_train": 0.8847195357833656, |
|
"eval_f1": 0.9385557615622165, |
|
"eval_loss": 0.36488327383995056, |
|
"eval_precision": 0.9444060841035706, |
|
"eval_recall": 0.9386829652996845, |
|
"eval_runtime": 14.5956, |
|
"eval_samples_per_second": 347.502, |
|
"eval_steps_per_second": 21.719, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.9629044988161011, |
|
"grad_norm": 0.0634990781545639, |
|
"learning_rate": 1.842744817726948e-05, |
|
"loss": 0.0267, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.978689818468824, |
|
"grad_norm": 0.029669882729649544, |
|
"learning_rate": 1.8284488920657615e-05, |
|
"loss": 0.0426, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.994475138121547, |
|
"grad_norm": 0.029255308210849762, |
|
"learning_rate": 1.8141529664045748e-05, |
|
"loss": 0.0556, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 1.01026045777427, |
|
"grad_norm": 0.0757206529378891, |
|
"learning_rate": 1.799857040743388e-05, |
|
"loss": 0.0458, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 1.0260457774269929, |
|
"grad_norm": 0.1398855596780777, |
|
"learning_rate": 1.7855611150822018e-05, |
|
"loss": 0.0641, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 1.0418310970797158, |
|
"grad_norm": 7.250219821929932, |
|
"learning_rate": 1.771265189421015e-05, |
|
"loss": 0.0204, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 1.057616416732439, |
|
"grad_norm": 0.022936342284083366, |
|
"learning_rate": 1.7569692637598285e-05, |
|
"loss": 0.002, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 1.0734017363851618, |
|
"grad_norm": 4.3541741371154785, |
|
"learning_rate": 1.742673338098642e-05, |
|
"loss": 0.0148, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 1.0891870560378847, |
|
"grad_norm": 0.024669496342539787, |
|
"learning_rate": 1.7283774124374555e-05, |
|
"loss": 0.062, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 1.1049723756906078, |
|
"grad_norm": 0.023665498942136765, |
|
"learning_rate": 1.7140814867762688e-05, |
|
"loss": 0.0245, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 1.1049723756906078, |
|
"eval_accuracy": 0.9702287066246057, |
|
"eval_accuracy_label_test": 0.9726578206674709, |
|
"eval_accuracy_label_train": 0.9678916827852998, |
|
"eval_f1": 0.9702301478266859, |
|
"eval_loss": 0.13387098908424377, |
|
"eval_precision": 0.9702485636354361, |
|
"eval_recall": 0.9702287066246057, |
|
"eval_runtime": 14.562, |
|
"eval_samples_per_second": 348.304, |
|
"eval_steps_per_second": 21.769, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 1.1207576953433307, |
|
"grad_norm": 20.071300506591797, |
|
"learning_rate": 1.6997855611150825e-05, |
|
"loss": 0.0867, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 1.1365430149960536, |
|
"grad_norm": 0.016291845589876175, |
|
"learning_rate": 1.6854896354538958e-05, |
|
"loss": 0.0167, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 1.1523283346487767, |
|
"grad_norm": 0.02355334907770157, |
|
"learning_rate": 1.671193709792709e-05, |
|
"loss": 0.003, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 1.1681136543014996, |
|
"grad_norm": 0.014460445381700993, |
|
"learning_rate": 1.6568977841315225e-05, |
|
"loss": 0.0137, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 1.1838989739542225, |
|
"grad_norm": 9.110311508178711, |
|
"learning_rate": 1.642601858470336e-05, |
|
"loss": 0.0592, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 1.1996842936069456, |
|
"grad_norm": 0.03517633676528931, |
|
"learning_rate": 1.6283059328091495e-05, |
|
"loss": 0.0356, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 1.2154696132596685, |
|
"grad_norm": 1.65547776222229, |
|
"learning_rate": 1.6140100071479628e-05, |
|
"loss": 0.0206, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 1.2312549329123914, |
|
"grad_norm": 0.044892311096191406, |
|
"learning_rate": 1.5997140814867765e-05, |
|
"loss": 0.0138, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 1.2470402525651145, |
|
"grad_norm": 16.877887725830078, |
|
"learning_rate": 1.5854181558255898e-05, |
|
"loss": 0.0201, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 1.2628255722178374, |
|
"grad_norm": 13.245466232299805, |
|
"learning_rate": 1.5711222301644035e-05, |
|
"loss": 0.0519, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 1.2628255722178374, |
|
"eval_accuracy": 0.9185725552050473, |
|
"eval_accuracy_label_test": 0.9991958182549256, |
|
"eval_accuracy_label_train": 0.8410058027079303, |
|
"eval_f1": 0.9181684151069668, |
|
"eval_loss": 0.4945215582847595, |
|
"eval_precision": 0.9299426884244839, |
|
"eval_recall": 0.9185725552050473, |
|
"eval_runtime": 14.6038, |
|
"eval_samples_per_second": 347.306, |
|
"eval_steps_per_second": 21.707, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 1.2786108918705603, |
|
"grad_norm": 12.54496955871582, |
|
"learning_rate": 1.5568263045032168e-05, |
|
"loss": 0.0505, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 1.2943962115232832, |
|
"grad_norm": 0.01659429259598255, |
|
"learning_rate": 1.54253037884203e-05, |
|
"loss": 0.0246, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 1.3101815311760063, |
|
"grad_norm": 0.03832576796412468, |
|
"learning_rate": 1.5282344531808435e-05, |
|
"loss": 0.0206, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 1.3259668508287292, |
|
"grad_norm": 4.976266384124756, |
|
"learning_rate": 1.513938527519657e-05, |
|
"loss": 0.011, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 1.3417521704814521, |
|
"grad_norm": 16.507583618164062, |
|
"learning_rate": 1.4996426018584705e-05, |
|
"loss": 0.0971, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 1.3575374901341752, |
|
"grad_norm": 9.206993103027344, |
|
"learning_rate": 1.485346676197284e-05, |
|
"loss": 0.0314, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 1.3733228097868981, |
|
"grad_norm": 0.040578652173280716, |
|
"learning_rate": 1.4710507505360973e-05, |
|
"loss": 0.0053, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 1.389108129439621, |
|
"grad_norm": 15.86640739440918, |
|
"learning_rate": 1.4567548248749108e-05, |
|
"loss": 0.0268, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 1.4048934490923441, |
|
"grad_norm": 1.9645894765853882, |
|
"learning_rate": 1.4424588992137243e-05, |
|
"loss": 0.0348, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 1.420678768745067, |
|
"grad_norm": 0.01832072250545025, |
|
"learning_rate": 1.4281629735525378e-05, |
|
"loss": 0.02, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 1.420678768745067, |
|
"eval_accuracy": 0.9548501577287066, |
|
"eval_accuracy_label_test": 0.995979091274628, |
|
"eval_accuracy_label_train": 0.9152804642166344, |
|
"eval_f1": 0.9548094220396917, |
|
"eval_loss": 0.26368606090545654, |
|
"eval_precision": 0.958024003205365, |
|
"eval_recall": 0.9548501577287066, |
|
"eval_runtime": 14.5839, |
|
"eval_samples_per_second": 347.781, |
|
"eval_steps_per_second": 21.736, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 1.43646408839779, |
|
"grad_norm": 2.4122228622436523, |
|
"learning_rate": 1.413867047891351e-05, |
|
"loss": 0.0431, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 1.452249408050513, |
|
"grad_norm": 0.04305430129170418, |
|
"learning_rate": 1.3995711222301644e-05, |
|
"loss": 0.0382, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 1.468034727703236, |
|
"grad_norm": 0.045583341270685196, |
|
"learning_rate": 1.385275196568978e-05, |
|
"loss": 0.0345, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 1.4838200473559588, |
|
"grad_norm": 0.08313434571027756, |
|
"learning_rate": 1.3709792709077914e-05, |
|
"loss": 0.0229, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 1.499605367008682, |
|
"grad_norm": 0.040190890431404114, |
|
"learning_rate": 1.3566833452466048e-05, |
|
"loss": 0.0022, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 1.5153906866614049, |
|
"grad_norm": 0.12089771777391434, |
|
"learning_rate": 1.3423874195854183e-05, |
|
"loss": 0.0126, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 1.5311760063141278, |
|
"grad_norm": 0.0291321761906147, |
|
"learning_rate": 1.3280914939242318e-05, |
|
"loss": 0.0246, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 1.5469613259668509, |
|
"grad_norm": 0.06409675627946854, |
|
"learning_rate": 1.3137955682630453e-05, |
|
"loss": 0.0611, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 1.5627466456195738, |
|
"grad_norm": 9.994110107421875, |
|
"learning_rate": 1.2994996426018586e-05, |
|
"loss": 0.0102, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 1.5785319652722967, |
|
"grad_norm": 10.403185844421387, |
|
"learning_rate": 1.285203716940672e-05, |
|
"loss": 0.0325, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 1.5785319652722967, |
|
"eval_accuracy": 0.9708201892744479, |
|
"eval_accuracy_label_test": 0.9851226377161239, |
|
"eval_accuracy_label_train": 0.9570599613152805, |
|
"eval_f1": 0.970822203818026, |
|
"eval_loss": 0.11654461175203323, |
|
"eval_precision": 0.9712375667689825, |
|
"eval_recall": 0.9708201892744479, |
|
"eval_runtime": 14.5989, |
|
"eval_samples_per_second": 347.423, |
|
"eval_steps_per_second": 21.714, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 1.5943172849250198, |
|
"grad_norm": 0.05680805817246437, |
|
"learning_rate": 1.2709077912794854e-05, |
|
"loss": 0.0028, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 1.6101026045777427, |
|
"grad_norm": 0.017462020739912987, |
|
"learning_rate": 1.2566118656182988e-05, |
|
"loss": 0.0357, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 1.6258879242304656, |
|
"grad_norm": 5.752472877502441, |
|
"learning_rate": 1.2423159399571123e-05, |
|
"loss": 0.0457, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 1.6416732438831887, |
|
"grad_norm": 0.14555960893630981, |
|
"learning_rate": 1.2280200142959258e-05, |
|
"loss": 0.0062, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 1.6574585635359116, |
|
"grad_norm": 0.019668666645884514, |
|
"learning_rate": 1.2137240886347393e-05, |
|
"loss": 0.0046, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 1.6732438831886345, |
|
"grad_norm": 0.01673435978591442, |
|
"learning_rate": 1.1994281629735528e-05, |
|
"loss": 0.0535, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 1.6890292028413576, |
|
"grad_norm": 4.553021430969238, |
|
"learning_rate": 1.1851322373123661e-05, |
|
"loss": 0.0767, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 1.7048145224940805, |
|
"grad_norm": 0.33790552616119385, |
|
"learning_rate": 1.1708363116511796e-05, |
|
"loss": 0.0433, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 1.7205998421468034, |
|
"grad_norm": 0.08579400181770325, |
|
"learning_rate": 1.156540385989993e-05, |
|
"loss": 0.0051, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 1.7363851617995265, |
|
"grad_norm": 0.24498964846134186, |
|
"learning_rate": 1.1422444603288063e-05, |
|
"loss": 0.016, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 1.7363851617995265, |
|
"eval_accuracy": 0.9692429022082019, |
|
"eval_accuracy_label_test": 0.9529553679131484, |
|
"eval_accuracy_label_train": 0.9849129593810445, |
|
"eval_f1": 0.9692264691880251, |
|
"eval_loss": 0.10065295547246933, |
|
"eval_precision": 0.9696695972871102, |
|
"eval_recall": 0.9692429022082019, |
|
"eval_runtime": 14.644, |
|
"eval_samples_per_second": 346.353, |
|
"eval_steps_per_second": 21.647, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 1.7521704814522494, |
|
"grad_norm": 3.6310765743255615, |
|
"learning_rate": 1.1279485346676198e-05, |
|
"loss": 0.0676, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 1.7679558011049723, |
|
"grad_norm": 0.26799193024635315, |
|
"learning_rate": 1.1136526090064333e-05, |
|
"loss": 0.011, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 1.7837411207576954, |
|
"grad_norm": 0.050716836005449295, |
|
"learning_rate": 1.0993566833452468e-05, |
|
"loss": 0.0474, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 1.7995264404104183, |
|
"grad_norm": 0.03631124272942543, |
|
"learning_rate": 1.0850607576840601e-05, |
|
"loss": 0.0013, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 1.8153117600631412, |
|
"grad_norm": 0.016972020268440247, |
|
"learning_rate": 1.0707648320228736e-05, |
|
"loss": 0.0098, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 1.8310970797158643, |
|
"grad_norm": 0.01586087793111801, |
|
"learning_rate": 1.0564689063616871e-05, |
|
"loss": 0.0013, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 1.8468823993685872, |
|
"grad_norm": 0.01447396818548441, |
|
"learning_rate": 1.0421729807005006e-05, |
|
"loss": 0.0148, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 1.8626677190213101, |
|
"grad_norm": 0.08018597215414047, |
|
"learning_rate": 1.0278770550393137e-05, |
|
"loss": 0.0067, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 1.8784530386740332, |
|
"grad_norm": 0.020933715626597404, |
|
"learning_rate": 1.0135811293781272e-05, |
|
"loss": 0.0177, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 1.8942383583267561, |
|
"grad_norm": 0.006128642242401838, |
|
"learning_rate": 9.992852037169407e-06, |
|
"loss": 0.0068, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 1.8942383583267561, |
|
"eval_accuracy": 0.9690457413249212, |
|
"eval_accuracy_label_test": 0.9871330920788098, |
|
"eval_accuracy_label_train": 0.9516441005802708, |
|
"eval_f1": 0.9690463008445059, |
|
"eval_loss": 0.16786406934261322, |
|
"eval_precision": 0.9696986578163762, |
|
"eval_recall": 0.9690457413249212, |
|
"eval_runtime": 14.5325, |
|
"eval_samples_per_second": 349.011, |
|
"eval_steps_per_second": 21.813, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 1.910023677979479, |
|
"grad_norm": 0.009371940977871418, |
|
"learning_rate": 9.849892780557542e-06, |
|
"loss": 0.024, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 1.9258089976322021, |
|
"grad_norm": 0.014078453183174133, |
|
"learning_rate": 9.706933523945676e-06, |
|
"loss": 0.0005, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 1.941594317284925, |
|
"grad_norm": 0.009515127167105675, |
|
"learning_rate": 9.56397426733381e-06, |
|
"loss": 0.0046, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 1.957379636937648, |
|
"grad_norm": 0.014615291729569435, |
|
"learning_rate": 9.421015010721944e-06, |
|
"loss": 0.0261, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 1.973164956590371, |
|
"grad_norm": 0.06883493065834045, |
|
"learning_rate": 9.278055754110079e-06, |
|
"loss": 0.0233, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 1.988950276243094, |
|
"grad_norm": 0.013247163034975529, |
|
"learning_rate": 9.135096497498214e-06, |
|
"loss": 0.0168, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 2.004735595895817, |
|
"grad_norm": 0.012575499713420868, |
|
"learning_rate": 8.992137240886349e-06, |
|
"loss": 0.0124, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 2.02052091554854, |
|
"grad_norm": 0.017276756465435028, |
|
"learning_rate": 8.849177984274482e-06, |
|
"loss": 0.0184, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 2.0363062352012626, |
|
"grad_norm": 0.03682788833975792, |
|
"learning_rate": 8.706218727662616e-06, |
|
"loss": 0.0092, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 2.0520915548539858, |
|
"grad_norm": 0.35351473093032837, |
|
"learning_rate": 8.56325947105075e-06, |
|
"loss": 0.0042, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 2.0520915548539858, |
|
"eval_accuracy": 0.9733832807570978, |
|
"eval_accuracy_label_test": 0.9722557297949337, |
|
"eval_accuracy_label_train": 0.9744680851063829, |
|
"eval_f1": 0.9733829671319272, |
|
"eval_loss": 0.11822589486837387, |
|
"eval_precision": 0.973383334857871, |
|
"eval_recall": 0.9733832807570978, |
|
"eval_runtime": 14.5763, |
|
"eval_samples_per_second": 347.961, |
|
"eval_steps_per_second": 21.748, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 2.067876874506709, |
|
"grad_norm": 0.012184061110019684, |
|
"learning_rate": 8.420300214438886e-06, |
|
"loss": 0.0441, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 2.0836621941594315, |
|
"grad_norm": 0.0278099924325943, |
|
"learning_rate": 8.27734095782702e-06, |
|
"loss": 0.008, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 2.0994475138121547, |
|
"grad_norm": 0.9140133261680603, |
|
"learning_rate": 8.134381701215154e-06, |
|
"loss": 0.0017, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 2.115232833464878, |
|
"grad_norm": 1.3029474020004272, |
|
"learning_rate": 7.991422444603289e-06, |
|
"loss": 0.0007, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 2.1310181531176005, |
|
"grad_norm": 0.010257584042847157, |
|
"learning_rate": 7.848463187991422e-06, |
|
"loss": 0.0158, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 2.1468034727703236, |
|
"grad_norm": 0.010051756165921688, |
|
"learning_rate": 7.705503931379557e-06, |
|
"loss": 0.0034, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 2.1625887924230467, |
|
"grad_norm": 0.10394269973039627, |
|
"learning_rate": 7.562544674767692e-06, |
|
"loss": 0.0404, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 2.1783741120757694, |
|
"grad_norm": 0.011765426024794579, |
|
"learning_rate": 7.419585418155826e-06, |
|
"loss": 0.0004, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 2.1941594317284925, |
|
"grad_norm": 1.1091564893722534, |
|
"learning_rate": 7.276626161543961e-06, |
|
"loss": 0.0006, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 2.2099447513812156, |
|
"grad_norm": 0.010108958929777145, |
|
"learning_rate": 7.133666904932095e-06, |
|
"loss": 0.0005, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 2.2099447513812156, |
|
"eval_accuracy": 0.9729889589905363, |
|
"eval_accuracy_label_test": 0.9798954563731403, |
|
"eval_accuracy_label_train": 0.9663442940038685, |
|
"eval_f1": 0.9729913291480798, |
|
"eval_loss": 0.1431707739830017, |
|
"eval_precision": 0.9730972588569067, |
|
"eval_recall": 0.9729889589905363, |
|
"eval_runtime": 14.631, |
|
"eval_samples_per_second": 346.662, |
|
"eval_steps_per_second": 21.666, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 2.2257300710339383, |
|
"grad_norm": 0.02260042168200016, |
|
"learning_rate": 6.99070764832023e-06, |
|
"loss": 0.0014, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 2.2415153906866614, |
|
"grad_norm": 0.0077365124598145485, |
|
"learning_rate": 6.847748391708363e-06, |
|
"loss": 0.0004, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 2.2573007103393845, |
|
"grad_norm": 0.008837452158331871, |
|
"learning_rate": 6.704789135096498e-06, |
|
"loss": 0.0153, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 2.273086029992107, |
|
"grad_norm": 0.021829022094607353, |
|
"learning_rate": 6.561829878484632e-06, |
|
"loss": 0.0003, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 2.2888713496448303, |
|
"grad_norm": 0.014965805225074291, |
|
"learning_rate": 6.418870621872767e-06, |
|
"loss": 0.0246, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 2.3046566692975534, |
|
"grad_norm": 0.013634726405143738, |
|
"learning_rate": 6.275911365260901e-06, |
|
"loss": 0.0339, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 2.320441988950276, |
|
"grad_norm": 6.8833794593811035, |
|
"learning_rate": 6.1329521086490355e-06, |
|
"loss": 0.0059, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 2.336227308602999, |
|
"grad_norm": 0.04814034327864647, |
|
"learning_rate": 5.98999285203717e-06, |
|
"loss": 0.0122, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 2.3520126282557223, |
|
"grad_norm": 0.026584748178720474, |
|
"learning_rate": 5.847033595425305e-06, |
|
"loss": 0.0022, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 2.367797947908445, |
|
"grad_norm": 0.015504554845392704, |
|
"learning_rate": 5.704074338813439e-06, |
|
"loss": 0.0182, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 2.367797947908445, |
|
"eval_accuracy": 0.9718059936908517, |
|
"eval_accuracy_label_test": 0.9871330920788098, |
|
"eval_accuracy_label_train": 0.9570599613152805, |
|
"eval_f1": 0.9718076387635299, |
|
"eval_loss": 0.14604002237319946, |
|
"eval_precision": 0.972280875829533, |
|
"eval_recall": 0.9718059936908517, |
|
"eval_runtime": 14.6031, |
|
"eval_samples_per_second": 347.324, |
|
"eval_steps_per_second": 21.708, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 2.383583267561168, |
|
"grad_norm": 0.07147639244794846, |
|
"learning_rate": 5.561115082201572e-06, |
|
"loss": 0.0005, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 2.3993685872138912, |
|
"grad_norm": 0.024004925042390823, |
|
"learning_rate": 5.418155825589707e-06, |
|
"loss": 0.0439, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 2.415153906866614, |
|
"grad_norm": 0.07405655831098557, |
|
"learning_rate": 5.275196568977842e-06, |
|
"loss": 0.0004, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 2.430939226519337, |
|
"grad_norm": 0.01881037838757038, |
|
"learning_rate": 5.132237312365976e-06, |
|
"loss": 0.0004, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 2.44672454617206, |
|
"grad_norm": 0.05772541090846062, |
|
"learning_rate": 4.98927805575411e-06, |
|
"loss": 0.0059, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 2.462509865824783, |
|
"grad_norm": 0.014262210577726364, |
|
"learning_rate": 4.846318799142245e-06, |
|
"loss": 0.0004, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 2.478295185477506, |
|
"grad_norm": 0.005270448978990316, |
|
"learning_rate": 4.703359542530379e-06, |
|
"loss": 0.0003, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 2.494080505130229, |
|
"grad_norm": 0.008271850645542145, |
|
"learning_rate": 4.560400285918514e-06, |
|
"loss": 0.0231, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 2.5098658247829517, |
|
"grad_norm": 0.013198798522353172, |
|
"learning_rate": 4.417441029306648e-06, |
|
"loss": 0.0034, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 2.525651144435675, |
|
"grad_norm": 0.07948605716228485, |
|
"learning_rate": 4.274481772694782e-06, |
|
"loss": 0.0004, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 2.525651144435675, |
|
"eval_accuracy": 0.973186119873817, |
|
"eval_accuracy_label_test": 0.9843184559710495, |
|
"eval_accuracy_label_train": 0.9624758220502901, |
|
"eval_f1": 0.973188538207199, |
|
"eval_loss": 0.13827118277549744, |
|
"eval_precision": 0.973445480376393, |
|
"eval_recall": 0.973186119873817, |
|
"eval_runtime": 14.6108, |
|
"eval_samples_per_second": 347.14, |
|
"eval_steps_per_second": 21.696, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 2.541436464088398, |
|
"grad_norm": 0.24930787086486816, |
|
"learning_rate": 4.131522516082916e-06, |
|
"loss": 0.0221, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 2.5572217837411206, |
|
"grad_norm": 0.011009340174496174, |
|
"learning_rate": 3.988563259471051e-06, |
|
"loss": 0.0003, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 2.5730071033938438, |
|
"grad_norm": 0.02251257933676243, |
|
"learning_rate": 3.845604002859185e-06, |
|
"loss": 0.0003, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 2.5887924230465664, |
|
"grad_norm": 0.004396820440888405, |
|
"learning_rate": 3.70264474624732e-06, |
|
"loss": 0.0332, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 2.6045777426992895, |
|
"grad_norm": 1.1856316328048706, |
|
"learning_rate": 3.5596854896354545e-06, |
|
"loss": 0.0006, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 2.6203630623520127, |
|
"grad_norm": 0.019586117938160896, |
|
"learning_rate": 3.4167262330235886e-06, |
|
"loss": 0.0132, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 2.636148382004736, |
|
"grad_norm": 0.010670648887753487, |
|
"learning_rate": 3.273766976411723e-06, |
|
"loss": 0.0007, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 2.6519337016574585, |
|
"grad_norm": 0.37919822335243225, |
|
"learning_rate": 3.130807719799857e-06, |
|
"loss": 0.0003, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 2.6677190213101816, |
|
"grad_norm": 0.2330523133277893, |
|
"learning_rate": 2.987848463187992e-06, |
|
"loss": 0.0004, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 2.6835043409629042, |
|
"grad_norm": 0.007127601653337479, |
|
"learning_rate": 2.8448892065761256e-06, |
|
"loss": 0.0003, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 2.6835043409629042, |
|
"eval_accuracy": 0.9743690851735016, |
|
"eval_accuracy_label_test": 0.9831121833534379, |
|
"eval_accuracy_label_train": 0.9659574468085106, |
|
"eval_f1": 0.9743714686586191, |
|
"eval_loss": 0.13812877237796783, |
|
"eval_precision": 0.9745341434373056, |
|
"eval_recall": 0.9743690851735016, |
|
"eval_runtime": 14.5972, |
|
"eval_samples_per_second": 347.464, |
|
"eval_steps_per_second": 21.717, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 2.6992896606156274, |
|
"grad_norm": 0.005769920535385609, |
|
"learning_rate": 2.7019299499642602e-06, |
|
"loss": 0.0005, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 2.7150749802683505, |
|
"grad_norm": 0.005288603249937296, |
|
"learning_rate": 2.5589706933523952e-06, |
|
"loss": 0.0006, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 2.7308602999210736, |
|
"grad_norm": 0.005266103427857161, |
|
"learning_rate": 2.416011436740529e-06, |
|
"loss": 0.0032, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 2.7466456195737963, |
|
"grad_norm": 0.005252339411526918, |
|
"learning_rate": 2.2730521801286635e-06, |
|
"loss": 0.0006, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 2.7624309392265194, |
|
"grad_norm": 0.004924137610942125, |
|
"learning_rate": 2.1300929235167977e-06, |
|
"loss": 0.0002, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 2.778216258879242, |
|
"grad_norm": 0.003662313334643841, |
|
"learning_rate": 1.9871336669049322e-06, |
|
"loss": 0.0017, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 2.794001578531965, |
|
"grad_norm": 0.00455325935035944, |
|
"learning_rate": 1.8441744102930666e-06, |
|
"loss": 0.0002, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 2.8097868981846883, |
|
"grad_norm": 0.0084315724670887, |
|
"learning_rate": 1.701215153681201e-06, |
|
"loss": 0.0007, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 2.8255722178374114, |
|
"grad_norm": 0.021251995116472244, |
|
"learning_rate": 1.5582558970693353e-06, |
|
"loss": 0.0002, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 2.841357537490134, |
|
"grad_norm": 0.005972001701593399, |
|
"learning_rate": 1.4152966404574697e-06, |
|
"loss": 0.0002, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 2.841357537490134, |
|
"eval_accuracy": 0.972397476340694, |
|
"eval_accuracy_label_test": 0.9863289103337354, |
|
"eval_accuracy_label_train": 0.9589941972920696, |
|
"eval_f1": 0.9723994850059051, |
|
"eval_loss": 0.1599443554878235, |
|
"eval_precision": 0.9727934306989875, |
|
"eval_recall": 0.972397476340694, |
|
"eval_runtime": 14.9428, |
|
"eval_samples_per_second": 339.427, |
|
"eval_steps_per_second": 21.214, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 2.857142857142857, |
|
"grad_norm": 0.005644650664180517, |
|
"learning_rate": 1.272337383845604e-06, |
|
"loss": 0.0002, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 2.87292817679558, |
|
"grad_norm": 0.010099658742547035, |
|
"learning_rate": 1.1293781272337384e-06, |
|
"loss": 0.0003, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 2.888713496448303, |
|
"grad_norm": 0.0057546221651136875, |
|
"learning_rate": 9.864188706218728e-07, |
|
"loss": 0.0002, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 2.904498816101026, |
|
"grad_norm": 0.010815597139298916, |
|
"learning_rate": 8.434596140100073e-07, |
|
"loss": 0.0104, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 2.9202841357537492, |
|
"grad_norm": 0.0059271338395774364, |
|
"learning_rate": 7.005003573981415e-07, |
|
"loss": 0.0002, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 2.936069455406472, |
|
"grad_norm": 0.004583253525197506, |
|
"learning_rate": 5.57541100786276e-07, |
|
"loss": 0.0033, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 2.951854775059195, |
|
"grad_norm": 0.003865251550450921, |
|
"learning_rate": 4.145818441744103e-07, |
|
"loss": 0.0102, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 2.9676400947119177, |
|
"grad_norm": 0.006317495368421078, |
|
"learning_rate": 2.716225875625447e-07, |
|
"loss": 0.0002, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 2.983425414364641, |
|
"grad_norm": 0.004333311691880226, |
|
"learning_rate": 1.2866333095067907e-07, |
|
"loss": 0.0003, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 2.997632202052092, |
|
"step": 1899, |
|
"total_flos": 7993457212661760.0, |
|
"train_loss": 0.08041724508642717, |
|
"train_runtime": 817.1898, |
|
"train_samples_per_second": 74.406, |
|
"train_steps_per_second": 2.324 |
|
}, |
|
{ |
|
"epoch": 2.997632202052092, |
|
"eval_accuracy": 0.9704258675078864, |
|
"eval_accuracy_label_test": 0.9879372738238842, |
|
"eval_accuracy_label_train": 0.9535783365570599, |
|
"eval_f1": 0.9704266952348471, |
|
"eval_loss": 0.1764293909072876, |
|
"eval_precision": 0.9710387924576857, |
|
"eval_recall": 0.9704258675078864, |
|
"eval_runtime": 14.6294, |
|
"eval_samples_per_second": 346.699, |
|
"eval_steps_per_second": 21.669, |
|
"step": 1899 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 1899, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 1000, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 7993457212661760.0, |
|
"train_batch_size": 16, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|