|
{ |
|
"best_metric": 0.9882, |
|
"best_model_checkpoint": "beit-base-patch16-224-finetuned-eurosat/checkpoint-2109", |
|
"epoch": 2.998933522929257, |
|
"eval_steps": 500, |
|
"global_step": 2109, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 42.045841217041016, |
|
"learning_rate": 2.3696682464454976e-06, |
|
"loss": 2.6733, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 19.79863166809082, |
|
"learning_rate": 4.739336492890995e-06, |
|
"loss": 2.5329, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 21.456212997436523, |
|
"learning_rate": 7.109004739336493e-06, |
|
"loss": 2.3177, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 24.893077850341797, |
|
"learning_rate": 9.47867298578199e-06, |
|
"loss": 2.0632, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 25.683191299438477, |
|
"learning_rate": 1.184834123222749e-05, |
|
"loss": 1.7585, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 27.140174865722656, |
|
"learning_rate": 1.4218009478672985e-05, |
|
"loss": 1.4911, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 17.97708511352539, |
|
"learning_rate": 1.6587677725118483e-05, |
|
"loss": 1.2294, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 36.129730224609375, |
|
"learning_rate": 1.895734597156398e-05, |
|
"loss": 1.0185, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 22.884685516357422, |
|
"learning_rate": 2.132701421800948e-05, |
|
"loss": 0.8923, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 19.778934478759766, |
|
"learning_rate": 2.369668246445498e-05, |
|
"loss": 0.7784, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 20.47595977783203, |
|
"learning_rate": 2.6066350710900477e-05, |
|
"loss": 0.6687, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 17.096622467041016, |
|
"learning_rate": 2.843601895734597e-05, |
|
"loss": 0.6549, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 18.601505279541016, |
|
"learning_rate": 3.080568720379147e-05, |
|
"loss": 0.6253, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 15.65523624420166, |
|
"learning_rate": 3.3175355450236966e-05, |
|
"loss": 0.6118, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 15.892389297485352, |
|
"learning_rate": 3.554502369668247e-05, |
|
"loss": 0.6227, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 12.784531593322754, |
|
"learning_rate": 3.791469194312796e-05, |
|
"loss": 0.4946, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 13.44602108001709, |
|
"learning_rate": 4.028436018957346e-05, |
|
"loss": 0.6287, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 17.538671493530273, |
|
"learning_rate": 4.265402843601896e-05, |
|
"loss": 0.5645, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 11.633671760559082, |
|
"learning_rate": 4.502369668246446e-05, |
|
"loss": 0.5169, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 13.391677856445312, |
|
"learning_rate": 4.739336492890996e-05, |
|
"loss": 0.6072, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 8.77043342590332, |
|
"learning_rate": 4.976303317535545e-05, |
|
"loss": 0.4768, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 11.799147605895996, |
|
"learning_rate": 4.976290832455216e-05, |
|
"loss": 0.4924, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 8.894740104675293, |
|
"learning_rate": 4.949947312961012e-05, |
|
"loss": 0.4062, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 11.966766357421875, |
|
"learning_rate": 4.923603793466807e-05, |
|
"loss": 0.5118, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 12.86320686340332, |
|
"learning_rate": 4.8972602739726034e-05, |
|
"loss": 0.5241, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 8.136141777038574, |
|
"learning_rate": 4.8709167544783986e-05, |
|
"loss": 0.4875, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 12.6289644241333, |
|
"learning_rate": 4.8445732349841945e-05, |
|
"loss": 0.482, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 11.868968963623047, |
|
"learning_rate": 4.8182297154899896e-05, |
|
"loss": 0.5488, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 11.955994606018066, |
|
"learning_rate": 4.791886195995785e-05, |
|
"loss": 0.4242, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 10.719066619873047, |
|
"learning_rate": 4.7655426765015806e-05, |
|
"loss": 0.4832, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 9.737288475036621, |
|
"learning_rate": 4.7391991570073765e-05, |
|
"loss": 0.4619, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 9.085290908813477, |
|
"learning_rate": 4.712855637513172e-05, |
|
"loss": 0.4583, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 5.268682956695557, |
|
"learning_rate": 4.6865121180189675e-05, |
|
"loss": 0.4926, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 8.957120895385742, |
|
"learning_rate": 4.6601685985247633e-05, |
|
"loss": 0.4615, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 5.488365650177002, |
|
"learning_rate": 4.6338250790305585e-05, |
|
"loss": 0.4487, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 9.455901145935059, |
|
"learning_rate": 4.6074815595363544e-05, |
|
"loss": 0.4429, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 10.715672492980957, |
|
"learning_rate": 4.58113804004215e-05, |
|
"loss": 0.4603, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 10.67427921295166, |
|
"learning_rate": 4.5547945205479454e-05, |
|
"loss": 0.3811, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 7.144228935241699, |
|
"learning_rate": 4.528451001053741e-05, |
|
"loss": 0.4789, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 10.277647972106934, |
|
"learning_rate": 4.5021074815595364e-05, |
|
"loss": 0.4265, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 9.004936218261719, |
|
"learning_rate": 4.4757639620653316e-05, |
|
"loss": 0.468, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 8.17403507232666, |
|
"learning_rate": 4.449420442571128e-05, |
|
"loss": 0.3897, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 7.23596715927124, |
|
"learning_rate": 4.423076923076923e-05, |
|
"loss": 0.3734, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 8.713665962219238, |
|
"learning_rate": 4.396733403582719e-05, |
|
"loss": 0.4557, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 9.222480773925781, |
|
"learning_rate": 4.370389884088514e-05, |
|
"loss": 0.3975, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 8.174620628356934, |
|
"learning_rate": 4.34404636459431e-05, |
|
"loss": 0.4742, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 10.548660278320312, |
|
"learning_rate": 4.317702845100105e-05, |
|
"loss": 0.4072, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 10.924005508422852, |
|
"learning_rate": 4.291359325605901e-05, |
|
"loss": 0.3547, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 7.682478427886963, |
|
"learning_rate": 4.265015806111697e-05, |
|
"loss": 0.3249, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 9.409101486206055, |
|
"learning_rate": 4.238672286617492e-05, |
|
"loss": 0.3852, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 8.597813606262207, |
|
"learning_rate": 4.212328767123288e-05, |
|
"loss": 0.3998, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 8.316597938537598, |
|
"learning_rate": 4.185985247629083e-05, |
|
"loss": 0.4554, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 7.785365581512451, |
|
"learning_rate": 4.159641728134879e-05, |
|
"loss": 0.4193, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 9.083528518676758, |
|
"learning_rate": 4.133298208640675e-05, |
|
"loss": 0.3247, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 6.537896633148193, |
|
"learning_rate": 4.10695468914647e-05, |
|
"loss": 0.3701, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 4.992751598358154, |
|
"learning_rate": 4.080611169652266e-05, |
|
"loss": 0.3539, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 9.362491607666016, |
|
"learning_rate": 4.054267650158061e-05, |
|
"loss": 0.3521, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 9.023208618164062, |
|
"learning_rate": 4.027924130663857e-05, |
|
"loss": 0.3028, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 11.851508140563965, |
|
"learning_rate": 4.001580611169653e-05, |
|
"loss": 0.3854, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 11.16539192199707, |
|
"learning_rate": 3.975237091675448e-05, |
|
"loss": 0.4021, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 7.1588311195373535, |
|
"learning_rate": 3.948893572181244e-05, |
|
"loss": 0.3865, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 8.23454761505127, |
|
"learning_rate": 3.922550052687039e-05, |
|
"loss": 0.3941, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 9.590554237365723, |
|
"learning_rate": 3.896206533192835e-05, |
|
"loss": 0.3831, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 7.980281829833984, |
|
"learning_rate": 3.86986301369863e-05, |
|
"loss": 0.4134, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 6.231551647186279, |
|
"learning_rate": 3.843519494204426e-05, |
|
"loss": 0.3746, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 8.723283767700195, |
|
"learning_rate": 3.8171759747102217e-05, |
|
"loss": 0.3349, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 7.533834934234619, |
|
"learning_rate": 3.790832455216017e-05, |
|
"loss": 0.3229, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 6.750338554382324, |
|
"learning_rate": 3.764488935721813e-05, |
|
"loss": 0.3163, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 11.813076972961426, |
|
"learning_rate": 3.738145416227608e-05, |
|
"loss": 0.4089, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 7.000370979309082, |
|
"learning_rate": 3.711801896733404e-05, |
|
"loss": 0.4327, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_accuracy": 0.9774, |
|
"eval_loss": 0.06401650607585907, |
|
"eval_runtime": 65.8888, |
|
"eval_samples_per_second": 75.885, |
|
"eval_steps_per_second": 4.75, |
|
"step": 703 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"grad_norm": 7.698484420776367, |
|
"learning_rate": 3.6854583772391995e-05, |
|
"loss": 0.359, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"grad_norm": 10.548850059509277, |
|
"learning_rate": 3.6591148577449954e-05, |
|
"loss": 0.3162, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"grad_norm": 10.149295806884766, |
|
"learning_rate": 3.6327713382507905e-05, |
|
"loss": 0.3265, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"grad_norm": 6.814464092254639, |
|
"learning_rate": 3.606427818756586e-05, |
|
"loss": 0.3536, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"grad_norm": 8.711482048034668, |
|
"learning_rate": 3.5800842992623816e-05, |
|
"loss": 0.379, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"grad_norm": 8.469487190246582, |
|
"learning_rate": 3.5537407797681774e-05, |
|
"loss": 0.3117, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"grad_norm": 9.90764045715332, |
|
"learning_rate": 3.527397260273973e-05, |
|
"loss": 0.3177, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"grad_norm": 7.835967063903809, |
|
"learning_rate": 3.5010537407797684e-05, |
|
"loss": 0.347, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"grad_norm": 5.563517093658447, |
|
"learning_rate": 3.4747102212855636e-05, |
|
"loss": 0.3244, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"grad_norm": 10.490474700927734, |
|
"learning_rate": 3.4483667017913594e-05, |
|
"loss": 0.3159, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"grad_norm": 7.063011169433594, |
|
"learning_rate": 3.4220231822971546e-05, |
|
"loss": 0.2711, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"grad_norm": 9.191926002502441, |
|
"learning_rate": 3.3956796628029505e-05, |
|
"loss": 0.3441, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"grad_norm": 5.757563591003418, |
|
"learning_rate": 3.369336143308746e-05, |
|
"loss": 0.3562, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"grad_norm": 5.849490642547607, |
|
"learning_rate": 3.342992623814542e-05, |
|
"loss": 0.2797, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"grad_norm": 5.61791467666626, |
|
"learning_rate": 3.316649104320337e-05, |
|
"loss": 0.3362, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"grad_norm": 7.7571916580200195, |
|
"learning_rate": 3.2903055848261325e-05, |
|
"loss": 0.2682, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"grad_norm": 9.108525276184082, |
|
"learning_rate": 3.2639620653319283e-05, |
|
"loss": 0.3755, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"grad_norm": 5.915694713592529, |
|
"learning_rate": 3.237618545837724e-05, |
|
"loss": 0.3061, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"grad_norm": 7.744335651397705, |
|
"learning_rate": 3.21127502634352e-05, |
|
"loss": 0.367, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"grad_norm": 9.851384162902832, |
|
"learning_rate": 3.184931506849315e-05, |
|
"loss": 0.3318, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"grad_norm": 6.104685306549072, |
|
"learning_rate": 3.1585879873551104e-05, |
|
"loss": 0.3477, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"grad_norm": 7.819393157958984, |
|
"learning_rate": 3.132244467860906e-05, |
|
"loss": 0.2994, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"grad_norm": 9.686634063720703, |
|
"learning_rate": 3.105900948366702e-05, |
|
"loss": 0.2885, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"grad_norm": 4.374415397644043, |
|
"learning_rate": 3.079557428872498e-05, |
|
"loss": 0.3053, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"grad_norm": 9.000449180603027, |
|
"learning_rate": 3.053213909378293e-05, |
|
"loss": 0.2896, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"grad_norm": 6.736356258392334, |
|
"learning_rate": 3.026870389884089e-05, |
|
"loss": 0.2897, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"grad_norm": 6.521255016326904, |
|
"learning_rate": 3.000526870389884e-05, |
|
"loss": 0.3365, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"grad_norm": 7.526865482330322, |
|
"learning_rate": 2.9741833508956796e-05, |
|
"loss": 0.2817, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"grad_norm": 3.9108359813690186, |
|
"learning_rate": 2.9478398314014755e-05, |
|
"loss": 0.3002, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"grad_norm": 9.110422134399414, |
|
"learning_rate": 2.921496311907271e-05, |
|
"loss": 0.3055, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"grad_norm": 5.4320573806762695, |
|
"learning_rate": 2.8951527924130668e-05, |
|
"loss": 0.3126, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"grad_norm": 8.311667442321777, |
|
"learning_rate": 2.868809272918862e-05, |
|
"loss": 0.3177, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"grad_norm": 8.694578170776367, |
|
"learning_rate": 2.842465753424658e-05, |
|
"loss": 0.2458, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"grad_norm": 6.780071258544922, |
|
"learning_rate": 2.8161222339304533e-05, |
|
"loss": 0.2171, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"grad_norm": 8.7374906539917, |
|
"learning_rate": 2.7897787144362485e-05, |
|
"loss": 0.2914, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"grad_norm": 8.430846214294434, |
|
"learning_rate": 2.7634351949420444e-05, |
|
"loss": 0.2981, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"grad_norm": 6.208935737609863, |
|
"learning_rate": 2.73709167544784e-05, |
|
"loss": 0.2859, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"grad_norm": 6.287723541259766, |
|
"learning_rate": 2.7107481559536357e-05, |
|
"loss": 0.2559, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"grad_norm": 9.158524513244629, |
|
"learning_rate": 2.6844046364594312e-05, |
|
"loss": 0.2544, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"grad_norm": 8.676222801208496, |
|
"learning_rate": 2.6580611169652264e-05, |
|
"loss": 0.2648, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"grad_norm": 5.815190315246582, |
|
"learning_rate": 2.6317175974710222e-05, |
|
"loss": 0.2794, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"grad_norm": 7.133479595184326, |
|
"learning_rate": 2.6053740779768177e-05, |
|
"loss": 0.3008, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"grad_norm": 9.465476989746094, |
|
"learning_rate": 2.5790305584826136e-05, |
|
"loss": 0.3032, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"grad_norm": 5.298926830291748, |
|
"learning_rate": 2.5526870389884088e-05, |
|
"loss": 0.2616, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"grad_norm": 3.1998720169067383, |
|
"learning_rate": 2.5263435194942046e-05, |
|
"loss": 0.3016, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"grad_norm": 9.716959953308105, |
|
"learning_rate": 2.5e-05, |
|
"loss": 0.2625, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"grad_norm": 9.040936470031738, |
|
"learning_rate": 2.4736564805057956e-05, |
|
"loss": 0.2634, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"grad_norm": 7.872269630432129, |
|
"learning_rate": 2.4473129610115915e-05, |
|
"loss": 0.2744, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"grad_norm": 4.666429042816162, |
|
"learning_rate": 2.420969441517387e-05, |
|
"loss": 0.2552, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"grad_norm": 5.423487663269043, |
|
"learning_rate": 2.394625922023182e-05, |
|
"loss": 0.3138, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"grad_norm": 5.613728046417236, |
|
"learning_rate": 2.368282402528978e-05, |
|
"loss": 0.3048, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"grad_norm": 5.959019660949707, |
|
"learning_rate": 2.3419388830347735e-05, |
|
"loss": 0.2784, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"grad_norm": 5.206528663635254, |
|
"learning_rate": 2.315595363540569e-05, |
|
"loss": 0.3245, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"grad_norm": 6.318498611450195, |
|
"learning_rate": 2.289251844046365e-05, |
|
"loss": 0.299, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"grad_norm": 5.1116437911987305, |
|
"learning_rate": 2.2629083245521604e-05, |
|
"loss": 0.2706, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"grad_norm": 9.792173385620117, |
|
"learning_rate": 2.236564805057956e-05, |
|
"loss": 0.2698, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"grad_norm": 5.214483737945557, |
|
"learning_rate": 2.2102212855637514e-05, |
|
"loss": 0.2968, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"grad_norm": 5.681826114654541, |
|
"learning_rate": 2.183877766069547e-05, |
|
"loss": 0.2893, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"grad_norm": 4.573737621307373, |
|
"learning_rate": 2.1575342465753427e-05, |
|
"loss": 0.2424, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"grad_norm": 5.404233932495117, |
|
"learning_rate": 2.1311907270811383e-05, |
|
"loss": 0.2956, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"grad_norm": 7.8233160972595215, |
|
"learning_rate": 2.1048472075869338e-05, |
|
"loss": 0.2992, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"grad_norm": 5.9538373947143555, |
|
"learning_rate": 2.0785036880927293e-05, |
|
"loss": 0.2843, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"grad_norm": 6.776533603668213, |
|
"learning_rate": 2.0521601685985248e-05, |
|
"loss": 0.2255, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"grad_norm": 4.279127597808838, |
|
"learning_rate": 2.0258166491043203e-05, |
|
"loss": 0.2791, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"grad_norm": 9.389575958251953, |
|
"learning_rate": 1.999473129610116e-05, |
|
"loss": 0.3064, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"grad_norm": 6.343210697174072, |
|
"learning_rate": 1.9731296101159116e-05, |
|
"loss": 0.215, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"grad_norm": 6.596449375152588, |
|
"learning_rate": 1.946786090621707e-05, |
|
"loss": 0.2429, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"grad_norm": 6.299315929412842, |
|
"learning_rate": 1.9204425711275027e-05, |
|
"loss": 0.3099, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"grad_norm": 7.726513385772705, |
|
"learning_rate": 1.894099051633298e-05, |
|
"loss": 0.2435, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"grad_norm": 7.626183986663818, |
|
"learning_rate": 1.8677555321390937e-05, |
|
"loss": 0.2814, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_accuracy": 0.9834, |
|
"eval_loss": 0.048424676060676575, |
|
"eval_runtime": 66.0101, |
|
"eval_samples_per_second": 75.746, |
|
"eval_steps_per_second": 4.742, |
|
"step": 1406 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 5.147545337677002, |
|
"learning_rate": 1.8414120126448895e-05, |
|
"loss": 0.2343, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"grad_norm": 3.9744009971618652, |
|
"learning_rate": 1.815068493150685e-05, |
|
"loss": 0.2542, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"grad_norm": 7.23543643951416, |
|
"learning_rate": 1.7887249736564805e-05, |
|
"loss": 0.2568, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"grad_norm": 4.5337395668029785, |
|
"learning_rate": 1.7623814541622764e-05, |
|
"loss": 0.2348, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"grad_norm": 7.225223541259766, |
|
"learning_rate": 1.7360379346680716e-05, |
|
"loss": 0.2861, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"grad_norm": 6.333385467529297, |
|
"learning_rate": 1.7096944151738674e-05, |
|
"loss": 0.2237, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"grad_norm": 9.622603416442871, |
|
"learning_rate": 1.683350895679663e-05, |
|
"loss": 0.2707, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"grad_norm": 3.6370978355407715, |
|
"learning_rate": 1.6570073761854584e-05, |
|
"loss": 0.2177, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"grad_norm": 5.092869758605957, |
|
"learning_rate": 1.630663856691254e-05, |
|
"loss": 0.174, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"grad_norm": 7.764402389526367, |
|
"learning_rate": 1.6043203371970498e-05, |
|
"loss": 0.2166, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"grad_norm": 5.907904624938965, |
|
"learning_rate": 1.577976817702845e-05, |
|
"loss": 0.219, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"grad_norm": 5.130086898803711, |
|
"learning_rate": 1.5516332982086408e-05, |
|
"loss": 0.1956, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"grad_norm": 7.971607685089111, |
|
"learning_rate": 1.5252897787144363e-05, |
|
"loss": 0.1871, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"grad_norm": 10.202670097351074, |
|
"learning_rate": 1.498946259220232e-05, |
|
"loss": 0.2444, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"grad_norm": 5.461312770843506, |
|
"learning_rate": 1.4726027397260275e-05, |
|
"loss": 0.2538, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"grad_norm": 8.133660316467285, |
|
"learning_rate": 1.4462592202318232e-05, |
|
"loss": 0.2324, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"grad_norm": 7.940431594848633, |
|
"learning_rate": 1.4199157007376185e-05, |
|
"loss": 0.2264, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"grad_norm": 6.688675403594971, |
|
"learning_rate": 1.3935721812434142e-05, |
|
"loss": 0.199, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"grad_norm": 7.409407138824463, |
|
"learning_rate": 1.3672286617492097e-05, |
|
"loss": 0.2311, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"grad_norm": 9.85133171081543, |
|
"learning_rate": 1.3408851422550054e-05, |
|
"loss": 0.2153, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"grad_norm": 6.340139865875244, |
|
"learning_rate": 1.3145416227608009e-05, |
|
"loss": 0.2717, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"grad_norm": 5.589156627655029, |
|
"learning_rate": 1.2881981032665966e-05, |
|
"loss": 0.2068, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"grad_norm": 8.83266830444336, |
|
"learning_rate": 1.2618545837723922e-05, |
|
"loss": 0.2459, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"grad_norm": 6.671909809112549, |
|
"learning_rate": 1.2355110642781877e-05, |
|
"loss": 0.2332, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"grad_norm": 8.771206855773926, |
|
"learning_rate": 1.209167544783983e-05, |
|
"loss": 0.2526, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"grad_norm": 9.015884399414062, |
|
"learning_rate": 1.1828240252897788e-05, |
|
"loss": 0.2239, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"grad_norm": 9.24109172821045, |
|
"learning_rate": 1.1564805057955744e-05, |
|
"loss": 0.2304, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"grad_norm": 6.8663330078125, |
|
"learning_rate": 1.1301369863013698e-05, |
|
"loss": 0.2107, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"grad_norm": 5.675321578979492, |
|
"learning_rate": 1.1037934668071655e-05, |
|
"loss": 0.2607, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"grad_norm": 6.900195598602295, |
|
"learning_rate": 1.0774499473129611e-05, |
|
"loss": 0.2323, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"grad_norm": 11.515998840332031, |
|
"learning_rate": 1.0511064278187566e-05, |
|
"loss": 0.2704, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"grad_norm": 7.85768461227417, |
|
"learning_rate": 1.0247629083245521e-05, |
|
"loss": 0.2175, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"grad_norm": 4.934202194213867, |
|
"learning_rate": 9.984193888303478e-06, |
|
"loss": 0.2069, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"grad_norm": 5.035346031188965, |
|
"learning_rate": 9.720758693361433e-06, |
|
"loss": 0.2145, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"grad_norm": 7.098711967468262, |
|
"learning_rate": 9.457323498419388e-06, |
|
"loss": 0.2642, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"grad_norm": 7.717360019683838, |
|
"learning_rate": 9.193888303477345e-06, |
|
"loss": 0.1801, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"grad_norm": 7.103116035461426, |
|
"learning_rate": 8.930453108535302e-06, |
|
"loss": 0.212, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"grad_norm": 4.945562839508057, |
|
"learning_rate": 8.667017913593255e-06, |
|
"loss": 0.2158, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"grad_norm": 8.597467422485352, |
|
"learning_rate": 8.403582718651212e-06, |
|
"loss": 0.233, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"grad_norm": 3.4780356884002686, |
|
"learning_rate": 8.140147523709169e-06, |
|
"loss": 0.2139, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"grad_norm": 8.698033332824707, |
|
"learning_rate": 7.876712328767124e-06, |
|
"loss": 0.2285, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"grad_norm": 8.690217971801758, |
|
"learning_rate": 7.613277133825079e-06, |
|
"loss": 0.221, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"grad_norm": 6.253328800201416, |
|
"learning_rate": 7.349841938883036e-06, |
|
"loss": 0.2156, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"grad_norm": 6.390902042388916, |
|
"learning_rate": 7.08640674394099e-06, |
|
"loss": 0.2134, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"grad_norm": 3.520571231842041, |
|
"learning_rate": 6.822971548998947e-06, |
|
"loss": 0.2177, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"grad_norm": 4.312754154205322, |
|
"learning_rate": 6.559536354056903e-06, |
|
"loss": 0.1913, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"grad_norm": 4.5632758140563965, |
|
"learning_rate": 6.296101159114858e-06, |
|
"loss": 0.2014, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"grad_norm": 8.447988510131836, |
|
"learning_rate": 6.032665964172814e-06, |
|
"loss": 0.2445, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"grad_norm": 4.897846698760986, |
|
"learning_rate": 5.76923076923077e-06, |
|
"loss": 0.1689, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"grad_norm": 6.561771392822266, |
|
"learning_rate": 5.505795574288726e-06, |
|
"loss": 0.2089, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"grad_norm": 8.186767578125, |
|
"learning_rate": 5.242360379346681e-06, |
|
"loss": 0.2337, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"grad_norm": 3.8493165969848633, |
|
"learning_rate": 4.978925184404637e-06, |
|
"loss": 0.2202, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"grad_norm": 4.461836814880371, |
|
"learning_rate": 4.715489989462593e-06, |
|
"loss": 0.2052, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"grad_norm": 6.122828483581543, |
|
"learning_rate": 4.452054794520548e-06, |
|
"loss": 0.2447, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"grad_norm": 7.432437419891357, |
|
"learning_rate": 4.188619599578504e-06, |
|
"loss": 0.2148, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"grad_norm": 5.491920471191406, |
|
"learning_rate": 3.92518440463646e-06, |
|
"loss": 0.2425, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"grad_norm": 5.547910690307617, |
|
"learning_rate": 3.661749209694415e-06, |
|
"loss": 0.2409, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"grad_norm": 7.146701335906982, |
|
"learning_rate": 3.398314014752371e-06, |
|
"loss": 0.2349, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"grad_norm": 9.67960262298584, |
|
"learning_rate": 3.1348788198103265e-06, |
|
"loss": 0.2119, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"grad_norm": 7.780117511749268, |
|
"learning_rate": 2.8714436248682825e-06, |
|
"loss": 0.225, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"grad_norm": 7.887894153594971, |
|
"learning_rate": 2.6080084299262384e-06, |
|
"loss": 0.1839, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"grad_norm": 3.4736483097076416, |
|
"learning_rate": 2.3445732349841943e-06, |
|
"loss": 0.1897, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"grad_norm": 3.9504001140594482, |
|
"learning_rate": 2.08113804004215e-06, |
|
"loss": 0.1665, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"grad_norm": 3.5273189544677734, |
|
"learning_rate": 1.8177028451001056e-06, |
|
"loss": 0.1694, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"grad_norm": 6.938018321990967, |
|
"learning_rate": 1.554267650158061e-06, |
|
"loss": 0.2071, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"grad_norm": 3.9890809059143066, |
|
"learning_rate": 1.290832455216017e-06, |
|
"loss": 0.1837, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"grad_norm": 5.470897197723389, |
|
"learning_rate": 1.0273972602739725e-06, |
|
"loss": 0.1977, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"grad_norm": 7.521297931671143, |
|
"learning_rate": 7.639620653319284e-07, |
|
"loss": 0.2391, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"grad_norm": 5.917148590087891, |
|
"learning_rate": 5.00526870389884e-07, |
|
"loss": 0.1858, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"grad_norm": 5.207727909088135, |
|
"learning_rate": 2.3709167544783985e-07, |
|
"loss": 0.1655, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_accuracy": 0.9882, |
|
"eval_loss": 0.03509841486811638, |
|
"eval_runtime": 65.8746, |
|
"eval_samples_per_second": 75.902, |
|
"eval_steps_per_second": 4.751, |
|
"step": 2109 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"step": 2109, |
|
"total_flos": 1.045459878817923e+19, |
|
"train_loss": 0.37951253058504525, |
|
"train_runtime": 4954.8349, |
|
"train_samples_per_second": 27.246, |
|
"train_steps_per_second": 0.426 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 2109, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 500, |
|
"total_flos": 1.045459878817923e+19, |
|
"train_batch_size": 16, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|