|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 5.0, |
|
"eval_steps": 500, |
|
"global_step": 28210, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.822757887274017e-05, |
|
"loss": 0.1689, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.645515774548033e-05, |
|
"loss": 0.1355, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.46827366182205e-05, |
|
"loss": 0.1347, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.291031549096066e-05, |
|
"loss": 0.1311, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 9.113789436370082e-05, |
|
"loss": 0.1305, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 8.936547323644098e-05, |
|
"loss": 0.1282, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 8.759305210918115e-05, |
|
"loss": 0.1283, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 8.582063098192131e-05, |
|
"loss": 0.124, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 8.404820985466147e-05, |
|
"loss": 0.1278, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 8.227578872740164e-05, |
|
"loss": 0.1157, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 8.05033676001418e-05, |
|
"loss": 0.1241, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 0.11910360306501389, |
|
"eval_runtime": 84.6024, |
|
"eval_samples_per_second": 33.344, |
|
"eval_steps_per_second": 8.345, |
|
"step": 5642 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 7.873094647288196e-05, |
|
"loss": 0.1108, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 7.695852534562212e-05, |
|
"loss": 0.1108, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 7.518610421836229e-05, |
|
"loss": 0.1086, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 7.341368309110245e-05, |
|
"loss": 0.1098, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 7.164126196384261e-05, |
|
"loss": 0.1072, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 6.986884083658278e-05, |
|
"loss": 0.1045, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 6.809641970932294e-05, |
|
"loss": 0.1069, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 6.632399858206309e-05, |
|
"loss": 0.104, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 6.455157745480327e-05, |
|
"loss": 0.1063, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 6.277915632754343e-05, |
|
"loss": 0.1041, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 6.100673520028359e-05, |
|
"loss": 0.1033, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 0.11260941624641418, |
|
"eval_runtime": 84.6202, |
|
"eval_samples_per_second": 33.337, |
|
"eval_steps_per_second": 8.343, |
|
"step": 11284 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 5.9234314073023755e-05, |
|
"loss": 0.0969, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 5.746189294576392e-05, |
|
"loss": 0.0906, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 5.568947181850408e-05, |
|
"loss": 0.0902, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 5.3917050691244244e-05, |
|
"loss": 0.0926, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 5.21446295639844e-05, |
|
"loss": 0.0901, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 5.0372208436724564e-05, |
|
"loss": 0.0917, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 4.8599787309464734e-05, |
|
"loss": 0.0927, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 4.68273661822049e-05, |
|
"loss": 0.0887, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 4.505494505494506e-05, |
|
"loss": 0.0918, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 4.328252392768522e-05, |
|
"loss": 0.0898, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 4.151010280042538e-05, |
|
"loss": 0.093, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_loss": 0.10974710434675217, |
|
"eval_runtime": 84.6092, |
|
"eval_samples_per_second": 33.342, |
|
"eval_steps_per_second": 8.344, |
|
"step": 16926 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"learning_rate": 3.973768167316554e-05, |
|
"loss": 0.0891, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 3.1, |
|
"learning_rate": 3.7965260545905705e-05, |
|
"loss": 0.0781, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 3.19, |
|
"learning_rate": 3.6192839418645875e-05, |
|
"loss": 0.0774, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 3.28, |
|
"learning_rate": 3.442041829138604e-05, |
|
"loss": 0.0805, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 3.37, |
|
"learning_rate": 3.26479971641262e-05, |
|
"loss": 0.0775, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 3.46, |
|
"learning_rate": 3.087557603686636e-05, |
|
"loss": 0.0793, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 3.54, |
|
"learning_rate": 2.9103154909606524e-05, |
|
"loss": 0.0808, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 3.63, |
|
"learning_rate": 2.7330733782346684e-05, |
|
"loss": 0.0785, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 3.72, |
|
"learning_rate": 2.5558312655086853e-05, |
|
"loss": 0.0819, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 3.81, |
|
"learning_rate": 2.3785891527827013e-05, |
|
"loss": 0.0804, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 3.9, |
|
"learning_rate": 2.2013470400567176e-05, |
|
"loss": 0.0793, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 3.99, |
|
"learning_rate": 2.024104927330734e-05, |
|
"loss": 0.0784, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_loss": 0.10898815095424652, |
|
"eval_runtime": 84.6147, |
|
"eval_samples_per_second": 33.339, |
|
"eval_steps_per_second": 8.344, |
|
"step": 22568 |
|
}, |
|
{ |
|
"epoch": 4.08, |
|
"learning_rate": 1.8468628146047502e-05, |
|
"loss": 0.0731, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 4.17, |
|
"learning_rate": 1.6696207018787665e-05, |
|
"loss": 0.0707, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 4.25, |
|
"learning_rate": 1.4923785891527828e-05, |
|
"loss": 0.0729, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 4.34, |
|
"learning_rate": 1.315136476426799e-05, |
|
"loss": 0.0706, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 4.43, |
|
"learning_rate": 1.1378943637008153e-05, |
|
"loss": 0.0729, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 4.52, |
|
"learning_rate": 9.606522509748318e-06, |
|
"loss": 0.0703, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 4.61, |
|
"learning_rate": 7.834101382488479e-06, |
|
"loss": 0.0714, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 4.7, |
|
"learning_rate": 6.061680255228642e-06, |
|
"loss": 0.0708, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 4.79, |
|
"learning_rate": 4.289259127968805e-06, |
|
"loss": 0.072, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 4.87, |
|
"learning_rate": 2.5168380007089687e-06, |
|
"loss": 0.0693, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 4.96, |
|
"learning_rate": 7.444168734491315e-07, |
|
"loss": 0.0726, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_loss": 0.1103164553642273, |
|
"eval_runtime": 84.6037, |
|
"eval_samples_per_second": 33.344, |
|
"eval_steps_per_second": 8.345, |
|
"step": 28210 |
|
} |
|
], |
|
"logging_steps": 500, |
|
"max_steps": 28210, |
|
"num_train_epochs": 5, |
|
"save_steps": 500, |
|
"total_flos": 5.89605470208e+16, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|