|
{ |
|
"best_metric": 1.8259689807891846, |
|
"best_model_checkpoint": "model_training/byt5_small/checkpoints-by_file-09-07-11-51/checkpoint-350", |
|
"epoch": 5.685279187817259, |
|
"eval_steps": 25, |
|
"global_step": 350, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.08121827411167512, |
|
"grad_norm": 1332.5194091796875, |
|
"learning_rate": 5.102040816326531e-05, |
|
"loss": 57.087, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.16243654822335024, |
|
"grad_norm": 73.31609344482422, |
|
"learning_rate": 0.00010204081632653062, |
|
"loss": 16.158, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.2436548223350254, |
|
"grad_norm": 7.558203220367432, |
|
"learning_rate": 0.00015306122448979594, |
|
"loss": 4.2458, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.3248730964467005, |
|
"grad_norm": 3.45859432220459, |
|
"learning_rate": 0.00020408163265306123, |
|
"loss": 3.5694, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.40609137055837563, |
|
"grad_norm": 1.5817078351974487, |
|
"learning_rate": 0.00025510204081632655, |
|
"loss": 2.5571, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.40609137055837563, |
|
"eval_loss": 1.86122727394104, |
|
"eval_runtime": 89.2369, |
|
"eval_samples_per_second": 11.206, |
|
"eval_steps_per_second": 1.401, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.4873096446700508, |
|
"grad_norm": 0.7394866943359375, |
|
"learning_rate": 0.0003061224489795919, |
|
"loss": 2.1392, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.5685279187817259, |
|
"grad_norm": 8.203688621520996, |
|
"learning_rate": 0.00035714285714285714, |
|
"loss": 2.0015, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.649746192893401, |
|
"grad_norm": 8.245688438415527, |
|
"learning_rate": 0.00040816326530612246, |
|
"loss": 1.9816, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.7309644670050761, |
|
"grad_norm": 1.8892487287521362, |
|
"learning_rate": 0.0004591836734693878, |
|
"loss": 2.3033, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.8121827411167513, |
|
"grad_norm": 6.5841898918151855, |
|
"learning_rate": 0.0004999935985425297, |
|
"loss": 3.3388, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.8121827411167513, |
|
"eval_loss": 2.1760518550872803, |
|
"eval_runtime": 88.9886, |
|
"eval_samples_per_second": 11.237, |
|
"eval_steps_per_second": 1.405, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.8934010152284264, |
|
"grad_norm": 7.12192440032959, |
|
"learning_rate": 0.0004997695819512612, |
|
"loss": 3.6302, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.9746192893401016, |
|
"grad_norm": 40.77882385253906, |
|
"learning_rate": 0.0004992258202402822, |
|
"loss": 2.7657, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 1.0558375634517767, |
|
"grad_norm": 2.32853102684021, |
|
"learning_rate": 0.0004983630095117843, |
|
"loss": 2.8533, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 1.1370558375634519, |
|
"grad_norm": 20.28560447692871, |
|
"learning_rate": 0.0004971822543018662, |
|
"loss": 2.6894, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 1.218274111675127, |
|
"grad_norm": 13.825627326965332, |
|
"learning_rate": 0.0004956850661665511, |
|
"loss": 2.6103, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 1.218274111675127, |
|
"eval_loss": 2.4702670574188232, |
|
"eval_runtime": 89.4229, |
|
"eval_samples_per_second": 11.183, |
|
"eval_steps_per_second": 1.398, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 1.299492385786802, |
|
"grad_norm": 10.716843605041504, |
|
"learning_rate": 0.0004938733617467517, |
|
"loss": 2.565, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 1.380710659898477, |
|
"grad_norm": 1.5047727823257446, |
|
"learning_rate": 0.0004917494603146632, |
|
"loss": 2.4888, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 1.4619289340101522, |
|
"grad_norm": 5.189760684967041, |
|
"learning_rate": 0.0004893160808047222, |
|
"loss": 2.4684, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 1.5431472081218274, |
|
"grad_norm": 1.0872949361801147, |
|
"learning_rate": 0.00048657633833293557, |
|
"loss": 2.4044, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 1.6243654822335025, |
|
"grad_norm": 0.9995412230491638, |
|
"learning_rate": 0.0004835337402090316, |
|
"loss": 2.3281, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 1.6243654822335025, |
|
"eval_loss": 2.164686918258667, |
|
"eval_runtime": 89.4619, |
|
"eval_samples_per_second": 11.178, |
|
"eval_steps_per_second": 1.397, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 1.7055837563451777, |
|
"grad_norm": 1.2759788036346436, |
|
"learning_rate": 0.0004801921814465414, |
|
"loss": 2.2785, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 1.7868020304568528, |
|
"grad_norm": 0.9360132217407227, |
|
"learning_rate": 0.00047655593977655674, |
|
"loss": 2.214, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 1.868020304568528, |
|
"grad_norm": 0.33227458596229553, |
|
"learning_rate": 0.0004726296701715489, |
|
"loss": 2.1815, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 1.9492385786802031, |
|
"grad_norm": 0.32235994935035706, |
|
"learning_rate": 0.00046841839888625623, |
|
"loss": 2.1927, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 2.030456852791878, |
|
"grad_norm": 0.5263159275054932, |
|
"learning_rate": 0.0004639275170232734, |
|
"loss": 2.1143, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 2.030456852791878, |
|
"eval_loss": 2.0160372257232666, |
|
"eval_runtime": 88.8689, |
|
"eval_samples_per_second": 11.253, |
|
"eval_steps_per_second": 1.407, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 2.1116751269035534, |
|
"grad_norm": 0.4549558162689209, |
|
"learning_rate": 0.0004591627736315743, |
|
"loss": 2.1318, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 2.1928934010152283, |
|
"grad_norm": 0.670222818851471, |
|
"learning_rate": 0.0004541302683468084, |
|
"loss": 2.1073, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 2.2741116751269037, |
|
"grad_norm": 0.536625862121582, |
|
"learning_rate": 0.0004488364435827881, |
|
"loss": 2.1021, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 2.3553299492385786, |
|
"grad_norm": 0.3629292845726013, |
|
"learning_rate": 0.00044328807628416644, |
|
"loss": 2.0923, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 2.436548223350254, |
|
"grad_norm": 0.2985295057296753, |
|
"learning_rate": 0.0004374922692508611, |
|
"loss": 2.0753, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 2.436548223350254, |
|
"eval_loss": 1.948063611984253, |
|
"eval_runtime": 88.8793, |
|
"eval_samples_per_second": 11.251, |
|
"eval_steps_per_second": 1.406, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 2.517766497461929, |
|
"grad_norm": 0.26300689578056335, |
|
"learning_rate": 0.0004314564420453311, |
|
"loss": 2.0447, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 2.598984771573604, |
|
"grad_norm": 0.1663161963224411, |
|
"learning_rate": 0.0004251883214943475, |
|
"loss": 2.0429, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 2.6802030456852792, |
|
"grad_norm": 0.6070680022239685, |
|
"learning_rate": 0.0004186959317974155, |
|
"loss": 2.0699, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 2.761421319796954, |
|
"grad_norm": 1.7647202014923096, |
|
"learning_rate": 0.00041198758425451266, |
|
"loss": 2.0304, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 2.8426395939086295, |
|
"grad_norm": 0.5461482405662537, |
|
"learning_rate": 0.00040507186662629185, |
|
"loss": 2.0449, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 2.8426395939086295, |
|
"eval_loss": 1.8957690000534058, |
|
"eval_runtime": 88.8966, |
|
"eval_samples_per_second": 11.249, |
|
"eval_steps_per_second": 1.406, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 2.9238578680203045, |
|
"grad_norm": 1.156761646270752, |
|
"learning_rate": 0.0003979576321403705, |
|
"loss": 1.9947, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 3.00507614213198, |
|
"grad_norm": 0.465610533952713, |
|
"learning_rate": 0.0003906539881577793, |
|
"loss": 2.0121, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 3.0862944162436547, |
|
"grad_norm": 0.1722594052553177, |
|
"learning_rate": 0.0003831702845140801, |
|
"loss": 2.0141, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 3.16751269035533, |
|
"grad_norm": 0.24845992028713226, |
|
"learning_rate": 0.00037551610155007613, |
|
"loss": 1.9536, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 3.248730964467005, |
|
"grad_norm": 0.4477495551109314, |
|
"learning_rate": 0.00036770123784744027, |
|
"loss": 2.0201, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 3.248730964467005, |
|
"eval_loss": 1.8864437341690063, |
|
"eval_runtime": 88.9059, |
|
"eval_samples_per_second": 11.248, |
|
"eval_steps_per_second": 1.406, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 3.3299492385786804, |
|
"grad_norm": 0.5112472176551819, |
|
"learning_rate": 0.00035973569768495855, |
|
"loss": 1.9846, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 3.4111675126903553, |
|
"grad_norm": 0.6808192133903503, |
|
"learning_rate": 0.0003516296782314491, |
|
"loss": 1.9839, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 3.4923857868020303, |
|
"grad_norm": 0.24290819466114044, |
|
"learning_rate": 0.00034339355649175095, |
|
"loss": 2.0036, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 3.5736040609137056, |
|
"grad_norm": 0.34100961685180664, |
|
"learning_rate": 0.00033503787602249364, |
|
"loss": 2.0178, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 3.6548223350253806, |
|
"grad_norm": 0.7999327182769775, |
|
"learning_rate": 0.00032657333343465356, |
|
"loss": 1.9947, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 3.6548223350253806, |
|
"eval_loss": 1.8772450685501099, |
|
"eval_runtime": 88.947, |
|
"eval_samples_per_second": 11.243, |
|
"eval_steps_per_second": 1.405, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 3.736040609137056, |
|
"grad_norm": 0.3293197453022003, |
|
"learning_rate": 0.0003180107647001769, |
|
"loss": 2.0078, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 3.817258883248731, |
|
"grad_norm": 0.19176721572875977, |
|
"learning_rate": 0.0003093611312801979, |
|
"loss": 1.9594, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 3.8984771573604062, |
|
"grad_norm": 0.24580970406532288, |
|
"learning_rate": 0.00030063550609261025, |
|
"loss": 1.9514, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 3.979695431472081, |
|
"grad_norm": 0.28957948088645935, |
|
"learning_rate": 0.000291845059336957, |
|
"loss": 1.9809, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 4.060913705583756, |
|
"grad_norm": 0.19108451902866364, |
|
"learning_rate": 0.0002830010441947834, |
|
"loss": 1.9295, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 4.060913705583756, |
|
"eval_loss": 1.8687773942947388, |
|
"eval_runtime": 88.8506, |
|
"eval_samples_per_second": 11.255, |
|
"eval_steps_per_second": 1.407, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 4.1421319796954315, |
|
"grad_norm": 0.18111436069011688, |
|
"learning_rate": 0.00027411478242376017, |
|
"loss": 1.9789, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 4.223350253807107, |
|
"grad_norm": 0.4630391299724579, |
|
"learning_rate": 0.00026519764986401774, |
|
"loss": 1.9657, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 4.304568527918782, |
|
"grad_norm": 0.2104698121547699, |
|
"learning_rate": 0.000256261061875247, |
|
"loss": 1.9403, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 4.385786802030457, |
|
"grad_norm": 0.4305352568626404, |
|
"learning_rate": 0.0002473164587232079, |
|
"loss": 1.9762, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 4.467005076142132, |
|
"grad_norm": 0.25098443031311035, |
|
"learning_rate": 0.0002383752909343547, |
|
"loss": 1.9564, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 4.467005076142132, |
|
"eval_loss": 1.8538275957107544, |
|
"eval_runtime": 88.8543, |
|
"eval_samples_per_second": 11.254, |
|
"eval_steps_per_second": 1.407, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 4.548223350253807, |
|
"grad_norm": 0.3380163908004761, |
|
"learning_rate": 0.0002294490046373259, |
|
"loss": 1.9096, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 4.629441624365482, |
|
"grad_norm": 0.18323899805545807, |
|
"learning_rate": 0.00022054902691006405, |
|
"loss": 1.9775, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 4.710659898477157, |
|
"grad_norm": 0.13836567103862762, |
|
"learning_rate": 0.00021168675115132315, |
|
"loss": 1.9888, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 4.791878172588833, |
|
"grad_norm": 0.20029225945472717, |
|
"learning_rate": 0.00020287352249529153, |
|
"loss": 1.9547, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 4.873096446700508, |
|
"grad_norm": 0.16367541253566742, |
|
"learning_rate": 0.00019412062328800044, |
|
"loss": 1.9243, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 4.873096446700508, |
|
"eval_loss": 1.8513847589492798, |
|
"eval_runtime": 88.8985, |
|
"eval_samples_per_second": 11.249, |
|
"eval_steps_per_second": 1.406, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 4.9543147208121825, |
|
"grad_norm": 0.14919321238994598, |
|
"learning_rate": 0.000185439258644112, |
|
"loss": 1.9703, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 5.035532994923858, |
|
"grad_norm": 0.20038799941539764, |
|
"learning_rate": 0.00017684054210257517, |
|
"loss": 1.9606, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 5.116751269035533, |
|
"grad_norm": 0.3811420798301697, |
|
"learning_rate": 0.00016833548139951395, |
|
"loss": 1.971, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 5.197969543147208, |
|
"grad_norm": 0.34453725814819336, |
|
"learning_rate": 0.0001599349643765599, |
|
"loss": 1.9342, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 5.279187817258883, |
|
"grad_norm": 0.22745971381664276, |
|
"learning_rate": 0.0001516497450426686, |
|
"loss": 1.958, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 5.279187817258883, |
|
"eval_loss": 1.8567323684692383, |
|
"eval_runtime": 88.9414, |
|
"eval_samples_per_second": 11.243, |
|
"eval_steps_per_second": 1.405, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 5.3604060913705585, |
|
"grad_norm": 0.3236718475818634, |
|
"learning_rate": 0.00014349042980726362, |
|
"loss": 1.927, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 5.441624365482234, |
|
"grad_norm": 0.5001394748687744, |
|
"learning_rate": 0.0001354674639023318, |
|
"loss": 1.9415, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 5.522842639593908, |
|
"grad_norm": 0.5372613668441772, |
|
"learning_rate": 0.00012759111801085066, |
|
"loss": 1.9346, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 5.604060913705584, |
|
"grad_norm": 0.39717650413513184, |
|
"learning_rate": 0.00011987147511866788, |
|
"loss": 1.9491, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 5.685279187817259, |
|
"grad_norm": 0.34586790204048157, |
|
"learning_rate": 0.00011231841760666186, |
|
"loss": 1.872, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 5.685279187817259, |
|
"eval_loss": 1.8259689807891846, |
|
"eval_runtime": 88.9255, |
|
"eval_samples_per_second": 11.245, |
|
"eval_steps_per_second": 1.406, |
|
"step": 350 |
|
} |
|
], |
|
"logging_steps": 5, |
|
"max_steps": 488, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 8, |
|
"save_steps": 25, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 3.435956091813888e+17, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|