|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 1.4149274849663955, |
|
"eval_steps": 1000, |
|
"global_step": 2000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.017686593562079942, |
|
"grad_norm": 28.95281982421875, |
|
"learning_rate": 4.6000000000000004e-07, |
|
"loss": 1.849, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.035373187124159884, |
|
"grad_norm": 21.0062255859375, |
|
"learning_rate": 9.600000000000001e-07, |
|
"loss": 1.7562, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.05305978068623983, |
|
"grad_norm": 21.639610290527344, |
|
"learning_rate": 1.46e-06, |
|
"loss": 1.6167, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.07074637424831977, |
|
"grad_norm": 19.797197341918945, |
|
"learning_rate": 1.9600000000000003e-06, |
|
"loss": 1.4966, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.08843296781039972, |
|
"grad_norm": 19.28240203857422, |
|
"learning_rate": 2.46e-06, |
|
"loss": 1.2549, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.10611956137247966, |
|
"grad_norm": 19.18793487548828, |
|
"learning_rate": 2.96e-06, |
|
"loss": 1.1787, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.1238061549345596, |
|
"grad_norm": 17.472681045532227, |
|
"learning_rate": 3.46e-06, |
|
"loss": 1.074, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.14149274849663954, |
|
"grad_norm": 17.52532386779785, |
|
"learning_rate": 3.96e-06, |
|
"loss": 1.1118, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.1591793420587195, |
|
"grad_norm": 17.355623245239258, |
|
"learning_rate": 4.4600000000000005e-06, |
|
"loss": 1.0218, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.17686593562079944, |
|
"grad_norm": 18.548181533813477, |
|
"learning_rate": 4.960000000000001e-06, |
|
"loss": 1.0106, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.19455252918287938, |
|
"grad_norm": 18.0971736907959, |
|
"learning_rate": 5.460000000000001e-06, |
|
"loss": 0.9914, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.21223912274495932, |
|
"grad_norm": 16.739173889160156, |
|
"learning_rate": 5.9600000000000005e-06, |
|
"loss": 0.96, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.22992571630703926, |
|
"grad_norm": 17.19851303100586, |
|
"learning_rate": 6.460000000000001e-06, |
|
"loss": 0.9472, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.2476123098691192, |
|
"grad_norm": 18.49357795715332, |
|
"learning_rate": 6.96e-06, |
|
"loss": 0.8836, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.26529890343119916, |
|
"grad_norm": 17.01639175415039, |
|
"learning_rate": 7.4600000000000006e-06, |
|
"loss": 0.8465, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.2829854969932791, |
|
"grad_norm": 16.33963966369629, |
|
"learning_rate": 7.960000000000002e-06, |
|
"loss": 0.8472, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.30067209055535904, |
|
"grad_norm": 16.123769760131836, |
|
"learning_rate": 8.46e-06, |
|
"loss": 0.8425, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.318358684117439, |
|
"grad_norm": 17.656047821044922, |
|
"learning_rate": 8.96e-06, |
|
"loss": 0.8385, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.3360452776795189, |
|
"grad_norm": 17.501123428344727, |
|
"learning_rate": 9.460000000000001e-06, |
|
"loss": 0.8041, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.3537318712415989, |
|
"grad_norm": 15.51560115814209, |
|
"learning_rate": 9.960000000000001e-06, |
|
"loss": 0.7809, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.3714184648036788, |
|
"grad_norm": 14.336594581604004, |
|
"learning_rate": 9.846666666666668e-06, |
|
"loss": 0.7417, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.38910505836575876, |
|
"grad_norm": 14.65346908569336, |
|
"learning_rate": 9.68e-06, |
|
"loss": 0.7605, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.40679165192783867, |
|
"grad_norm": 15.94804573059082, |
|
"learning_rate": 9.513333333333334e-06, |
|
"loss": 0.7472, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.42447824548991864, |
|
"grad_norm": 17.208477020263672, |
|
"learning_rate": 9.346666666666666e-06, |
|
"loss": 0.7179, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.4421648390519986, |
|
"grad_norm": 16.141422271728516, |
|
"learning_rate": 9.180000000000002e-06, |
|
"loss": 0.7047, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 0.4598514326140785, |
|
"grad_norm": 14.764686584472656, |
|
"learning_rate": 9.013333333333334e-06, |
|
"loss": 0.705, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.4775380261761585, |
|
"grad_norm": 15.155953407287598, |
|
"learning_rate": 8.846666666666668e-06, |
|
"loss": 0.6948, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 0.4952246197382384, |
|
"grad_norm": 15.866467475891113, |
|
"learning_rate": 8.68e-06, |
|
"loss": 0.6712, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.5129112133003184, |
|
"grad_norm": 16.8061466217041, |
|
"learning_rate": 8.513333333333335e-06, |
|
"loss": 0.6726, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 0.5305978068623983, |
|
"grad_norm": 14.077337265014648, |
|
"learning_rate": 8.346666666666668e-06, |
|
"loss": 0.6666, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.5482844004244782, |
|
"grad_norm": 13.416068077087402, |
|
"learning_rate": 8.18e-06, |
|
"loss": 0.6808, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 0.5659709939865581, |
|
"grad_norm": 14.89547348022461, |
|
"learning_rate": 8.013333333333333e-06, |
|
"loss": 0.6492, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.5836575875486382, |
|
"grad_norm": 13.067695617675781, |
|
"learning_rate": 7.846666666666667e-06, |
|
"loss": 0.651, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 0.6013441811107181, |
|
"grad_norm": 14.657025337219238, |
|
"learning_rate": 7.680000000000001e-06, |
|
"loss": 0.6188, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.619030774672798, |
|
"grad_norm": 16.345060348510742, |
|
"learning_rate": 7.513333333333334e-06, |
|
"loss": 0.6519, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 0.636717368234878, |
|
"grad_norm": 17.713294982910156, |
|
"learning_rate": 7.346666666666668e-06, |
|
"loss": 0.604, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.6544039617969579, |
|
"grad_norm": 14.218528747558594, |
|
"learning_rate": 7.180000000000001e-06, |
|
"loss": 0.6272, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 0.6720905553590378, |
|
"grad_norm": 12.941822052001953, |
|
"learning_rate": 7.0133333333333345e-06, |
|
"loss": 0.6271, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.6897771489211177, |
|
"grad_norm": 15.65915298461914, |
|
"learning_rate": 6.846666666666667e-06, |
|
"loss": 0.6139, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 0.7074637424831978, |
|
"grad_norm": 15.025140762329102, |
|
"learning_rate": 6.680000000000001e-06, |
|
"loss": 0.6185, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.7074637424831978, |
|
"eval_loss": 0.6446258425712585, |
|
"eval_runtime": 1231.4066, |
|
"eval_samples_per_second": 8.74, |
|
"eval_steps_per_second": 0.547, |
|
"eval_wer": 0.3809160603927091, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.7251503360452777, |
|
"grad_norm": 15.671004295349121, |
|
"learning_rate": 6.513333333333333e-06, |
|
"loss": 0.5958, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 0.7428369296073576, |
|
"grad_norm": 14.718060493469238, |
|
"learning_rate": 6.346666666666668e-06, |
|
"loss": 0.5905, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.7605235231694376, |
|
"grad_norm": 14.305146217346191, |
|
"learning_rate": 6.18e-06, |
|
"loss": 0.6024, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 0.7782101167315175, |
|
"grad_norm": 14.678905487060547, |
|
"learning_rate": 6.013333333333335e-06, |
|
"loss": 0.6035, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.7958967102935974, |
|
"grad_norm": 13.787155151367188, |
|
"learning_rate": 5.846666666666667e-06, |
|
"loss": 0.5738, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 0.8135833038556773, |
|
"grad_norm": 13.068892478942871, |
|
"learning_rate": 5.68e-06, |
|
"loss": 0.5771, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.8312698974177574, |
|
"grad_norm": 14.032684326171875, |
|
"learning_rate": 5.513333333333334e-06, |
|
"loss": 0.5921, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 0.8489564909798373, |
|
"grad_norm": 13.227592468261719, |
|
"learning_rate": 5.346666666666667e-06, |
|
"loss": 0.5848, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.8666430845419172, |
|
"grad_norm": 15.441591262817383, |
|
"learning_rate": 5.18e-06, |
|
"loss": 0.5607, |
|
"step": 1225 |
|
}, |
|
{ |
|
"epoch": 0.8843296781039972, |
|
"grad_norm": 13.670654296875, |
|
"learning_rate": 5.013333333333333e-06, |
|
"loss": 0.5968, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.9020162716660771, |
|
"grad_norm": 14.419194221496582, |
|
"learning_rate": 4.846666666666667e-06, |
|
"loss": 0.533, |
|
"step": 1275 |
|
}, |
|
{ |
|
"epoch": 0.919702865228157, |
|
"grad_norm": 14.973267555236816, |
|
"learning_rate": 4.680000000000001e-06, |
|
"loss": 0.5584, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.937389458790237, |
|
"grad_norm": 13.328692436218262, |
|
"learning_rate": 4.513333333333333e-06, |
|
"loss": 0.5475, |
|
"step": 1325 |
|
}, |
|
{ |
|
"epoch": 0.955076052352317, |
|
"grad_norm": 15.836326599121094, |
|
"learning_rate": 4.346666666666667e-06, |
|
"loss": 0.5422, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.9727626459143969, |
|
"grad_norm": 12.22904109954834, |
|
"learning_rate": 4.18e-06, |
|
"loss": 0.574, |
|
"step": 1375 |
|
}, |
|
{ |
|
"epoch": 0.9904492394764768, |
|
"grad_norm": 16.11627769470215, |
|
"learning_rate": 4.013333333333334e-06, |
|
"loss": 0.5559, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 1.0081358330385568, |
|
"grad_norm": 12.527290344238281, |
|
"learning_rate": 3.8466666666666665e-06, |
|
"loss": 0.5176, |
|
"step": 1425 |
|
}, |
|
{ |
|
"epoch": 1.0258224266006368, |
|
"grad_norm": 11.869111061096191, |
|
"learning_rate": 3.6800000000000003e-06, |
|
"loss": 0.5112, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 1.0435090201627166, |
|
"grad_norm": 13.095253944396973, |
|
"learning_rate": 3.5133333333333337e-06, |
|
"loss": 0.483, |
|
"step": 1475 |
|
}, |
|
{ |
|
"epoch": 1.0611956137247966, |
|
"grad_norm": 11.897124290466309, |
|
"learning_rate": 3.346666666666667e-06, |
|
"loss": 0.4822, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 1.0788822072868764, |
|
"grad_norm": 15.847147941589355, |
|
"learning_rate": 3.1800000000000005e-06, |
|
"loss": 0.4396, |
|
"step": 1525 |
|
}, |
|
{ |
|
"epoch": 1.0965688008489565, |
|
"grad_norm": 12.42253303527832, |
|
"learning_rate": 3.013333333333334e-06, |
|
"loss": 0.4731, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 1.1142553944110365, |
|
"grad_norm": 13.371386528015137, |
|
"learning_rate": 2.8466666666666672e-06, |
|
"loss": 0.4623, |
|
"step": 1575 |
|
}, |
|
{ |
|
"epoch": 1.1319419879731163, |
|
"grad_norm": 13.145184516906738, |
|
"learning_rate": 2.68e-06, |
|
"loss": 0.4499, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 1.1496285815351963, |
|
"grad_norm": 13.22415542602539, |
|
"learning_rate": 2.5133333333333336e-06, |
|
"loss": 0.4716, |
|
"step": 1625 |
|
}, |
|
{ |
|
"epoch": 1.1673151750972763, |
|
"grad_norm": 10.703633308410645, |
|
"learning_rate": 2.346666666666667e-06, |
|
"loss": 0.4536, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 1.1850017686593561, |
|
"grad_norm": 15.410745620727539, |
|
"learning_rate": 2.1800000000000003e-06, |
|
"loss": 0.4616, |
|
"step": 1675 |
|
}, |
|
{ |
|
"epoch": 1.2026883622214362, |
|
"grad_norm": 14.111817359924316, |
|
"learning_rate": 2.0133333333333337e-06, |
|
"loss": 0.4498, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 1.2203749557835162, |
|
"grad_norm": 12.579985618591309, |
|
"learning_rate": 1.8466666666666668e-06, |
|
"loss": 0.464, |
|
"step": 1725 |
|
}, |
|
{ |
|
"epoch": 1.238061549345596, |
|
"grad_norm": 12.897194862365723, |
|
"learning_rate": 1.6800000000000002e-06, |
|
"loss": 0.4678, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 1.255748142907676, |
|
"grad_norm": 11.709689140319824, |
|
"learning_rate": 1.5133333333333334e-06, |
|
"loss": 0.4221, |
|
"step": 1775 |
|
}, |
|
{ |
|
"epoch": 1.2734347364697558, |
|
"grad_norm": 14.162332534790039, |
|
"learning_rate": 1.3466666666666668e-06, |
|
"loss": 0.4625, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 1.2911213300318358, |
|
"grad_norm": 12.656081199645996, |
|
"learning_rate": 1.1800000000000001e-06, |
|
"loss": 0.4524, |
|
"step": 1825 |
|
}, |
|
{ |
|
"epoch": 1.3088079235939158, |
|
"grad_norm": 12.56009292602539, |
|
"learning_rate": 1.0133333333333333e-06, |
|
"loss": 0.4386, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 1.3264945171559956, |
|
"grad_norm": 11.350845336914062, |
|
"learning_rate": 8.466666666666668e-07, |
|
"loss": 0.4516, |
|
"step": 1875 |
|
}, |
|
{ |
|
"epoch": 1.3441811107180757, |
|
"grad_norm": 11.870651245117188, |
|
"learning_rate": 6.800000000000001e-07, |
|
"loss": 0.4226, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 1.3618677042801557, |
|
"grad_norm": 11.236252784729004, |
|
"learning_rate": 5.133333333333334e-07, |
|
"loss": 0.442, |
|
"step": 1925 |
|
}, |
|
{ |
|
"epoch": 1.3795542978422355, |
|
"grad_norm": 12.496944427490234, |
|
"learning_rate": 3.466666666666667e-07, |
|
"loss": 0.4464, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 1.3972408914043155, |
|
"grad_norm": 11.83437728881836, |
|
"learning_rate": 1.8e-07, |
|
"loss": 0.4552, |
|
"step": 1975 |
|
}, |
|
{ |
|
"epoch": 1.4149274849663955, |
|
"grad_norm": 12.044703483581543, |
|
"learning_rate": 1.3333333333333334e-08, |
|
"loss": 0.4497, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.4149274849663955, |
|
"eval_loss": 0.5812480449676514, |
|
"eval_runtime": 1144.5779, |
|
"eval_samples_per_second": 9.403, |
|
"eval_steps_per_second": 0.588, |
|
"eval_wer": 0.3526167566298991, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.4149274849663955, |
|
"step": 2000, |
|
"total_flos": 4.15045601427456e+18, |
|
"train_loss": 0.6969459619522095, |
|
"train_runtime": 9919.8131, |
|
"train_samples_per_second": 6.452, |
|
"train_steps_per_second": 0.202 |
|
} |
|
], |
|
"logging_steps": 25, |
|
"max_steps": 2000, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 2, |
|
"save_steps": 1000, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 4.15045601427456e+18, |
|
"train_batch_size": 16, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|