|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 1.8382916143264194, |
|
"global_step": 60000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 5.999999999999999e-06, |
|
"loss": 0.9345, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 1.1999999999999999e-05, |
|
"loss": 0.7514, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"eval_loss": 0.9052248597145081, |
|
"eval_runtime": 0.5073, |
|
"eval_samples_per_second": 1971.284, |
|
"eval_steps_per_second": 31.541, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.7999999999999997e-05, |
|
"loss": 0.7408, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 2.3999999999999997e-05, |
|
"loss": 0.74, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"eval_loss": 0.9058064818382263, |
|
"eval_runtime": 0.5123, |
|
"eval_samples_per_second": 1952.096, |
|
"eval_steps_per_second": 31.234, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 2.9999999999999997e-05, |
|
"loss": 0.7398, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 3.5999999999999994e-05, |
|
"loss": 0.7395, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"eval_loss": 0.9068936705589294, |
|
"eval_runtime": 0.499, |
|
"eval_samples_per_second": 2003.855, |
|
"eval_steps_per_second": 32.062, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 4.2e-05, |
|
"loss": 0.7394, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 4.7999999999999994e-05, |
|
"loss": 0.7392, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"eval_loss": 0.9032273888587952, |
|
"eval_runtime": 0.5086, |
|
"eval_samples_per_second": 1966.372, |
|
"eval_steps_per_second": 31.462, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 5.399999999999999e-05, |
|
"loss": 0.7389, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 5.9999999999999995e-05, |
|
"loss": 0.7386, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"eval_loss": 0.9034351706504822, |
|
"eval_runtime": 0.5101, |
|
"eval_samples_per_second": 1960.24, |
|
"eval_steps_per_second": 31.364, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 6.599999999999999e-05, |
|
"loss": 0.7382, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 7.199999999999999e-05, |
|
"loss": 0.7377, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"eval_loss": 0.8661078810691833, |
|
"eval_runtime": 0.5119, |
|
"eval_samples_per_second": 1953.649, |
|
"eval_steps_per_second": 31.258, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 7.8e-05, |
|
"loss": 0.7375, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 8.4e-05, |
|
"loss": 0.7373, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"eval_loss": 0.8658460974693298, |
|
"eval_runtime": 0.5198, |
|
"eval_samples_per_second": 1923.698, |
|
"eval_steps_per_second": 30.779, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 8.999999999999999e-05, |
|
"loss": 0.7367, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.599999999999999e-05, |
|
"loss": 0.7207, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"eval_loss": 0.8675529956817627, |
|
"eval_runtime": 0.5261, |
|
"eval_samples_per_second": 1900.917, |
|
"eval_steps_per_second": 30.415, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.000102, |
|
"loss": 0.6905, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.00010799999999999998, |
|
"loss": 0.6746, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"eval_loss": 0.8773286938667297, |
|
"eval_runtime": 0.5159, |
|
"eval_samples_per_second": 1938.548, |
|
"eval_steps_per_second": 31.017, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.00011399999999999999, |
|
"loss": 0.6617, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00011999999999999999, |
|
"loss": 0.6406, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"eval_loss": 0.8803548812866211, |
|
"eval_runtime": 0.5084, |
|
"eval_samples_per_second": 1966.828, |
|
"eval_steps_per_second": 31.469, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.00012599999999999997, |
|
"loss": 0.6275, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.00013199999999999998, |
|
"loss": 0.6079, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"eval_loss": 0.8707549571990967, |
|
"eval_runtime": 0.5111, |
|
"eval_samples_per_second": 1956.647, |
|
"eval_steps_per_second": 31.306, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.000138, |
|
"loss": 0.5977, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.00014399999999999998, |
|
"loss": 0.5907, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"eval_loss": 0.8676372766494751, |
|
"eval_runtime": 0.5179, |
|
"eval_samples_per_second": 1930.992, |
|
"eval_steps_per_second": 30.896, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.00015, |
|
"loss": 0.5834, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.000156, |
|
"loss": 0.5764, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"eval_loss": 0.8676320910453796, |
|
"eval_runtime": 0.5238, |
|
"eval_samples_per_second": 1909.011, |
|
"eval_steps_per_second": 30.544, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.000162, |
|
"loss": 0.5696, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.000168, |
|
"loss": 0.5648, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"eval_loss": 0.860458254814148, |
|
"eval_runtime": 0.5396, |
|
"eval_samples_per_second": 1853.225, |
|
"eval_steps_per_second": 29.652, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.00017399999999999997, |
|
"loss": 0.5602, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00017999999999999998, |
|
"loss": 0.556, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"eval_loss": 0.8584501147270203, |
|
"eval_runtime": 0.5323, |
|
"eval_samples_per_second": 1878.593, |
|
"eval_steps_per_second": 30.057, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.000186, |
|
"loss": 0.5519, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00019199999999999998, |
|
"loss": 0.5459, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"eval_loss": 0.8554501533508301, |
|
"eval_runtime": 0.5107, |
|
"eval_samples_per_second": 1957.988, |
|
"eval_steps_per_second": 31.328, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.000198, |
|
"loss": 0.5412, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.000204, |
|
"loss": 0.5374, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"eval_loss": 0.8554509282112122, |
|
"eval_runtime": 0.5145, |
|
"eval_samples_per_second": 1943.819, |
|
"eval_steps_per_second": 31.101, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.00020999999999999998, |
|
"loss": 0.5338, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00021599999999999996, |
|
"loss": 0.5305, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"eval_loss": 0.8557196259498596, |
|
"eval_runtime": 0.5163, |
|
"eval_samples_per_second": 1936.818, |
|
"eval_steps_per_second": 30.989, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 0.00022199999999999998, |
|
"loss": 0.5273, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 0.00022799999999999999, |
|
"loss": 0.5243, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"eval_loss": 0.8606428503990173, |
|
"eval_runtime": 0.5087, |
|
"eval_samples_per_second": 1965.911, |
|
"eval_steps_per_second": 31.455, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 0.000234, |
|
"loss": 0.5215, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 0.00023999999999999998, |
|
"loss": 0.5188, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"eval_loss": 0.8626545071601868, |
|
"eval_runtime": 0.4996, |
|
"eval_samples_per_second": 2001.549, |
|
"eval_steps_per_second": 32.025, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 0.00024599999999999996, |
|
"loss": 0.516, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 0.00025199999999999995, |
|
"loss": 0.5136, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"eval_loss": 0.8466567993164062, |
|
"eval_runtime": 0.521, |
|
"eval_samples_per_second": 1919.499, |
|
"eval_steps_per_second": 30.712, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 0.000258, |
|
"loss": 0.5112, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 0.00026399999999999997, |
|
"loss": 0.5089, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"eval_loss": 0.8465701937675476, |
|
"eval_runtime": 0.5015, |
|
"eval_samples_per_second": 1994.167, |
|
"eval_steps_per_second": 31.907, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 0.00027, |
|
"loss": 0.5067, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 0.000276, |
|
"loss": 0.5048, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"eval_loss": 0.8367106914520264, |
|
"eval_runtime": 0.5157, |
|
"eval_samples_per_second": 1939.06, |
|
"eval_steps_per_second": 31.025, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 0.00028199999999999997, |
|
"loss": 0.5029, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 0.00028799999999999995, |
|
"loss": 0.5009, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"eval_loss": 0.8282895088195801, |
|
"eval_runtime": 0.5191, |
|
"eval_samples_per_second": 1926.314, |
|
"eval_steps_per_second": 30.821, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 0.000294, |
|
"loss": 0.4986, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 0.0003, |
|
"loss": 0.4968, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"eval_loss": 0.8393005132675171, |
|
"eval_runtime": 0.5106, |
|
"eval_samples_per_second": 1958.371, |
|
"eval_steps_per_second": 31.334, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 0.00029999920715161553, |
|
"loss": 0.495, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 0.0002999968286151326, |
|
"loss": 0.4932, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"eval_loss": 0.8276752829551697, |
|
"eval_runtime": 0.5157, |
|
"eval_samples_per_second": 1939.179, |
|
"eval_steps_per_second": 31.027, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 0.0002999928644165624, |
|
"loss": 0.4916, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 0.0002999873145992569, |
|
"loss": 0.4899, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"eval_loss": 0.8151644468307495, |
|
"eval_runtime": 0.5093, |
|
"eval_samples_per_second": 1963.529, |
|
"eval_steps_per_second": 31.416, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 0.000299980179223908, |
|
"loss": 0.4883, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 0.0002999714583685469, |
|
"loss": 0.4868, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"eval_loss": 0.8103233575820923, |
|
"eval_runtime": 0.5126, |
|
"eval_samples_per_second": 1950.786, |
|
"eval_steps_per_second": 31.213, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 0.00029996115212854366, |
|
"loss": 0.4855, |
|
"step": 28500 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 0.00029994926061660554, |
|
"loss": 0.4841, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"eval_loss": 0.8084909319877625, |
|
"eval_runtime": 0.5298, |
|
"eval_samples_per_second": 1887.571, |
|
"eval_steps_per_second": 30.201, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 0.0002999357839627762, |
|
"loss": 0.4823, |
|
"step": 29500 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 0.00029992072231443425, |
|
"loss": 0.4806, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"eval_loss": 0.8131279945373535, |
|
"eval_runtime": 0.5066, |
|
"eval_samples_per_second": 1974.102, |
|
"eval_steps_per_second": 31.586, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 0.0002999040758362914, |
|
"loss": 0.4791, |
|
"step": 30500 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 0.00029988584471039094, |
|
"loss": 0.4776, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"eval_loss": 0.8034030795097351, |
|
"eval_runtime": 0.4996, |
|
"eval_samples_per_second": 2001.487, |
|
"eval_steps_per_second": 32.024, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 0.0002998660291361054, |
|
"loss": 0.476, |
|
"step": 31500 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 0.0002998446293301349, |
|
"loss": 0.4741, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"eval_loss": 0.8003847599029541, |
|
"eval_runtime": 0.5033, |
|
"eval_samples_per_second": 1986.864, |
|
"eval_steps_per_second": 31.79, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 0.0002998216455265042, |
|
"loss": 0.4724, |
|
"step": 32500 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 0.00029979707797656046, |
|
"loss": 0.4709, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"eval_loss": 0.7997156977653503, |
|
"eval_runtime": 0.5174, |
|
"eval_samples_per_second": 1932.776, |
|
"eval_steps_per_second": 30.924, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 0.00029977092694897053, |
|
"loss": 0.4691, |
|
"step": 33500 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 0.0002997431927297178, |
|
"loss": 0.467, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"eval_loss": 0.803820013999939, |
|
"eval_runtime": 0.5367, |
|
"eval_samples_per_second": 1863.102, |
|
"eval_steps_per_second": 29.81, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 0.00029971387562209936, |
|
"loss": 0.4653, |
|
"step": 34500 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 0.00029968297594672226, |
|
"loss": 0.4638, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"eval_loss": 0.8011023998260498, |
|
"eval_runtime": 0.5429, |
|
"eval_samples_per_second": 1842.037, |
|
"eval_steps_per_second": 29.473, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 0.0002996504940415005, |
|
"loss": 0.462, |
|
"step": 35500 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 0.00029961643026165096, |
|
"loss": 0.4601, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"eval_loss": 0.8029471635818481, |
|
"eval_runtime": 0.5067, |
|
"eval_samples_per_second": 1973.42, |
|
"eval_steps_per_second": 31.575, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 0.00029958078497968973, |
|
"loss": 0.4587, |
|
"step": 36500 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 0.0002995435585854278, |
|
"loss": 0.457, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"eval_loss": 0.7997317910194397, |
|
"eval_runtime": 0.5216, |
|
"eval_samples_per_second": 1917.135, |
|
"eval_steps_per_second": 30.674, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 0.0002995047514859671, |
|
"loss": 0.4556, |
|
"step": 37500 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 0.0002994643641056959, |
|
"loss": 0.4541, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"eval_loss": 0.8004159927368164, |
|
"eval_runtime": 0.5088, |
|
"eval_samples_per_second": 1965.38, |
|
"eval_steps_per_second": 31.446, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 0.000299422396886284, |
|
"loss": 0.4524, |
|
"step": 38500 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 0.0002993788502866783, |
|
"loss": 0.4514, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"eval_loss": 0.797309160232544, |
|
"eval_runtime": 0.5211, |
|
"eval_samples_per_second": 1918.834, |
|
"eval_steps_per_second": 30.701, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 0.00029933372478309746, |
|
"loss": 0.45, |
|
"step": 39500 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 0.00029928702086902664, |
|
"loss": 0.4485, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"eval_loss": 0.7876560688018799, |
|
"eval_runtime": 0.5381, |
|
"eval_samples_per_second": 1858.547, |
|
"eval_steps_per_second": 29.737, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 0.00029923873905521244, |
|
"loss": 0.4471, |
|
"step": 40500 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 0.000299188879869657, |
|
"loss": 0.446, |
|
"step": 41000 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"eval_loss": 0.7949715852737427, |
|
"eval_runtime": 0.5157, |
|
"eval_samples_per_second": 1939.022, |
|
"eval_steps_per_second": 31.024, |
|
"step": 41000 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 0.00029913744385761244, |
|
"loss": 0.4446, |
|
"step": 41500 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 0.00029908443158157465, |
|
"loss": 0.4437, |
|
"step": 42000 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"eval_loss": 0.7941656112670898, |
|
"eval_runtime": 0.5261, |
|
"eval_samples_per_second": 1900.807, |
|
"eval_steps_per_second": 30.413, |
|
"step": 42000 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 0.0002990298436212775, |
|
"loss": 0.4422, |
|
"step": 42500 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 0.0002989736805736861, |
|
"loss": 0.4413, |
|
"step": 43000 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"eval_loss": 0.7861095666885376, |
|
"eval_runtime": 0.5221, |
|
"eval_samples_per_second": 1915.408, |
|
"eval_steps_per_second": 30.647, |
|
"step": 43000 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 0.00029891594305299065, |
|
"loss": 0.44, |
|
"step": 43500 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 0.00029885663169059926, |
|
"loss": 0.4391, |
|
"step": 44000 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"eval_loss": 0.7935870289802551, |
|
"eval_runtime": 0.5329, |
|
"eval_samples_per_second": 1876.586, |
|
"eval_steps_per_second": 30.025, |
|
"step": 44000 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 0.0002987957471351316, |
|
"loss": 0.4378, |
|
"step": 44500 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 0.00029873329005241137, |
|
"loss": 0.4366, |
|
"step": 45000 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"eval_loss": 0.7878534197807312, |
|
"eval_runtime": 0.5149, |
|
"eval_samples_per_second": 1941.994, |
|
"eval_steps_per_second": 31.072, |
|
"step": 45000 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 0.00029866926112545925, |
|
"loss": 0.4355, |
|
"step": 45500 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 0.00029860366105448534, |
|
"loss": 0.4342, |
|
"step": 46000 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"eval_loss": 0.7861126661300659, |
|
"eval_runtime": 0.5143, |
|
"eval_samples_per_second": 1944.264, |
|
"eval_steps_per_second": 31.108, |
|
"step": 46000 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 0.00029853649055688143, |
|
"loss": 0.4333, |
|
"step": 46500 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 0.00029846775036721337, |
|
"loss": 0.432, |
|
"step": 47000 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"eval_loss": 0.7865832448005676, |
|
"eval_runtime": 0.5183, |
|
"eval_samples_per_second": 1929.412, |
|
"eval_steps_per_second": 30.871, |
|
"step": 47000 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 0.0002983974412372129, |
|
"loss": 0.4308, |
|
"step": 47500 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 0.00029832556393576934, |
|
"loss": 0.4297, |
|
"step": 48000 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"eval_loss": 0.777747392654419, |
|
"eval_runtime": 0.527, |
|
"eval_samples_per_second": 1897.653, |
|
"eval_steps_per_second": 30.362, |
|
"step": 48000 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 0.0002982521192489214, |
|
"loss": 0.4288, |
|
"step": 48500 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 0.0002981771079798483, |
|
"loss": 0.4278, |
|
"step": 49000 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"eval_loss": 0.7763716578483582, |
|
"eval_runtime": 0.5211, |
|
"eval_samples_per_second": 1919.193, |
|
"eval_steps_per_second": 30.707, |
|
"step": 49000 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 0.00029810053094886136, |
|
"loss": 0.4269, |
|
"step": 49500 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 0.00029802238899339473, |
|
"loss": 0.4258, |
|
"step": 50000 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"eval_loss": 0.7902368307113647, |
|
"eval_runtime": 0.5206, |
|
"eval_samples_per_second": 1920.725, |
|
"eval_steps_per_second": 30.732, |
|
"step": 50000 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 0.0002979426829679962, |
|
"loss": 0.4247, |
|
"step": 50500 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 0.0002978614137443183, |
|
"loss": 0.424, |
|
"step": 51000 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"eval_loss": 0.7871042490005493, |
|
"eval_runtime": 0.5261, |
|
"eval_samples_per_second": 1900.754, |
|
"eval_steps_per_second": 30.412, |
|
"step": 51000 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 0.000297778582211108, |
|
"loss": 0.4232, |
|
"step": 51500 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 0.00029769418927419786, |
|
"loss": 0.422, |
|
"step": 52000 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"eval_loss": 0.7916954159736633, |
|
"eval_runtime": 0.5311, |
|
"eval_samples_per_second": 1882.884, |
|
"eval_steps_per_second": 30.126, |
|
"step": 52000 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 0.0002976082358564954, |
|
"loss": 0.4211, |
|
"step": 52500 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 0.00029752072289797353, |
|
"loss": 0.4202, |
|
"step": 53000 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"eval_loss": 0.7859019637107849, |
|
"eval_runtime": 0.5358, |
|
"eval_samples_per_second": 1866.426, |
|
"eval_steps_per_second": 29.863, |
|
"step": 53000 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 0.00029743165135565986, |
|
"loss": 0.4194, |
|
"step": 53500 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 0.00029734102220362654, |
|
"loss": 0.4184, |
|
"step": 54000 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"eval_loss": 0.7869405150413513, |
|
"eval_runtime": 0.5363, |
|
"eval_samples_per_second": 1864.609, |
|
"eval_steps_per_second": 29.834, |
|
"step": 54000 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 0.00029724883643297937, |
|
"loss": 0.4177, |
|
"step": 54500 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 0.0002971550950518473, |
|
"loss": 0.4165, |
|
"step": 55000 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"eval_loss": 0.7873055338859558, |
|
"eval_runtime": 0.5249, |
|
"eval_samples_per_second": 1905.112, |
|
"eval_steps_per_second": 30.482, |
|
"step": 55000 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 0.000297059799085371, |
|
"loss": 0.4156, |
|
"step": 55500 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 0.00029696294957569196, |
|
"loss": 0.4149, |
|
"step": 56000 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"eval_loss": 0.7851352095603943, |
|
"eval_runtime": 0.5685, |
|
"eval_samples_per_second": 1758.943, |
|
"eval_steps_per_second": 28.143, |
|
"step": 56000 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 0.00029686454758194076, |
|
"loss": 0.4141, |
|
"step": 56500 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 0.00029676459418022594, |
|
"loss": 0.4136, |
|
"step": 57000 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"eval_loss": 0.7847647666931152, |
|
"eval_runtime": 0.5348, |
|
"eval_samples_per_second": 1869.958, |
|
"eval_steps_per_second": 29.919, |
|
"step": 57000 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 0.0002966630904636219, |
|
"loss": 0.4126, |
|
"step": 57500 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 0.0002965600375421569, |
|
"loss": 0.4119, |
|
"step": 58000 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"eval_loss": 0.7846025228500366, |
|
"eval_runtime": 0.5441, |
|
"eval_samples_per_second": 1837.795, |
|
"eval_steps_per_second": 29.405, |
|
"step": 58000 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 0.0002964554365428013, |
|
"loss": 0.4107, |
|
"step": 58500 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 0.00029634928860945486, |
|
"loss": 0.4103, |
|
"step": 59000 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"eval_loss": 0.7957924604415894, |
|
"eval_runtime": 0.5171, |
|
"eval_samples_per_second": 1934.049, |
|
"eval_steps_per_second": 30.945, |
|
"step": 59000 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 0.0002962415949029343, |
|
"loss": 0.4099, |
|
"step": 59500 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 0.00029613235660096084, |
|
"loss": 0.409, |
|
"step": 60000 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"eval_loss": 0.7877106070518494, |
|
"eval_runtime": 0.5384, |
|
"eval_samples_per_second": 1857.26, |
|
"eval_steps_per_second": 29.716, |
|
"step": 60000 |
|
} |
|
], |
|
"max_steps": 500000, |
|
"num_train_epochs": 16, |
|
"total_flos": 1.916914844695024e+21, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|