|
{ |
|
"best_metric": 1.42880117893219, |
|
"best_model_checkpoint": "outputs/checkpoint-828", |
|
"epoch": 17.878542510121456, |
|
"eval_steps": 500, |
|
"global_step": 828, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.4318488529014845, |
|
"grad_norm": 0.9647712111473083, |
|
"learning_rate": 2.0000000000000003e-06, |
|
"loss": 2.3464, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.863697705802969, |
|
"grad_norm": 0.7620949149131775, |
|
"learning_rate": 4.000000000000001e-06, |
|
"loss": 2.3457, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.9932523616734144, |
|
"eval_loss": 2.172492504119873, |
|
"eval_runtime": 12.5552, |
|
"eval_samples_per_second": 29.629, |
|
"eval_steps_per_second": 3.743, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 1.2955465587044535, |
|
"grad_norm": 0.7905983924865723, |
|
"learning_rate": 6e-06, |
|
"loss": 2.3256, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 1.7273954116059378, |
|
"grad_norm": 0.8448758721351624, |
|
"learning_rate": 8.000000000000001e-06, |
|
"loss": 2.1999, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 1.9865047233468287, |
|
"eval_loss": 1.9573273658752441, |
|
"eval_runtime": 12.2982, |
|
"eval_samples_per_second": 30.248, |
|
"eval_steps_per_second": 3.822, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 2.1592442645074224, |
|
"grad_norm": 0.665162980556488, |
|
"learning_rate": 1e-05, |
|
"loss": 2.1343, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 2.591093117408907, |
|
"grad_norm": 0.5304886698722839, |
|
"learning_rate": 9.981389099710132e-06, |
|
"loss": 2.0325, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 2.979757085020243, |
|
"eval_loss": 1.8101614713668823, |
|
"eval_runtime": 12.0422, |
|
"eval_samples_per_second": 30.891, |
|
"eval_steps_per_second": 3.903, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 3.0229419703103915, |
|
"grad_norm": 0.6249691247940063, |
|
"learning_rate": 9.925694945084369e-06, |
|
"loss": 2.038, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 3.454790823211876, |
|
"grad_norm": 0.549341082572937, |
|
"learning_rate": 9.833332143466099e-06, |
|
"loss": 1.9539, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 3.8866396761133606, |
|
"grad_norm": 0.5345794558525085, |
|
"learning_rate": 9.704988276811883e-06, |
|
"loss": 1.9295, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 3.9946018893387314, |
|
"eval_loss": 1.7296996116638184, |
|
"eval_runtime": 12.0445, |
|
"eval_samples_per_second": 30.885, |
|
"eval_steps_per_second": 3.902, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 4.318488529014845, |
|
"grad_norm": 0.6139589548110962, |
|
"learning_rate": 9.54161878308377e-06, |
|
"loss": 1.8949, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 4.75033738191633, |
|
"grad_norm": 0.6864662766456604, |
|
"learning_rate": 9.344439843625034e-06, |
|
"loss": 1.8976, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 4.987854251012146, |
|
"eval_loss": 1.6704739332199097, |
|
"eval_runtime": 12.0592, |
|
"eval_samples_per_second": 30.848, |
|
"eval_steps_per_second": 3.897, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 5.182186234817814, |
|
"grad_norm": 0.5886803865432739, |
|
"learning_rate": 9.114919329468283e-06, |
|
"loss": 1.8686, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 5.614035087719298, |
|
"grad_norm": 0.6739407181739807, |
|
"learning_rate": 8.854765873974898e-06, |
|
"loss": 1.8265, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 5.98110661268556, |
|
"eval_loss": 1.6191655397415161, |
|
"eval_runtime": 12.0582, |
|
"eval_samples_per_second": 30.85, |
|
"eval_steps_per_second": 3.898, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 6.045883940620783, |
|
"grad_norm": 0.7047673463821411, |
|
"learning_rate": 8.565916153152982e-06, |
|
"loss": 1.8201, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 6.477732793522267, |
|
"grad_norm": 0.6909105777740479, |
|
"learning_rate": 8.250520468343722e-06, |
|
"loss": 1.7656, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 6.909581646423752, |
|
"grad_norm": 0.7460944056510925, |
|
"learning_rate": 7.910926738603855e-06, |
|
"loss": 1.8038, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 6.995951417004049, |
|
"eval_loss": 1.5762709379196167, |
|
"eval_runtime": 12.0686, |
|
"eval_samples_per_second": 30.824, |
|
"eval_steps_per_second": 3.894, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 7.341430499325236, |
|
"grad_norm": 0.8885225653648376, |
|
"learning_rate": 7.5496630219506805e-06, |
|
"loss": 1.7428, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 7.77327935222672, |
|
"grad_norm": 0.771353542804718, |
|
"learning_rate": 7.169418695587791e-06, |
|
"loss": 1.7608, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 7.989203778677463, |
|
"eval_loss": 1.5418975353240967, |
|
"eval_runtime": 12.0558, |
|
"eval_samples_per_second": 30.857, |
|
"eval_steps_per_second": 3.899, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 8.205128205128204, |
|
"grad_norm": 0.7849822640419006, |
|
"learning_rate": 6.773024435212678e-06, |
|
"loss": 1.7396, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 8.63697705802969, |
|
"grad_norm": 0.997053325176239, |
|
"learning_rate": 6.363431142447469e-06, |
|
"loss": 1.7123, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 8.982456140350877, |
|
"eval_loss": 1.5107336044311523, |
|
"eval_runtime": 12.0487, |
|
"eval_samples_per_second": 30.875, |
|
"eval_steps_per_second": 3.901, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 9.068825910931174, |
|
"grad_norm": 0.8057714104652405, |
|
"learning_rate": 5.943687977264584e-06, |
|
"loss": 1.7143, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 9.50067476383266, |
|
"grad_norm": 1.0519981384277344, |
|
"learning_rate": 5.51691965894185e-06, |
|
"loss": 1.6754, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 9.932523616734143, |
|
"grad_norm": 0.8790378570556641, |
|
"learning_rate": 5.0863032045269435e-06, |
|
"loss": 1.7078, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 9.997300944669366, |
|
"eval_loss": 1.4861911535263062, |
|
"eval_runtime": 12.0616, |
|
"eval_samples_per_second": 30.842, |
|
"eval_steps_per_second": 3.897, |
|
"step": 463 |
|
}, |
|
{ |
|
"epoch": 10.364372469635628, |
|
"grad_norm": 0.9965023994445801, |
|
"learning_rate": 4.6550442779783755e-06, |
|
"loss": 1.6729, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 10.796221322537113, |
|
"grad_norm": 0.9378894567489624, |
|
"learning_rate": 4.226353326048594e-06, |
|
"loss": 1.6375, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 10.99055330634278, |
|
"eval_loss": 1.4666177034378052, |
|
"eval_runtime": 12.0581, |
|
"eval_samples_per_second": 30.851, |
|
"eval_steps_per_second": 3.898, |
|
"step": 509 |
|
}, |
|
{ |
|
"epoch": 11.228070175438596, |
|
"grad_norm": 1.088391900062561, |
|
"learning_rate": 3.803421678562213e-06, |
|
"loss": 1.6817, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 11.65991902834008, |
|
"grad_norm": 1.2648347616195679, |
|
"learning_rate": 3.389397791007548e-06, |
|
"loss": 1.6618, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 11.983805668016194, |
|
"eval_loss": 1.4524654150009155, |
|
"eval_runtime": 12.0523, |
|
"eval_samples_per_second": 30.866, |
|
"eval_steps_per_second": 3.9, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 12.091767881241566, |
|
"grad_norm": 1.1444419622421265, |
|
"learning_rate": 2.9873638063001633e-06, |
|
"loss": 1.637, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 12.523616734143049, |
|
"grad_norm": 0.8730684518814087, |
|
"learning_rate": 2.6003126102010696e-06, |
|
"loss": 1.6219, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 12.955465587044534, |
|
"grad_norm": 1.3365355730056763, |
|
"learning_rate": 2.2311255511973347e-06, |
|
"loss": 1.629, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 12.998650472334683, |
|
"eval_loss": 1.4416956901550293, |
|
"eval_runtime": 12.0642, |
|
"eval_samples_per_second": 30.835, |
|
"eval_steps_per_second": 3.896, |
|
"step": 602 |
|
}, |
|
{ |
|
"epoch": 13.387314439946019, |
|
"grad_norm": 1.1396286487579346, |
|
"learning_rate": 1.8825509907063328e-06, |
|
"loss": 1.5987, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 13.819163292847504, |
|
"grad_norm": 1.1069096326828003, |
|
"learning_rate": 1.557183843283614e-06, |
|
"loss": 1.6634, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 13.991902834008098, |
|
"eval_loss": 1.4354872703552246, |
|
"eval_runtime": 12.0604, |
|
"eval_samples_per_second": 30.845, |
|
"eval_steps_per_second": 3.897, |
|
"step": 648 |
|
}, |
|
{ |
|
"epoch": 14.251012145748987, |
|
"grad_norm": 0.929169237613678, |
|
"learning_rate": 1.257446259144494e-06, |
|
"loss": 1.6245, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 14.682860998650472, |
|
"grad_norm": 1.1899679899215698, |
|
"learning_rate": 9.85569592805588e-07, |
|
"loss": 1.6097, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 14.98515519568151, |
|
"eval_loss": 1.4315454959869385, |
|
"eval_runtime": 12.0693, |
|
"eval_samples_per_second": 30.822, |
|
"eval_steps_per_second": 3.894, |
|
"step": 694 |
|
}, |
|
{ |
|
"epoch": 15.114709851551957, |
|
"grad_norm": 1.0818545818328857, |
|
"learning_rate": 7.435777920782444e-07, |
|
"loss": 1.6349, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 15.54655870445344, |
|
"grad_norm": 1.077989935874939, |
|
"learning_rate": 5.332723310721855e-07, |
|
"loss": 1.6035, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 15.978407557354926, |
|
"grad_norm": 1.1727399826049805, |
|
"learning_rate": 3.5621879937348836e-07, |
|
"loss": 1.6489, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"eval_loss": 1.4295289516448975, |
|
"eval_runtime": 12.0333, |
|
"eval_samples_per_second": 30.914, |
|
"eval_steps_per_second": 3.906, |
|
"step": 741 |
|
}, |
|
{ |
|
"epoch": 16.41025641025641, |
|
"grad_norm": 1.0742014646530151, |
|
"learning_rate": 2.137352472319215e-07, |
|
"loss": 1.5994, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 16.842105263157894, |
|
"grad_norm": 1.0748684406280518, |
|
"learning_rate": 1.0688237352022346e-07, |
|
"loss": 1.6251, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 16.993252361673413, |
|
"eval_loss": 1.4288655519485474, |
|
"eval_runtime": 11.987, |
|
"eval_samples_per_second": 31.034, |
|
"eval_steps_per_second": 3.921, |
|
"step": 787 |
|
}, |
|
{ |
|
"epoch": 17.27395411605938, |
|
"grad_norm": 1.1126863956451416, |
|
"learning_rate": 3.645562950973014e-08, |
|
"loss": 1.6199, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 17.705802968960864, |
|
"grad_norm": 1.1560289859771729, |
|
"learning_rate": 2.9792972446479605e-09, |
|
"loss": 1.5742, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 17.878542510121456, |
|
"eval_loss": 1.42880117893219, |
|
"eval_runtime": 11.9734, |
|
"eval_samples_per_second": 31.069, |
|
"eval_steps_per_second": 3.925, |
|
"step": 828 |
|
} |
|
], |
|
"logging_steps": 20, |
|
"max_steps": 828, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 18, |
|
"save_steps": 500, |
|
"total_flos": 7.361856849884774e+16, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|