|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 11.0, |
|
"eval_steps": 500, |
|
"global_step": 3498, |
|
"is_hyper_param_search": true, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.9968553459119497, |
|
"grad_norm": 0.429341197013855, |
|
"learning_rate": 0.00013794880138427262, |
|
"loss": 0.3323, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_accuracy": 0.8664516129032258, |
|
"eval_loss": 0.07744947820901871, |
|
"eval_runtime": 2.7326, |
|
"eval_samples_per_second": 1134.466, |
|
"eval_steps_per_second": 23.787, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 1.9937106918238994, |
|
"grad_norm": 0.2267828732728958, |
|
"learning_rate": 0.00014653290645786366, |
|
"loss": 0.0632, |
|
"step": 634 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_accuracy": 0.9290322580645162, |
|
"eval_loss": 0.034380555152893066, |
|
"eval_runtime": 2.6579, |
|
"eval_samples_per_second": 1166.326, |
|
"eval_steps_per_second": 24.455, |
|
"step": 636 |
|
}, |
|
{ |
|
"epoch": 2.990566037735849, |
|
"grad_norm": 0.17225514352321625, |
|
"learning_rate": 0.00013031400584782778, |
|
"loss": 0.0332, |
|
"step": 951 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_accuracy": 0.937741935483871, |
|
"eval_loss": 0.025218434631824493, |
|
"eval_runtime": 2.6227, |
|
"eval_samples_per_second": 1181.985, |
|
"eval_steps_per_second": 24.784, |
|
"step": 954 |
|
}, |
|
{ |
|
"epoch": 3.9874213836477987, |
|
"grad_norm": 0.10644444078207016, |
|
"learning_rate": 0.0001140951052377919, |
|
"loss": 0.0254, |
|
"step": 1268 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_accuracy": 0.9448387096774193, |
|
"eval_loss": 0.02074083872139454, |
|
"eval_runtime": 2.6543, |
|
"eval_samples_per_second": 1167.909, |
|
"eval_steps_per_second": 24.488, |
|
"step": 1272 |
|
}, |
|
{ |
|
"epoch": 4.984276729559748, |
|
"grad_norm": 0.08869536221027374, |
|
"learning_rate": 9.787620462775599e-05, |
|
"loss": 0.0223, |
|
"step": 1585 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_accuracy": 0.9412903225806452, |
|
"eval_loss": 0.020231781527400017, |
|
"eval_runtime": 2.6557, |
|
"eval_samples_per_second": 1167.32, |
|
"eval_steps_per_second": 24.476, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 5.981132075471698, |
|
"grad_norm": 0.08641496300697327, |
|
"learning_rate": 8.16573040177201e-05, |
|
"loss": 0.0201, |
|
"step": 1902 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_accuracy": 0.9441935483870968, |
|
"eval_loss": 0.01878916658461094, |
|
"eval_runtime": 2.635, |
|
"eval_samples_per_second": 1176.481, |
|
"eval_steps_per_second": 24.668, |
|
"step": 1908 |
|
}, |
|
{ |
|
"epoch": 6.977987421383648, |
|
"grad_norm": 0.07254982739686966, |
|
"learning_rate": 6.543840340768423e-05, |
|
"loss": 0.0183, |
|
"step": 2219 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"eval_accuracy": 0.9467741935483871, |
|
"eval_loss": 0.017883770167827606, |
|
"eval_runtime": 2.651, |
|
"eval_samples_per_second": 1169.38, |
|
"eval_steps_per_second": 24.519, |
|
"step": 2226 |
|
}, |
|
{ |
|
"epoch": 7.9748427672955975, |
|
"grad_norm": 0.0713103711605072, |
|
"learning_rate": 4.9219502797648335e-05, |
|
"loss": 0.0171, |
|
"step": 2536 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_accuracy": 0.9493548387096774, |
|
"eval_loss": 0.017295770347118378, |
|
"eval_runtime": 2.6543, |
|
"eval_samples_per_second": 1167.916, |
|
"eval_steps_per_second": 24.489, |
|
"step": 2544 |
|
}, |
|
{ |
|
"epoch": 8.971698113207546, |
|
"grad_norm": 0.0737316906452179, |
|
"learning_rate": 3.3000602187612454e-05, |
|
"loss": 0.0162, |
|
"step": 2853 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"eval_accuracy": 0.9480645161290323, |
|
"eval_loss": 0.01655430532991886, |
|
"eval_runtime": 2.6551, |
|
"eval_samples_per_second": 1167.55, |
|
"eval_steps_per_second": 24.481, |
|
"step": 2862 |
|
}, |
|
{ |
|
"epoch": 9.968553459119496, |
|
"grad_norm": 0.06824547797441483, |
|
"learning_rate": 1.6781701577576563e-05, |
|
"loss": 0.0155, |
|
"step": 3170 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_accuracy": 0.947741935483871, |
|
"eval_loss": 0.016286808997392654, |
|
"eval_runtime": 2.6473, |
|
"eval_samples_per_second": 1170.992, |
|
"eval_steps_per_second": 24.553, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 10.965408805031446, |
|
"grad_norm": 0.058784905821084976, |
|
"learning_rate": 5.628009675406775e-07, |
|
"loss": 0.0149, |
|
"step": 3487 |
|
} |
|
], |
|
"logging_steps": 317, |
|
"max_steps": 3498, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 11, |
|
"save_steps": 1000000000.0, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 907183422801936.0, |
|
"train_batch_size": 48, |
|
"trial_name": null, |
|
"trial_params": { |
|
"alpha": 0.5336013283534605, |
|
"fp16": true, |
|
"learning_rate": 0.00016014245712748368, |
|
"lr_scheduler": "cosine", |
|
"num_train_epochs": 11, |
|
"temperature": 9, |
|
"warmup_steps": 368, |
|
"weight_decay": 0.06096677657018104 |
|
} |
|
} |
|
|