|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 1.0940170940170941, |
|
"eval_steps": 100, |
|
"global_step": 1400, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.07814407814407814, |
|
"grad_norm": 0.389704167842865, |
|
"learning_rate": 4.000000000000001e-06, |
|
"loss": 2.4124, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.07814407814407814, |
|
"eval_loss": 2.2181031703948975, |
|
"eval_runtime": 3367.0243, |
|
"eval_samples_per_second": 0.75, |
|
"eval_steps_per_second": 0.094, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.1562881562881563, |
|
"grad_norm": 0.27927058935165405, |
|
"learning_rate": 8.000000000000001e-06, |
|
"loss": 2.2928, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.1562881562881563, |
|
"eval_loss": 2.0591659545898438, |
|
"eval_runtime": 3366.419, |
|
"eval_samples_per_second": 0.75, |
|
"eval_steps_per_second": 0.094, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.23443223443223443, |
|
"grad_norm": 0.22626566886901855, |
|
"learning_rate": 1.2e-05, |
|
"loss": 2.1104, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.23443223443223443, |
|
"eval_loss": 1.8923490047454834, |
|
"eval_runtime": 3360.9485, |
|
"eval_samples_per_second": 0.752, |
|
"eval_steps_per_second": 0.094, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.3125763125763126, |
|
"grad_norm": 0.210145503282547, |
|
"learning_rate": 1.6000000000000003e-05, |
|
"loss": 1.9296, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.3125763125763126, |
|
"eval_loss": 1.7505871057510376, |
|
"eval_runtime": 3357.6413, |
|
"eval_samples_per_second": 0.752, |
|
"eval_steps_per_second": 0.094, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.3907203907203907, |
|
"grad_norm": 0.2200106978416443, |
|
"learning_rate": 2e-05, |
|
"loss": 1.8345, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.3907203907203907, |
|
"eval_loss": 1.7071024179458618, |
|
"eval_runtime": 3361.7985, |
|
"eval_samples_per_second": 0.751, |
|
"eval_steps_per_second": 0.094, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.46886446886446886, |
|
"grad_norm": 0.22381120920181274, |
|
"learning_rate": 1.9980028422948323e-05, |
|
"loss": 1.8103, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.46886446886446886, |
|
"eval_loss": 1.6838117837905884, |
|
"eval_runtime": 3360.7216, |
|
"eval_samples_per_second": 0.752, |
|
"eval_steps_per_second": 0.094, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.5470085470085471, |
|
"grad_norm": 0.2300158590078354, |
|
"learning_rate": 1.9920193464571277e-05, |
|
"loss": 1.7827, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.5470085470085471, |
|
"eval_loss": 1.6691471338272095, |
|
"eval_runtime": 3362.2, |
|
"eval_samples_per_second": 0.751, |
|
"eval_steps_per_second": 0.094, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.6251526251526252, |
|
"grad_norm": 0.25236302614212036, |
|
"learning_rate": 1.982073412456518e-05, |
|
"loss": 1.765, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.6251526251526252, |
|
"eval_loss": 1.6575521230697632, |
|
"eval_runtime": 3363.3293, |
|
"eval_samples_per_second": 0.751, |
|
"eval_steps_per_second": 0.094, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.7032967032967034, |
|
"grad_norm": 0.26079344749450684, |
|
"learning_rate": 1.9682047674904527e-05, |
|
"loss": 1.7425, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.7032967032967034, |
|
"eval_loss": 1.6489633321762085, |
|
"eval_runtime": 3359.1797, |
|
"eval_samples_per_second": 0.752, |
|
"eval_steps_per_second": 0.094, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.7814407814407814, |
|
"grad_norm": 0.2565068304538727, |
|
"learning_rate": 1.9504688073012397e-05, |
|
"loss": 1.7521, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.7814407814407814, |
|
"eval_loss": 1.6409879922866821, |
|
"eval_runtime": 3358.9412, |
|
"eval_samples_per_second": 0.752, |
|
"eval_steps_per_second": 0.094, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.8595848595848596, |
|
"grad_norm": 0.26189491152763367, |
|
"learning_rate": 1.9289363749079798e-05, |
|
"loss": 1.7326, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.8595848595848596, |
|
"eval_loss": 1.6336146593093872, |
|
"eval_runtime": 3363.4368, |
|
"eval_samples_per_second": 0.751, |
|
"eval_steps_per_second": 0.094, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.9377289377289377, |
|
"grad_norm": 0.2778639495372772, |
|
"learning_rate": 1.903693477637204e-05, |
|
"loss": 1.7298, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.9377289377289377, |
|
"eval_loss": 1.627331256866455, |
|
"eval_runtime": 3361.7893, |
|
"eval_samples_per_second": 0.751, |
|
"eval_steps_per_second": 0.094, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 1.0158730158730158, |
|
"grad_norm": 0.2954370975494385, |
|
"learning_rate": 1.874840943582482e-05, |
|
"loss": 1.7128, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 1.0158730158730158, |
|
"eval_loss": 1.62191641330719, |
|
"eval_runtime": 3358.2318, |
|
"eval_samples_per_second": 0.752, |
|
"eval_steps_per_second": 0.094, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 1.0940170940170941, |
|
"grad_norm": 0.2985123097896576, |
|
"learning_rate": 1.842494018865216e-05, |
|
"loss": 1.7265, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 1.0940170940170941, |
|
"eval_loss": 1.616808295249939, |
|
"eval_runtime": 3366.219, |
|
"eval_samples_per_second": 0.75, |
|
"eval_steps_per_second": 0.094, |
|
"step": 1400 |
|
} |
|
], |
|
"logging_steps": 100, |
|
"max_steps": 5470, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 5, |
|
"save_steps": 100, |
|
"total_flos": 5.70672180486144e+17, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|