|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.020415662896574253, |
|
"eval_steps": 63, |
|
"global_step": 250, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 8.166265158629701e-05, |
|
"eval_loss": 10.840862274169922, |
|
"eval_runtime": 7.8991, |
|
"eval_samples_per_second": 326.366, |
|
"eval_steps_per_second": 163.183, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0008166265158629701, |
|
"grad_norm": 0.3761455714702606, |
|
"learning_rate": 0.00019979453927503364, |
|
"loss": 10.8359, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.0016332530317259401, |
|
"grad_norm": 0.394729346036911, |
|
"learning_rate": 0.00019815591569910654, |
|
"loss": 10.815, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.0024498795475889103, |
|
"grad_norm": 0.30855050683021545, |
|
"learning_rate": 0.00019490557470106686, |
|
"loss": 10.7786, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.0032665060634518802, |
|
"grad_norm": 0.42155948281288147, |
|
"learning_rate": 0.0001900968867902419, |
|
"loss": 10.7552, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.00408313257931485, |
|
"grad_norm": 0.2492990493774414, |
|
"learning_rate": 0.00018380881048918405, |
|
"loss": 10.7528, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.0048997590951778205, |
|
"grad_norm": 0.2550627887248993, |
|
"learning_rate": 0.00017614459583691346, |
|
"loss": 10.7484, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.005144747049936712, |
|
"eval_loss": 10.749500274658203, |
|
"eval_runtime": 7.3007, |
|
"eval_samples_per_second": 353.116, |
|
"eval_steps_per_second": 176.558, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.005716385611040791, |
|
"grad_norm": 0.2560204565525055, |
|
"learning_rate": 0.0001672300890261317, |
|
"loss": 10.7501, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.0065330121269037604, |
|
"grad_norm": 0.189836785197258, |
|
"learning_rate": 0.00015721166601221698, |
|
"loss": 10.7455, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.007349638642766731, |
|
"grad_norm": 0.23742568492889404, |
|
"learning_rate": 0.00014625382902408356, |
|
"loss": 10.7458, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.0081662651586297, |
|
"grad_norm": 0.2744472026824951, |
|
"learning_rate": 0.00013453650544213076, |
|
"loss": 10.7482, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.008982891674492672, |
|
"grad_norm": 0.3114999532699585, |
|
"learning_rate": 0.00012225209339563145, |
|
"loss": 10.7417, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.009799518190355641, |
|
"grad_norm": 0.24193626642227173, |
|
"learning_rate": 0.00010960230259076818, |
|
"loss": 10.7392, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.010289494099873423, |
|
"eval_loss": 10.737412452697754, |
|
"eval_runtime": 7.1792, |
|
"eval_samples_per_second": 359.091, |
|
"eval_steps_per_second": 179.545, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.01061614470621861, |
|
"grad_norm": 0.310713529586792, |
|
"learning_rate": 9.679484224283449e-05, |
|
"loss": 10.7422, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.011432771222081582, |
|
"grad_norm": 0.2092239260673523, |
|
"learning_rate": 8.404001049666211e-05, |
|
"loss": 10.7337, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.012249397737944551, |
|
"grad_norm": 0.19805459678173065, |
|
"learning_rate": 7.154724133689677e-05, |
|
"loss": 10.7418, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.013066024253807521, |
|
"grad_norm": 0.179554283618927, |
|
"learning_rate": 5.952166568776062e-05, |
|
"loss": 10.7348, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.01388265076967049, |
|
"grad_norm": 0.22178767621517181, |
|
"learning_rate": 4.8160743168947496e-05, |
|
"loss": 10.7329, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.014699277285533462, |
|
"grad_norm": 0.23290759325027466, |
|
"learning_rate": 3.7651019814126654e-05, |
|
"loss": 10.7324, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.015434241149810134, |
|
"eval_loss": 10.731443405151367, |
|
"eval_runtime": 7.1741, |
|
"eval_samples_per_second": 359.346, |
|
"eval_steps_per_second": 179.673, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.015515903801396431, |
|
"grad_norm": 0.2047327607870102, |
|
"learning_rate": 2.8165064990227252e-05, |
|
"loss": 10.7352, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.0163325303172594, |
|
"grad_norm": 0.20196911692619324, |
|
"learning_rate": 1.985863781320435e-05, |
|
"loss": 10.7285, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.017149156833122372, |
|
"grad_norm": 0.21181835234165192, |
|
"learning_rate": 1.286812958766106e-05, |
|
"loss": 10.7256, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.017965783348985343, |
|
"grad_norm": 0.23975443840026855, |
|
"learning_rate": 7.308324265397836e-06, |
|
"loss": 10.7394, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.01878240986484831, |
|
"grad_norm": 0.21740056574344635, |
|
"learning_rate": 3.270513696097055e-06, |
|
"loss": 10.7338, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.019599036380711282, |
|
"grad_norm": 0.2698234021663666, |
|
"learning_rate": 8.209986176753948e-07, |
|
"loss": 10.7245, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.020415662896574253, |
|
"grad_norm": 0.2873691916465759, |
|
"learning_rate": 0.0, |
|
"loss": 10.729, |
|
"step": 250 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 250, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 63, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 16031529566208.0, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|