|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 3.963109354413702, |
|
"eval_steps": 500, |
|
"global_step": 376, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 476.0, |
|
"learning_rate": 5.263157894736842e-06, |
|
"loss": 13.5076, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 27.125, |
|
"learning_rate": 2.6315789473684212e-05, |
|
"loss": 10.3382, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 7.09375, |
|
"learning_rate": 5.2631578947368424e-05, |
|
"loss": 6.5621, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 5.09375, |
|
"learning_rate": 7.894736842105263e-05, |
|
"loss": 5.4048, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 13.625, |
|
"learning_rate": 0.00010526315789473685, |
|
"loss": 3.8714, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 4.3125, |
|
"learning_rate": 0.00013157894736842108, |
|
"loss": 1.907, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 2.484375, |
|
"learning_rate": 0.00015789473684210527, |
|
"loss": 1.7156, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 1.71875, |
|
"learning_rate": 0.00018421052631578948, |
|
"loss": 1.5905, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 1.4375, |
|
"learning_rate": 0.00019998272238333606, |
|
"loss": 1.4996, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 1.2109375, |
|
"learning_rate": 0.00019978841775475367, |
|
"loss": 1.424, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 2.515625, |
|
"learning_rate": 0.00019937863245275304, |
|
"loss": 1.3526, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 1.6953125, |
|
"learning_rate": 0.0001987542513577122, |
|
"loss": 1.3334, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 2.3125, |
|
"learning_rate": 0.00019791662274292637, |
|
"loss": 1.2863, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 0.91015625, |
|
"learning_rate": 0.00019686755536317945, |
|
"loss": 1.2384, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 0.99609375, |
|
"learning_rate": 0.00019560931454896298, |
|
"loss": 1.2248, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 0.73046875, |
|
"learning_rate": 0.000194144617314777, |
|
"loss": 1.1772, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 0.58203125, |
|
"learning_rate": 0.0001924766264920751, |
|
"loss": 1.1654, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 0.74609375, |
|
"learning_rate": 0.00019060894389952328, |
|
"loss": 1.1382, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 0.78515625, |
|
"learning_rate": 0.000188545602565321, |
|
"loss": 1.1537, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"eval_loss": 1.0988351106643677, |
|
"eval_runtime": 16.9422, |
|
"eval_samples_per_second": 16.822, |
|
"eval_steps_per_second": 2.125, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.65625, |
|
"learning_rate": 0.00018629105801837818, |
|
"loss": 1.1162, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"grad_norm": 0.84375, |
|
"learning_rate": 0.00018385017866715507, |
|
"loss": 0.9838, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"grad_norm": 0.61328125, |
|
"learning_rate": 0.00018122823528693966, |
|
"loss": 0.9895, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"grad_norm": 0.8359375, |
|
"learning_rate": 0.00017843088963826435, |
|
"loss": 0.9588, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"grad_norm": 0.74609375, |
|
"learning_rate": 0.00017546418224103838, |
|
"loss": 0.9304, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"grad_norm": 0.63671875, |
|
"learning_rate": 0.00017233451933079664, |
|
"loss": 0.9315, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"grad_norm": 0.53125, |
|
"learning_rate": 0.00016904865902523096, |
|
"loss": 0.9869, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"grad_norm": 0.796875, |
|
"learning_rate": 0.00016561369673087588, |
|
"loss": 0.9399, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"grad_norm": 0.75390625, |
|
"learning_rate": 0.00016203704982146073, |
|
"loss": 0.9061, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"grad_norm": 2.296875, |
|
"learning_rate": 0.00015832644162101417, |
|
"loss": 0.9537, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"grad_norm": 0.6875, |
|
"learning_rate": 0.00015448988472630654, |
|
"loss": 0.9107, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"grad_norm": 0.61328125, |
|
"learning_rate": 0.00015053566370464415, |
|
"loss": 0.9483, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"grad_norm": 0.7265625, |
|
"learning_rate": 0.00014647231720437686, |
|
"loss": 0.9075, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"grad_norm": 0.53515625, |
|
"learning_rate": 0.00014230861951674913, |
|
"loss": 0.9148, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 0.0001380535616289099, |
|
"loss": 0.9142, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"grad_norm": 0.5703125, |
|
"learning_rate": 0.00013371633180899416, |
|
"loss": 0.8966, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"grad_norm": 0.53515625, |
|
"learning_rate": 0.00012930629576520132, |
|
"loss": 0.8795, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"grad_norm": 0.609375, |
|
"learning_rate": 0.00012483297642171333, |
|
"loss": 0.8548, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"grad_norm": 0.53515625, |
|
"learning_rate": 0.00012030603335512468, |
|
"loss": 0.9028, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"eval_loss": 0.8055516481399536, |
|
"eval_runtime": 16.4906, |
|
"eval_samples_per_second": 17.283, |
|
"eval_steps_per_second": 2.183, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 0.58203125, |
|
"learning_rate": 0.00011573524193578863, |
|
"loss": 0.8422, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"grad_norm": 0.57421875, |
|
"learning_rate": 0.00011113047221912096, |
|
"loss": 0.7173, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 0.0001065016676324433, |
|
"loss": 0.6956, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"grad_norm": 0.53125, |
|
"learning_rate": 0.0001018588235033888, |
|
"loss": 0.6894, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"grad_norm": 0.54296875, |
|
"learning_rate": 9.721196547623584e-05, |
|
"loss": 0.6804, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"grad_norm": 0.4765625, |
|
"learning_rate": 9.257112786277631e-05, |
|
"loss": 0.6327, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"grad_norm": 0.51953125, |
|
"learning_rate": 8.79463319744677e-05, |
|
"loss": 0.637, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"grad_norm": 0.53125, |
|
"learning_rate": 8.334756448265781e-05, |
|
"loss": 0.6765, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"grad_norm": 0.51953125, |
|
"learning_rate": 7.878475585361045e-05, |
|
"loss": 0.6518, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"grad_norm": 0.5390625, |
|
"learning_rate": 7.42677589048989e-05, |
|
"loss": 0.6535, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 6.98063275294722e-05, |
|
"loss": 0.6311, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 6.54100956333369e-05, |
|
"loss": 0.6438, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"grad_norm": 0.55859375, |
|
"learning_rate": 6.108855633233546e-05, |
|
"loss": 0.6366, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 5.6851041452943646e-05, |
|
"loss": 0.627, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 5.270670138135234e-05, |
|
"loss": 0.6359, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 4.866448530434692e-05, |
|
"loss": 0.6727, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 4.4733121884651664e-05, |
|
"loss": 0.6406, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"grad_norm": 0.54296875, |
|
"learning_rate": 4.092110041246865e-05, |
|
"loss": 0.6535, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 3.7236652473911814e-05, |
|
"loss": 0.6553, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"eval_loss": 0.6577403545379639, |
|
"eval_runtime": 16.6273, |
|
"eval_samples_per_second": 17.141, |
|
"eval_steps_per_second": 2.165, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"grad_norm": 0.578125, |
|
"learning_rate": 3.36877341759205e-05, |
|
"loss": 0.6292, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 3.06, |
|
"grad_norm": 0.7421875, |
|
"learning_rate": 3.0282008966036646e-05, |
|
"loss": 0.4973, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 3.11, |
|
"grad_norm": 0.54296875, |
|
"learning_rate": 2.7026831084143255e-05, |
|
"loss": 0.4967, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 3.16, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 2.3929229681898003e-05, |
|
"loss": 0.5102, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 3.21, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 2.0995893644155008e-05, |
|
"loss": 0.4894, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 3.27, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 1.823315714515018e-05, |
|
"loss": 0.5026, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 3.32, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 1.5646985970639717e-05, |
|
"loss": 0.4853, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 3.37, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 1.324296463552821e-05, |
|
"loss": 0.483, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 3.43, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 1.1026284324803494e-05, |
|
"loss": 0.5077, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 3.48, |
|
"grad_norm": 0.51953125, |
|
"learning_rate": 9.001731683818337e-06, |
|
"loss": 0.5099, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 3.53, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 7.17367848212539e-06, |
|
"loss": 0.4787, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 3.58, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 5.546072173184791e-06, |
|
"loss": 0.4859, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 3.64, |
|
"grad_norm": 0.5, |
|
"learning_rate": 4.1224273703294514e-06, |
|
"loss": 0.5011, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 3.69, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 2.905818257394799e-06, |
|
"loss": 0.5044, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 3.74, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 1.8988719504013374e-06, |
|
"loss": 0.4728, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 3.79, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 1.103762824624377e-06, |
|
"loss": 0.5035, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 3.85, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 5.222078193011126e-07, |
|
"loss": 0.4899, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 3.9, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 1.554627301140199e-07, |
|
"loss": 0.4906, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 3.95, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 4.319497456273247e-09, |
|
"loss": 0.4936, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 3.96, |
|
"eval_loss": 0.6385390162467957, |
|
"eval_runtime": 16.4691, |
|
"eval_samples_per_second": 17.305, |
|
"eval_steps_per_second": 2.186, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 3.96, |
|
"step": 376, |
|
"total_flos": 2.3511765822965023e+18, |
|
"train_loss": 1.1451922758938151, |
|
"train_runtime": 4455.6354, |
|
"train_samples_per_second": 5.447, |
|
"train_steps_per_second": 0.084 |
|
} |
|
], |
|
"logging_steps": 5, |
|
"max_steps": 376, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 4, |
|
"save_steps": 500, |
|
"total_flos": 2.3511765822965023e+18, |
|
"train_batch_size": 1, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|