sedrickkeh's picture
End of training
3e551dd verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 2.9941520467836256,
"eval_steps": 500,
"global_step": 576,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.05198180636777128,
"grad_norm": 3.0779116014443817,
"learning_rate": 5e-06,
"loss": 0.7363,
"step": 10
},
{
"epoch": 0.10396361273554255,
"grad_norm": 0.8378025538905853,
"learning_rate": 5e-06,
"loss": 0.6574,
"step": 20
},
{
"epoch": 0.15594541910331383,
"grad_norm": 1.0326593461431761,
"learning_rate": 5e-06,
"loss": 0.6354,
"step": 30
},
{
"epoch": 0.2079272254710851,
"grad_norm": 0.8798298291096642,
"learning_rate": 5e-06,
"loss": 0.6258,
"step": 40
},
{
"epoch": 0.2599090318388564,
"grad_norm": 0.7652583221522038,
"learning_rate": 5e-06,
"loss": 0.6081,
"step": 50
},
{
"epoch": 0.31189083820662766,
"grad_norm": 0.8092758063286689,
"learning_rate": 5e-06,
"loss": 0.5957,
"step": 60
},
{
"epoch": 0.36387264457439894,
"grad_norm": 0.6377593975168986,
"learning_rate": 5e-06,
"loss": 0.5951,
"step": 70
},
{
"epoch": 0.4158544509421702,
"grad_norm": 0.8604320761158406,
"learning_rate": 5e-06,
"loss": 0.5853,
"step": 80
},
{
"epoch": 0.4678362573099415,
"grad_norm": 0.5659478868871771,
"learning_rate": 5e-06,
"loss": 0.5736,
"step": 90
},
{
"epoch": 0.5198180636777128,
"grad_norm": 0.6666095975840195,
"learning_rate": 5e-06,
"loss": 0.5681,
"step": 100
},
{
"epoch": 0.571799870045484,
"grad_norm": 0.7108362162785513,
"learning_rate": 5e-06,
"loss": 0.5752,
"step": 110
},
{
"epoch": 0.6237816764132553,
"grad_norm": 0.5055975881819947,
"learning_rate": 5e-06,
"loss": 0.5661,
"step": 120
},
{
"epoch": 0.6757634827810266,
"grad_norm": 0.5122594507537827,
"learning_rate": 5e-06,
"loss": 0.5724,
"step": 130
},
{
"epoch": 0.7277452891487979,
"grad_norm": 0.6645034900787495,
"learning_rate": 5e-06,
"loss": 0.5642,
"step": 140
},
{
"epoch": 0.7797270955165692,
"grad_norm": 0.4688383993640717,
"learning_rate": 5e-06,
"loss": 0.563,
"step": 150
},
{
"epoch": 0.8317089018843404,
"grad_norm": 0.5311685960170432,
"learning_rate": 5e-06,
"loss": 0.5636,
"step": 160
},
{
"epoch": 0.8836907082521117,
"grad_norm": 0.5275216737667017,
"learning_rate": 5e-06,
"loss": 0.5622,
"step": 170
},
{
"epoch": 0.935672514619883,
"grad_norm": 0.49741573059254773,
"learning_rate": 5e-06,
"loss": 0.5604,
"step": 180
},
{
"epoch": 0.9876543209876543,
"grad_norm": 0.6752670080708038,
"learning_rate": 5e-06,
"loss": 0.5606,
"step": 190
},
{
"epoch": 0.9980506822612085,
"eval_loss": 0.5604754090309143,
"eval_runtime": 134.8008,
"eval_samples_per_second": 38.457,
"eval_steps_per_second": 0.601,
"step": 192
},
{
"epoch": 1.0396361273554255,
"grad_norm": 0.4942854982934459,
"learning_rate": 5e-06,
"loss": 0.5581,
"step": 200
},
{
"epoch": 1.0916179337231968,
"grad_norm": 0.47133396673557604,
"learning_rate": 5e-06,
"loss": 0.5131,
"step": 210
},
{
"epoch": 1.143599740090968,
"grad_norm": 0.6706210343356649,
"learning_rate": 5e-06,
"loss": 0.5108,
"step": 220
},
{
"epoch": 1.1955815464587394,
"grad_norm": 0.4990475384686441,
"learning_rate": 5e-06,
"loss": 0.5127,
"step": 230
},
{
"epoch": 1.2475633528265107,
"grad_norm": 0.6948432011959483,
"learning_rate": 5e-06,
"loss": 0.516,
"step": 240
},
{
"epoch": 1.299545159194282,
"grad_norm": 0.4617478562618637,
"learning_rate": 5e-06,
"loss": 0.5124,
"step": 250
},
{
"epoch": 1.3515269655620532,
"grad_norm": 0.40496266862562025,
"learning_rate": 5e-06,
"loss": 0.5169,
"step": 260
},
{
"epoch": 1.4035087719298245,
"grad_norm": 0.4441934839800834,
"learning_rate": 5e-06,
"loss": 0.5153,
"step": 270
},
{
"epoch": 1.4554905782975958,
"grad_norm": 0.5551575728453572,
"learning_rate": 5e-06,
"loss": 0.5153,
"step": 280
},
{
"epoch": 1.507472384665367,
"grad_norm": 0.6109195782274622,
"learning_rate": 5e-06,
"loss": 0.5095,
"step": 290
},
{
"epoch": 1.5594541910331383,
"grad_norm": 0.501987852874291,
"learning_rate": 5e-06,
"loss": 0.5134,
"step": 300
},
{
"epoch": 1.6114359974009096,
"grad_norm": 0.42511170270357557,
"learning_rate": 5e-06,
"loss": 0.5165,
"step": 310
},
{
"epoch": 1.6634178037686809,
"grad_norm": 0.49289651016151736,
"learning_rate": 5e-06,
"loss": 0.5096,
"step": 320
},
{
"epoch": 1.7153996101364521,
"grad_norm": 0.4308251737050951,
"learning_rate": 5e-06,
"loss": 0.5138,
"step": 330
},
{
"epoch": 1.7673814165042234,
"grad_norm": 0.5157909639859894,
"learning_rate": 5e-06,
"loss": 0.5171,
"step": 340
},
{
"epoch": 1.8193632228719947,
"grad_norm": 0.5188996520268809,
"learning_rate": 5e-06,
"loss": 0.5038,
"step": 350
},
{
"epoch": 1.871345029239766,
"grad_norm": 0.4609663557546815,
"learning_rate": 5e-06,
"loss": 0.5135,
"step": 360
},
{
"epoch": 1.9233268356075373,
"grad_norm": 0.5130233449439461,
"learning_rate": 5e-06,
"loss": 0.4969,
"step": 370
},
{
"epoch": 1.9753086419753085,
"grad_norm": 0.47208230393580913,
"learning_rate": 5e-06,
"loss": 0.5116,
"step": 380
},
{
"epoch": 1.996101364522417,
"eval_loss": 0.5507273077964783,
"eval_runtime": 131.5009,
"eval_samples_per_second": 39.422,
"eval_steps_per_second": 0.616,
"step": 384
},
{
"epoch": 2.02729044834308,
"grad_norm": 0.6127372201164472,
"learning_rate": 5e-06,
"loss": 0.5155,
"step": 390
},
{
"epoch": 2.079272254710851,
"grad_norm": 0.7411594983025063,
"learning_rate": 5e-06,
"loss": 0.4578,
"step": 400
},
{
"epoch": 2.1312540610786224,
"grad_norm": 0.7428780446595316,
"learning_rate": 5e-06,
"loss": 0.4704,
"step": 410
},
{
"epoch": 2.1832358674463936,
"grad_norm": 0.5200617783769791,
"learning_rate": 5e-06,
"loss": 0.4637,
"step": 420
},
{
"epoch": 2.235217673814165,
"grad_norm": 0.6498653019719489,
"learning_rate": 5e-06,
"loss": 0.4622,
"step": 430
},
{
"epoch": 2.287199480181936,
"grad_norm": 0.46276007763796656,
"learning_rate": 5e-06,
"loss": 0.4607,
"step": 440
},
{
"epoch": 2.3391812865497075,
"grad_norm": 0.5996673377937245,
"learning_rate": 5e-06,
"loss": 0.4622,
"step": 450
},
{
"epoch": 2.3911630929174787,
"grad_norm": 0.49488878939584074,
"learning_rate": 5e-06,
"loss": 0.4682,
"step": 460
},
{
"epoch": 2.44314489928525,
"grad_norm": 0.4643784900227937,
"learning_rate": 5e-06,
"loss": 0.4594,
"step": 470
},
{
"epoch": 2.4951267056530213,
"grad_norm": 0.44786233784344404,
"learning_rate": 5e-06,
"loss": 0.4675,
"step": 480
},
{
"epoch": 2.5471085120207926,
"grad_norm": 0.6126679422066866,
"learning_rate": 5e-06,
"loss": 0.4659,
"step": 490
},
{
"epoch": 2.599090318388564,
"grad_norm": 0.530097732347591,
"learning_rate": 5e-06,
"loss": 0.4588,
"step": 500
},
{
"epoch": 2.651072124756335,
"grad_norm": 0.5516810761121376,
"learning_rate": 5e-06,
"loss": 0.4634,
"step": 510
},
{
"epoch": 2.7030539311241064,
"grad_norm": 0.5437842249870648,
"learning_rate": 5e-06,
"loss": 0.4642,
"step": 520
},
{
"epoch": 2.7550357374918777,
"grad_norm": 0.5248669258621385,
"learning_rate": 5e-06,
"loss": 0.4658,
"step": 530
},
{
"epoch": 2.807017543859649,
"grad_norm": 0.4756943098859436,
"learning_rate": 5e-06,
"loss": 0.4712,
"step": 540
},
{
"epoch": 2.8589993502274202,
"grad_norm": 0.5165218818933938,
"learning_rate": 5e-06,
"loss": 0.4672,
"step": 550
},
{
"epoch": 2.9109811565951915,
"grad_norm": 0.5304020568782208,
"learning_rate": 5e-06,
"loss": 0.4682,
"step": 560
},
{
"epoch": 2.962962962962963,
"grad_norm": 0.5954622072015786,
"learning_rate": 5e-06,
"loss": 0.4569,
"step": 570
},
{
"epoch": 2.9941520467836256,
"eval_loss": 0.5534155368804932,
"eval_runtime": 134.2301,
"eval_samples_per_second": 38.62,
"eval_steps_per_second": 0.603,
"step": 576
},
{
"epoch": 2.9941520467836256,
"step": 576,
"total_flos": 964612073717760.0,
"train_loss": 0.5243237420088716,
"train_runtime": 19210.653,
"train_samples_per_second": 15.379,
"train_steps_per_second": 0.03
}
],
"logging_steps": 10,
"max_steps": 576,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 964612073717760.0,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}