gemma-sft-qlora / trainer_state.json
satpalsr's picture
Model save
e1b689f verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 3.963109354413702,
"eval_steps": 500,
"global_step": 376,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.01,
"grad_norm": 476.0,
"learning_rate": 5.263157894736842e-06,
"loss": 13.5076,
"step": 1
},
{
"epoch": 0.05,
"grad_norm": 27.125,
"learning_rate": 2.6315789473684212e-05,
"loss": 10.3382,
"step": 5
},
{
"epoch": 0.11,
"grad_norm": 7.09375,
"learning_rate": 5.2631578947368424e-05,
"loss": 6.5621,
"step": 10
},
{
"epoch": 0.16,
"grad_norm": 5.09375,
"learning_rate": 7.894736842105263e-05,
"loss": 5.4048,
"step": 15
},
{
"epoch": 0.21,
"grad_norm": 13.625,
"learning_rate": 0.00010526315789473685,
"loss": 3.8714,
"step": 20
},
{
"epoch": 0.26,
"grad_norm": 4.3125,
"learning_rate": 0.00013157894736842108,
"loss": 1.907,
"step": 25
},
{
"epoch": 0.32,
"grad_norm": 2.484375,
"learning_rate": 0.00015789473684210527,
"loss": 1.7156,
"step": 30
},
{
"epoch": 0.37,
"grad_norm": 1.71875,
"learning_rate": 0.00018421052631578948,
"loss": 1.5905,
"step": 35
},
{
"epoch": 0.42,
"grad_norm": 1.4375,
"learning_rate": 0.00019998272238333606,
"loss": 1.4996,
"step": 40
},
{
"epoch": 0.47,
"grad_norm": 1.2109375,
"learning_rate": 0.00019978841775475367,
"loss": 1.424,
"step": 45
},
{
"epoch": 0.53,
"grad_norm": 2.515625,
"learning_rate": 0.00019937863245275304,
"loss": 1.3526,
"step": 50
},
{
"epoch": 0.58,
"grad_norm": 1.6953125,
"learning_rate": 0.0001987542513577122,
"loss": 1.3334,
"step": 55
},
{
"epoch": 0.63,
"grad_norm": 2.3125,
"learning_rate": 0.00019791662274292637,
"loss": 1.2863,
"step": 60
},
{
"epoch": 0.69,
"grad_norm": 0.91015625,
"learning_rate": 0.00019686755536317945,
"loss": 1.2384,
"step": 65
},
{
"epoch": 0.74,
"grad_norm": 0.99609375,
"learning_rate": 0.00019560931454896298,
"loss": 1.2248,
"step": 70
},
{
"epoch": 0.79,
"grad_norm": 0.73046875,
"learning_rate": 0.000194144617314777,
"loss": 1.1772,
"step": 75
},
{
"epoch": 0.84,
"grad_norm": 0.58203125,
"learning_rate": 0.0001924766264920751,
"loss": 1.1654,
"step": 80
},
{
"epoch": 0.9,
"grad_norm": 0.74609375,
"learning_rate": 0.00019060894389952328,
"loss": 1.1382,
"step": 85
},
{
"epoch": 0.95,
"grad_norm": 0.78515625,
"learning_rate": 0.000188545602565321,
"loss": 1.1537,
"step": 90
},
{
"epoch": 0.99,
"eval_loss": 1.0988351106643677,
"eval_runtime": 16.9422,
"eval_samples_per_second": 16.822,
"eval_steps_per_second": 2.125,
"step": 94
},
{
"epoch": 1.0,
"grad_norm": 0.65625,
"learning_rate": 0.00018629105801837818,
"loss": 1.1162,
"step": 95
},
{
"epoch": 1.05,
"grad_norm": 0.84375,
"learning_rate": 0.00018385017866715507,
"loss": 0.9838,
"step": 100
},
{
"epoch": 1.11,
"grad_norm": 0.61328125,
"learning_rate": 0.00018122823528693966,
"loss": 0.9895,
"step": 105
},
{
"epoch": 1.16,
"grad_norm": 0.8359375,
"learning_rate": 0.00017843088963826435,
"loss": 0.9588,
"step": 110
},
{
"epoch": 1.21,
"grad_norm": 0.74609375,
"learning_rate": 0.00017546418224103838,
"loss": 0.9304,
"step": 115
},
{
"epoch": 1.26,
"grad_norm": 0.63671875,
"learning_rate": 0.00017233451933079664,
"loss": 0.9315,
"step": 120
},
{
"epoch": 1.32,
"grad_norm": 0.53125,
"learning_rate": 0.00016904865902523096,
"loss": 0.9869,
"step": 125
},
{
"epoch": 1.37,
"grad_norm": 0.796875,
"learning_rate": 0.00016561369673087588,
"loss": 0.9399,
"step": 130
},
{
"epoch": 1.42,
"grad_norm": 0.75390625,
"learning_rate": 0.00016203704982146073,
"loss": 0.9061,
"step": 135
},
{
"epoch": 1.48,
"grad_norm": 2.296875,
"learning_rate": 0.00015832644162101417,
"loss": 0.9537,
"step": 140
},
{
"epoch": 1.53,
"grad_norm": 0.6875,
"learning_rate": 0.00015448988472630654,
"loss": 0.9107,
"step": 145
},
{
"epoch": 1.58,
"grad_norm": 0.61328125,
"learning_rate": 0.00015053566370464415,
"loss": 0.9483,
"step": 150
},
{
"epoch": 1.63,
"grad_norm": 0.7265625,
"learning_rate": 0.00014647231720437686,
"loss": 0.9075,
"step": 155
},
{
"epoch": 1.69,
"grad_norm": 0.53515625,
"learning_rate": 0.00014230861951674913,
"loss": 0.9148,
"step": 160
},
{
"epoch": 1.74,
"grad_norm": 0.51171875,
"learning_rate": 0.0001380535616289099,
"loss": 0.9142,
"step": 165
},
{
"epoch": 1.79,
"grad_norm": 0.5703125,
"learning_rate": 0.00013371633180899416,
"loss": 0.8966,
"step": 170
},
{
"epoch": 1.84,
"grad_norm": 0.53515625,
"learning_rate": 0.00012930629576520132,
"loss": 0.8795,
"step": 175
},
{
"epoch": 1.9,
"grad_norm": 0.609375,
"learning_rate": 0.00012483297642171333,
"loss": 0.8548,
"step": 180
},
{
"epoch": 1.95,
"grad_norm": 0.53515625,
"learning_rate": 0.00012030603335512468,
"loss": 0.9028,
"step": 185
},
{
"epoch": 1.99,
"eval_loss": 0.8055516481399536,
"eval_runtime": 16.4906,
"eval_samples_per_second": 17.283,
"eval_steps_per_second": 2.183,
"step": 189
},
{
"epoch": 2.0,
"grad_norm": 0.58203125,
"learning_rate": 0.00011573524193578863,
"loss": 0.8422,
"step": 190
},
{
"epoch": 2.06,
"grad_norm": 0.57421875,
"learning_rate": 0.00011113047221912096,
"loss": 0.7173,
"step": 195
},
{
"epoch": 2.11,
"grad_norm": 0.52734375,
"learning_rate": 0.0001065016676324433,
"loss": 0.6956,
"step": 200
},
{
"epoch": 2.16,
"grad_norm": 0.53125,
"learning_rate": 0.0001018588235033888,
"loss": 0.6894,
"step": 205
},
{
"epoch": 2.21,
"grad_norm": 0.54296875,
"learning_rate": 9.721196547623584e-05,
"loss": 0.6804,
"step": 210
},
{
"epoch": 2.27,
"grad_norm": 0.4765625,
"learning_rate": 9.257112786277631e-05,
"loss": 0.6327,
"step": 215
},
{
"epoch": 2.32,
"grad_norm": 0.51953125,
"learning_rate": 8.79463319744677e-05,
"loss": 0.637,
"step": 220
},
{
"epoch": 2.37,
"grad_norm": 0.53125,
"learning_rate": 8.334756448265781e-05,
"loss": 0.6765,
"step": 225
},
{
"epoch": 2.42,
"grad_norm": 0.51953125,
"learning_rate": 7.878475585361045e-05,
"loss": 0.6518,
"step": 230
},
{
"epoch": 2.48,
"grad_norm": 0.5390625,
"learning_rate": 7.42677589048989e-05,
"loss": 0.6535,
"step": 235
},
{
"epoch": 2.53,
"grad_norm": 0.5078125,
"learning_rate": 6.98063275294722e-05,
"loss": 0.6311,
"step": 240
},
{
"epoch": 2.58,
"grad_norm": 0.52734375,
"learning_rate": 6.54100956333369e-05,
"loss": 0.6438,
"step": 245
},
{
"epoch": 2.64,
"grad_norm": 0.55859375,
"learning_rate": 6.108855633233546e-05,
"loss": 0.6366,
"step": 250
},
{
"epoch": 2.69,
"grad_norm": 0.51171875,
"learning_rate": 5.6851041452943646e-05,
"loss": 0.627,
"step": 255
},
{
"epoch": 2.74,
"grad_norm": 0.498046875,
"learning_rate": 5.270670138135234e-05,
"loss": 0.6359,
"step": 260
},
{
"epoch": 2.79,
"grad_norm": 0.49609375,
"learning_rate": 4.866448530434692e-05,
"loss": 0.6727,
"step": 265
},
{
"epoch": 2.85,
"grad_norm": 0.49609375,
"learning_rate": 4.4733121884651664e-05,
"loss": 0.6406,
"step": 270
},
{
"epoch": 2.9,
"grad_norm": 0.54296875,
"learning_rate": 4.092110041246865e-05,
"loss": 0.6535,
"step": 275
},
{
"epoch": 2.95,
"grad_norm": 0.48046875,
"learning_rate": 3.7236652473911814e-05,
"loss": 0.6553,
"step": 280
},
{
"epoch": 2.99,
"eval_loss": 0.6577403545379639,
"eval_runtime": 16.6273,
"eval_samples_per_second": 17.141,
"eval_steps_per_second": 2.165,
"step": 284
},
{
"epoch": 3.0,
"grad_norm": 0.578125,
"learning_rate": 3.36877341759205e-05,
"loss": 0.6292,
"step": 285
},
{
"epoch": 3.06,
"grad_norm": 0.7421875,
"learning_rate": 3.0282008966036646e-05,
"loss": 0.4973,
"step": 290
},
{
"epoch": 3.11,
"grad_norm": 0.54296875,
"learning_rate": 2.7026831084143255e-05,
"loss": 0.4967,
"step": 295
},
{
"epoch": 3.16,
"grad_norm": 0.5234375,
"learning_rate": 2.3929229681898003e-05,
"loss": 0.5102,
"step": 300
},
{
"epoch": 3.21,
"grad_norm": 0.50390625,
"learning_rate": 2.0995893644155008e-05,
"loss": 0.4894,
"step": 305
},
{
"epoch": 3.27,
"grad_norm": 0.515625,
"learning_rate": 1.823315714515018e-05,
"loss": 0.5026,
"step": 310
},
{
"epoch": 3.32,
"grad_norm": 0.49609375,
"learning_rate": 1.5646985970639717e-05,
"loss": 0.4853,
"step": 315
},
{
"epoch": 3.37,
"grad_norm": 0.50390625,
"learning_rate": 1.324296463552821e-05,
"loss": 0.483,
"step": 320
},
{
"epoch": 3.43,
"grad_norm": 0.4921875,
"learning_rate": 1.1026284324803494e-05,
"loss": 0.5077,
"step": 325
},
{
"epoch": 3.48,
"grad_norm": 0.51953125,
"learning_rate": 9.001731683818337e-06,
"loss": 0.5099,
"step": 330
},
{
"epoch": 3.53,
"grad_norm": 0.50390625,
"learning_rate": 7.17367848212539e-06,
"loss": 0.4787,
"step": 335
},
{
"epoch": 3.58,
"grad_norm": 0.5078125,
"learning_rate": 5.546072173184791e-06,
"loss": 0.4859,
"step": 340
},
{
"epoch": 3.64,
"grad_norm": 0.5,
"learning_rate": 4.1224273703294514e-06,
"loss": 0.5011,
"step": 345
},
{
"epoch": 3.69,
"grad_norm": 0.5078125,
"learning_rate": 2.905818257394799e-06,
"loss": 0.5044,
"step": 350
},
{
"epoch": 3.74,
"grad_norm": 0.5078125,
"learning_rate": 1.8988719504013374e-06,
"loss": 0.4728,
"step": 355
},
{
"epoch": 3.79,
"grad_norm": 0.51171875,
"learning_rate": 1.103762824624377e-06,
"loss": 0.5035,
"step": 360
},
{
"epoch": 3.85,
"grad_norm": 0.498046875,
"learning_rate": 5.222078193011126e-07,
"loss": 0.4899,
"step": 365
},
{
"epoch": 3.9,
"grad_norm": 0.50390625,
"learning_rate": 1.554627301140199e-07,
"loss": 0.4906,
"step": 370
},
{
"epoch": 3.95,
"grad_norm": 0.4921875,
"learning_rate": 4.319497456273247e-09,
"loss": 0.4936,
"step": 375
},
{
"epoch": 3.96,
"eval_loss": 0.6385390162467957,
"eval_runtime": 16.4691,
"eval_samples_per_second": 17.305,
"eval_steps_per_second": 2.186,
"step": 376
},
{
"epoch": 3.96,
"step": 376,
"total_flos": 2.3511765822965023e+18,
"train_loss": 1.1451922758938151,
"train_runtime": 4455.6354,
"train_samples_per_second": 5.447,
"train_steps_per_second": 0.084
}
],
"logging_steps": 5,
"max_steps": 376,
"num_input_tokens_seen": 0,
"num_train_epochs": 4,
"save_steps": 500,
"total_flos": 2.3511765822965023e+18,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}