GaetanMichelet's picture
Model save
fa15631 verified
raw
history blame contribute delete
No virus
50.3 kB
{
"best_metric": 0.9015458226203918,
"best_model_checkpoint": "data/Llama-31-8B_task-1_120-samples_config-3_full/checkpoint-396",
"epoch": 43.0,
"eval_steps": 500,
"global_step": 473,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.09090909090909091,
"grad_norm": 2.021212100982666,
"learning_rate": 6.060606060606061e-08,
"loss": 2.4844,
"step": 1
},
{
"epoch": 0.18181818181818182,
"grad_norm": 1.8868809938430786,
"learning_rate": 1.2121212121212122e-07,
"loss": 2.5082,
"step": 2
},
{
"epoch": 0.36363636363636365,
"grad_norm": 1.9227168560028076,
"learning_rate": 2.4242424242424244e-07,
"loss": 2.5149,
"step": 4
},
{
"epoch": 0.5454545454545454,
"grad_norm": 2.164538860321045,
"learning_rate": 3.6363636363636366e-07,
"loss": 2.4938,
"step": 6
},
{
"epoch": 0.7272727272727273,
"grad_norm": 1.9525693655014038,
"learning_rate": 4.848484848484849e-07,
"loss": 2.4402,
"step": 8
},
{
"epoch": 0.9090909090909091,
"grad_norm": 1.973608136177063,
"learning_rate": 6.060606060606061e-07,
"loss": 2.4681,
"step": 10
},
{
"epoch": 1.0,
"eval_loss": 2.453854560852051,
"eval_runtime": 9.6392,
"eval_samples_per_second": 2.49,
"eval_steps_per_second": 2.49,
"step": 11
},
{
"epoch": 1.0909090909090908,
"grad_norm": 1.9462313652038574,
"learning_rate": 7.272727272727273e-07,
"loss": 2.4751,
"step": 12
},
{
"epoch": 1.2727272727272727,
"grad_norm": 1.8247989416122437,
"learning_rate": 8.484848484848486e-07,
"loss": 2.523,
"step": 14
},
{
"epoch": 1.4545454545454546,
"grad_norm": 1.8403159379959106,
"learning_rate": 9.696969696969698e-07,
"loss": 2.4192,
"step": 16
},
{
"epoch": 1.6363636363636362,
"grad_norm": 1.6670445203781128,
"learning_rate": 1.090909090909091e-06,
"loss": 2.5069,
"step": 18
},
{
"epoch": 1.8181818181818183,
"grad_norm": 1.802819013595581,
"learning_rate": 1.2121212121212122e-06,
"loss": 2.4854,
"step": 20
},
{
"epoch": 2.0,
"grad_norm": 1.4742677211761475,
"learning_rate": 1.3333333333333334e-06,
"loss": 2.3894,
"step": 22
},
{
"epoch": 2.0,
"eval_loss": 2.4259960651397705,
"eval_runtime": 9.6188,
"eval_samples_per_second": 2.495,
"eval_steps_per_second": 2.495,
"step": 22
},
{
"epoch": 2.1818181818181817,
"grad_norm": 1.6892441511154175,
"learning_rate": 1.4545454545454546e-06,
"loss": 2.4004,
"step": 24
},
{
"epoch": 2.3636363636363638,
"grad_norm": 1.5024837255477905,
"learning_rate": 1.5757575757575759e-06,
"loss": 2.4339,
"step": 26
},
{
"epoch": 2.5454545454545454,
"grad_norm": 1.6032872200012207,
"learning_rate": 1.6969696969696973e-06,
"loss": 2.4351,
"step": 28
},
{
"epoch": 2.7272727272727275,
"grad_norm": 1.526031732559204,
"learning_rate": 1.8181818181818183e-06,
"loss": 2.3962,
"step": 30
},
{
"epoch": 2.909090909090909,
"grad_norm": 1.6687572002410889,
"learning_rate": 1.9393939393939395e-06,
"loss": 2.4746,
"step": 32
},
{
"epoch": 3.0,
"eval_loss": 2.382659912109375,
"eval_runtime": 9.6201,
"eval_samples_per_second": 2.495,
"eval_steps_per_second": 2.495,
"step": 33
},
{
"epoch": 3.090909090909091,
"grad_norm": 1.5408605337142944,
"learning_rate": 2.0606060606060607e-06,
"loss": 2.4125,
"step": 34
},
{
"epoch": 3.2727272727272725,
"grad_norm": 1.6628339290618896,
"learning_rate": 2.181818181818182e-06,
"loss": 2.3574,
"step": 36
},
{
"epoch": 3.4545454545454546,
"grad_norm": 1.6864051818847656,
"learning_rate": 2.303030303030303e-06,
"loss": 2.3033,
"step": 38
},
{
"epoch": 3.6363636363636362,
"grad_norm": 1.705551266670227,
"learning_rate": 2.4242424242424244e-06,
"loss": 2.4227,
"step": 40
},
{
"epoch": 3.8181818181818183,
"grad_norm": 1.6931661367416382,
"learning_rate": 2.5454545454545456e-06,
"loss": 2.3626,
"step": 42
},
{
"epoch": 4.0,
"grad_norm": 1.910327434539795,
"learning_rate": 2.666666666666667e-06,
"loss": 2.4177,
"step": 44
},
{
"epoch": 4.0,
"eval_loss": 2.313791036605835,
"eval_runtime": 9.6408,
"eval_samples_per_second": 2.489,
"eval_steps_per_second": 2.489,
"step": 44
},
{
"epoch": 4.181818181818182,
"grad_norm": 1.9009881019592285,
"learning_rate": 2.7878787878787885e-06,
"loss": 2.2818,
"step": 46
},
{
"epoch": 4.363636363636363,
"grad_norm": 1.9625577926635742,
"learning_rate": 2.9090909090909093e-06,
"loss": 2.4051,
"step": 48
},
{
"epoch": 4.545454545454545,
"grad_norm": 1.4953864812850952,
"learning_rate": 3.0303030303030305e-06,
"loss": 2.2829,
"step": 50
},
{
"epoch": 4.7272727272727275,
"grad_norm": 1.5853980779647827,
"learning_rate": 3.1515151515151517e-06,
"loss": 2.2705,
"step": 52
},
{
"epoch": 4.909090909090909,
"grad_norm": 1.5361964702606201,
"learning_rate": 3.272727272727273e-06,
"loss": 2.1959,
"step": 54
},
{
"epoch": 5.0,
"eval_loss": 2.226909875869751,
"eval_runtime": 9.6204,
"eval_samples_per_second": 2.495,
"eval_steps_per_second": 2.495,
"step": 55
},
{
"epoch": 5.090909090909091,
"grad_norm": 1.919047236442566,
"learning_rate": 3.3939393939393946e-06,
"loss": 2.2901,
"step": 56
},
{
"epoch": 5.2727272727272725,
"grad_norm": 2.0125391483306885,
"learning_rate": 3.5151515151515154e-06,
"loss": 2.2584,
"step": 58
},
{
"epoch": 5.454545454545454,
"grad_norm": 1.8322199583053589,
"learning_rate": 3.6363636363636366e-06,
"loss": 2.1565,
"step": 60
},
{
"epoch": 5.636363636363637,
"grad_norm": 1.318441390991211,
"learning_rate": 3.757575757575758e-06,
"loss": 2.2066,
"step": 62
},
{
"epoch": 5.818181818181818,
"grad_norm": 1.2064168453216553,
"learning_rate": 3.878787878787879e-06,
"loss": 2.1475,
"step": 64
},
{
"epoch": 6.0,
"grad_norm": 1.1312469244003296,
"learning_rate": 4.000000000000001e-06,
"loss": 2.16,
"step": 66
},
{
"epoch": 6.0,
"eval_loss": 2.1177315711975098,
"eval_runtime": 9.6288,
"eval_samples_per_second": 2.493,
"eval_steps_per_second": 2.493,
"step": 66
},
{
"epoch": 6.181818181818182,
"grad_norm": 1.0877413749694824,
"learning_rate": 4.1212121212121215e-06,
"loss": 2.0577,
"step": 68
},
{
"epoch": 6.363636363636363,
"grad_norm": 1.1328318119049072,
"learning_rate": 4.242424242424243e-06,
"loss": 2.1584,
"step": 70
},
{
"epoch": 6.545454545454545,
"grad_norm": 1.1114661693572998,
"learning_rate": 4.363636363636364e-06,
"loss": 2.1214,
"step": 72
},
{
"epoch": 6.7272727272727275,
"grad_norm": 1.0736863613128662,
"learning_rate": 4.4848484848484855e-06,
"loss": 2.0338,
"step": 74
},
{
"epoch": 6.909090909090909,
"grad_norm": 1.1103681325912476,
"learning_rate": 4.606060606060606e-06,
"loss": 2.0388,
"step": 76
},
{
"epoch": 7.0,
"eval_loss": 1.9843875169754028,
"eval_runtime": 9.6304,
"eval_samples_per_second": 2.492,
"eval_steps_per_second": 2.492,
"step": 77
},
{
"epoch": 7.090909090909091,
"grad_norm": 1.0429500341415405,
"learning_rate": 4.727272727272728e-06,
"loss": 1.9515,
"step": 78
},
{
"epoch": 7.2727272727272725,
"grad_norm": 1.0388323068618774,
"learning_rate": 4.848484848484849e-06,
"loss": 2.0436,
"step": 80
},
{
"epoch": 7.454545454545454,
"grad_norm": 0.9600175023078918,
"learning_rate": 4.9696969696969696e-06,
"loss": 1.8726,
"step": 82
},
{
"epoch": 7.636363636363637,
"grad_norm": 0.9380725026130676,
"learning_rate": 5.090909090909091e-06,
"loss": 1.9307,
"step": 84
},
{
"epoch": 7.818181818181818,
"grad_norm": 0.8361033797264099,
"learning_rate": 5.212121212121213e-06,
"loss": 1.869,
"step": 86
},
{
"epoch": 8.0,
"grad_norm": 0.8874518275260925,
"learning_rate": 5.333333333333334e-06,
"loss": 1.8932,
"step": 88
},
{
"epoch": 8.0,
"eval_loss": 1.8441652059555054,
"eval_runtime": 9.6269,
"eval_samples_per_second": 2.493,
"eval_steps_per_second": 2.493,
"step": 88
},
{
"epoch": 8.181818181818182,
"grad_norm": 0.8483957052230835,
"learning_rate": 5.4545454545454545e-06,
"loss": 1.8148,
"step": 90
},
{
"epoch": 8.363636363636363,
"grad_norm": 0.9277289509773254,
"learning_rate": 5.575757575757577e-06,
"loss": 1.8286,
"step": 92
},
{
"epoch": 8.545454545454545,
"grad_norm": 0.9386701583862305,
"learning_rate": 5.696969696969698e-06,
"loss": 1.7935,
"step": 94
},
{
"epoch": 8.727272727272727,
"grad_norm": 0.9653158783912659,
"learning_rate": 5.8181818181818185e-06,
"loss": 1.7714,
"step": 96
},
{
"epoch": 8.909090909090908,
"grad_norm": 0.929076075553894,
"learning_rate": 5.93939393939394e-06,
"loss": 1.7199,
"step": 98
},
{
"epoch": 9.0,
"eval_loss": 1.6830488443374634,
"eval_runtime": 9.6281,
"eval_samples_per_second": 2.493,
"eval_steps_per_second": 2.493,
"step": 99
},
{
"epoch": 9.090909090909092,
"grad_norm": 0.9527233839035034,
"learning_rate": 6.060606060606061e-06,
"loss": 1.7054,
"step": 100
},
{
"epoch": 9.272727272727273,
"grad_norm": 1.1190400123596191,
"learning_rate": 6.181818181818182e-06,
"loss": 1.623,
"step": 102
},
{
"epoch": 9.454545454545455,
"grad_norm": 1.1004133224487305,
"learning_rate": 6.303030303030303e-06,
"loss": 1.6696,
"step": 104
},
{
"epoch": 9.636363636363637,
"grad_norm": 1.036831259727478,
"learning_rate": 6.424242424242425e-06,
"loss": 1.5955,
"step": 106
},
{
"epoch": 9.818181818181818,
"grad_norm": 0.9210329055786133,
"learning_rate": 6.545454545454546e-06,
"loss": 1.5252,
"step": 108
},
{
"epoch": 10.0,
"grad_norm": 1.0149990320205688,
"learning_rate": 6.666666666666667e-06,
"loss": 1.4973,
"step": 110
},
{
"epoch": 10.0,
"eval_loss": 1.4929395914077759,
"eval_runtime": 9.6201,
"eval_samples_per_second": 2.495,
"eval_steps_per_second": 2.495,
"step": 110
},
{
"epoch": 10.181818181818182,
"grad_norm": 0.9690227508544922,
"learning_rate": 6.787878787878789e-06,
"loss": 1.4956,
"step": 112
},
{
"epoch": 10.363636363636363,
"grad_norm": 0.8670191168785095,
"learning_rate": 6.90909090909091e-06,
"loss": 1.4899,
"step": 114
},
{
"epoch": 10.545454545454545,
"grad_norm": 0.9516677260398865,
"learning_rate": 7.030303030303031e-06,
"loss": 1.3702,
"step": 116
},
{
"epoch": 10.727272727272727,
"grad_norm": 0.987469494342804,
"learning_rate": 7.151515151515152e-06,
"loss": 1.4109,
"step": 118
},
{
"epoch": 10.909090909090908,
"grad_norm": 0.9726764559745789,
"learning_rate": 7.272727272727273e-06,
"loss": 1.2726,
"step": 120
},
{
"epoch": 11.0,
"eval_loss": 1.2980190515518188,
"eval_runtime": 9.6227,
"eval_samples_per_second": 2.494,
"eval_steps_per_second": 2.494,
"step": 121
},
{
"epoch": 11.090909090909092,
"grad_norm": 0.925747811794281,
"learning_rate": 7.393939393939395e-06,
"loss": 1.2873,
"step": 122
},
{
"epoch": 11.272727272727273,
"grad_norm": 1.0416390895843506,
"learning_rate": 7.515151515151516e-06,
"loss": 1.2906,
"step": 124
},
{
"epoch": 11.454545454545455,
"grad_norm": 0.8796542286872864,
"learning_rate": 7.636363636363638e-06,
"loss": 1.2497,
"step": 126
},
{
"epoch": 11.636363636363637,
"grad_norm": 0.7352049350738525,
"learning_rate": 7.757575757575758e-06,
"loss": 1.2023,
"step": 128
},
{
"epoch": 11.818181818181818,
"grad_norm": 0.7811641693115234,
"learning_rate": 7.87878787878788e-06,
"loss": 1.1205,
"step": 130
},
{
"epoch": 12.0,
"grad_norm": 0.6574164032936096,
"learning_rate": 8.000000000000001e-06,
"loss": 1.204,
"step": 132
},
{
"epoch": 12.0,
"eval_loss": 1.1553701162338257,
"eval_runtime": 9.6262,
"eval_samples_per_second": 2.493,
"eval_steps_per_second": 2.493,
"step": 132
},
{
"epoch": 12.181818181818182,
"grad_norm": 0.7204536199569702,
"learning_rate": 8.121212121212121e-06,
"loss": 1.134,
"step": 134
},
{
"epoch": 12.363636363636363,
"grad_norm": 0.7305138111114502,
"learning_rate": 8.242424242424243e-06,
"loss": 1.046,
"step": 136
},
{
"epoch": 12.545454545454545,
"grad_norm": 0.7755628228187561,
"learning_rate": 8.363636363636365e-06,
"loss": 1.1237,
"step": 138
},
{
"epoch": 12.727272727272727,
"grad_norm": 0.8240987062454224,
"learning_rate": 8.484848484848486e-06,
"loss": 1.1215,
"step": 140
},
{
"epoch": 12.909090909090908,
"grad_norm": 0.716643750667572,
"learning_rate": 8.606060606060606e-06,
"loss": 1.0597,
"step": 142
},
{
"epoch": 13.0,
"eval_loss": 1.0772147178649902,
"eval_runtime": 9.6246,
"eval_samples_per_second": 2.494,
"eval_steps_per_second": 2.494,
"step": 143
},
{
"epoch": 13.090909090909092,
"grad_norm": 0.6686062812805176,
"learning_rate": 8.727272727272728e-06,
"loss": 1.09,
"step": 144
},
{
"epoch": 13.272727272727273,
"grad_norm": 0.5461896061897278,
"learning_rate": 8.84848484848485e-06,
"loss": 1.0405,
"step": 146
},
{
"epoch": 13.454545454545455,
"grad_norm": 0.5121557712554932,
"learning_rate": 8.969696969696971e-06,
"loss": 1.0367,
"step": 148
},
{
"epoch": 13.636363636363637,
"grad_norm": 0.5350603461265564,
"learning_rate": 9.090909090909091e-06,
"loss": 1.0327,
"step": 150
},
{
"epoch": 13.818181818181818,
"grad_norm": 0.48038652539253235,
"learning_rate": 9.212121212121213e-06,
"loss": 1.0409,
"step": 152
},
{
"epoch": 14.0,
"grad_norm": 0.5158300399780273,
"learning_rate": 9.333333333333334e-06,
"loss": 1.0642,
"step": 154
},
{
"epoch": 14.0,
"eval_loss": 1.0425277948379517,
"eval_runtime": 9.6218,
"eval_samples_per_second": 2.494,
"eval_steps_per_second": 2.494,
"step": 154
},
{
"epoch": 14.181818181818182,
"grad_norm": 0.5284631252288818,
"learning_rate": 9.454545454545456e-06,
"loss": 1.0258,
"step": 156
},
{
"epoch": 14.363636363636363,
"grad_norm": 0.5030417442321777,
"learning_rate": 9.575757575757576e-06,
"loss": 1.0207,
"step": 158
},
{
"epoch": 14.545454545454545,
"grad_norm": 0.5023478269577026,
"learning_rate": 9.696969696969698e-06,
"loss": 0.9857,
"step": 160
},
{
"epoch": 14.727272727272727,
"grad_norm": 0.4412921071052551,
"learning_rate": 9.81818181818182e-06,
"loss": 1.006,
"step": 162
},
{
"epoch": 14.909090909090908,
"grad_norm": 0.4654785096645355,
"learning_rate": 9.939393939393939e-06,
"loss": 1.0466,
"step": 164
},
{
"epoch": 15.0,
"eval_loss": 1.020145297050476,
"eval_runtime": 9.6194,
"eval_samples_per_second": 2.495,
"eval_steps_per_second": 2.495,
"step": 165
},
{
"epoch": 15.090909090909092,
"grad_norm": 0.4788985848426819,
"learning_rate": 9.999988811118232e-06,
"loss": 0.9711,
"step": 166
},
{
"epoch": 15.272727272727273,
"grad_norm": 0.505489706993103,
"learning_rate": 9.999899300364534e-06,
"loss": 0.9704,
"step": 168
},
{
"epoch": 15.454545454545455,
"grad_norm": 0.5644559860229492,
"learning_rate": 9.999720280459576e-06,
"loss": 1.0248,
"step": 170
},
{
"epoch": 15.636363636363637,
"grad_norm": 0.4749113619327545,
"learning_rate": 9.999451754608208e-06,
"loss": 0.9284,
"step": 172
},
{
"epoch": 15.818181818181818,
"grad_norm": 0.658306896686554,
"learning_rate": 9.99909372761763e-06,
"loss": 1.0245,
"step": 174
},
{
"epoch": 16.0,
"grad_norm": 0.5014739036560059,
"learning_rate": 9.99864620589731e-06,
"loss": 1.0044,
"step": 176
},
{
"epoch": 16.0,
"eval_loss": 1.0009642839431763,
"eval_runtime": 9.6199,
"eval_samples_per_second": 2.495,
"eval_steps_per_second": 2.495,
"step": 176
},
{
"epoch": 16.181818181818183,
"grad_norm": 0.5191582441329956,
"learning_rate": 9.998109197458865e-06,
"loss": 0.9723,
"step": 178
},
{
"epoch": 16.363636363636363,
"grad_norm": 0.5656840205192566,
"learning_rate": 9.997482711915926e-06,
"loss": 0.9901,
"step": 180
},
{
"epoch": 16.545454545454547,
"grad_norm": 0.49254581332206726,
"learning_rate": 9.996766760483955e-06,
"loss": 0.9321,
"step": 182
},
{
"epoch": 16.727272727272727,
"grad_norm": 0.500410795211792,
"learning_rate": 9.995961355980052e-06,
"loss": 0.9564,
"step": 184
},
{
"epoch": 16.90909090909091,
"grad_norm": 0.6280571222305298,
"learning_rate": 9.99506651282272e-06,
"loss": 0.9967,
"step": 186
},
{
"epoch": 17.0,
"eval_loss": 0.9866144061088562,
"eval_runtime": 9.621,
"eval_samples_per_second": 2.495,
"eval_steps_per_second": 2.495,
"step": 187
},
{
"epoch": 17.09090909090909,
"grad_norm": 0.578583300113678,
"learning_rate": 9.994082247031613e-06,
"loss": 0.9784,
"step": 188
},
{
"epoch": 17.272727272727273,
"grad_norm": 0.5218169093132019,
"learning_rate": 9.993008576227248e-06,
"loss": 0.8817,
"step": 190
},
{
"epoch": 17.454545454545453,
"grad_norm": 0.6009635329246521,
"learning_rate": 9.991845519630679e-06,
"loss": 0.9515,
"step": 192
},
{
"epoch": 17.636363636363637,
"grad_norm": 0.5707380771636963,
"learning_rate": 9.99059309806317e-06,
"loss": 0.9694,
"step": 194
},
{
"epoch": 17.818181818181817,
"grad_norm": 0.6348333358764648,
"learning_rate": 9.989251333945813e-06,
"loss": 0.9498,
"step": 196
},
{
"epoch": 18.0,
"grad_norm": 0.5231234431266785,
"learning_rate": 9.987820251299121e-06,
"loss": 0.9863,
"step": 198
},
{
"epoch": 18.0,
"eval_loss": 0.9735579490661621,
"eval_runtime": 9.6192,
"eval_samples_per_second": 2.495,
"eval_steps_per_second": 2.495,
"step": 198
},
{
"epoch": 18.181818181818183,
"grad_norm": 0.6517296433448792,
"learning_rate": 9.986299875742612e-06,
"loss": 0.9591,
"step": 200
},
{
"epoch": 18.363636363636363,
"grad_norm": 0.6254110932350159,
"learning_rate": 9.984690234494338e-06,
"loss": 0.9296,
"step": 202
},
{
"epoch": 18.545454545454547,
"grad_norm": 0.7348714470863342,
"learning_rate": 9.982991356370404e-06,
"loss": 0.9482,
"step": 204
},
{
"epoch": 18.727272727272727,
"grad_norm": 0.5867448449134827,
"learning_rate": 9.98120327178445e-06,
"loss": 0.9276,
"step": 206
},
{
"epoch": 18.90909090909091,
"grad_norm": 0.6433852910995483,
"learning_rate": 9.979326012747106e-06,
"loss": 0.9065,
"step": 208
},
{
"epoch": 19.0,
"eval_loss": 0.9643786549568176,
"eval_runtime": 9.647,
"eval_samples_per_second": 2.488,
"eval_steps_per_second": 2.488,
"step": 209
},
{
"epoch": 19.09090909090909,
"grad_norm": 0.650622546672821,
"learning_rate": 9.977359612865424e-06,
"loss": 0.8947,
"step": 210
},
{
"epoch": 19.272727272727273,
"grad_norm": 0.7716985940933228,
"learning_rate": 9.975304107342268e-06,
"loss": 0.929,
"step": 212
},
{
"epoch": 19.454545454545453,
"grad_norm": 0.5945254564285278,
"learning_rate": 9.973159532975691e-06,
"loss": 0.9743,
"step": 214
},
{
"epoch": 19.636363636363637,
"grad_norm": 0.6686264872550964,
"learning_rate": 9.970925928158275e-06,
"loss": 0.9398,
"step": 216
},
{
"epoch": 19.818181818181817,
"grad_norm": 0.6576296091079712,
"learning_rate": 9.968603332876435e-06,
"loss": 0.9251,
"step": 218
},
{
"epoch": 20.0,
"grad_norm": 0.5890394449234009,
"learning_rate": 9.966191788709716e-06,
"loss": 0.8669,
"step": 220
},
{
"epoch": 20.0,
"eval_loss": 0.9538940787315369,
"eval_runtime": 9.6225,
"eval_samples_per_second": 2.494,
"eval_steps_per_second": 2.494,
"step": 220
},
{
"epoch": 20.181818181818183,
"grad_norm": 0.7149389386177063,
"learning_rate": 9.963691338830045e-06,
"loss": 0.9322,
"step": 222
},
{
"epoch": 20.363636363636363,
"grad_norm": 0.6404098272323608,
"learning_rate": 9.961102028000948e-06,
"loss": 0.8914,
"step": 224
},
{
"epoch": 20.545454545454547,
"grad_norm": 0.6958956122398376,
"learning_rate": 9.958423902576764e-06,
"loss": 0.9489,
"step": 226
},
{
"epoch": 20.727272727272727,
"grad_norm": 0.6515837907791138,
"learning_rate": 9.955657010501807e-06,
"loss": 0.8477,
"step": 228
},
{
"epoch": 20.90909090909091,
"grad_norm": 0.8464462161064148,
"learning_rate": 9.952801401309504e-06,
"loss": 0.9253,
"step": 230
},
{
"epoch": 21.0,
"eval_loss": 0.9454106688499451,
"eval_runtime": 9.6279,
"eval_samples_per_second": 2.493,
"eval_steps_per_second": 2.493,
"step": 231
},
{
"epoch": 21.09090909090909,
"grad_norm": 0.7701188325881958,
"learning_rate": 9.949857126121519e-06,
"loss": 0.9027,
"step": 232
},
{
"epoch": 21.272727272727273,
"grad_norm": 0.6819570660591125,
"learning_rate": 9.946824237646823e-06,
"loss": 0.8545,
"step": 234
},
{
"epoch": 21.454545454545453,
"grad_norm": 0.7400511503219604,
"learning_rate": 9.94370279018077e-06,
"loss": 0.9229,
"step": 236
},
{
"epoch": 21.636363636363637,
"grad_norm": 0.6731954216957092,
"learning_rate": 9.940492839604103e-06,
"loss": 0.8596,
"step": 238
},
{
"epoch": 21.818181818181817,
"grad_norm": 0.7251449227333069,
"learning_rate": 9.937194443381972e-06,
"loss": 0.9052,
"step": 240
},
{
"epoch": 22.0,
"grad_norm": 0.7023078799247742,
"learning_rate": 9.933807660562898e-06,
"loss": 0.872,
"step": 242
},
{
"epoch": 22.0,
"eval_loss": 0.9398066997528076,
"eval_runtime": 9.6205,
"eval_samples_per_second": 2.495,
"eval_steps_per_second": 2.495,
"step": 242
},
{
"epoch": 22.181818181818183,
"grad_norm": 0.6639079451560974,
"learning_rate": 9.930332551777709e-06,
"loss": 0.8602,
"step": 244
},
{
"epoch": 22.363636363636363,
"grad_norm": 0.8378622531890869,
"learning_rate": 9.926769179238467e-06,
"loss": 0.9034,
"step": 246
},
{
"epoch": 22.545454545454547,
"grad_norm": 0.8481977581977844,
"learning_rate": 9.923117606737347e-06,
"loss": 0.8493,
"step": 248
},
{
"epoch": 22.727272727272727,
"grad_norm": 0.8623536825180054,
"learning_rate": 9.919377899645497e-06,
"loss": 0.8883,
"step": 250
},
{
"epoch": 22.90909090909091,
"grad_norm": 0.8115867376327515,
"learning_rate": 9.915550124911866e-06,
"loss": 0.8824,
"step": 252
},
{
"epoch": 23.0,
"eval_loss": 0.9327845573425293,
"eval_runtime": 9.6274,
"eval_samples_per_second": 2.493,
"eval_steps_per_second": 2.493,
"step": 253
},
{
"epoch": 23.09090909090909,
"grad_norm": 0.7263345122337341,
"learning_rate": 9.91163435106201e-06,
"loss": 0.8292,
"step": 254
},
{
"epoch": 23.272727272727273,
"grad_norm": 0.8343206644058228,
"learning_rate": 9.907630648196857e-06,
"loss": 0.9044,
"step": 256
},
{
"epoch": 23.454545454545453,
"grad_norm": 0.7706195116043091,
"learning_rate": 9.903539087991462e-06,
"loss": 0.8363,
"step": 258
},
{
"epoch": 23.636363636363637,
"grad_norm": 0.7953319549560547,
"learning_rate": 9.899359743693715e-06,
"loss": 0.8528,
"step": 260
},
{
"epoch": 23.818181818181817,
"grad_norm": 0.8244236707687378,
"learning_rate": 9.895092690123036e-06,
"loss": 0.8787,
"step": 262
},
{
"epoch": 24.0,
"grad_norm": 0.9236502647399902,
"learning_rate": 9.890738003669029e-06,
"loss": 0.8582,
"step": 264
},
{
"epoch": 24.0,
"eval_loss": 0.9282767176628113,
"eval_runtime": 9.6299,
"eval_samples_per_second": 2.492,
"eval_steps_per_second": 2.492,
"step": 264
},
{
"epoch": 24.181818181818183,
"grad_norm": 0.7912314534187317,
"learning_rate": 9.886295762290125e-06,
"loss": 0.8291,
"step": 266
},
{
"epoch": 24.363636363636363,
"grad_norm": 0.8443674445152283,
"learning_rate": 9.881766045512176e-06,
"loss": 0.8599,
"step": 268
},
{
"epoch": 24.545454545454547,
"grad_norm": 0.9810454249382019,
"learning_rate": 9.877148934427037e-06,
"loss": 0.844,
"step": 270
},
{
"epoch": 24.727272727272727,
"grad_norm": 0.827754020690918,
"learning_rate": 9.872444511691108e-06,
"loss": 0.8447,
"step": 272
},
{
"epoch": 24.90909090909091,
"grad_norm": 0.9015039801597595,
"learning_rate": 9.867652861523866e-06,
"loss": 0.8763,
"step": 274
},
{
"epoch": 25.0,
"eval_loss": 0.9221246838569641,
"eval_runtime": 9.6209,
"eval_samples_per_second": 2.495,
"eval_steps_per_second": 2.495,
"step": 275
},
{
"epoch": 25.09090909090909,
"grad_norm": 0.8841392993927002,
"learning_rate": 9.862774069706346e-06,
"loss": 0.8184,
"step": 276
},
{
"epoch": 25.272727272727273,
"grad_norm": 0.8848661780357361,
"learning_rate": 9.85780822357961e-06,
"loss": 0.8496,
"step": 278
},
{
"epoch": 25.454545454545453,
"grad_norm": 0.7846701145172119,
"learning_rate": 9.85275541204318e-06,
"loss": 0.8165,
"step": 280
},
{
"epoch": 25.636363636363637,
"grad_norm": 1.0641032457351685,
"learning_rate": 9.847615725553457e-06,
"loss": 0.8485,
"step": 282
},
{
"epoch": 25.818181818181817,
"grad_norm": 0.9293933510780334,
"learning_rate": 9.842389256122086e-06,
"loss": 0.829,
"step": 284
},
{
"epoch": 26.0,
"grad_norm": 0.8131471872329712,
"learning_rate": 9.83707609731432e-06,
"loss": 0.8199,
"step": 286
},
{
"epoch": 26.0,
"eval_loss": 0.917729377746582,
"eval_runtime": 9.6285,
"eval_samples_per_second": 2.493,
"eval_steps_per_second": 2.493,
"step": 286
},
{
"epoch": 26.181818181818183,
"grad_norm": 0.8052296042442322,
"learning_rate": 9.831676344247343e-06,
"loss": 0.8356,
"step": 288
},
{
"epoch": 26.363636363636363,
"grad_norm": 0.9019801020622253,
"learning_rate": 9.826190093588564e-06,
"loss": 0.8427,
"step": 290
},
{
"epoch": 26.545454545454547,
"grad_norm": 0.8782145380973816,
"learning_rate": 9.820617443553889e-06,
"loss": 0.8175,
"step": 292
},
{
"epoch": 26.727272727272727,
"grad_norm": 1.0153470039367676,
"learning_rate": 9.814958493905962e-06,
"loss": 0.8059,
"step": 294
},
{
"epoch": 26.90909090909091,
"grad_norm": 0.9501240253448486,
"learning_rate": 9.80921334595238e-06,
"loss": 0.7986,
"step": 296
},
{
"epoch": 27.0,
"eval_loss": 0.9146122336387634,
"eval_runtime": 9.6275,
"eval_samples_per_second": 2.493,
"eval_steps_per_second": 2.493,
"step": 297
},
{
"epoch": 27.09090909090909,
"grad_norm": 0.8864769339561462,
"learning_rate": 9.80338210254388e-06,
"loss": 0.7819,
"step": 298
},
{
"epoch": 27.272727272727273,
"grad_norm": 0.8944150805473328,
"learning_rate": 9.797464868072489e-06,
"loss": 0.7875,
"step": 300
},
{
"epoch": 27.454545454545453,
"grad_norm": 0.9001815915107727,
"learning_rate": 9.791461748469669e-06,
"loss": 0.8734,
"step": 302
},
{
"epoch": 27.636363636363637,
"grad_norm": 0.9148157835006714,
"learning_rate": 9.785372851204415e-06,
"loss": 0.8076,
"step": 304
},
{
"epoch": 27.818181818181817,
"grad_norm": 1.0519267320632935,
"learning_rate": 9.779198285281326e-06,
"loss": 0.7918,
"step": 306
},
{
"epoch": 28.0,
"grad_norm": 0.8914878964424133,
"learning_rate": 9.77293816123866e-06,
"loss": 0.7754,
"step": 308
},
{
"epoch": 28.0,
"eval_loss": 0.9141943454742432,
"eval_runtime": 9.6378,
"eval_samples_per_second": 2.49,
"eval_steps_per_second": 2.49,
"step": 308
},
{
"epoch": 28.181818181818183,
"grad_norm": 0.9089798331260681,
"learning_rate": 9.766592591146353e-06,
"loss": 0.7795,
"step": 310
},
{
"epoch": 28.363636363636363,
"grad_norm": 1.054513931274414,
"learning_rate": 9.760161688604008e-06,
"loss": 0.7935,
"step": 312
},
{
"epoch": 28.545454545454547,
"grad_norm": 0.9446949362754822,
"learning_rate": 9.753645568738872e-06,
"loss": 0.797,
"step": 314
},
{
"epoch": 28.727272727272727,
"grad_norm": 0.9865514039993286,
"learning_rate": 9.747044348203766e-06,
"loss": 0.7824,
"step": 316
},
{
"epoch": 28.90909090909091,
"grad_norm": 1.0658093690872192,
"learning_rate": 9.740358145174999e-06,
"loss": 0.7893,
"step": 318
},
{
"epoch": 29.0,
"eval_loss": 0.9086329340934753,
"eval_runtime": 9.6391,
"eval_samples_per_second": 2.49,
"eval_steps_per_second": 2.49,
"step": 319
},
{
"epoch": 29.09090909090909,
"grad_norm": 0.9681562185287476,
"learning_rate": 9.733587079350254e-06,
"loss": 0.7736,
"step": 320
},
{
"epoch": 29.272727272727273,
"grad_norm": 0.9801912903785706,
"learning_rate": 9.72673127194644e-06,
"loss": 0.8747,
"step": 322
},
{
"epoch": 29.454545454545453,
"grad_norm": 1.076412558555603,
"learning_rate": 9.719790845697534e-06,
"loss": 0.7737,
"step": 324
},
{
"epoch": 29.636363636363637,
"grad_norm": 1.0102134943008423,
"learning_rate": 9.71276592485237e-06,
"loss": 0.7632,
"step": 326
},
{
"epoch": 29.818181818181817,
"grad_norm": 1.1170337200164795,
"learning_rate": 9.705656635172418e-06,
"loss": 0.7417,
"step": 328
},
{
"epoch": 30.0,
"grad_norm": 1.0282917022705078,
"learning_rate": 9.698463103929542e-06,
"loss": 0.7312,
"step": 330
},
{
"epoch": 30.0,
"eval_loss": 0.9086909294128418,
"eval_runtime": 9.6293,
"eval_samples_per_second": 2.492,
"eval_steps_per_second": 2.492,
"step": 330
},
{
"epoch": 30.181818181818183,
"grad_norm": 1.050240397453308,
"learning_rate": 9.69118545990371e-06,
"loss": 0.7415,
"step": 332
},
{
"epoch": 30.363636363636363,
"grad_norm": 1.090142011642456,
"learning_rate": 9.683823833380692e-06,
"loss": 0.7346,
"step": 334
},
{
"epoch": 30.545454545454547,
"grad_norm": 1.1241227388381958,
"learning_rate": 9.676378356149733e-06,
"loss": 0.7492,
"step": 336
},
{
"epoch": 30.727272727272727,
"grad_norm": 1.0831716060638428,
"learning_rate": 9.668849161501186e-06,
"loss": 0.7854,
"step": 338
},
{
"epoch": 30.90909090909091,
"grad_norm": 0.9916213154792786,
"learning_rate": 9.66123638422413e-06,
"loss": 0.7431,
"step": 340
},
{
"epoch": 31.0,
"eval_loss": 0.9049713611602783,
"eval_runtime": 9.6277,
"eval_samples_per_second": 2.493,
"eval_steps_per_second": 2.493,
"step": 341
},
{
"epoch": 31.09090909090909,
"grad_norm": 0.9737184047698975,
"learning_rate": 9.653540160603956e-06,
"loss": 0.8,
"step": 342
},
{
"epoch": 31.272727272727273,
"grad_norm": 1.002685308456421,
"learning_rate": 9.64576062841993e-06,
"loss": 0.7644,
"step": 344
},
{
"epoch": 31.454545454545453,
"grad_norm": 1.2203502655029297,
"learning_rate": 9.637897926942716e-06,
"loss": 0.7335,
"step": 346
},
{
"epoch": 31.636363636363637,
"grad_norm": 1.2266589403152466,
"learning_rate": 9.629952196931902e-06,
"loss": 0.7724,
"step": 348
},
{
"epoch": 31.818181818181817,
"grad_norm": 1.3024228811264038,
"learning_rate": 9.621923580633462e-06,
"loss": 0.7164,
"step": 350
},
{
"epoch": 32.0,
"grad_norm": 1.1638389825820923,
"learning_rate": 9.613812221777212e-06,
"loss": 0.7103,
"step": 352
},
{
"epoch": 32.0,
"eval_loss": 0.9036614298820496,
"eval_runtime": 9.6223,
"eval_samples_per_second": 2.494,
"eval_steps_per_second": 2.494,
"step": 352
},
{
"epoch": 32.18181818181818,
"grad_norm": 1.1456729173660278,
"learning_rate": 9.60561826557425e-06,
"loss": 0.7032,
"step": 354
},
{
"epoch": 32.36363636363637,
"grad_norm": 1.0486174821853638,
"learning_rate": 9.597341858714344e-06,
"loss": 0.7216,
"step": 356
},
{
"epoch": 32.54545454545455,
"grad_norm": 1.2181857824325562,
"learning_rate": 9.588983149363307e-06,
"loss": 0.7591,
"step": 358
},
{
"epoch": 32.72727272727273,
"grad_norm": 1.0791609287261963,
"learning_rate": 9.580542287160348e-06,
"loss": 0.7719,
"step": 360
},
{
"epoch": 32.90909090909091,
"grad_norm": 1.1914441585540771,
"learning_rate": 9.572019423215395e-06,
"loss": 0.6967,
"step": 362
},
{
"epoch": 33.0,
"eval_loss": 0.9091615080833435,
"eval_runtime": 9.619,
"eval_samples_per_second": 2.495,
"eval_steps_per_second": 2.495,
"step": 363
},
{
"epoch": 33.09090909090909,
"grad_norm": 1.206595778465271,
"learning_rate": 9.563414710106382e-06,
"loss": 0.6713,
"step": 364
},
{
"epoch": 33.27272727272727,
"grad_norm": 1.2539595365524292,
"learning_rate": 9.554728301876525e-06,
"loss": 0.7171,
"step": 366
},
{
"epoch": 33.45454545454545,
"grad_norm": 1.135501503944397,
"learning_rate": 9.545960354031564e-06,
"loss": 0.7324,
"step": 368
},
{
"epoch": 33.63636363636363,
"grad_norm": 1.2056708335876465,
"learning_rate": 9.537111023536973e-06,
"loss": 0.7101,
"step": 370
},
{
"epoch": 33.81818181818182,
"grad_norm": 1.1787803173065186,
"learning_rate": 9.528180468815155e-06,
"loss": 0.7468,
"step": 372
},
{
"epoch": 34.0,
"grad_norm": 1.490050196647644,
"learning_rate": 9.519168849742603e-06,
"loss": 0.6502,
"step": 374
},
{
"epoch": 34.0,
"eval_loss": 0.9070794582366943,
"eval_runtime": 9.63,
"eval_samples_per_second": 2.492,
"eval_steps_per_second": 2.492,
"step": 374
},
{
"epoch": 34.18181818181818,
"grad_norm": 1.2995131015777588,
"learning_rate": 9.510076327647043e-06,
"loss": 0.6144,
"step": 376
},
{
"epoch": 34.36363636363637,
"grad_norm": 1.3441383838653564,
"learning_rate": 9.50090306530454e-06,
"loss": 0.7332,
"step": 378
},
{
"epoch": 34.54545454545455,
"grad_norm": 1.1878114938735962,
"learning_rate": 9.491649226936586e-06,
"loss": 0.7056,
"step": 380
},
{
"epoch": 34.72727272727273,
"grad_norm": 1.3715388774871826,
"learning_rate": 9.48231497820716e-06,
"loss": 0.67,
"step": 382
},
{
"epoch": 34.90909090909091,
"grad_norm": 1.466930866241455,
"learning_rate": 9.47290048621977e-06,
"loss": 0.6659,
"step": 384
},
{
"epoch": 35.0,
"eval_loss": 0.9019081592559814,
"eval_runtime": 9.6208,
"eval_samples_per_second": 2.495,
"eval_steps_per_second": 2.495,
"step": 385
},
{
"epoch": 35.09090909090909,
"grad_norm": 1.4354052543640137,
"learning_rate": 9.46340591951444e-06,
"loss": 0.7589,
"step": 386
},
{
"epoch": 35.27272727272727,
"grad_norm": 1.2188067436218262,
"learning_rate": 9.453831448064717e-06,
"loss": 0.6215,
"step": 388
},
{
"epoch": 35.45454545454545,
"grad_norm": 1.3039608001708984,
"learning_rate": 9.444177243274619e-06,
"loss": 0.6502,
"step": 390
},
{
"epoch": 35.63636363636363,
"grad_norm": 1.3067864179611206,
"learning_rate": 9.434443477975557e-06,
"loss": 0.6884,
"step": 392
},
{
"epoch": 35.81818181818182,
"grad_norm": 1.33348548412323,
"learning_rate": 9.42463032642326e-06,
"loss": 0.6852,
"step": 394
},
{
"epoch": 36.0,
"grad_norm": 1.5278242826461792,
"learning_rate": 9.414737964294636e-06,
"loss": 0.7003,
"step": 396
},
{
"epoch": 36.0,
"eval_loss": 0.9015458226203918,
"eval_runtime": 9.6251,
"eval_samples_per_second": 2.493,
"eval_steps_per_second": 2.493,
"step": 396
},
{
"epoch": 36.18181818181818,
"grad_norm": 1.3526606559753418,
"learning_rate": 9.40476656868464e-06,
"loss": 0.6329,
"step": 398
},
{
"epoch": 36.36363636363637,
"grad_norm": 1.6108198165893555,
"learning_rate": 9.394716318103098e-06,
"loss": 0.6626,
"step": 400
},
{
"epoch": 36.54545454545455,
"grad_norm": 1.5193760395050049,
"learning_rate": 9.384587392471516e-06,
"loss": 0.6816,
"step": 402
},
{
"epoch": 36.72727272727273,
"grad_norm": 1.2977081537246704,
"learning_rate": 9.37437997311985e-06,
"loss": 0.6422,
"step": 404
},
{
"epoch": 36.90909090909091,
"grad_norm": 1.4797639846801758,
"learning_rate": 9.364094242783272e-06,
"loss": 0.629,
"step": 406
},
{
"epoch": 37.0,
"eval_loss": 0.9018394351005554,
"eval_runtime": 9.6212,
"eval_samples_per_second": 2.494,
"eval_steps_per_second": 2.494,
"step": 407
},
{
"epoch": 37.09090909090909,
"grad_norm": 1.2966375350952148,
"learning_rate": 9.353730385598887e-06,
"loss": 0.6654,
"step": 408
},
{
"epoch": 37.27272727272727,
"grad_norm": 1.3821908235549927,
"learning_rate": 9.343288587102444e-06,
"loss": 0.6054,
"step": 410
},
{
"epoch": 37.45454545454545,
"grad_norm": 1.3437285423278809,
"learning_rate": 9.332769034225012e-06,
"loss": 0.6449,
"step": 412
},
{
"epoch": 37.63636363636363,
"grad_norm": 1.4815727472305298,
"learning_rate": 9.322171915289635e-06,
"loss": 0.658,
"step": 414
},
{
"epoch": 37.81818181818182,
"grad_norm": 1.273409366607666,
"learning_rate": 9.311497420007955e-06,
"loss": 0.6629,
"step": 416
},
{
"epoch": 38.0,
"grad_norm": 1.4128069877624512,
"learning_rate": 9.30074573947683e-06,
"loss": 0.6299,
"step": 418
},
{
"epoch": 38.0,
"eval_loss": 0.9080850481987,
"eval_runtime": 9.6337,
"eval_samples_per_second": 2.491,
"eval_steps_per_second": 2.491,
"step": 418
},
{
"epoch": 38.18181818181818,
"grad_norm": 1.467628836631775,
"learning_rate": 9.289917066174887e-06,
"loss": 0.6516,
"step": 420
},
{
"epoch": 38.36363636363637,
"grad_norm": 1.4847270250320435,
"learning_rate": 9.279011593959107e-06,
"loss": 0.6093,
"step": 422
},
{
"epoch": 38.54545454545455,
"grad_norm": 1.7236661911010742,
"learning_rate": 9.268029518061335e-06,
"loss": 0.6272,
"step": 424
},
{
"epoch": 38.72727272727273,
"grad_norm": 1.563481330871582,
"learning_rate": 9.256971035084786e-06,
"loss": 0.6293,
"step": 426
},
{
"epoch": 38.90909090909091,
"grad_norm": 1.4896105527877808,
"learning_rate": 9.245836343000534e-06,
"loss": 0.6259,
"step": 428
},
{
"epoch": 39.0,
"eval_loss": 0.9161927103996277,
"eval_runtime": 9.6218,
"eval_samples_per_second": 2.494,
"eval_steps_per_second": 2.494,
"step": 429
},
{
"epoch": 39.09090909090909,
"grad_norm": 1.36057448387146,
"learning_rate": 9.234625641143962e-06,
"loss": 0.5801,
"step": 430
},
{
"epoch": 39.27272727272727,
"grad_norm": 1.5864324569702148,
"learning_rate": 9.223339130211194e-06,
"loss": 0.5785,
"step": 432
},
{
"epoch": 39.45454545454545,
"grad_norm": 1.609424352645874,
"learning_rate": 9.211977012255497e-06,
"loss": 0.5777,
"step": 434
},
{
"epoch": 39.63636363636363,
"grad_norm": 1.4554574489593506,
"learning_rate": 9.200539490683682e-06,
"loss": 0.5927,
"step": 436
},
{
"epoch": 39.81818181818182,
"grad_norm": 1.5687090158462524,
"learning_rate": 9.189026770252437e-06,
"loss": 0.6421,
"step": 438
},
{
"epoch": 40.0,
"grad_norm": 1.4782568216323853,
"learning_rate": 9.177439057064684e-06,
"loss": 0.6262,
"step": 440
},
{
"epoch": 40.0,
"eval_loss": 0.9212110638618469,
"eval_runtime": 9.6258,
"eval_samples_per_second": 2.493,
"eval_steps_per_second": 2.493,
"step": 440
},
{
"epoch": 40.18181818181818,
"grad_norm": 1.4512465000152588,
"learning_rate": 9.16577655856587e-06,
"loss": 0.5827,
"step": 442
},
{
"epoch": 40.36363636363637,
"grad_norm": 2.0493810176849365,
"learning_rate": 9.154039483540273e-06,
"loss": 0.5539,
"step": 444
},
{
"epoch": 40.54545454545455,
"grad_norm": 1.6594229936599731,
"learning_rate": 9.142228042107248e-06,
"loss": 0.5907,
"step": 446
},
{
"epoch": 40.72727272727273,
"grad_norm": 1.4589492082595825,
"learning_rate": 9.130342445717474e-06,
"loss": 0.6161,
"step": 448
},
{
"epoch": 40.90909090909091,
"grad_norm": 1.7103596925735474,
"learning_rate": 9.118382907149164e-06,
"loss": 0.5707,
"step": 450
},
{
"epoch": 41.0,
"eval_loss": 0.9212350845336914,
"eval_runtime": 9.6173,
"eval_samples_per_second": 2.495,
"eval_steps_per_second": 2.495,
"step": 451
},
{
"epoch": 41.09090909090909,
"grad_norm": 1.3275471925735474,
"learning_rate": 9.10634964050426e-06,
"loss": 0.6536,
"step": 452
},
{
"epoch": 41.27272727272727,
"grad_norm": 1.6759672164916992,
"learning_rate": 9.094242861204598e-06,
"loss": 0.5257,
"step": 454
},
{
"epoch": 41.45454545454545,
"grad_norm": 1.9791761636734009,
"learning_rate": 9.08206278598805e-06,
"loss": 0.5604,
"step": 456
},
{
"epoch": 41.63636363636363,
"grad_norm": 1.740043044090271,
"learning_rate": 9.069809632904647e-06,
"loss": 0.544,
"step": 458
},
{
"epoch": 41.81818181818182,
"grad_norm": 1.7312424182891846,
"learning_rate": 9.057483621312671e-06,
"loss": 0.5987,
"step": 460
},
{
"epoch": 42.0,
"grad_norm": 1.5688552856445312,
"learning_rate": 9.045084971874738e-06,
"loss": 0.5749,
"step": 462
},
{
"epoch": 42.0,
"eval_loss": 0.927391529083252,
"eval_runtime": 9.6244,
"eval_samples_per_second": 2.494,
"eval_steps_per_second": 2.494,
"step": 462
},
{
"epoch": 42.18181818181818,
"grad_norm": 1.69509756565094,
"learning_rate": 9.032613906553833e-06,
"loss": 0.6091,
"step": 464
},
{
"epoch": 42.36363636363637,
"grad_norm": 2.032390832901001,
"learning_rate": 9.020070648609347e-06,
"loss": 0.5022,
"step": 466
},
{
"epoch": 42.54545454545455,
"grad_norm": 1.9728387594223022,
"learning_rate": 9.007455422593077e-06,
"loss": 0.5395,
"step": 468
},
{
"epoch": 42.72727272727273,
"grad_norm": 1.7745451927185059,
"learning_rate": 8.994768454345207e-06,
"loss": 0.5662,
"step": 470
},
{
"epoch": 42.90909090909091,
"grad_norm": 1.9252541065216064,
"learning_rate": 8.982009970990262e-06,
"loss": 0.533,
"step": 472
},
{
"epoch": 43.0,
"eval_loss": 0.9368504881858826,
"eval_runtime": 9.6216,
"eval_samples_per_second": 2.494,
"eval_steps_per_second": 2.494,
"step": 473
},
{
"epoch": 43.0,
"step": 473,
"total_flos": 5.808369757165978e+16,
"train_loss": 1.1408403902678863,
"train_runtime": 4431.3714,
"train_samples_per_second": 2.979,
"train_steps_per_second": 0.372
}
],
"logging_steps": 2,
"max_steps": 1650,
"num_input_tokens_seen": 0,
"num_train_epochs": 150,
"save_steps": 25,
"stateful_callbacks": {
"EarlyStoppingCallback": {
"args": {
"early_stopping_patience": 7,
"early_stopping_threshold": 0.0
},
"attributes": {
"early_stopping_patience_counter": 0
}
},
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 5.808369757165978e+16,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}