|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 1.0, |
|
"eval_steps": 100, |
|
"global_step": 487, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.002053388090349076, |
|
"grad_norm": 14.228013038635254, |
|
"learning_rate": 4.0816326530612243e-07, |
|
"loss": 2.0477, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.01026694045174538, |
|
"grad_norm": 12.610700607299805, |
|
"learning_rate": 2.0408163265306125e-06, |
|
"loss": 2.1254, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.02053388090349076, |
|
"grad_norm": 8.27630615234375, |
|
"learning_rate": 4.081632653061225e-06, |
|
"loss": 1.8104, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.030800821355236138, |
|
"grad_norm": 3.9820382595062256, |
|
"learning_rate": 6.122448979591837e-06, |
|
"loss": 1.4185, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.04106776180698152, |
|
"grad_norm": 2.656637668609619, |
|
"learning_rate": 8.16326530612245e-06, |
|
"loss": 1.2615, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.0513347022587269, |
|
"grad_norm": 2.308213949203491, |
|
"learning_rate": 1.0204081632653063e-05, |
|
"loss": 1.1945, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.061601642710472276, |
|
"grad_norm": 2.196038007736206, |
|
"learning_rate": 1.2244897959183674e-05, |
|
"loss": 1.118, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.07186858316221766, |
|
"grad_norm": 2.158684730529785, |
|
"learning_rate": 1.4285714285714287e-05, |
|
"loss": 1.1065, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.08213552361396304, |
|
"grad_norm": 2.052507162094116, |
|
"learning_rate": 1.63265306122449e-05, |
|
"loss": 1.0728, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.09240246406570841, |
|
"grad_norm": 2.01088285446167, |
|
"learning_rate": 1.836734693877551e-05, |
|
"loss": 1.0644, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.1026694045174538, |
|
"grad_norm": 2.1248321533203125, |
|
"learning_rate": 1.999974277115551e-05, |
|
"loss": 1.0271, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.11293634496919917, |
|
"grad_norm": 1.90983247756958, |
|
"learning_rate": 1.9990741151022302e-05, |
|
"loss": 0.99, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.12320328542094455, |
|
"grad_norm": 1.873200535774231, |
|
"learning_rate": 1.9968891318940332e-05, |
|
"loss": 0.9706, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.13347022587268995, |
|
"grad_norm": 2.0408380031585693, |
|
"learning_rate": 1.9934221374053538e-05, |
|
"loss": 0.9673, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.1437371663244353, |
|
"grad_norm": 1.8692070245742798, |
|
"learning_rate": 1.9886775902323405e-05, |
|
"loss": 0.9476, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.1540041067761807, |
|
"grad_norm": 1.8879870176315308, |
|
"learning_rate": 1.9826615919190886e-05, |
|
"loss": 0.8964, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.16427104722792607, |
|
"grad_norm": 2.0121207237243652, |
|
"learning_rate": 1.975381879110983e-05, |
|
"loss": 0.9109, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.17453798767967146, |
|
"grad_norm": 2.071239948272705, |
|
"learning_rate": 1.9668478136052776e-05, |
|
"loss": 0.8292, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.18480492813141683, |
|
"grad_norm": 1.7859665155410767, |
|
"learning_rate": 1.957070370311717e-05, |
|
"loss": 0.8361, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.19507186858316222, |
|
"grad_norm": 2.460575819015503, |
|
"learning_rate": 1.946062123138668e-05, |
|
"loss": 0.7816, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.2053388090349076, |
|
"grad_norm": 2.230203866958618, |
|
"learning_rate": 1.9338372288229253e-05, |
|
"loss": 0.8011, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.2053388090349076, |
|
"eval_loss": 1.2018204927444458, |
|
"eval_runtime": 19.1301, |
|
"eval_samples_per_second": 23.314, |
|
"eval_steps_per_second": 0.732, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.21560574948665298, |
|
"grad_norm": 1.8656007051467896, |
|
"learning_rate": 1.9204114087239806e-05, |
|
"loss": 0.726, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.22587268993839835, |
|
"grad_norm": 1.8082348108291626, |
|
"learning_rate": 1.9058019286061662e-05, |
|
"loss": 0.7246, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.23613963039014374, |
|
"grad_norm": 2.0330450534820557, |
|
"learning_rate": 1.890027576434677e-05, |
|
"loss": 0.7142, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.2464065708418891, |
|
"grad_norm": 1.9140502214431763, |
|
"learning_rate": 1.8731086382140228e-05, |
|
"loss": 0.7093, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.25667351129363447, |
|
"grad_norm": 2.4732396602630615, |
|
"learning_rate": 1.8550668718999873e-05, |
|
"loss": 0.6693, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.2669404517453799, |
|
"grad_norm": 2.632209062576294, |
|
"learning_rate": 1.8359254794186368e-05, |
|
"loss": 0.7145, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.27720739219712526, |
|
"grad_norm": 2.557025909423828, |
|
"learning_rate": 1.815709076828368e-05, |
|
"loss": 0.6378, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.2874743326488706, |
|
"grad_norm": 1.9707200527191162, |
|
"learning_rate": 1.7944436626633625e-05, |
|
"loss": 0.5618, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.29774127310061604, |
|
"grad_norm": 13.473517417907715, |
|
"learning_rate": 1.7721565844991643e-05, |
|
"loss": 0.5915, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.3080082135523614, |
|
"grad_norm": 3.342193603515625, |
|
"learning_rate": 1.748876503783373e-05, |
|
"loss": 0.584, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.3182751540041068, |
|
"grad_norm": 1.8444448709487915, |
|
"learning_rate": 1.7246333589766786e-05, |
|
"loss": 0.5936, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.32854209445585214, |
|
"grad_norm": 1.684865117073059, |
|
"learning_rate": 1.699458327051647e-05, |
|
"loss": 0.5379, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.33880903490759756, |
|
"grad_norm": 2.2392663955688477, |
|
"learning_rate": 1.6733837833987634e-05, |
|
"loss": 0.533, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.3490759753593429, |
|
"grad_norm": 1.850683569908142, |
|
"learning_rate": 1.6464432601912914e-05, |
|
"loss": 0.5305, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.3593429158110883, |
|
"grad_norm": 1.9396766424179077, |
|
"learning_rate": 1.6186714032625036e-05, |
|
"loss": 0.4685, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.36960985626283366, |
|
"grad_norm": 1.9056674242019653, |
|
"learning_rate": 1.5901039275507247e-05, |
|
"loss": 0.4402, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.3798767967145791, |
|
"grad_norm": 1.9246410131454468, |
|
"learning_rate": 1.560777571169498e-05, |
|
"loss": 0.4796, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.39014373716632444, |
|
"grad_norm": 1.6989043951034546, |
|
"learning_rate": 1.5307300481619334e-05, |
|
"loss": 0.4851, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.4004106776180698, |
|
"grad_norm": 2.2395427227020264, |
|
"learning_rate": 1.5000000000000002e-05, |
|
"loss": 0.436, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.4106776180698152, |
|
"grad_norm": 1.839769721031189, |
|
"learning_rate": 1.4686269458911333e-05, |
|
"loss": 0.4058, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.4106776180698152, |
|
"eval_loss": 1.28135347366333, |
|
"eval_runtime": 19.6018, |
|
"eval_samples_per_second": 22.753, |
|
"eval_steps_per_second": 0.714, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.4209445585215606, |
|
"grad_norm": 2.4135115146636963, |
|
"learning_rate": 1.4366512319560642e-05, |
|
"loss": 0.3935, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.43121149897330596, |
|
"grad_norm": 1.8663309812545776, |
|
"learning_rate": 1.4041139793432274e-05, |
|
"loss": 0.3793, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.4414784394250513, |
|
"grad_norm": 1.6783833503723145, |
|
"learning_rate": 1.3710570313464778e-05, |
|
"loss": 0.3641, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.4517453798767967, |
|
"grad_norm": 1.7463020086288452, |
|
"learning_rate": 1.3375228995941135e-05, |
|
"loss": 0.366, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.4620123203285421, |
|
"grad_norm": 1.7100086212158203, |
|
"learning_rate": 1.3035547093784187e-05, |
|
"loss": 0.3338, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.4722792607802875, |
|
"grad_norm": 1.5141621828079224, |
|
"learning_rate": 1.2691961441960238e-05, |
|
"loss": 0.3076, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.48254620123203285, |
|
"grad_norm": 1.7006425857543945, |
|
"learning_rate": 1.2344913895704099e-05, |
|
"loss": 0.3131, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.4928131416837782, |
|
"grad_norm": 1.575813889503479, |
|
"learning_rate": 1.1994850762287988e-05, |
|
"loss": 0.2957, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.5030800821355236, |
|
"grad_norm": 1.5434437990188599, |
|
"learning_rate": 1.164222222706509e-05, |
|
"loss": 0.2995, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.5133470225872689, |
|
"grad_norm": 1.7136874198913574, |
|
"learning_rate": 1.128748177452581e-05, |
|
"loss": 0.2937, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.5236139630390144, |
|
"grad_norm": 1.378823161125183, |
|
"learning_rate": 1.0931085605111354e-05, |
|
"loss": 0.2605, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.5338809034907598, |
|
"grad_norm": 1.536054015159607, |
|
"learning_rate": 1.0573492048534515e-05, |
|
"loss": 0.27, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.5441478439425051, |
|
"grad_norm": 1.396335482597351, |
|
"learning_rate": 1.0215160974362224e-05, |
|
"loss": 0.2791, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.5544147843942505, |
|
"grad_norm": 3.4254772663116455, |
|
"learning_rate": 9.856553200617805e-06, |
|
"loss": 0.2783, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.5646817248459959, |
|
"grad_norm": 1.5541990995407104, |
|
"learning_rate": 9.49812990116353e-06, |
|
"loss": 0.269, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.5749486652977412, |
|
"grad_norm": 1.3364843130111694, |
|
"learning_rate": 9.140352012625538e-06, |
|
"loss": 0.225, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.5852156057494866, |
|
"grad_norm": 1.2611546516418457, |
|
"learning_rate": 8.783679641623845e-06, |
|
"loss": 0.209, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.5954825462012321, |
|
"grad_norm": 1.3054827451705933, |
|
"learning_rate": 8.428571473069775e-06, |
|
"loss": 0.1905, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.6057494866529775, |
|
"grad_norm": 1.3551859855651855, |
|
"learning_rate": 8.075484180291702e-06, |
|
"loss": 0.2282, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.6160164271047228, |
|
"grad_norm": 1.5083565711975098, |
|
"learning_rate": 7.724871837747708e-06, |
|
"loss": 0.2067, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.6160164271047228, |
|
"eval_loss": 1.3847053050994873, |
|
"eval_runtime": 19.5099, |
|
"eval_samples_per_second": 22.86, |
|
"eval_steps_per_second": 0.718, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.6262833675564682, |
|
"grad_norm": 1.1781014204025269, |
|
"learning_rate": 7.377185337080443e-06, |
|
"loss": 0.1895, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.6365503080082136, |
|
"grad_norm": 1.429604411125183, |
|
"learning_rate": 7.032871807265097e-06, |
|
"loss": 0.1929, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.6468172484599589, |
|
"grad_norm": 1.2146339416503906, |
|
"learning_rate": 6.692374039596241e-06, |
|
"loss": 0.1842, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.6570841889117043, |
|
"grad_norm": 1.354224681854248, |
|
"learning_rate": 6.356129918252927e-06, |
|
"loss": 0.1798, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.6673511293634496, |
|
"grad_norm": 1.2704695463180542, |
|
"learning_rate": 6.024571857174443e-06, |
|
"loss": 0.1636, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.6776180698151951, |
|
"grad_norm": 1.0078154802322388, |
|
"learning_rate": 5.698126243970845e-06, |
|
"loss": 0.1472, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.6878850102669405, |
|
"grad_norm": 1.2621588706970215, |
|
"learning_rate": 5.377212891583419e-06, |
|
"loss": 0.1716, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.6981519507186859, |
|
"grad_norm": 1.163734793663025, |
|
"learning_rate": 5.062244498400228e-06, |
|
"loss": 0.1595, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.7084188911704312, |
|
"grad_norm": 1.1806224584579468, |
|
"learning_rate": 4.753626117521103e-06, |
|
"loss": 0.1674, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.7186858316221766, |
|
"grad_norm": 1.2119107246398926, |
|
"learning_rate": 4.4517546358545175e-06, |
|
"loss": 0.1453, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.728952772073922, |
|
"grad_norm": 1.1235816478729248, |
|
"learning_rate": 4.1570182637163155e-06, |
|
"loss": 0.15, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.7392197125256673, |
|
"grad_norm": 1.1243553161621094, |
|
"learning_rate": 3.869796035586625e-06, |
|
"loss": 0.1426, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.7494866529774127, |
|
"grad_norm": 1.1293926239013672, |
|
"learning_rate": 3.590457322666997e-06, |
|
"loss": 0.1401, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.7597535934291582, |
|
"grad_norm": 1.0441910028457642, |
|
"learning_rate": 3.3193613578646633e-06, |
|
"loss": 0.1359, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.7700205338809035, |
|
"grad_norm": 1.05973219871521, |
|
"learning_rate": 3.0568567738147505e-06, |
|
"loss": 0.1491, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.7802874743326489, |
|
"grad_norm": 1.1299043893814087, |
|
"learning_rate": 2.8032811545345294e-06, |
|
"loss": 0.1342, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.7905544147843943, |
|
"grad_norm": 1.086897373199463, |
|
"learning_rate": 2.5589606012863968e-06, |
|
"loss": 0.1379, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.8008213552361396, |
|
"grad_norm": 1.0586342811584473, |
|
"learning_rate": 2.324209313207736e-06, |
|
"loss": 0.1281, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.811088295687885, |
|
"grad_norm": 1.0056428909301758, |
|
"learning_rate": 2.099329183247126e-06, |
|
"loss": 0.1157, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.8213552361396304, |
|
"grad_norm": 0.9998786449432373, |
|
"learning_rate": 1.8846094099263911e-06, |
|
"loss": 0.1183, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.8213552361396304, |
|
"eval_loss": 1.4414706230163574, |
|
"eval_runtime": 19.2041, |
|
"eval_samples_per_second": 23.224, |
|
"eval_steps_per_second": 0.729, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.8316221765913757, |
|
"grad_norm": 0.9717850089073181, |
|
"learning_rate": 1.6803261254278635e-06, |
|
"loss": 0.1143, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.8418891170431212, |
|
"grad_norm": 1.070147156715393, |
|
"learning_rate": 1.4867420404851306e-06, |
|
"loss": 0.1197, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.8521560574948666, |
|
"grad_norm": 0.9136067628860474, |
|
"learning_rate": 1.30410610653389e-06, |
|
"loss": 0.1166, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 0.8624229979466119, |
|
"grad_norm": 1.1148228645324707, |
|
"learning_rate": 1.1326531955574526e-06, |
|
"loss": 0.1182, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.8726899383983573, |
|
"grad_norm": 1.0000271797180176, |
|
"learning_rate": 9.72603798038574e-07, |
|
"loss": 0.1209, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.8829568788501027, |
|
"grad_norm": 0.8131821751594543, |
|
"learning_rate": 8.241637394060619e-07, |
|
"loss": 0.1107, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.893223819301848, |
|
"grad_norm": 0.8392049074172974, |
|
"learning_rate": 6.875239153408541e-07, |
|
"loss": 0.1053, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.9034907597535934, |
|
"grad_norm": 1.0311695337295532, |
|
"learning_rate": 5.628600462818668e-07, |
|
"loss": 0.1053, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.9137577002053389, |
|
"grad_norm": 0.8752442002296448, |
|
"learning_rate": 4.503324514474483e-07, |
|
"loss": 0.1013, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 0.9240246406570842, |
|
"grad_norm": 0.9375615119934082, |
|
"learning_rate": 3.500858426629439e-07, |
|
"loss": 0.1033, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.9342915811088296, |
|
"grad_norm": 0.8648166060447693, |
|
"learning_rate": 2.6224913825956933e-07, |
|
"loss": 0.1098, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 0.944558521560575, |
|
"grad_norm": 0.9131202697753906, |
|
"learning_rate": 1.869352972839067e-07, |
|
"loss": 0.1073, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.9548254620123203, |
|
"grad_norm": 0.9665296077728271, |
|
"learning_rate": 1.242411742312233e-07, |
|
"loss": 0.0966, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 0.9650924024640657, |
|
"grad_norm": 0.8326826691627502, |
|
"learning_rate": 7.42473944894384e-08, |
|
"loss": 0.1107, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.9753593429158111, |
|
"grad_norm": 0.8301469683647156, |
|
"learning_rate": 3.701825065392184e-08, |
|
"loss": 0.1123, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.9856262833675564, |
|
"grad_norm": 0.9057932496070862, |
|
"learning_rate": 1.2601619846444035e-08, |
|
"loss": 0.0985, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.9958932238193019, |
|
"grad_norm": 0.7701483964920044, |
|
"learning_rate": 1.0289021446308057e-09, |
|
"loss": 0.0917, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"step": 487, |
|
"total_flos": 1.4371645474744566e+18, |
|
"train_loss": 0.45526161300328233, |
|
"train_runtime": 3331.5803, |
|
"train_samples_per_second": 4.672, |
|
"train_steps_per_second": 0.146 |
|
} |
|
], |
|
"logging_steps": 5, |
|
"max_steps": 487, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1.4371645474744566e+18, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|