{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.4342162396873643, "eval_steps": 500, "global_step": 1000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0004342162396873643, "grad_norm": 0.9606487154960632, "learning_rate": 5e-05, "loss": 9.188, "step": 1 }, { "epoch": 0.0008684324793747286, "grad_norm": 0.9193044304847717, "learning_rate": 0.0001, "loss": 9.1897, "step": 2 }, { "epoch": 0.0013026487190620929, "grad_norm": 0.9406704306602478, "learning_rate": 0.00015, "loss": 9.0726, "step": 3 }, { "epoch": 0.0017368649587494573, "grad_norm": 0.9401040077209473, "learning_rate": 0.0002, "loss": 8.7129, "step": 4 }, { "epoch": 0.0021710811984368217, "grad_norm": 0.8919561505317688, "learning_rate": 0.00025, "loss": 8.1364, "step": 5 }, { "epoch": 0.0026052974381241857, "grad_norm": 0.8310884237289429, "learning_rate": 0.0003, "loss": 7.4088, "step": 6 }, { "epoch": 0.00303951367781155, "grad_norm": 0.748382031917572, "learning_rate": 0.00035, "loss": 6.6918, "step": 7 }, { "epoch": 0.0034737299174989146, "grad_norm": 0.6353039145469666, "learning_rate": 0.0004, "loss": 5.8753, "step": 8 }, { "epoch": 0.0039079461571862786, "grad_norm": 0.5517627000808716, "learning_rate": 0.00045000000000000004, "loss": 5.4347, "step": 9 }, { "epoch": 0.004342162396873643, "grad_norm": 0.5389806628227234, "learning_rate": 0.0005, "loss": 5.0181, "step": 10 }, { "epoch": 0.004776378636561007, "grad_norm": 0.6018754243850708, "learning_rate": 0.00055, "loss": 4.9487, "step": 11 }, { "epoch": 0.005210594876248371, "grad_norm": 0.6734316349029541, "learning_rate": 0.0006, "loss": 4.9233, "step": 12 }, { "epoch": 0.005644811115935736, "grad_norm": 0.6842392086982727, "learning_rate": 0.0006500000000000001, "loss": 4.8078, "step": 13 }, { "epoch": 0.0060790273556231, "grad_norm": 0.6519020199775696, "learning_rate": 0.0007, "loss": 4.6065, "step": 14 }, { "epoch": 0.006513243595310464, "grad_norm": 0.5788787603378296, "learning_rate": 0.00075, "loss": 4.3967, "step": 15 }, { "epoch": 0.006947459834997829, "grad_norm": 0.5370550751686096, "learning_rate": 0.0008, "loss": 4.1971, "step": 16 }, { "epoch": 0.007381676074685193, "grad_norm": 0.4862042963504791, "learning_rate": 0.00085, "loss": 3.8583, "step": 17 }, { "epoch": 0.007815892314372557, "grad_norm": 0.4636250436306, "learning_rate": 0.0009000000000000001, "loss": 3.6372, "step": 18 }, { "epoch": 0.008250108554059922, "grad_norm": 0.4613461494445801, "learning_rate": 0.00095, "loss": 3.5642, "step": 19 }, { "epoch": 0.008684324793747287, "grad_norm": 0.4593052566051483, "learning_rate": 0.001, "loss": 3.4463, "step": 20 }, { "epoch": 0.00911854103343465, "grad_norm": 0.4367975890636444, "learning_rate": 0.0009999974308631955, "loss": 3.2621, "step": 21 }, { "epoch": 0.009552757273122015, "grad_norm": 0.41558611392974854, "learning_rate": 0.000999989723479183, "loss": 3.0598, "step": 22 }, { "epoch": 0.00998697351280938, "grad_norm": 0.3967452049255371, "learning_rate": 0.0009999768779271685, "loss": 2.936, "step": 23 }, { "epoch": 0.010421189752496743, "grad_norm": 0.3664601147174835, "learning_rate": 0.0009999588943391596, "loss": 2.6913, "step": 24 }, { "epoch": 0.010855405992184108, "grad_norm": 0.359183669090271, "learning_rate": 0.0009999357728999656, "loss": 2.5639, "step": 25 }, { "epoch": 0.011289622231871473, "grad_norm": 0.3342216908931732, "learning_rate": 0.000999907513847195, "loss": 2.401, "step": 26 }, { "epoch": 0.011723838471558836, "grad_norm": 0.36653760075569153, "learning_rate": 0.0009998741174712534, "loss": 2.6216, "step": 27 }, { "epoch": 0.0121580547112462, "grad_norm": 0.30254194140434265, "learning_rate": 0.00099983558411534, "loss": 2.2086, "step": 28 }, { "epoch": 0.012592270950933565, "grad_norm": 0.2917495369911194, "learning_rate": 0.0009997919141754449, "loss": 2.1274, "step": 29 }, { "epoch": 0.013026487190620929, "grad_norm": 0.2690434455871582, "learning_rate": 0.000999743108100344, "loss": 2.0581, "step": 30 }, { "epoch": 0.013460703430308293, "grad_norm": 0.26215028762817383, "learning_rate": 0.0009996891663915954, "loss": 1.9764, "step": 31 }, { "epoch": 0.013894919669995658, "grad_norm": 0.2453673779964447, "learning_rate": 0.0009996300896035338, "loss": 1.9219, "step": 32 }, { "epoch": 0.014329135909683021, "grad_norm": 0.24287743866443634, "learning_rate": 0.0009995658783432644, "loss": 1.8433, "step": 33 }, { "epoch": 0.014763352149370386, "grad_norm": 0.2258252203464508, "learning_rate": 0.0009994965332706573, "loss": 1.7392, "step": 34 }, { "epoch": 0.015197568389057751, "grad_norm": 0.21900461614131927, "learning_rate": 0.0009994220550983404, "loss": 1.7185, "step": 35 }, { "epoch": 0.015631784628745114, "grad_norm": 0.21046680212020874, "learning_rate": 0.0009993424445916923, "loss": 1.6541, "step": 36 }, { "epoch": 0.01606600086843248, "grad_norm": 0.19736959040164948, "learning_rate": 0.0009992577025688338, "loss": 1.6321, "step": 37 }, { "epoch": 0.016500217108119844, "grad_norm": 0.19068720936775208, "learning_rate": 0.0009991678299006205, "loss": 1.6236, "step": 38 }, { "epoch": 0.01693443334780721, "grad_norm": 0.18818727135658264, "learning_rate": 0.000999072827510633, "loss": 1.5791, "step": 39 }, { "epoch": 0.017368649587494574, "grad_norm": 0.17485196888446808, "learning_rate": 0.0009989726963751681, "loss": 1.6086, "step": 40 }, { "epoch": 0.017802865827181935, "grad_norm": 0.1762186735868454, "learning_rate": 0.000998867437523228, "loss": 1.5328, "step": 41 }, { "epoch": 0.0182370820668693, "grad_norm": 0.19658854603767395, "learning_rate": 0.0009987570520365104, "loss": 1.4491, "step": 42 }, { "epoch": 0.018671298306556665, "grad_norm": 0.1639322191476822, "learning_rate": 0.0009986415410493965, "loss": 1.4447, "step": 43 }, { "epoch": 0.01910551454624403, "grad_norm": 0.1552889496088028, "learning_rate": 0.000998520905748941, "loss": 1.4345, "step": 44 }, { "epoch": 0.019539730785931395, "grad_norm": 0.14960208535194397, "learning_rate": 0.0009983951473748578, "loss": 1.364, "step": 45 }, { "epoch": 0.01997394702561876, "grad_norm": 0.14552171528339386, "learning_rate": 0.0009982642672195091, "loss": 1.3422, "step": 46 }, { "epoch": 0.02040816326530612, "grad_norm": 0.14638382196426392, "learning_rate": 0.0009981282666278908, "loss": 1.325, "step": 47 }, { "epoch": 0.020842379504993486, "grad_norm": 0.1372004598379135, "learning_rate": 0.0009979871469976197, "loss": 1.2685, "step": 48 }, { "epoch": 0.02127659574468085, "grad_norm": 0.12978093326091766, "learning_rate": 0.0009978409097789176, "loss": 1.3268, "step": 49 }, { "epoch": 0.021710811984368215, "grad_norm": 0.13076020777225494, "learning_rate": 0.0009976895564745992, "loss": 1.2211, "step": 50 }, { "epoch": 0.02214502822405558, "grad_norm": 0.1264386773109436, "learning_rate": 0.000997533088640053, "loss": 1.2597, "step": 51 }, { "epoch": 0.022579244463742945, "grad_norm": 0.12510421872138977, "learning_rate": 0.0009973715078832287, "loss": 1.178, "step": 52 }, { "epoch": 0.02301346070343031, "grad_norm": 0.13376790285110474, "learning_rate": 0.0009972048158646183, "loss": 1.1614, "step": 53 }, { "epoch": 0.02344767694311767, "grad_norm": 0.11965133249759674, "learning_rate": 0.00099703301429724, "loss": 1.1549, "step": 54 }, { "epoch": 0.023881893182805036, "grad_norm": 0.10872318595647812, "learning_rate": 0.0009968561049466214, "loss": 1.143, "step": 55 }, { "epoch": 0.0243161094224924, "grad_norm": 0.12483104318380356, "learning_rate": 0.000996674089630779, "loss": 1.2124, "step": 56 }, { "epoch": 0.024750325662179766, "grad_norm": 0.10931334644556046, "learning_rate": 0.0009964869702202021, "loss": 1.1176, "step": 57 }, { "epoch": 0.02518454190186713, "grad_norm": 0.12575986981391907, "learning_rate": 0.0009962947486378325, "loss": 1.0858, "step": 58 }, { "epoch": 0.025618758141554496, "grad_norm": 0.11589810997247696, "learning_rate": 0.0009960974268590438, "loss": 1.0867, "step": 59 }, { "epoch": 0.026052974381241857, "grad_norm": 0.10477310419082642, "learning_rate": 0.000995895006911623, "loss": 1.1162, "step": 60 }, { "epoch": 0.026487190620929222, "grad_norm": 0.09671472758054733, "learning_rate": 0.0009956874908757481, "loss": 1.0662, "step": 61 }, { "epoch": 0.026921406860616587, "grad_norm": 0.10130201280117035, "learning_rate": 0.0009954748808839674, "loss": 1.0826, "step": 62 }, { "epoch": 0.02735562310030395, "grad_norm": 0.10421687364578247, "learning_rate": 0.0009952571791211776, "loss": 1.0302, "step": 63 }, { "epoch": 0.027789839339991317, "grad_norm": 0.09363257884979248, "learning_rate": 0.0009950343878246009, "loss": 1.0769, "step": 64 }, { "epoch": 0.02822405557967868, "grad_norm": 0.10000266879796982, "learning_rate": 0.000994806509283763, "loss": 1.0048, "step": 65 }, { "epoch": 0.028658271819366043, "grad_norm": 0.091400146484375, "learning_rate": 0.0009945735458404682, "loss": 0.9903, "step": 66 }, { "epoch": 0.029092488059053408, "grad_norm": 0.09233607351779938, "learning_rate": 0.0009943354998887763, "loss": 1.0563, "step": 67 }, { "epoch": 0.029526704298740773, "grad_norm": 0.0991148054599762, "learning_rate": 0.0009940923738749779, "loss": 1.0117, "step": 68 }, { "epoch": 0.029960920538428137, "grad_norm": 0.08926694095134735, "learning_rate": 0.0009938441702975688, "loss": 0.974, "step": 69 }, { "epoch": 0.030395136778115502, "grad_norm": 0.08934075385332108, "learning_rate": 0.0009935908917072251, "loss": 1.0283, "step": 70 }, { "epoch": 0.030829353017802867, "grad_norm": 0.11343932151794434, "learning_rate": 0.000993332540706776, "loss": 0.998, "step": 71 }, { "epoch": 0.03126356925749023, "grad_norm": 0.11343089491128922, "learning_rate": 0.0009930691199511775, "loss": 0.993, "step": 72 }, { "epoch": 0.0316977854971776, "grad_norm": 0.09531175345182419, "learning_rate": 0.000992800632147486, "loss": 1.0179, "step": 73 }, { "epoch": 0.03213200173686496, "grad_norm": 0.13536381721496582, "learning_rate": 0.0009925270800548284, "loss": 0.9539, "step": 74 }, { "epoch": 0.03256621797655232, "grad_norm": 0.09377475082874298, "learning_rate": 0.0009922484664843763, "loss": 0.9811, "step": 75 }, { "epoch": 0.03300043421623969, "grad_norm": 0.09711369127035141, "learning_rate": 0.0009919647942993148, "loss": 0.9223, "step": 76 }, { "epoch": 0.03343465045592705, "grad_norm": 0.11353003978729248, "learning_rate": 0.0009916760664148148, "loss": 0.9736, "step": 77 }, { "epoch": 0.03386886669561442, "grad_norm": 0.09144185483455658, "learning_rate": 0.0009913822857980019, "loss": 0.9576, "step": 78 }, { "epoch": 0.03430308293530178, "grad_norm": 0.08733920753002167, "learning_rate": 0.0009910834554679266, "loss": 0.9016, "step": 79 }, { "epoch": 0.03473729917498915, "grad_norm": 0.08889693766832352, "learning_rate": 0.0009907795784955327, "loss": 0.8981, "step": 80 }, { "epoch": 0.03517151541467651, "grad_norm": 0.081461600959301, "learning_rate": 0.0009904706580036264, "loss": 0.9246, "step": 81 }, { "epoch": 0.03560573165436387, "grad_norm": 0.08032266050577164, "learning_rate": 0.0009901566971668436, "loss": 0.901, "step": 82 }, { "epoch": 0.03603994789405124, "grad_norm": 0.08888774365186691, "learning_rate": 0.0009898376992116178, "loss": 0.916, "step": 83 }, { "epoch": 0.0364741641337386, "grad_norm": 0.08942156285047531, "learning_rate": 0.0009895136674161465, "loss": 0.9122, "step": 84 }, { "epoch": 0.03690838037342597, "grad_norm": 0.08162346482276917, "learning_rate": 0.0009891846051103576, "loss": 0.8893, "step": 85 }, { "epoch": 0.03734259661311333, "grad_norm": 0.08491238951683044, "learning_rate": 0.0009888505156758758, "loss": 0.8572, "step": 86 }, { "epoch": 0.0377768128528007, "grad_norm": 0.07837095111608505, "learning_rate": 0.0009885114025459864, "loss": 0.851, "step": 87 }, { "epoch": 0.03821102909248806, "grad_norm": 0.08016610145568848, "learning_rate": 0.0009881672692056021, "loss": 0.8985, "step": 88 }, { "epoch": 0.03864524533217542, "grad_norm": 0.07894952595233917, "learning_rate": 0.000987818119191225, "loss": 0.8719, "step": 89 }, { "epoch": 0.03907946157186279, "grad_norm": 0.07604355365037918, "learning_rate": 0.0009874639560909118, "loss": 0.883, "step": 90 }, { "epoch": 0.03951367781155015, "grad_norm": 0.07936527580022812, "learning_rate": 0.0009871047835442364, "loss": 0.8385, "step": 91 }, { "epoch": 0.03994789405123752, "grad_norm": 0.08259500563144684, "learning_rate": 0.0009867406052422524, "loss": 0.8862, "step": 92 }, { "epoch": 0.04038211029092488, "grad_norm": 0.09410865604877472, "learning_rate": 0.0009863714249274552, "loss": 0.8654, "step": 93 }, { "epoch": 0.04081632653061224, "grad_norm": 0.08195285499095917, "learning_rate": 0.000985997246393744, "loss": 0.8616, "step": 94 }, { "epoch": 0.04125054277029961, "grad_norm": 0.07515832781791687, "learning_rate": 0.000985618073486382, "loss": 0.858, "step": 95 }, { "epoch": 0.04168475900998697, "grad_norm": 0.07317376881837845, "learning_rate": 0.0009852339101019574, "loss": 0.8699, "step": 96 }, { "epoch": 0.04211897524967434, "grad_norm": 0.088593028485775, "learning_rate": 0.0009848447601883434, "loss": 0.8326, "step": 97 }, { "epoch": 0.0425531914893617, "grad_norm": 0.07599225640296936, "learning_rate": 0.0009844506277446577, "loss": 0.7984, "step": 98 }, { "epoch": 0.04298740772904907, "grad_norm": 0.08955889940261841, "learning_rate": 0.0009840515168212207, "loss": 0.8489, "step": 99 }, { "epoch": 0.04342162396873643, "grad_norm": 0.07896170765161514, "learning_rate": 0.0009836474315195148, "loss": 0.8133, "step": 100 }, { "epoch": 0.04385584020842379, "grad_norm": 0.08088196814060211, "learning_rate": 0.0009832383759921415, "loss": 0.8356, "step": 101 }, { "epoch": 0.04429005644811116, "grad_norm": 0.08699057996273041, "learning_rate": 0.0009828243544427796, "loss": 0.8226, "step": 102 }, { "epoch": 0.04472427268779852, "grad_norm": 0.08965849131345749, "learning_rate": 0.0009824053711261404, "loss": 0.8137, "step": 103 }, { "epoch": 0.04515848892748589, "grad_norm": 0.09515663236379623, "learning_rate": 0.0009819814303479266, "loss": 0.7937, "step": 104 }, { "epoch": 0.04559270516717325, "grad_norm": 0.06896204501390457, "learning_rate": 0.0009815525364647853, "loss": 0.8056, "step": 105 }, { "epoch": 0.04602692140686062, "grad_norm": 0.08081001043319702, "learning_rate": 0.0009811186938842645, "loss": 0.8465, "step": 106 }, { "epoch": 0.04646113764654798, "grad_norm": 0.07806787639856339, "learning_rate": 0.000980679907064768, "loss": 0.7897, "step": 107 }, { "epoch": 0.04689535388623534, "grad_norm": 0.07646839320659637, "learning_rate": 0.0009802361805155097, "loss": 0.7929, "step": 108 }, { "epoch": 0.04732957012592271, "grad_norm": 0.08139129728078842, "learning_rate": 0.000979787518796466, "loss": 0.8097, "step": 109 }, { "epoch": 0.04776378636561007, "grad_norm": 0.07595842331647873, "learning_rate": 0.0009793339265183304, "loss": 0.8028, "step": 110 }, { "epoch": 0.04819800260529744, "grad_norm": 0.07882706820964813, "learning_rate": 0.0009788754083424652, "loss": 0.7825, "step": 111 }, { "epoch": 0.0486322188449848, "grad_norm": 0.09278842061758041, "learning_rate": 0.0009784119689808543, "loss": 0.7897, "step": 112 }, { "epoch": 0.049066435084672164, "grad_norm": 0.07860872894525528, "learning_rate": 0.0009779436131960543, "loss": 0.8193, "step": 113 }, { "epoch": 0.04950065132435953, "grad_norm": 0.09767181426286697, "learning_rate": 0.0009774703458011453, "loss": 0.7653, "step": 114 }, { "epoch": 0.04993486756404689, "grad_norm": 0.09374384582042694, "learning_rate": 0.000976992171659682, "loss": 0.7541, "step": 115 }, { "epoch": 0.05036908380373426, "grad_norm": 0.07632128894329071, "learning_rate": 0.0009765090956856436, "loss": 0.7518, "step": 116 }, { "epoch": 0.05080330004342162, "grad_norm": 0.07544126361608505, "learning_rate": 0.0009760211228433832, "loss": 0.8217, "step": 117 }, { "epoch": 0.05123751628310899, "grad_norm": 0.07379148155450821, "learning_rate": 0.0009755282581475768, "loss": 0.7902, "step": 118 }, { "epoch": 0.05167173252279635, "grad_norm": 0.07656335830688477, "learning_rate": 0.0009750305066631716, "loss": 0.7357, "step": 119 }, { "epoch": 0.052105948762483714, "grad_norm": 0.07699162513017654, "learning_rate": 0.0009745278735053344, "loss": 0.7329, "step": 120 }, { "epoch": 0.05254016500217108, "grad_norm": 0.07921908050775528, "learning_rate": 0.0009740203638393983, "loss": 0.7368, "step": 121 }, { "epoch": 0.052974381241858444, "grad_norm": 0.06974276900291443, "learning_rate": 0.0009735079828808107, "loss": 0.7593, "step": 122 }, { "epoch": 0.05340859748154581, "grad_norm": 0.06961283832788467, "learning_rate": 0.0009729907358950785, "loss": 0.7899, "step": 123 }, { "epoch": 0.053842813721233174, "grad_norm": 0.07279238104820251, "learning_rate": 0.0009724686281977146, "loss": 0.7455, "step": 124 }, { "epoch": 0.054277029960920535, "grad_norm": 0.07039442658424377, "learning_rate": 0.0009719416651541838, "loss": 0.7755, "step": 125 }, { "epoch": 0.0547112462006079, "grad_norm": 0.07618307322263718, "learning_rate": 0.0009714098521798464, "loss": 0.7631, "step": 126 }, { "epoch": 0.055145462440295265, "grad_norm": 0.06525314599275589, "learning_rate": 0.0009708731947399038, "loss": 0.7605, "step": 127 }, { "epoch": 0.05557967867998263, "grad_norm": 0.07441145926713943, "learning_rate": 0.0009703316983493413, "loss": 0.7402, "step": 128 }, { "epoch": 0.056013894919669995, "grad_norm": 0.06484754383563995, "learning_rate": 0.0009697853685728721, "loss": 0.7238, "step": 129 }, { "epoch": 0.05644811115935736, "grad_norm": 0.07414862513542175, "learning_rate": 0.0009692342110248802, "loss": 0.7633, "step": 130 }, { "epoch": 0.056882327399044724, "grad_norm": 0.07470294088125229, "learning_rate": 0.0009686782313693621, "loss": 0.7467, "step": 131 }, { "epoch": 0.057316543638732086, "grad_norm": 0.07042668759822845, "learning_rate": 0.0009681174353198686, "loss": 0.7309, "step": 132 }, { "epoch": 0.057750759878419454, "grad_norm": 0.07646633684635162, "learning_rate": 0.0009675518286394472, "loss": 0.722, "step": 133 }, { "epoch": 0.058184976118106815, "grad_norm": 0.07386059314012527, "learning_rate": 0.0009669814171405816, "loss": 0.7028, "step": 134 }, { "epoch": 0.058619192357794184, "grad_norm": 0.07046350091695786, "learning_rate": 0.0009664062066851325, "loss": 0.711, "step": 135 }, { "epoch": 0.059053408597481545, "grad_norm": 0.07790421694517136, "learning_rate": 0.000965826203184277, "loss": 0.7183, "step": 136 }, { "epoch": 0.05948762483716891, "grad_norm": 0.08006292581558228, "learning_rate": 0.000965241412598449, "loss": 0.7534, "step": 137 }, { "epoch": 0.059921841076856275, "grad_norm": 0.06924112141132355, "learning_rate": 0.0009646518409372759, "loss": 0.7401, "step": 138 }, { "epoch": 0.060356057316543636, "grad_norm": 0.07095640152692795, "learning_rate": 0.0009640574942595195, "loss": 0.7056, "step": 139 }, { "epoch": 0.060790273556231005, "grad_norm": 0.0727936401963234, "learning_rate": 0.0009634583786730109, "loss": 0.6811, "step": 140 }, { "epoch": 0.061224489795918366, "grad_norm": 0.0759989321231842, "learning_rate": 0.0009628545003345899, "loss": 0.719, "step": 141 }, { "epoch": 0.061658706035605734, "grad_norm": 0.07433141767978668, "learning_rate": 0.0009622458654500408, "loss": 0.724, "step": 142 }, { "epoch": 0.062092922275293096, "grad_norm": 0.06923756748437881, "learning_rate": 0.0009616324802740286, "loss": 0.7562, "step": 143 }, { "epoch": 0.06252713851498046, "grad_norm": 0.10080388188362122, "learning_rate": 0.0009610143511100353, "loss": 0.7116, "step": 144 }, { "epoch": 0.06296135475466783, "grad_norm": 0.07774697989225388, "learning_rate": 0.0009603914843102941, "loss": 0.7277, "step": 145 }, { "epoch": 0.0633955709943552, "grad_norm": 0.07297523319721222, "learning_rate": 0.0009597638862757254, "loss": 0.6817, "step": 146 }, { "epoch": 0.06382978723404255, "grad_norm": 0.07876230776309967, "learning_rate": 0.0009591315634558697, "loss": 0.7015, "step": 147 }, { "epoch": 0.06426400347372992, "grad_norm": 0.07196466624736786, "learning_rate": 0.0009584945223488226, "loss": 0.6995, "step": 148 }, { "epoch": 0.06469821971341728, "grad_norm": 0.07307913899421692, "learning_rate": 0.0009578527695011669, "loss": 0.6931, "step": 149 }, { "epoch": 0.06513243595310464, "grad_norm": 0.0770203247666359, "learning_rate": 0.0009572063115079062, "loss": 0.7135, "step": 150 }, { "epoch": 0.06556665219279201, "grad_norm": 0.06902311742305756, "learning_rate": 0.0009565551550123967, "loss": 0.6906, "step": 151 }, { "epoch": 0.06600086843247938, "grad_norm": 0.10059111565351486, "learning_rate": 0.0009558993067062785, "loss": 0.7237, "step": 152 }, { "epoch": 0.06643508467216674, "grad_norm": 0.07180771976709366, "learning_rate": 0.000955238773329408, "loss": 0.6929, "step": 153 }, { "epoch": 0.0668693009118541, "grad_norm": 0.07852017134428024, "learning_rate": 0.0009545735616697875, "loss": 0.6531, "step": 154 }, { "epoch": 0.06730351715154147, "grad_norm": 0.07011571526527405, "learning_rate": 0.0009539036785634961, "loss": 0.6949, "step": 155 }, { "epoch": 0.06773773339122884, "grad_norm": 0.07384242117404938, "learning_rate": 0.000953229130894619, "loss": 0.688, "step": 156 }, { "epoch": 0.06817194963091619, "grad_norm": 0.07481442391872406, "learning_rate": 0.0009525499255951775, "loss": 0.6969, "step": 157 }, { "epoch": 0.06860616587060356, "grad_norm": 0.06441578269004822, "learning_rate": 0.0009518660696450568, "loss": 0.6478, "step": 158 }, { "epoch": 0.06904038211029093, "grad_norm": 0.06720245629549026, "learning_rate": 0.0009511775700719346, "loss": 0.677, "step": 159 }, { "epoch": 0.0694745983499783, "grad_norm": 0.08833543956279755, "learning_rate": 0.0009504844339512095, "loss": 0.6912, "step": 160 }, { "epoch": 0.06990881458966565, "grad_norm": 0.08247049152851105, "learning_rate": 0.0009497866684059277, "loss": 0.6833, "step": 161 }, { "epoch": 0.07034303082935302, "grad_norm": 0.06711483001708984, "learning_rate": 0.0009490842806067095, "loss": 0.6695, "step": 162 }, { "epoch": 0.07077724706904039, "grad_norm": 0.06512755900621414, "learning_rate": 0.0009483772777716766, "loss": 0.695, "step": 163 }, { "epoch": 0.07121146330872774, "grad_norm": 0.06651882827281952, "learning_rate": 0.0009476656671663766, "loss": 0.6608, "step": 164 }, { "epoch": 0.07164567954841511, "grad_norm": 0.06416794657707214, "learning_rate": 0.0009469494561037098, "loss": 0.6918, "step": 165 }, { "epoch": 0.07207989578810248, "grad_norm": 0.06808705627918243, "learning_rate": 0.000946228651943853, "loss": 0.6959, "step": 166 }, { "epoch": 0.07251411202778985, "grad_norm": 0.077076256275177, "learning_rate": 0.0009455032620941839, "loss": 0.6737, "step": 167 }, { "epoch": 0.0729483282674772, "grad_norm": 0.09289199858903885, "learning_rate": 0.000944773294009206, "loss": 0.651, "step": 168 }, { "epoch": 0.07338254450716457, "grad_norm": 0.07365357875823975, "learning_rate": 0.0009440387551904703, "loss": 0.6778, "step": 169 }, { "epoch": 0.07381676074685194, "grad_norm": 0.0635438859462738, "learning_rate": 0.0009432996531865001, "loss": 0.7081, "step": 170 }, { "epoch": 0.07425097698653929, "grad_norm": 0.07002374529838562, "learning_rate": 0.0009425559955927117, "loss": 0.6505, "step": 171 }, { "epoch": 0.07468519322622666, "grad_norm": 0.08253403007984161, "learning_rate": 0.0009418077900513376, "loss": 0.6752, "step": 172 }, { "epoch": 0.07511940946591403, "grad_norm": 0.06753169745206833, "learning_rate": 0.0009410550442513475, "loss": 0.6695, "step": 173 }, { "epoch": 0.0755536257056014, "grad_norm": 0.06907963007688522, "learning_rate": 0.000940297765928369, "loss": 0.6441, "step": 174 }, { "epoch": 0.07598784194528875, "grad_norm": 0.0881882980465889, "learning_rate": 0.0009395359628646086, "loss": 0.6456, "step": 175 }, { "epoch": 0.07642205818497612, "grad_norm": 0.06529023498296738, "learning_rate": 0.0009387696428887715, "loss": 0.6689, "step": 176 }, { "epoch": 0.07685627442466349, "grad_norm": 0.06664346903562546, "learning_rate": 0.0009379988138759809, "loss": 0.6485, "step": 177 }, { "epoch": 0.07729049066435084, "grad_norm": 0.06951016932725906, "learning_rate": 0.0009372234837476979, "loss": 0.68, "step": 178 }, { "epoch": 0.07772470690403821, "grad_norm": 0.060999054461717606, "learning_rate": 0.0009364436604716389, "loss": 0.6706, "step": 179 }, { "epoch": 0.07815892314372558, "grad_norm": 0.07197262346744537, "learning_rate": 0.0009356593520616947, "loss": 0.6277, "step": 180 }, { "epoch": 0.07859313938341293, "grad_norm": 0.06676187366247177, "learning_rate": 0.0009348705665778478, "loss": 0.6492, "step": 181 }, { "epoch": 0.0790273556231003, "grad_norm": 0.07378092408180237, "learning_rate": 0.0009340773121260893, "loss": 0.6375, "step": 182 }, { "epoch": 0.07946157186278767, "grad_norm": 0.06367610394954681, "learning_rate": 0.000933279596858336, "loss": 0.6221, "step": 183 }, { "epoch": 0.07989578810247504, "grad_norm": 0.0669325664639473, "learning_rate": 0.0009324774289723468, "loss": 0.6294, "step": 184 }, { "epoch": 0.08033000434216239, "grad_norm": 0.06983687728643417, "learning_rate": 0.0009316708167116377, "loss": 0.6487, "step": 185 }, { "epoch": 0.08076422058184976, "grad_norm": 0.06569714099168777, "learning_rate": 0.0009308597683653976, "loss": 0.651, "step": 186 }, { "epoch": 0.08119843682153713, "grad_norm": 0.06825820356607437, "learning_rate": 0.0009300442922684032, "loss": 0.6193, "step": 187 }, { "epoch": 0.08163265306122448, "grad_norm": 0.07318625599145889, "learning_rate": 0.000929224396800933, "loss": 0.7024, "step": 188 }, { "epoch": 0.08206686930091185, "grad_norm": 0.1047971323132515, "learning_rate": 0.0009284000903886818, "loss": 0.6637, "step": 189 }, { "epoch": 0.08250108554059922, "grad_norm": 0.06583527475595474, "learning_rate": 0.0009275713815026732, "loss": 0.6419, "step": 190 }, { "epoch": 0.08293530178028659, "grad_norm": 0.09143144637346268, "learning_rate": 0.000926738278659173, "loss": 0.6295, "step": 191 }, { "epoch": 0.08336951801997394, "grad_norm": 0.06756502389907837, "learning_rate": 0.0009259007904196022, "loss": 0.6718, "step": 192 }, { "epoch": 0.08380373425966131, "grad_norm": 0.06964406371116638, "learning_rate": 0.000925058925390448, "loss": 0.6542, "step": 193 }, { "epoch": 0.08423795049934868, "grad_norm": 0.07476377487182617, "learning_rate": 0.0009242126922231762, "loss": 0.6266, "step": 194 }, { "epoch": 0.08467216673903603, "grad_norm": 0.07210344076156616, "learning_rate": 0.0009233620996141421, "loss": 0.6721, "step": 195 }, { "epoch": 0.0851063829787234, "grad_norm": 0.07828934490680695, "learning_rate": 0.0009225071563045006, "loss": 0.6223, "step": 196 }, { "epoch": 0.08554059921841077, "grad_norm": 0.07536309212446213, "learning_rate": 0.000921647871080117, "loss": 0.6563, "step": 197 }, { "epoch": 0.08597481545809814, "grad_norm": 0.07678169757127762, "learning_rate": 0.0009207842527714766, "loss": 0.6646, "step": 198 }, { "epoch": 0.0864090316977855, "grad_norm": 0.07808982580900192, "learning_rate": 0.0009199163102535937, "loss": 0.6497, "step": 199 }, { "epoch": 0.08684324793747286, "grad_norm": 0.07775306701660156, "learning_rate": 0.0009190440524459203, "loss": 0.6083, "step": 200 }, { "epoch": 0.08727746417716023, "grad_norm": 0.06569099426269531, "learning_rate": 0.0009181674883122553, "loss": 0.6424, "step": 201 }, { "epoch": 0.08771168041684758, "grad_norm": 0.08122856914997101, "learning_rate": 0.0009172866268606513, "loss": 0.6054, "step": 202 }, { "epoch": 0.08814589665653495, "grad_norm": 0.08045452833175659, "learning_rate": 0.0009164014771433226, "loss": 0.6372, "step": 203 }, { "epoch": 0.08858011289622232, "grad_norm": 0.0744997188448906, "learning_rate": 0.000915512048256552, "loss": 0.6341, "step": 204 }, { "epoch": 0.08901432913590969, "grad_norm": 0.07247951626777649, "learning_rate": 0.0009146183493405975, "loss": 0.5922, "step": 205 }, { "epoch": 0.08944854537559704, "grad_norm": 0.08652166277170181, "learning_rate": 0.0009137203895795982, "loss": 0.6111, "step": 206 }, { "epoch": 0.08988276161528441, "grad_norm": 0.06636743992567062, "learning_rate": 0.0009128181782014801, "loss": 0.6287, "step": 207 }, { "epoch": 0.09031697785497178, "grad_norm": 0.08643815666437149, "learning_rate": 0.0009119117244778608, "loss": 0.6091, "step": 208 }, { "epoch": 0.09075119409465914, "grad_norm": 0.08439099788665771, "learning_rate": 0.0009110010377239551, "loss": 0.6735, "step": 209 }, { "epoch": 0.0911854103343465, "grad_norm": 0.0863455981016159, "learning_rate": 0.0009100861272984779, "loss": 0.6049, "step": 210 }, { "epoch": 0.09161962657403387, "grad_norm": 0.06976283341646194, "learning_rate": 0.0009091670026035499, "loss": 0.5934, "step": 211 }, { "epoch": 0.09205384281372124, "grad_norm": 0.10158892720937729, "learning_rate": 0.0009082436730845993, "loss": 0.6314, "step": 212 }, { "epoch": 0.0924880590534086, "grad_norm": 0.06954232603311539, "learning_rate": 0.0009073161482302654, "loss": 0.6461, "step": 213 }, { "epoch": 0.09292227529309596, "grad_norm": 0.06838048249483109, "learning_rate": 0.0009063844375723014, "loss": 0.5993, "step": 214 }, { "epoch": 0.09335649153278333, "grad_norm": 0.06874913722276688, "learning_rate": 0.0009054485506854755, "loss": 0.6035, "step": 215 }, { "epoch": 0.09379070777247069, "grad_norm": 0.07478207349777222, "learning_rate": 0.0009045084971874737, "loss": 0.6123, "step": 216 }, { "epoch": 0.09422492401215805, "grad_norm": 0.06408224254846573, "learning_rate": 0.0009035642867388002, "loss": 0.6242, "step": 217 }, { "epoch": 0.09465914025184542, "grad_norm": 0.07433444261550903, "learning_rate": 0.0009026159290426781, "loss": 0.6055, "step": 218 }, { "epoch": 0.09509335649153278, "grad_norm": 0.06954223662614822, "learning_rate": 0.0009016634338449503, "loss": 0.6257, "step": 219 }, { "epoch": 0.09552757273122015, "grad_norm": 0.06509328633546829, "learning_rate": 0.0009007068109339783, "loss": 0.5749, "step": 220 }, { "epoch": 0.09596178897090751, "grad_norm": 0.08184398710727692, "learning_rate": 0.000899746070140543, "loss": 0.6166, "step": 221 }, { "epoch": 0.09639600521059488, "grad_norm": 0.06200498715043068, "learning_rate": 0.0008987812213377423, "loss": 0.615, "step": 222 }, { "epoch": 0.09683022145028224, "grad_norm": 0.06602746993303299, "learning_rate": 0.0008978122744408905, "loss": 0.6087, "step": 223 }, { "epoch": 0.0972644376899696, "grad_norm": 0.0869906097650528, "learning_rate": 0.0008968392394074163, "loss": 0.5725, "step": 224 }, { "epoch": 0.09769865392965697, "grad_norm": 0.064040407538414, "learning_rate": 0.0008958621262367599, "loss": 0.6116, "step": 225 }, { "epoch": 0.09813287016934433, "grad_norm": 0.0657719075679779, "learning_rate": 0.0008948809449702712, "loss": 0.6219, "step": 226 }, { "epoch": 0.0985670864090317, "grad_norm": 0.07608803361654282, "learning_rate": 0.0008938957056911057, "loss": 0.6018, "step": 227 }, { "epoch": 0.09900130264871906, "grad_norm": 0.07249824702739716, "learning_rate": 0.0008929064185241212, "loss": 0.6027, "step": 228 }, { "epoch": 0.09943551888840643, "grad_norm": 0.06748249381780624, "learning_rate": 0.0008919130936357742, "loss": 0.6032, "step": 229 }, { "epoch": 0.09986973512809379, "grad_norm": 0.08020893484354019, "learning_rate": 0.000890915741234015, "loss": 0.607, "step": 230 }, { "epoch": 0.10030395136778116, "grad_norm": 0.0669545903801918, "learning_rate": 0.0008899143715681822, "loss": 0.5974, "step": 231 }, { "epoch": 0.10073816760746852, "grad_norm": 0.06186607852578163, "learning_rate": 0.0008889089949288987, "loss": 0.6186, "step": 232 }, { "epoch": 0.10117238384715588, "grad_norm": 0.06386630982160568, "learning_rate": 0.0008878996216479651, "loss": 0.5995, "step": 233 }, { "epoch": 0.10160660008684325, "grad_norm": 0.06557858735322952, "learning_rate": 0.0008868862620982534, "loss": 0.6066, "step": 234 }, { "epoch": 0.10204081632653061, "grad_norm": 0.06579785794019699, "learning_rate": 0.0008858689266936008, "loss": 0.6144, "step": 235 }, { "epoch": 0.10247503256621798, "grad_norm": 0.06238356605172157, "learning_rate": 0.0008848476258887031, "loss": 0.5899, "step": 236 }, { "epoch": 0.10290924880590534, "grad_norm": 0.061628226190805435, "learning_rate": 0.0008838223701790055, "loss": 0.6092, "step": 237 }, { "epoch": 0.1033434650455927, "grad_norm": 0.06545189768075943, "learning_rate": 0.0008827931701005973, "loss": 0.6014, "step": 238 }, { "epoch": 0.10377768128528007, "grad_norm": 0.056564707309007645, "learning_rate": 0.0008817600362301017, "loss": 0.5959, "step": 239 }, { "epoch": 0.10421189752496743, "grad_norm": 0.0715203583240509, "learning_rate": 0.0008807229791845672, "loss": 0.5942, "step": 240 }, { "epoch": 0.1046461137646548, "grad_norm": 0.06504496932029724, "learning_rate": 0.00087968200962136, "loss": 0.5652, "step": 241 }, { "epoch": 0.10508033000434217, "grad_norm": 0.06740335375070572, "learning_rate": 0.0008786371382380527, "loss": 0.6159, "step": 242 }, { "epoch": 0.10551454624402953, "grad_norm": 0.06163683906197548, "learning_rate": 0.0008775883757723155, "loss": 0.5913, "step": 243 }, { "epoch": 0.10594876248371689, "grad_norm": 0.07384572178125381, "learning_rate": 0.0008765357330018055, "loss": 0.5829, "step": 244 }, { "epoch": 0.10638297872340426, "grad_norm": 0.06352236866950989, "learning_rate": 0.0008754792207440556, "loss": 0.6069, "step": 245 }, { "epoch": 0.10681719496309162, "grad_norm": 0.0605890229344368, "learning_rate": 0.0008744188498563641, "loss": 0.5685, "step": 246 }, { "epoch": 0.10725141120277898, "grad_norm": 0.06994660943746567, "learning_rate": 0.0008733546312356824, "loss": 0.581, "step": 247 }, { "epoch": 0.10768562744246635, "grad_norm": 0.08128967881202698, "learning_rate": 0.0008722865758185036, "loss": 0.5754, "step": 248 }, { "epoch": 0.10811984368215372, "grad_norm": 0.06497329473495483, "learning_rate": 0.0008712146945807493, "loss": 0.598, "step": 249 }, { "epoch": 0.10855405992184107, "grad_norm": 0.06091492623090744, "learning_rate": 0.0008701389985376577, "loss": 0.5807, "step": 250 }, { "epoch": 0.10898827616152844, "grad_norm": 0.06555697321891785, "learning_rate": 0.0008690594987436704, "loss": 0.5768, "step": 251 }, { "epoch": 0.1094224924012158, "grad_norm": 0.07119280844926834, "learning_rate": 0.0008679762062923176, "loss": 0.6088, "step": 252 }, { "epoch": 0.10985670864090318, "grad_norm": 0.06837710738182068, "learning_rate": 0.0008668891323161053, "loss": 0.5905, "step": 253 }, { "epoch": 0.11029092488059053, "grad_norm": 0.067451111972332, "learning_rate": 0.0008657982879864007, "loss": 0.5974, "step": 254 }, { "epoch": 0.1107251411202779, "grad_norm": 0.06649015098810196, "learning_rate": 0.0008647036845133172, "loss": 0.5684, "step": 255 }, { "epoch": 0.11115935735996527, "grad_norm": 0.06311678886413574, "learning_rate": 0.0008636053331455986, "loss": 0.58, "step": 256 }, { "epoch": 0.11159357359965262, "grad_norm": 0.06087026745080948, "learning_rate": 0.0008625032451705052, "loss": 0.5735, "step": 257 }, { "epoch": 0.11202778983933999, "grad_norm": 0.061150964349508286, "learning_rate": 0.0008613974319136957, "loss": 0.6044, "step": 258 }, { "epoch": 0.11246200607902736, "grad_norm": 0.06023947522044182, "learning_rate": 0.0008602879047391126, "loss": 0.551, "step": 259 }, { "epoch": 0.11289622231871473, "grad_norm": 0.06919502466917038, "learning_rate": 0.0008591746750488638, "loss": 0.5744, "step": 260 }, { "epoch": 0.11333043855840208, "grad_norm": 0.06585050374269485, "learning_rate": 0.0008580577542831072, "loss": 0.6053, "step": 261 }, { "epoch": 0.11376465479808945, "grad_norm": 0.06967171281576157, "learning_rate": 0.0008569371539199316, "loss": 0.5896, "step": 262 }, { "epoch": 0.11419887103777682, "grad_norm": 0.062438324093818665, "learning_rate": 0.0008558128854752396, "loss": 0.585, "step": 263 }, { "epoch": 0.11463308727746417, "grad_norm": 0.06600722670555115, "learning_rate": 0.0008546849605026289, "loss": 0.5809, "step": 264 }, { "epoch": 0.11506730351715154, "grad_norm": 0.06594278663396835, "learning_rate": 0.0008535533905932737, "loss": 0.5854, "step": 265 }, { "epoch": 0.11550151975683891, "grad_norm": 0.06402723491191864, "learning_rate": 0.0008524181873758059, "loss": 0.5612, "step": 266 }, { "epoch": 0.11593573599652628, "grad_norm": 0.06593722105026245, "learning_rate": 0.0008512793625161946, "loss": 0.5825, "step": 267 }, { "epoch": 0.11636995223621363, "grad_norm": 0.057960718870162964, "learning_rate": 0.0008501369277176275, "loss": 0.5972, "step": 268 }, { "epoch": 0.116804168475901, "grad_norm": 0.06126386299729347, "learning_rate": 0.0008489908947203897, "loss": 0.5675, "step": 269 }, { "epoch": 0.11723838471558837, "grad_norm": 0.06619458645582199, "learning_rate": 0.0008478412753017432, "loss": 0.5724, "step": 270 }, { "epoch": 0.11767260095527572, "grad_norm": 0.05950764939188957, "learning_rate": 0.0008466880812758065, "loss": 0.5732, "step": 271 }, { "epoch": 0.11810681719496309, "grad_norm": 0.0596248060464859, "learning_rate": 0.0008455313244934324, "loss": 0.5844, "step": 272 }, { "epoch": 0.11854103343465046, "grad_norm": 0.06557915359735489, "learning_rate": 0.0008443710168420865, "loss": 0.5706, "step": 273 }, { "epoch": 0.11897524967433783, "grad_norm": 0.06171106547117233, "learning_rate": 0.0008432071702457253, "loss": 0.561, "step": 274 }, { "epoch": 0.11940946591402518, "grad_norm": 0.06009051948785782, "learning_rate": 0.0008420397966646731, "loss": 0.5764, "step": 275 }, { "epoch": 0.11984368215371255, "grad_norm": 0.06843849271535873, "learning_rate": 0.0008408689080954998, "loss": 0.5833, "step": 276 }, { "epoch": 0.12027789839339992, "grad_norm": 0.06061152368783951, "learning_rate": 0.0008396945165708972, "loss": 0.5962, "step": 277 }, { "epoch": 0.12071211463308727, "grad_norm": 0.05997680127620697, "learning_rate": 0.0008385166341595548, "loss": 0.5738, "step": 278 }, { "epoch": 0.12114633087277464, "grad_norm": 0.06045431271195412, "learning_rate": 0.0008373352729660373, "loss": 0.5738, "step": 279 }, { "epoch": 0.12158054711246201, "grad_norm": 0.06007954850792885, "learning_rate": 0.0008361504451306584, "loss": 0.5753, "step": 280 }, { "epoch": 0.12201476335214936, "grad_norm": 0.05806737020611763, "learning_rate": 0.0008349621628293577, "loss": 0.5703, "step": 281 }, { "epoch": 0.12244897959183673, "grad_norm": 0.06177588179707527, "learning_rate": 0.0008337704382735741, "loss": 0.5533, "step": 282 }, { "epoch": 0.1228831958315241, "grad_norm": 0.06487903743982315, "learning_rate": 0.0008325752837101213, "loss": 0.5655, "step": 283 }, { "epoch": 0.12331741207121147, "grad_norm": 0.06531866639852524, "learning_rate": 0.0008313767114210615, "loss": 0.5694, "step": 284 }, { "epoch": 0.12375162831089882, "grad_norm": 0.08110279589891434, "learning_rate": 0.0008301747337235797, "loss": 0.5517, "step": 285 }, { "epoch": 0.12418584455058619, "grad_norm": 0.06367824971675873, "learning_rate": 0.0008289693629698564, "loss": 0.5474, "step": 286 }, { "epoch": 0.12462006079027356, "grad_norm": 0.07917732000350952, "learning_rate": 0.0008277606115469409, "loss": 0.5685, "step": 287 }, { "epoch": 0.12505427702996091, "grad_norm": 0.06585974991321564, "learning_rate": 0.0008265484918766243, "loss": 0.5547, "step": 288 }, { "epoch": 0.12548849326964828, "grad_norm": 0.0627543181180954, "learning_rate": 0.0008253330164153117, "loss": 0.5795, "step": 289 }, { "epoch": 0.12592270950933565, "grad_norm": 0.06639768183231354, "learning_rate": 0.0008241141976538943, "loss": 0.5669, "step": 290 }, { "epoch": 0.12635692574902302, "grad_norm": 0.06217719987034798, "learning_rate": 0.0008228920481176202, "loss": 0.5552, "step": 291 }, { "epoch": 0.1267911419887104, "grad_norm": 0.06806106120347977, "learning_rate": 0.0008216665803659671, "loss": 0.5553, "step": 292 }, { "epoch": 0.12722535822839776, "grad_norm": 0.058773159980773926, "learning_rate": 0.000820437806992512, "loss": 0.5678, "step": 293 }, { "epoch": 0.1276595744680851, "grad_norm": 0.06487017124891281, "learning_rate": 0.0008192057406248028, "loss": 0.5493, "step": 294 }, { "epoch": 0.12809379070777246, "grad_norm": 0.05514706298708916, "learning_rate": 0.0008179703939242276, "loss": 0.5834, "step": 295 }, { "epoch": 0.12852800694745983, "grad_norm": 0.05893951281905174, "learning_rate": 0.0008167317795858851, "loss": 0.5668, "step": 296 }, { "epoch": 0.1289622231871472, "grad_norm": 0.06080016866326332, "learning_rate": 0.0008154899103384537, "loss": 0.6112, "step": 297 }, { "epoch": 0.12939643942683457, "grad_norm": 0.06676195561885834, "learning_rate": 0.0008142447989440618, "loss": 0.5897, "step": 298 }, { "epoch": 0.12983065566652194, "grad_norm": 0.055440496653318405, "learning_rate": 0.0008129964581981553, "loss": 0.5199, "step": 299 }, { "epoch": 0.13026487190620928, "grad_norm": 0.0642077848315239, "learning_rate": 0.0008117449009293668, "loss": 0.5688, "step": 300 }, { "epoch": 0.13069908814589665, "grad_norm": 0.056085068732500076, "learning_rate": 0.0008104901399993836, "loss": 0.5626, "step": 301 }, { "epoch": 0.13113330438558402, "grad_norm": 0.07240674644708633, "learning_rate": 0.0008092321883028157, "loss": 0.5705, "step": 302 }, { "epoch": 0.13156752062527138, "grad_norm": 0.06402257084846497, "learning_rate": 0.0008079710587670633, "loss": 0.5608, "step": 303 }, { "epoch": 0.13200173686495875, "grad_norm": 0.05889822542667389, "learning_rate": 0.0008067067643521834, "loss": 0.5552, "step": 304 }, { "epoch": 0.13243595310464612, "grad_norm": 0.07787332683801651, "learning_rate": 0.0008054393180507572, "loss": 0.5526, "step": 305 }, { "epoch": 0.1328701693443335, "grad_norm": 0.06866924464702606, "learning_rate": 0.0008041687328877566, "loss": 0.5726, "step": 306 }, { "epoch": 0.13330438558402083, "grad_norm": 0.06494660675525665, "learning_rate": 0.00080289502192041, "loss": 0.5605, "step": 307 }, { "epoch": 0.1337386018237082, "grad_norm": 0.06333691626787186, "learning_rate": 0.0008016181982380681, "loss": 0.5668, "step": 308 }, { "epoch": 0.13417281806339557, "grad_norm": 0.05925397574901581, "learning_rate": 0.0008003382749620702, "loss": 0.539, "step": 309 }, { "epoch": 0.13460703430308293, "grad_norm": 0.06211424991488457, "learning_rate": 0.000799055265245608, "loss": 0.5581, "step": 310 }, { "epoch": 0.1350412505427703, "grad_norm": 0.06610213220119476, "learning_rate": 0.0007977691822735914, "loss": 0.5726, "step": 311 }, { "epoch": 0.13547546678245767, "grad_norm": 0.06418436765670776, "learning_rate": 0.0007964800392625129, "loss": 0.5616, "step": 312 }, { "epoch": 0.13590968302214504, "grad_norm": 0.05993996933102608, "learning_rate": 0.0007951878494603115, "loss": 0.5392, "step": 313 }, { "epoch": 0.13634389926183238, "grad_norm": 0.05776641145348549, "learning_rate": 0.0007938926261462366, "loss": 0.5402, "step": 314 }, { "epoch": 0.13677811550151975, "grad_norm": 0.06915484368801117, "learning_rate": 0.0007925943826307118, "loss": 0.5336, "step": 315 }, { "epoch": 0.13721233174120712, "grad_norm": 0.06394239515066147, "learning_rate": 0.0007912931322551981, "loss": 0.5647, "step": 316 }, { "epoch": 0.13764654798089448, "grad_norm": 0.05997829511761665, "learning_rate": 0.000789988888392056, "loss": 0.5588, "step": 317 }, { "epoch": 0.13808076422058185, "grad_norm": 0.07337518781423569, "learning_rate": 0.0007886816644444098, "loss": 0.5825, "step": 318 }, { "epoch": 0.13851498046026922, "grad_norm": 0.07288537919521332, "learning_rate": 0.0007873714738460075, "loss": 0.5615, "step": 319 }, { "epoch": 0.1389491966999566, "grad_norm": 0.06601582467556, "learning_rate": 0.0007860583300610849, "loss": 0.5545, "step": 320 }, { "epoch": 0.13938341293964393, "grad_norm": 0.06303299218416214, "learning_rate": 0.000784742246584226, "loss": 0.5579, "step": 321 }, { "epoch": 0.1398176291793313, "grad_norm": 0.06565733999013901, "learning_rate": 0.000783423236940225, "loss": 0.5728, "step": 322 }, { "epoch": 0.14025184541901867, "grad_norm": 0.06590873748064041, "learning_rate": 0.0007821013146839467, "loss": 0.5552, "step": 323 }, { "epoch": 0.14068606165870604, "grad_norm": 0.06782017648220062, "learning_rate": 0.0007807764934001874, "loss": 0.5299, "step": 324 }, { "epoch": 0.1411202778983934, "grad_norm": 0.06369118392467499, "learning_rate": 0.0007794487867035358, "loss": 0.524, "step": 325 }, { "epoch": 0.14155449413808077, "grad_norm": 0.06337650865316391, "learning_rate": 0.0007781182082382324, "loss": 0.5802, "step": 326 }, { "epoch": 0.14198871037776814, "grad_norm": 0.07011637836694717, "learning_rate": 0.0007767847716780296, "loss": 0.5569, "step": 327 }, { "epoch": 0.14242292661745548, "grad_norm": 0.06176460534334183, "learning_rate": 0.0007754484907260512, "loss": 0.5266, "step": 328 }, { "epoch": 0.14285714285714285, "grad_norm": 0.06650613993406296, "learning_rate": 0.0007741093791146516, "loss": 0.5628, "step": 329 }, { "epoch": 0.14329135909683022, "grad_norm": 0.07864295691251755, "learning_rate": 0.0007727674506052743, "loss": 0.5683, "step": 330 }, { "epoch": 0.14372557533651759, "grad_norm": 0.06307731568813324, "learning_rate": 0.0007714227189883112, "loss": 0.5705, "step": 331 }, { "epoch": 0.14415979157620495, "grad_norm": 0.07688986510038376, "learning_rate": 0.0007700751980829601, "loss": 0.5714, "step": 332 }, { "epoch": 0.14459400781589232, "grad_norm": 0.06042364984750748, "learning_rate": 0.0007687249017370832, "loss": 0.5629, "step": 333 }, { "epoch": 0.1450282240555797, "grad_norm": 0.06250355392694473, "learning_rate": 0.0007673718438270648, "loss": 0.5318, "step": 334 }, { "epoch": 0.14546244029526703, "grad_norm": 0.06661161035299301, "learning_rate": 0.0007660160382576683, "loss": 0.5261, "step": 335 }, { "epoch": 0.1458966565349544, "grad_norm": 0.07277819514274597, "learning_rate": 0.0007646574989618937, "loss": 0.5427, "step": 336 }, { "epoch": 0.14633087277464177, "grad_norm": 0.05938050523400307, "learning_rate": 0.0007632962399008341, "loss": 0.5469, "step": 337 }, { "epoch": 0.14676508901432914, "grad_norm": 0.0626835823059082, "learning_rate": 0.0007619322750635327, "loss": 0.5321, "step": 338 }, { "epoch": 0.1471993052540165, "grad_norm": 0.06853507459163666, "learning_rate": 0.0007605656184668384, "loss": 0.5319, "step": 339 }, { "epoch": 0.14763352149370387, "grad_norm": 0.06294318288564682, "learning_rate": 0.0007591962841552626, "loss": 0.5735, "step": 340 }, { "epoch": 0.14806773773339124, "grad_norm": 0.06769460439682007, "learning_rate": 0.0007578242862008336, "loss": 0.5734, "step": 341 }, { "epoch": 0.14850195397307858, "grad_norm": 0.06945870816707611, "learning_rate": 0.0007564496387029531, "loss": 0.5486, "step": 342 }, { "epoch": 0.14893617021276595, "grad_norm": 0.055839743465185165, "learning_rate": 0.0007550723557882513, "loss": 0.5646, "step": 343 }, { "epoch": 0.14937038645245332, "grad_norm": 0.05766444280743599, "learning_rate": 0.0007536924516104411, "loss": 0.5671, "step": 344 }, { "epoch": 0.1498046026921407, "grad_norm": 0.05658617615699768, "learning_rate": 0.000752309940350173, "loss": 0.5573, "step": 345 }, { "epoch": 0.15023881893182806, "grad_norm": 0.06456328183412552, "learning_rate": 0.0007509248362148888, "loss": 0.5413, "step": 346 }, { "epoch": 0.15067303517151542, "grad_norm": 0.05971672758460045, "learning_rate": 0.000749537153438677, "loss": 0.528, "step": 347 }, { "epoch": 0.1511072514112028, "grad_norm": 0.059805817902088165, "learning_rate": 0.0007481469062821251, "loss": 0.5517, "step": 348 }, { "epoch": 0.15154146765089013, "grad_norm": 0.05796041712164879, "learning_rate": 0.0007467541090321735, "loss": 0.5651, "step": 349 }, { "epoch": 0.1519756838905775, "grad_norm": 0.0807691365480423, "learning_rate": 0.000745358776001969, "loss": 0.5632, "step": 350 }, { "epoch": 0.15240990013026487, "grad_norm": 0.056193944066762924, "learning_rate": 0.0007439609215307173, "loss": 0.5544, "step": 351 }, { "epoch": 0.15284411636995224, "grad_norm": 0.06092338636517525, "learning_rate": 0.000742560559983536, "loss": 0.5378, "step": 352 }, { "epoch": 0.1532783326096396, "grad_norm": 0.06231565400958061, "learning_rate": 0.0007411577057513066, "loss": 0.5584, "step": 353 }, { "epoch": 0.15371254884932697, "grad_norm": 0.06663519144058228, "learning_rate": 0.000739752373250527, "loss": 0.5246, "step": 354 }, { "epoch": 0.15414676508901434, "grad_norm": 0.060423221439123154, "learning_rate": 0.0007383445769231627, "loss": 0.5203, "step": 355 }, { "epoch": 0.15458098132870168, "grad_norm": 0.06870024651288986, "learning_rate": 0.0007369343312364993, "loss": 0.52, "step": 356 }, { "epoch": 0.15501519756838905, "grad_norm": 0.06956614553928375, "learning_rate": 0.0007355216506829932, "loss": 0.5816, "step": 357 }, { "epoch": 0.15544941380807642, "grad_norm": 0.060565389692783356, "learning_rate": 0.0007341065497801229, "loss": 0.5508, "step": 358 }, { "epoch": 0.1558836300477638, "grad_norm": 0.06745211035013199, "learning_rate": 0.0007326890430702396, "loss": 0.5589, "step": 359 }, { "epoch": 0.15631784628745116, "grad_norm": 0.06009297072887421, "learning_rate": 0.0007312691451204177, "loss": 0.5259, "step": 360 }, { "epoch": 0.15675206252713852, "grad_norm": 0.05682894587516785, "learning_rate": 0.0007298468705223059, "loss": 0.5689, "step": 361 }, { "epoch": 0.15718627876682587, "grad_norm": 0.06062963977456093, "learning_rate": 0.0007284222338919758, "loss": 0.5339, "step": 362 }, { "epoch": 0.15762049500651323, "grad_norm": 0.058002885431051254, "learning_rate": 0.0007269952498697733, "loss": 0.5714, "step": 363 }, { "epoch": 0.1580547112462006, "grad_norm": 0.056898582726716995, "learning_rate": 0.0007255659331201672, "loss": 0.5811, "step": 364 }, { "epoch": 0.15848892748588797, "grad_norm": 0.06246396526694298, "learning_rate": 0.0007241342983315984, "loss": 0.5554, "step": 365 }, { "epoch": 0.15892314372557534, "grad_norm": 0.05486922711133957, "learning_rate": 0.0007227003602163296, "loss": 0.5567, "step": 366 }, { "epoch": 0.1593573599652627, "grad_norm": 0.057243555784225464, "learning_rate": 0.0007212641335102932, "loss": 0.5193, "step": 367 }, { "epoch": 0.15979157620495008, "grad_norm": 0.05640234053134918, "learning_rate": 0.0007198256329729411, "loss": 0.5098, "step": 368 }, { "epoch": 0.16022579244463742, "grad_norm": 0.05819845199584961, "learning_rate": 0.0007183848733870917, "loss": 0.5366, "step": 369 }, { "epoch": 0.16066000868432478, "grad_norm": 0.05703158304095268, "learning_rate": 0.0007169418695587791, "loss": 0.5374, "step": 370 }, { "epoch": 0.16109422492401215, "grad_norm": 0.05363311246037483, "learning_rate": 0.0007154966363171003, "loss": 0.555, "step": 371 }, { "epoch": 0.16152844116369952, "grad_norm": 0.060146309435367584, "learning_rate": 0.0007140491885140629, "loss": 0.5395, "step": 372 }, { "epoch": 0.1619626574033869, "grad_norm": 0.05488353967666626, "learning_rate": 0.0007125995410244324, "loss": 0.5117, "step": 373 }, { "epoch": 0.16239687364307426, "grad_norm": 0.05473377928137779, "learning_rate": 0.00071114770874558, "loss": 0.5463, "step": 374 }, { "epoch": 0.16283108988276163, "grad_norm": 0.05373527109622955, "learning_rate": 0.0007096937065973285, "loss": 0.522, "step": 375 }, { "epoch": 0.16326530612244897, "grad_norm": 0.054411571472883224, "learning_rate": 0.0007082375495217995, "loss": 0.519, "step": 376 }, { "epoch": 0.16369952236213633, "grad_norm": 0.059973325580358505, "learning_rate": 0.0007067792524832604, "loss": 0.5546, "step": 377 }, { "epoch": 0.1641337386018237, "grad_norm": 0.07386507093906403, "learning_rate": 0.000705318830467969, "loss": 0.548, "step": 378 }, { "epoch": 0.16456795484151107, "grad_norm": 0.05318816378712654, "learning_rate": 0.0007038562984840216, "loss": 0.5325, "step": 379 }, { "epoch": 0.16500217108119844, "grad_norm": 0.06362178921699524, "learning_rate": 0.0007023916715611969, "loss": 0.5537, "step": 380 }, { "epoch": 0.1654363873208858, "grad_norm": 0.05186508968472481, "learning_rate": 0.0007009249647508028, "loss": 0.5534, "step": 381 }, { "epoch": 0.16587060356057318, "grad_norm": 0.05456273630261421, "learning_rate": 0.0006994561931255209, "loss": 0.533, "step": 382 }, { "epoch": 0.16630481980026052, "grad_norm": 0.07069144397974014, "learning_rate": 0.0006979853717792523, "loss": 0.5591, "step": 383 }, { "epoch": 0.16673903603994789, "grad_norm": 0.13737140595912933, "learning_rate": 0.0006965125158269618, "loss": 0.8425, "step": 384 }, { "epoch": 0.16717325227963525, "grad_norm": 0.08646462857723236, "learning_rate": 0.0006950376404045235, "loss": 0.5356, "step": 385 }, { "epoch": 0.16760746851932262, "grad_norm": 0.059485238045454025, "learning_rate": 0.0006935607606685641, "loss": 0.571, "step": 386 }, { "epoch": 0.16804168475901, "grad_norm": 0.07368376851081848, "learning_rate": 0.000692081891796308, "loss": 0.5346, "step": 387 }, { "epoch": 0.16847590099869736, "grad_norm": 0.0615500770509243, "learning_rate": 0.0006906010489854209, "loss": 0.5257, "step": 388 }, { "epoch": 0.16891011723838473, "grad_norm": 0.064671091735363, "learning_rate": 0.0006891182474538539, "loss": 0.5492, "step": 389 }, { "epoch": 0.16934433347807207, "grad_norm": 0.06681560724973679, "learning_rate": 0.0006876335024396872, "loss": 0.5595, "step": 390 }, { "epoch": 0.16977854971775944, "grad_norm": 0.0575130432844162, "learning_rate": 0.0006861468292009726, "loss": 0.5895, "step": 391 }, { "epoch": 0.1702127659574468, "grad_norm": 0.05460560694336891, "learning_rate": 0.0006846582430155782, "loss": 0.5374, "step": 392 }, { "epoch": 0.17064698219713417, "grad_norm": 0.06030447408556938, "learning_rate": 0.0006831677591810301, "loss": 0.6073, "step": 393 }, { "epoch": 0.17108119843682154, "grad_norm": 0.0582847036421299, "learning_rate": 0.0006816753930143557, "loss": 0.5347, "step": 394 }, { "epoch": 0.1715154146765089, "grad_norm": 0.05572697892785072, "learning_rate": 0.0006801811598519267, "loss": 0.5042, "step": 395 }, { "epoch": 0.17194963091619628, "grad_norm": 0.05724466219544411, "learning_rate": 0.0006786850750493006, "loss": 0.5613, "step": 396 }, { "epoch": 0.17238384715588362, "grad_norm": 0.06665828824043274, "learning_rate": 0.0006771871539810632, "loss": 0.5841, "step": 397 }, { "epoch": 0.172818063395571, "grad_norm": 0.05465397238731384, "learning_rate": 0.0006756874120406714, "loss": 0.5636, "step": 398 }, { "epoch": 0.17325227963525835, "grad_norm": 0.07157056033611298, "learning_rate": 0.000674185864640294, "loss": 0.561, "step": 399 }, { "epoch": 0.17368649587494572, "grad_norm": 0.05561946704983711, "learning_rate": 0.0006726825272106538, "loss": 0.5351, "step": 400 }, { "epoch": 0.1741207121146331, "grad_norm": 0.05475137010216713, "learning_rate": 0.0006711774152008689, "loss": 0.5445, "step": 401 }, { "epoch": 0.17455492835432046, "grad_norm": 0.07090818136930466, "learning_rate": 0.0006696705440782938, "loss": 0.5279, "step": 402 }, { "epoch": 0.17498914459400783, "grad_norm": 0.051751721650362015, "learning_rate": 0.0006681619293283609, "loss": 0.4934, "step": 403 }, { "epoch": 0.17542336083369517, "grad_norm": 0.06089828908443451, "learning_rate": 0.0006666515864544209, "loss": 0.5413, "step": 404 }, { "epoch": 0.17585757707338254, "grad_norm": 0.052899301052093506, "learning_rate": 0.0006651395309775837, "loss": 0.5187, "step": 405 }, { "epoch": 0.1762917933130699, "grad_norm": 0.05137433856725693, "learning_rate": 0.0006636257784365584, "loss": 0.5636, "step": 406 }, { "epoch": 0.17672600955275727, "grad_norm": 0.060585830360651016, "learning_rate": 0.0006621103443874949, "loss": 0.5311, "step": 407 }, { "epoch": 0.17716022579244464, "grad_norm": 0.052542008459568024, "learning_rate": 0.0006605932444038228, "loss": 0.5344, "step": 408 }, { "epoch": 0.177594442032132, "grad_norm": 0.052282754331827164, "learning_rate": 0.0006590744940760914, "loss": 0.5616, "step": 409 }, { "epoch": 0.17802865827181938, "grad_norm": 0.055943962186574936, "learning_rate": 0.0006575541090118104, "loss": 0.5657, "step": 410 }, { "epoch": 0.17846287451150672, "grad_norm": 0.05422157794237137, "learning_rate": 0.0006560321048352886, "loss": 0.5482, "step": 411 }, { "epoch": 0.1788970907511941, "grad_norm": 0.053238000720739365, "learning_rate": 0.0006545084971874737, "loss": 0.5477, "step": 412 }, { "epoch": 0.17933130699088146, "grad_norm": 0.05666307359933853, "learning_rate": 0.0006529833017257919, "loss": 0.5593, "step": 413 }, { "epoch": 0.17976552323056882, "grad_norm": 0.05702383071184158, "learning_rate": 0.000651456534123986, "loss": 0.5229, "step": 414 }, { "epoch": 0.1801997394702562, "grad_norm": 0.05041234940290451, "learning_rate": 0.0006499282100719558, "loss": 0.5197, "step": 415 }, { "epoch": 0.18063395570994356, "grad_norm": 0.05691606178879738, "learning_rate": 0.0006483983452755952, "loss": 0.5282, "step": 416 }, { "epoch": 0.18106817194963093, "grad_norm": 0.05236973240971565, "learning_rate": 0.0006468669554566324, "loss": 0.574, "step": 417 }, { "epoch": 0.18150238818931827, "grad_norm": 0.05194111168384552, "learning_rate": 0.0006453340563524669, "loss": 0.521, "step": 418 }, { "epoch": 0.18193660442900564, "grad_norm": 0.05307863652706146, "learning_rate": 0.0006437996637160086, "loss": 0.5474, "step": 419 }, { "epoch": 0.182370820668693, "grad_norm": 0.05186399817466736, "learning_rate": 0.0006422637933155162, "loss": 0.5584, "step": 420 }, { "epoch": 0.18280503690838038, "grad_norm": 0.053380340337753296, "learning_rate": 0.0006407264609344343, "loss": 0.5309, "step": 421 }, { "epoch": 0.18323925314806774, "grad_norm": 0.05926523730158806, "learning_rate": 0.0006391876823712317, "loss": 0.5362, "step": 422 }, { "epoch": 0.1836734693877551, "grad_norm": 0.053086958825588226, "learning_rate": 0.0006376474734392387, "loss": 0.5345, "step": 423 }, { "epoch": 0.18410768562744248, "grad_norm": 0.05506280064582825, "learning_rate": 0.0006361058499664855, "loss": 0.5301, "step": 424 }, { "epoch": 0.18454190186712982, "grad_norm": 0.0567692294716835, "learning_rate": 0.0006345628277955385, "loss": 0.5336, "step": 425 }, { "epoch": 0.1849761181068172, "grad_norm": 0.04989070072770119, "learning_rate": 0.0006330184227833376, "loss": 0.53, "step": 426 }, { "epoch": 0.18541033434650456, "grad_norm": 0.05129670351743698, "learning_rate": 0.000631472650801034, "loss": 0.523, "step": 427 }, { "epoch": 0.18584455058619193, "grad_norm": 0.052287109196186066, "learning_rate": 0.0006299255277338265, "loss": 0.5309, "step": 428 }, { "epoch": 0.1862787668258793, "grad_norm": 0.051617227494716644, "learning_rate": 0.0006283770694807982, "loss": 0.5406, "step": 429 }, { "epoch": 0.18671298306556666, "grad_norm": 0.06112902611494064, "learning_rate": 0.0006268272919547536, "loss": 0.5148, "step": 430 }, { "epoch": 0.187147199305254, "grad_norm": 0.04904035106301308, "learning_rate": 0.0006252762110820547, "loss": 0.5277, "step": 431 }, { "epoch": 0.18758141554494137, "grad_norm": 0.05905110388994217, "learning_rate": 0.0006237238428024572, "loss": 0.5148, "step": 432 }, { "epoch": 0.18801563178462874, "grad_norm": 0.05047300457954407, "learning_rate": 0.000622170203068947, "loss": 0.5119, "step": 433 }, { "epoch": 0.1884498480243161, "grad_norm": 0.0518287792801857, "learning_rate": 0.0006206153078475762, "loss": 0.5276, "step": 434 }, { "epoch": 0.18888406426400348, "grad_norm": 0.04822736606001854, "learning_rate": 0.0006190591731172991, "loss": 0.5347, "step": 435 }, { "epoch": 0.18931828050369084, "grad_norm": 0.052807264029979706, "learning_rate": 0.0006175018148698076, "loss": 0.5343, "step": 436 }, { "epoch": 0.1897524967433782, "grad_norm": 0.054650746285915375, "learning_rate": 0.0006159432491093672, "loss": 0.5079, "step": 437 }, { "epoch": 0.19018671298306555, "grad_norm": 0.047450825572013855, "learning_rate": 0.0006143834918526527, "loss": 0.5532, "step": 438 }, { "epoch": 0.19062092922275292, "grad_norm": 0.047222550958395004, "learning_rate": 0.0006128225591285831, "loss": 0.5376, "step": 439 }, { "epoch": 0.1910551454624403, "grad_norm": 0.05076510086655617, "learning_rate": 0.0006112604669781572, "loss": 0.5152, "step": 440 }, { "epoch": 0.19148936170212766, "grad_norm": 0.06690964847803116, "learning_rate": 0.0006096972314542889, "loss": 0.5395, "step": 441 }, { "epoch": 0.19192357794181503, "grad_norm": 0.049380868673324585, "learning_rate": 0.0006081328686216418, "loss": 0.4996, "step": 442 }, { "epoch": 0.1923577941815024, "grad_norm": 0.06767404824495316, "learning_rate": 0.0006065673945564642, "loss": 0.5217, "step": 443 }, { "epoch": 0.19279201042118976, "grad_norm": 0.054420776665210724, "learning_rate": 0.0006050008253464246, "loss": 0.5587, "step": 444 }, { "epoch": 0.1932262266608771, "grad_norm": 0.0557849258184433, "learning_rate": 0.0006034331770904454, "loss": 0.5441, "step": 445 }, { "epoch": 0.19366044290056447, "grad_norm": 0.06514862924814224, "learning_rate": 0.0006018644658985379, "loss": 0.5666, "step": 446 }, { "epoch": 0.19409465914025184, "grad_norm": 0.061305928975343704, "learning_rate": 0.0006002947078916364, "loss": 0.5576, "step": 447 }, { "epoch": 0.1945288753799392, "grad_norm": 0.05214373767375946, "learning_rate": 0.0005987239192014335, "loss": 0.5424, "step": 448 }, { "epoch": 0.19496309161962658, "grad_norm": 0.05650056526064873, "learning_rate": 0.0005971521159702136, "loss": 0.5361, "step": 449 }, { "epoch": 0.19539730785931395, "grad_norm": 0.05745421350002289, "learning_rate": 0.0005955793143506863, "loss": 0.5273, "step": 450 }, { "epoch": 0.19583152409900131, "grad_norm": 0.05364866554737091, "learning_rate": 0.0005940055305058219, "loss": 0.5026, "step": 451 }, { "epoch": 0.19626574033868865, "grad_norm": 0.047814685851335526, "learning_rate": 0.0005924307806086844, "loss": 0.5439, "step": 452 }, { "epoch": 0.19669995657837602, "grad_norm": 0.051681190729141235, "learning_rate": 0.0005908550808422655, "loss": 0.5271, "step": 453 }, { "epoch": 0.1971341728180634, "grad_norm": 0.05405402183532715, "learning_rate": 0.0005892784473993184, "loss": 0.4996, "step": 454 }, { "epoch": 0.19756838905775076, "grad_norm": 0.04617345333099365, "learning_rate": 0.0005877008964821908, "loss": 0.5409, "step": 455 }, { "epoch": 0.19800260529743813, "grad_norm": 0.05039280652999878, "learning_rate": 0.0005861224443026595, "loss": 0.5433, "step": 456 }, { "epoch": 0.1984368215371255, "grad_norm": 0.05184314772486687, "learning_rate": 0.0005845431070817626, "loss": 0.5422, "step": 457 }, { "epoch": 0.19887103777681286, "grad_norm": 0.056363269686698914, "learning_rate": 0.000582962901049634, "loss": 0.5202, "step": 458 }, { "epoch": 0.1993052540165002, "grad_norm": 0.04747706279158592, "learning_rate": 0.0005813818424453351, "loss": 0.5287, "step": 459 }, { "epoch": 0.19973947025618757, "grad_norm": 0.05112398415803909, "learning_rate": 0.0005797999475166897, "loss": 0.563, "step": 460 }, { "epoch": 0.20017368649587494, "grad_norm": 0.050983451306819916, "learning_rate": 0.0005782172325201155, "loss": 0.5309, "step": 461 }, { "epoch": 0.2006079027355623, "grad_norm": 0.05590641498565674, "learning_rate": 0.000576633713720458, "loss": 0.54, "step": 462 }, { "epoch": 0.20104211897524968, "grad_norm": 0.04705791547894478, "learning_rate": 0.000575049407390823, "loss": 0.5291, "step": 463 }, { "epoch": 0.20147633521493705, "grad_norm": 0.05234816297888756, "learning_rate": 0.000573464329812409, "loss": 0.5331, "step": 464 }, { "epoch": 0.20191055145462442, "grad_norm": 0.04806197062134743, "learning_rate": 0.0005718784972743409, "loss": 0.5141, "step": 465 }, { "epoch": 0.20234476769431176, "grad_norm": 0.0483107753098011, "learning_rate": 0.0005702919260735014, "loss": 0.5066, "step": 466 }, { "epoch": 0.20277898393399912, "grad_norm": 0.05739190801978111, "learning_rate": 0.0005687046325143647, "loss": 0.5282, "step": 467 }, { "epoch": 0.2032132001736865, "grad_norm": 0.04906076937913895, "learning_rate": 0.0005671166329088278, "loss": 0.5257, "step": 468 }, { "epoch": 0.20364741641337386, "grad_norm": 0.04784254729747772, "learning_rate": 0.0005655279435760435, "loss": 0.5193, "step": 469 }, { "epoch": 0.20408163265306123, "grad_norm": 0.06473550200462341, "learning_rate": 0.000563938580842253, "loss": 0.5257, "step": 470 }, { "epoch": 0.2045158488927486, "grad_norm": 0.047775350511074066, "learning_rate": 0.0005623485610406173, "loss": 0.5008, "step": 471 }, { "epoch": 0.20495006513243597, "grad_norm": 0.05449296161532402, "learning_rate": 0.0005607579005110502, "loss": 0.5316, "step": 472 }, { "epoch": 0.2053842813721233, "grad_norm": 0.05296491086483002, "learning_rate": 0.0005591666156000494, "loss": 0.5493, "step": 473 }, { "epoch": 0.20581849761181067, "grad_norm": 0.045328423380851746, "learning_rate": 0.0005575747226605297, "loss": 0.508, "step": 474 }, { "epoch": 0.20625271385149804, "grad_norm": 0.05045296251773834, "learning_rate": 0.0005559822380516539, "loss": 0.5118, "step": 475 }, { "epoch": 0.2066869300911854, "grad_norm": 0.04571797326207161, "learning_rate": 0.0005543891781386656, "loss": 0.5168, "step": 476 }, { "epoch": 0.20712114633087278, "grad_norm": 0.05250009894371033, "learning_rate": 0.0005527955592927197, "loss": 0.5175, "step": 477 }, { "epoch": 0.20755536257056015, "grad_norm": 0.048049163073301315, "learning_rate": 0.0005512013978907157, "loss": 0.5279, "step": 478 }, { "epoch": 0.20798957881024752, "grad_norm": 0.04621399939060211, "learning_rate": 0.0005496067103151288, "loss": 0.5304, "step": 479 }, { "epoch": 0.20842379504993486, "grad_norm": 0.04643552750349045, "learning_rate": 0.0005480115129538409, "loss": 0.5137, "step": 480 }, { "epoch": 0.20885801128962223, "grad_norm": 0.05422172695398331, "learning_rate": 0.0005464158221999731, "loss": 0.525, "step": 481 }, { "epoch": 0.2092922275293096, "grad_norm": 0.04593772441148758, "learning_rate": 0.0005448196544517168, "loss": 0.5229, "step": 482 }, { "epoch": 0.20972644376899696, "grad_norm": 0.05278801918029785, "learning_rate": 0.0005432230261121651, "loss": 0.5256, "step": 483 }, { "epoch": 0.21016066000868433, "grad_norm": 0.04598787799477577, "learning_rate": 0.0005416259535891447, "loss": 0.5275, "step": 484 }, { "epoch": 0.2105948762483717, "grad_norm": 0.054117828607559204, "learning_rate": 0.0005400284532950467, "loss": 0.5179, "step": 485 }, { "epoch": 0.21102909248805907, "grad_norm": 0.046420540660619736, "learning_rate": 0.0005384305416466584, "loss": 0.5399, "step": 486 }, { "epoch": 0.2114633087277464, "grad_norm": 0.048214301466941833, "learning_rate": 0.0005368322350649942, "loss": 0.5202, "step": 487 }, { "epoch": 0.21189752496743378, "grad_norm": 0.05419298633933067, "learning_rate": 0.0005352335499751269, "loss": 0.4925, "step": 488 }, { "epoch": 0.21233174120712114, "grad_norm": 0.04759565740823746, "learning_rate": 0.0005336345028060199, "loss": 0.5162, "step": 489 }, { "epoch": 0.2127659574468085, "grad_norm": 0.05174829810857773, "learning_rate": 0.0005320351099903565, "loss": 0.5043, "step": 490 }, { "epoch": 0.21320017368649588, "grad_norm": 0.06595347821712494, "learning_rate": 0.0005304353879643726, "loss": 0.5029, "step": 491 }, { "epoch": 0.21363438992618325, "grad_norm": 0.04908424988389015, "learning_rate": 0.0005288353531676872, "loss": 0.5043, "step": 492 }, { "epoch": 0.21406860616587062, "grad_norm": 0.05055024474859238, "learning_rate": 0.0005272350220431334, "loss": 0.5192, "step": 493 }, { "epoch": 0.21450282240555796, "grad_norm": 0.049286942929029465, "learning_rate": 0.0005256344110365896, "loss": 0.5041, "step": 494 }, { "epoch": 0.21493703864524533, "grad_norm": 0.04700294882059097, "learning_rate": 0.0005240335365968104, "loss": 0.5258, "step": 495 }, { "epoch": 0.2153712548849327, "grad_norm": 0.0560920424759388, "learning_rate": 0.0005224324151752575, "loss": 0.5388, "step": 496 }, { "epoch": 0.21580547112462006, "grad_norm": 0.05053974315524101, "learning_rate": 0.0005208310632259308, "loss": 0.5045, "step": 497 }, { "epoch": 0.21623968736430743, "grad_norm": 0.06237909570336342, "learning_rate": 0.0005192294972051992, "loss": 0.463, "step": 498 }, { "epoch": 0.2166739036039948, "grad_norm": 0.05128193646669388, "learning_rate": 0.0005176277335716317, "loss": 0.5209, "step": 499 }, { "epoch": 0.21710811984368214, "grad_norm": 0.04433543235063553, "learning_rate": 0.0005160257887858277, "loss": 0.5046, "step": 500 }, { "epoch": 0.2175423360833695, "grad_norm": 0.05125072970986366, "learning_rate": 0.0005144236793102484, "loss": 0.4894, "step": 501 }, { "epoch": 0.21797655232305688, "grad_norm": 0.04465992748737335, "learning_rate": 0.0005128214216090478, "loss": 0.4987, "step": 502 }, { "epoch": 0.21841076856274425, "grad_norm": 0.04684034362435341, "learning_rate": 0.0005112190321479025, "loss": 0.5498, "step": 503 }, { "epoch": 0.2188449848024316, "grad_norm": 0.04876565933227539, "learning_rate": 0.0005096165273938436, "loss": 0.5071, "step": 504 }, { "epoch": 0.21927920104211898, "grad_norm": 0.04982059448957443, "learning_rate": 0.0005080139238150869, "loss": 0.5084, "step": 505 }, { "epoch": 0.21971341728180635, "grad_norm": 0.052202560007572174, "learning_rate": 0.0005064112378808637, "loss": 0.5233, "step": 506 }, { "epoch": 0.2201476335214937, "grad_norm": 0.04747169837355614, "learning_rate": 0.0005048084860612516, "loss": 0.5263, "step": 507 }, { "epoch": 0.22058184976118106, "grad_norm": 0.0462409108877182, "learning_rate": 0.0005032056848270056, "loss": 0.5237, "step": 508 }, { "epoch": 0.22101606600086843, "grad_norm": 0.04612501338124275, "learning_rate": 0.000501602850649388, "loss": 0.5507, "step": 509 }, { "epoch": 0.2214502822405558, "grad_norm": 0.04581817239522934, "learning_rate": 0.0005, "loss": 0.4986, "step": 510 }, { "epoch": 0.22188449848024316, "grad_norm": 0.0477612242102623, "learning_rate": 0.0004983971493506121, "loss": 0.51, "step": 511 }, { "epoch": 0.22231871471993053, "grad_norm": 0.05176355689764023, "learning_rate": 0.0004967943151729944, "loss": 0.5115, "step": 512 }, { "epoch": 0.2227529309596179, "grad_norm": 0.04731719568371773, "learning_rate": 0.0004951915139387483, "loss": 0.5564, "step": 513 }, { "epoch": 0.22318714719930524, "grad_norm": 0.04438960924744606, "learning_rate": 0.0004935887621191363, "loss": 0.5502, "step": 514 }, { "epoch": 0.2236213634389926, "grad_norm": 0.06116556003689766, "learning_rate": 0.0004919860761849132, "loss": 0.5417, "step": 515 }, { "epoch": 0.22405557967867998, "grad_norm": 0.046353522688150406, "learning_rate": 0.0004903834726061564, "loss": 0.5285, "step": 516 }, { "epoch": 0.22448979591836735, "grad_norm": 0.04643326997756958, "learning_rate": 0.0004887809678520976, "loss": 0.4774, "step": 517 }, { "epoch": 0.22492401215805471, "grad_norm": 0.0530591681599617, "learning_rate": 0.0004871785783909523, "loss": 0.5076, "step": 518 }, { "epoch": 0.22535822839774208, "grad_norm": 0.04789597541093826, "learning_rate": 0.0004855763206897516, "loss": 0.4989, "step": 519 }, { "epoch": 0.22579244463742945, "grad_norm": 0.047394026070833206, "learning_rate": 0.0004839742112141724, "loss": 0.5069, "step": 520 }, { "epoch": 0.2262266608771168, "grad_norm": 0.044425006955862045, "learning_rate": 0.0004823722664283684, "loss": 0.5212, "step": 521 }, { "epoch": 0.22666087711680416, "grad_norm": 0.044534265995025635, "learning_rate": 0.0004807705027948008, "loss": 0.5184, "step": 522 }, { "epoch": 0.22709509335649153, "grad_norm": 0.04293319582939148, "learning_rate": 0.0004791689367740692, "loss": 0.4807, "step": 523 }, { "epoch": 0.2275293095961789, "grad_norm": 0.04485393315553665, "learning_rate": 0.0004775675848247427, "loss": 0.4746, "step": 524 }, { "epoch": 0.22796352583586627, "grad_norm": 0.043829191476106644, "learning_rate": 0.0004759664634031897, "loss": 0.4936, "step": 525 }, { "epoch": 0.22839774207555363, "grad_norm": 0.05110118165612221, "learning_rate": 0.00047436558896341046, "loss": 0.5849, "step": 526 }, { "epoch": 0.228831958315241, "grad_norm": 0.04783101752400398, "learning_rate": 0.0004727649779568666, "loss": 0.5034, "step": 527 }, { "epoch": 0.22926617455492834, "grad_norm": 0.047072965651750565, "learning_rate": 0.00047116464683231285, "loss": 0.5155, "step": 528 }, { "epoch": 0.2297003907946157, "grad_norm": 0.0427449531853199, "learning_rate": 0.0004695646120356274, "loss": 0.5138, "step": 529 }, { "epoch": 0.23013460703430308, "grad_norm": 0.04664051905274391, "learning_rate": 0.0004679648900096436, "loss": 0.4727, "step": 530 }, { "epoch": 0.23056882327399045, "grad_norm": 0.04555191844701767, "learning_rate": 0.0004663654971939802, "loss": 0.5386, "step": 531 }, { "epoch": 0.23100303951367782, "grad_norm": 0.05113929882645607, "learning_rate": 0.00046476645002487297, "loss": 0.5018, "step": 532 }, { "epoch": 0.23143725575336518, "grad_norm": 0.04245986044406891, "learning_rate": 0.00046316776493500613, "loss": 0.4949, "step": 533 }, { "epoch": 0.23187147199305255, "grad_norm": 0.047864172607660294, "learning_rate": 0.0004615694583533418, "loss": 0.5032, "step": 534 }, { "epoch": 0.2323056882327399, "grad_norm": 0.06898235529661179, "learning_rate": 0.0004599715467049534, "loss": 0.4922, "step": 535 }, { "epoch": 0.23273990447242726, "grad_norm": 0.04230332002043724, "learning_rate": 0.0004583740464108554, "loss": 0.5164, "step": 536 }, { "epoch": 0.23317412071211463, "grad_norm": 0.043854959309101105, "learning_rate": 0.00045677697388783495, "loss": 0.4894, "step": 537 }, { "epoch": 0.233608336951802, "grad_norm": 0.05595362186431885, "learning_rate": 0.0004551803455482833, "loss": 0.5107, "step": 538 }, { "epoch": 0.23404255319148937, "grad_norm": 0.05336504802107811, "learning_rate": 0.0004535841778000269, "loss": 0.4845, "step": 539 }, { "epoch": 0.23447676943117673, "grad_norm": 0.052477382123470306, "learning_rate": 0.0004519884870461591, "loss": 0.5215, "step": 540 }, { "epoch": 0.2349109856708641, "grad_norm": 0.04965275526046753, "learning_rate": 0.00045039328968487125, "loss": 0.499, "step": 541 }, { "epoch": 0.23534520191055144, "grad_norm": 0.07106837630271912, "learning_rate": 0.00044879860210928434, "loss": 0.5159, "step": 542 }, { "epoch": 0.2357794181502388, "grad_norm": 0.06256967037916183, "learning_rate": 0.0004472044407072805, "loss": 0.505, "step": 543 }, { "epoch": 0.23621363438992618, "grad_norm": 0.04907793179154396, "learning_rate": 0.0004456108218613346, "loss": 0.5117, "step": 544 }, { "epoch": 0.23664785062961355, "grad_norm": 0.06353382021188736, "learning_rate": 0.0004440177619483461, "loss": 0.4941, "step": 545 }, { "epoch": 0.23708206686930092, "grad_norm": 0.05265835300087929, "learning_rate": 0.0004424252773394704, "loss": 0.5179, "step": 546 }, { "epoch": 0.23751628310898829, "grad_norm": 0.05525769665837288, "learning_rate": 0.0004408333843999506, "loss": 0.4938, "step": 547 }, { "epoch": 0.23795049934867565, "grad_norm": 0.06166142225265503, "learning_rate": 0.00043924209948894985, "loss": 0.5287, "step": 548 }, { "epoch": 0.238384715588363, "grad_norm": 0.04812607169151306, "learning_rate": 0.0004376514389593826, "loss": 0.5401, "step": 549 }, { "epoch": 0.23881893182805036, "grad_norm": 0.048272911459207535, "learning_rate": 0.00043606141915774693, "loss": 0.5258, "step": 550 }, { "epoch": 0.23925314806773773, "grad_norm": 0.04806293919682503, "learning_rate": 0.00043447205642395664, "loss": 0.5023, "step": 551 }, { "epoch": 0.2396873643074251, "grad_norm": 0.04957396164536476, "learning_rate": 0.0004328833670911724, "loss": 0.5015, "step": 552 }, { "epoch": 0.24012158054711247, "grad_norm": 0.04413224756717682, "learning_rate": 0.00043129536748563545, "loss": 0.5336, "step": 553 }, { "epoch": 0.24055579678679984, "grad_norm": 0.04619096592068672, "learning_rate": 0.0004297080739264987, "loss": 0.4832, "step": 554 }, { "epoch": 0.2409900130264872, "grad_norm": 0.05157000198960304, "learning_rate": 0.0004281215027256592, "loss": 0.529, "step": 555 }, { "epoch": 0.24142422926617454, "grad_norm": 0.04697950556874275, "learning_rate": 0.00042653567018759105, "loss": 0.5034, "step": 556 }, { "epoch": 0.2418584455058619, "grad_norm": 0.04245249181985855, "learning_rate": 0.0004249505926091771, "loss": 0.4839, "step": 557 }, { "epoch": 0.24229266174554928, "grad_norm": 0.042621515691280365, "learning_rate": 0.000423366286279542, "loss": 0.5639, "step": 558 }, { "epoch": 0.24272687798523665, "grad_norm": 0.054264895617961884, "learning_rate": 0.0004217827674798845, "loss": 0.5077, "step": 559 }, { "epoch": 0.24316109422492402, "grad_norm": 0.04171125963330269, "learning_rate": 0.0004202000524833105, "loss": 0.5001, "step": 560 }, { "epoch": 0.2435953104646114, "grad_norm": 0.047443173825740814, "learning_rate": 0.00041861815755466506, "loss": 0.5153, "step": 561 }, { "epoch": 0.24402952670429873, "grad_norm": 0.04359521344304085, "learning_rate": 0.0004170370989503662, "loss": 0.516, "step": 562 }, { "epoch": 0.2444637429439861, "grad_norm": 0.042067334055900574, "learning_rate": 0.0004154568929182374, "loss": 0.4783, "step": 563 }, { "epoch": 0.24489795918367346, "grad_norm": 0.042515527456998825, "learning_rate": 0.00041387755569734057, "loss": 0.5112, "step": 564 }, { "epoch": 0.24533217542336083, "grad_norm": 0.04880267009139061, "learning_rate": 0.00041229910351780926, "loss": 0.5108, "step": 565 }, { "epoch": 0.2457663916630482, "grad_norm": 0.048264361917972565, "learning_rate": 0.0004107215526006817, "loss": 0.4802, "step": 566 }, { "epoch": 0.24620060790273557, "grad_norm": 0.0446770153939724, "learning_rate": 0.00040914491915773453, "loss": 0.5461, "step": 567 }, { "epoch": 0.24663482414242294, "grad_norm": 0.05128175765275955, "learning_rate": 0.00040756921939131565, "loss": 0.5288, "step": 568 }, { "epoch": 0.24706904038211028, "grad_norm": 0.05804765224456787, "learning_rate": 0.0004059944694941783, "loss": 0.4981, "step": 569 }, { "epoch": 0.24750325662179765, "grad_norm": 0.04847874119877815, "learning_rate": 0.00040442068564931397, "loss": 0.5185, "step": 570 }, { "epoch": 0.24793747286148501, "grad_norm": 0.04028409346938133, "learning_rate": 0.0004028478840297866, "loss": 0.493, "step": 571 }, { "epoch": 0.24837168910117238, "grad_norm": 0.04243363067507744, "learning_rate": 0.00040127608079856646, "loss": 0.5079, "step": 572 }, { "epoch": 0.24880590534085975, "grad_norm": 0.04515422508120537, "learning_rate": 0.00039970529210836363, "loss": 0.5121, "step": 573 }, { "epoch": 0.24924012158054712, "grad_norm": 0.05250799283385277, "learning_rate": 0.00039813553410146226, "loss": 0.4975, "step": 574 }, { "epoch": 0.2496743378202345, "grad_norm": 0.043340008705854416, "learning_rate": 0.00039656682290955457, "loss": 0.5094, "step": 575 }, { "epoch": 0.25010855405992183, "grad_norm": 0.04954323545098305, "learning_rate": 0.00039499917465357534, "loss": 0.5327, "step": 576 }, { "epoch": 0.2505427702996092, "grad_norm": 0.04032951593399048, "learning_rate": 0.00039343260544353573, "loss": 0.5035, "step": 577 }, { "epoch": 0.25097698653929656, "grad_norm": 0.04386662319302559, "learning_rate": 0.0003918671313783583, "loss": 0.5142, "step": 578 }, { "epoch": 0.25141120277898393, "grad_norm": 0.04662942886352539, "learning_rate": 0.00039030276854571115, "loss": 0.4996, "step": 579 }, { "epoch": 0.2518454190186713, "grad_norm": 0.046797946095466614, "learning_rate": 0.00038873953302184284, "loss": 0.4728, "step": 580 }, { "epoch": 0.25227963525835867, "grad_norm": 0.041140951216220856, "learning_rate": 0.000387177440871417, "loss": 0.546, "step": 581 }, { "epoch": 0.25271385149804604, "grad_norm": 0.04658018797636032, "learning_rate": 0.00038561650814734736, "loss": 0.4804, "step": 582 }, { "epoch": 0.2531480677377334, "grad_norm": 0.04261363297700882, "learning_rate": 0.0003840567508906328, "loss": 0.4833, "step": 583 }, { "epoch": 0.2535822839774208, "grad_norm": 0.046267736703157425, "learning_rate": 0.0003824981851301924, "loss": 0.4809, "step": 584 }, { "epoch": 0.25401650021710814, "grad_norm": 0.04030013084411621, "learning_rate": 0.0003809408268827009, "loss": 0.4878, "step": 585 }, { "epoch": 0.2544507164567955, "grad_norm": 0.041512105613946915, "learning_rate": 0.0003793846921524237, "loss": 0.5165, "step": 586 }, { "epoch": 0.2548849326964828, "grad_norm": 0.04772355034947395, "learning_rate": 0.00037782979693105293, "loss": 0.5193, "step": 587 }, { "epoch": 0.2553191489361702, "grad_norm": 0.040986839681863785, "learning_rate": 0.00037627615719754295, "loss": 0.5028, "step": 588 }, { "epoch": 0.25575336517585756, "grad_norm": 0.04139196127653122, "learning_rate": 0.00037472378891794533, "loss": 0.4998, "step": 589 }, { "epoch": 0.25618758141554493, "grad_norm": 0.04017603024840355, "learning_rate": 0.0003731727080452464, "loss": 0.5239, "step": 590 }, { "epoch": 0.2566217976552323, "grad_norm": 0.04039851576089859, "learning_rate": 0.00037162293051920184, "loss": 0.4739, "step": 591 }, { "epoch": 0.25705601389491967, "grad_norm": 0.04436818137764931, "learning_rate": 0.0003700744722661736, "loss": 0.5153, "step": 592 }, { "epoch": 0.25749023013460703, "grad_norm": 0.03977528214454651, "learning_rate": 0.0003685273491989661, "loss": 0.5217, "step": 593 }, { "epoch": 0.2579244463742944, "grad_norm": 0.038483258336782455, "learning_rate": 0.0003669815772166625, "loss": 0.4904, "step": 594 }, { "epoch": 0.25835866261398177, "grad_norm": 0.04520029202103615, "learning_rate": 0.0003654371722044616, "loss": 0.5303, "step": 595 }, { "epoch": 0.25879287885366914, "grad_norm": 0.04642047360539436, "learning_rate": 0.00036389415003351444, "loss": 0.4783, "step": 596 }, { "epoch": 0.2592270950933565, "grad_norm": 0.037170182913541794, "learning_rate": 0.00036235252656076133, "loss": 0.5254, "step": 597 }, { "epoch": 0.2596613113330439, "grad_norm": 0.037870265543460846, "learning_rate": 0.00036081231762876846, "loss": 0.4933, "step": 598 }, { "epoch": 0.26009552757273124, "grad_norm": 0.03957228735089302, "learning_rate": 0.0003592735390655658, "loss": 0.5194, "step": 599 }, { "epoch": 0.26052974381241856, "grad_norm": 0.04471513256430626, "learning_rate": 0.0003577362066844838, "loss": 0.4737, "step": 600 }, { "epoch": 0.2609639600521059, "grad_norm": 0.0370667465031147, "learning_rate": 0.0003562003362839914, "loss": 0.5072, "step": 601 }, { "epoch": 0.2613981762917933, "grad_norm": 0.04277738928794861, "learning_rate": 0.00035466594364753326, "loss": 0.5093, "step": 602 }, { "epoch": 0.26183239253148066, "grad_norm": 0.04622077941894531, "learning_rate": 0.00035313304454336763, "loss": 0.5008, "step": 603 }, { "epoch": 0.26226660877116803, "grad_norm": 0.0380844846367836, "learning_rate": 0.0003516016547244047, "loss": 0.4899, "step": 604 }, { "epoch": 0.2627008250108554, "grad_norm": 0.03964554890990257, "learning_rate": 0.0003500717899280442, "loss": 0.544, "step": 605 }, { "epoch": 0.26313504125054277, "grad_norm": 0.041507843881845474, "learning_rate": 0.00034854346587601397, "loss": 0.5341, "step": 606 }, { "epoch": 0.26356925749023014, "grad_norm": 0.04041410982608795, "learning_rate": 0.0003470166982742082, "loss": 0.5325, "step": 607 }, { "epoch": 0.2640034737299175, "grad_norm": 0.03915121778845787, "learning_rate": 0.00034549150281252633, "loss": 0.5164, "step": 608 }, { "epoch": 0.26443768996960487, "grad_norm": 0.04141729697585106, "learning_rate": 0.0003439678951647115, "loss": 0.4988, "step": 609 }, { "epoch": 0.26487190620929224, "grad_norm": 0.043251313269138336, "learning_rate": 0.0003424458909881897, "loss": 0.5319, "step": 610 }, { "epoch": 0.2653061224489796, "grad_norm": 0.045615166425704956, "learning_rate": 0.00034092550592390857, "loss": 0.4668, "step": 611 }, { "epoch": 0.265740338688667, "grad_norm": 0.03943486511707306, "learning_rate": 0.00033940675559617726, "loss": 0.4991, "step": 612 }, { "epoch": 0.26617455492835435, "grad_norm": 0.0382043793797493, "learning_rate": 0.000337889655612505, "loss": 0.5117, "step": 613 }, { "epoch": 0.26660877116804166, "grad_norm": 0.039549313485622406, "learning_rate": 0.00033637422156344155, "loss": 0.5029, "step": 614 }, { "epoch": 0.267042987407729, "grad_norm": 0.03972559794783592, "learning_rate": 0.00033486046902241664, "loss": 0.5073, "step": 615 }, { "epoch": 0.2674772036474164, "grad_norm": 0.03974078595638275, "learning_rate": 0.0003333484135455792, "loss": 0.5083, "step": 616 }, { "epoch": 0.26791141988710376, "grad_norm": 0.039242375642061234, "learning_rate": 0.00033183807067163916, "loss": 0.4979, "step": 617 }, { "epoch": 0.26834563612679113, "grad_norm": 0.045767784118652344, "learning_rate": 0.0003303294559217063, "loss": 0.4835, "step": 618 }, { "epoch": 0.2687798523664785, "grad_norm": 0.03927744925022125, "learning_rate": 0.0003288225847991312, "loss": 0.5103, "step": 619 }, { "epoch": 0.26921406860616587, "grad_norm": 0.03815273195505142, "learning_rate": 0.0003273174727893463, "loss": 0.5097, "step": 620 }, { "epoch": 0.26964828484585324, "grad_norm": 0.0421292670071125, "learning_rate": 0.00032581413535970593, "loss": 0.4996, "step": 621 }, { "epoch": 0.2700825010855406, "grad_norm": 0.0482180081307888, "learning_rate": 0.0003243125879593286, "loss": 0.4785, "step": 622 }, { "epoch": 0.270516717325228, "grad_norm": 0.03935736045241356, "learning_rate": 0.0003228128460189368, "loss": 0.5063, "step": 623 }, { "epoch": 0.27095093356491534, "grad_norm": 0.037722837179899216, "learning_rate": 0.0003213149249506997, "loss": 0.4979, "step": 624 }, { "epoch": 0.2713851498046027, "grad_norm": 0.03764335811138153, "learning_rate": 0.0003198188401480734, "loss": 0.4818, "step": 625 }, { "epoch": 0.2718193660442901, "grad_norm": 0.04136700555682182, "learning_rate": 0.0003183246069856443, "loss": 0.5144, "step": 626 }, { "epoch": 0.27225358228397745, "grad_norm": 0.04152214527130127, "learning_rate": 0.00031683224081897, "loss": 0.5217, "step": 627 }, { "epoch": 0.27268779852366476, "grad_norm": 0.036686960607767105, "learning_rate": 0.00031534175698442194, "loss": 0.4938, "step": 628 }, { "epoch": 0.27312201476335213, "grad_norm": 0.0376245342195034, "learning_rate": 0.00031385317079902743, "loss": 0.5255, "step": 629 }, { "epoch": 0.2735562310030395, "grad_norm": 0.04143936559557915, "learning_rate": 0.000312366497560313, "loss": 0.5005, "step": 630 }, { "epoch": 0.27399044724272686, "grad_norm": 0.036747369915246964, "learning_rate": 0.00031088175254614616, "loss": 0.4621, "step": 631 }, { "epoch": 0.27442466348241423, "grad_norm": 0.039392732083797455, "learning_rate": 0.00030939895101457916, "loss": 0.5333, "step": 632 }, { "epoch": 0.2748588797221016, "grad_norm": 0.03783806785941124, "learning_rate": 0.0003079181082036922, "loss": 0.5075, "step": 633 }, { "epoch": 0.27529309596178897, "grad_norm": 0.03799246624112129, "learning_rate": 0.000306439239331436, "loss": 0.5086, "step": 634 }, { "epoch": 0.27572731220147634, "grad_norm": 0.040941022336483, "learning_rate": 0.0003049623595954766, "loss": 0.5125, "step": 635 }, { "epoch": 0.2761615284411637, "grad_norm": 0.045769453048706055, "learning_rate": 0.0003034874841730382, "loss": 0.4913, "step": 636 }, { "epoch": 0.2765957446808511, "grad_norm": 0.044418562203645706, "learning_rate": 0.00030201462822074786, "loss": 0.5225, "step": 637 }, { "epoch": 0.27702996092053844, "grad_norm": 0.036289479583501816, "learning_rate": 0.0003005438068744792, "loss": 0.5016, "step": 638 }, { "epoch": 0.2774641771602258, "grad_norm": 0.03927507996559143, "learning_rate": 0.0002990750352491973, "loss": 0.497, "step": 639 }, { "epoch": 0.2778983933999132, "grad_norm": 0.052069611847400665, "learning_rate": 0.0002976083284388031, "loss": 0.5026, "step": 640 }, { "epoch": 0.27833260963960055, "grad_norm": 0.0408078134059906, "learning_rate": 0.00029614370151597835, "loss": 0.5069, "step": 641 }, { "epoch": 0.27876682587928786, "grad_norm": 0.040615539997816086, "learning_rate": 0.0002946811695320311, "loss": 0.4878, "step": 642 }, { "epoch": 0.27920104211897523, "grad_norm": 0.04145561158657074, "learning_rate": 0.00029322074751673977, "loss": 0.4547, "step": 643 }, { "epoch": 0.2796352583586626, "grad_norm": 0.044274091720581055, "learning_rate": 0.00029176245047820063, "loss": 0.5051, "step": 644 }, { "epoch": 0.28006947459834997, "grad_norm": 0.055839963257312775, "learning_rate": 0.0002903062934026716, "loss": 0.495, "step": 645 }, { "epoch": 0.28050369083803733, "grad_norm": 0.038132019340991974, "learning_rate": 0.0002888522912544202, "loss": 0.4746, "step": 646 }, { "epoch": 0.2809379070777247, "grad_norm": 0.039124827831983566, "learning_rate": 0.00028740045897556767, "loss": 0.4978, "step": 647 }, { "epoch": 0.28137212331741207, "grad_norm": 0.03960909694433212, "learning_rate": 0.0002859508114859374, "loss": 0.5043, "step": 648 }, { "epoch": 0.28180633955709944, "grad_norm": 0.03705562651157379, "learning_rate": 0.00028450336368289974, "loss": 0.5041, "step": 649 }, { "epoch": 0.2822405557967868, "grad_norm": 0.03920963406562805, "learning_rate": 0.00028305813044122096, "loss": 0.5055, "step": 650 }, { "epoch": 0.2826747720364742, "grad_norm": 0.04196755215525627, "learning_rate": 0.00028161512661290845, "loss": 0.4962, "step": 651 }, { "epoch": 0.28310898827616154, "grad_norm": 0.03515279293060303, "learning_rate": 0.00028017436702705903, "loss": 0.4708, "step": 652 }, { "epoch": 0.2835432045158489, "grad_norm": 0.042391590774059296, "learning_rate": 0.0002787358664897068, "loss": 0.5268, "step": 653 }, { "epoch": 0.2839774207555363, "grad_norm": 0.03707970678806305, "learning_rate": 0.0002772996397836704, "loss": 0.5252, "step": 654 }, { "epoch": 0.28441163699522365, "grad_norm": 0.039679594337940216, "learning_rate": 0.0002758657016684015, "loss": 0.4954, "step": 655 }, { "epoch": 0.28484585323491096, "grad_norm": 0.03491951525211334, "learning_rate": 0.0002744340668798326, "loss": 0.4859, "step": 656 }, { "epoch": 0.28528006947459833, "grad_norm": 0.033995699137449265, "learning_rate": 0.00027300475013022663, "loss": 0.4917, "step": 657 }, { "epoch": 0.2857142857142857, "grad_norm": 0.0408964566886425, "learning_rate": 0.00027157776610802414, "loss": 0.5176, "step": 658 }, { "epoch": 0.28614850195397307, "grad_norm": 0.035840053111314774, "learning_rate": 0.0002701531294776943, "loss": 0.4757, "step": 659 }, { "epoch": 0.28658271819366044, "grad_norm": 0.03554424270987511, "learning_rate": 0.0002687308548795825, "loss": 0.5043, "step": 660 }, { "epoch": 0.2870169344333478, "grad_norm": 0.03549895063042641, "learning_rate": 0.00026731095692976073, "loss": 0.4924, "step": 661 }, { "epoch": 0.28745115067303517, "grad_norm": 0.0438251756131649, "learning_rate": 0.00026589345021987723, "loss": 0.4986, "step": 662 }, { "epoch": 0.28788536691272254, "grad_norm": 0.037839341908693314, "learning_rate": 0.00026447834931700686, "loss": 0.4716, "step": 663 }, { "epoch": 0.2883195831524099, "grad_norm": 0.03682737424969673, "learning_rate": 0.0002630656687635007, "loss": 0.4991, "step": 664 }, { "epoch": 0.2887537993920973, "grad_norm": 0.03697109594941139, "learning_rate": 0.0002616554230768374, "loss": 0.5104, "step": 665 }, { "epoch": 0.28918801563178465, "grad_norm": 0.036725230515003204, "learning_rate": 0.0002602476267494731, "loss": 0.5163, "step": 666 }, { "epoch": 0.289622231871472, "grad_norm": 0.04177143797278404, "learning_rate": 0.0002588422942486932, "loss": 0.4567, "step": 667 }, { "epoch": 0.2900564481111594, "grad_norm": 0.04078909382224083, "learning_rate": 0.0002574394400164639, "loss": 0.4946, "step": 668 }, { "epoch": 0.2904906643508467, "grad_norm": 0.04146347567439079, "learning_rate": 0.0002560390784692828, "loss": 0.4947, "step": 669 }, { "epoch": 0.29092488059053406, "grad_norm": 0.03871840611100197, "learning_rate": 0.00025464122399803123, "loss": 0.5238, "step": 670 }, { "epoch": 0.29135909683022143, "grad_norm": 0.04148663207888603, "learning_rate": 0.00025324589096782657, "loss": 0.5246, "step": 671 }, { "epoch": 0.2917933130699088, "grad_norm": 0.03910430520772934, "learning_rate": 0.0002518530937178751, "loss": 0.4846, "step": 672 }, { "epoch": 0.29222752930959617, "grad_norm": 0.036824680864810944, "learning_rate": 0.000250462846561323, "loss": 0.4654, "step": 673 }, { "epoch": 0.29266174554928354, "grad_norm": 0.036250337958335876, "learning_rate": 0.00024907516378511136, "loss": 0.5106, "step": 674 }, { "epoch": 0.2930959617889709, "grad_norm": 0.03372564911842346, "learning_rate": 0.00024769005964982713, "loss": 0.5222, "step": 675 }, { "epoch": 0.2935301780286583, "grad_norm": 0.03601228445768356, "learning_rate": 0.000246307548389559, "loss": 0.5099, "step": 676 }, { "epoch": 0.29396439426834564, "grad_norm": 0.039937473833560944, "learning_rate": 0.0002449276442117486, "loss": 0.5262, "step": 677 }, { "epoch": 0.294398610508033, "grad_norm": 0.03719143569469452, "learning_rate": 0.000243550361297047, "loss": 0.5089, "step": 678 }, { "epoch": 0.2948328267477204, "grad_norm": 0.034734416753053665, "learning_rate": 0.00024217571379916668, "loss": 0.4962, "step": 679 }, { "epoch": 0.29526704298740775, "grad_norm": 0.03447417914867401, "learning_rate": 0.00024080371584473748, "loss": 0.4952, "step": 680 }, { "epoch": 0.2957012592270951, "grad_norm": 0.03631613776087761, "learning_rate": 0.00023943438153316155, "loss": 0.497, "step": 681 }, { "epoch": 0.2961354754667825, "grad_norm": 0.03350459039211273, "learning_rate": 0.00023806772493646723, "loss": 0.4775, "step": 682 }, { "epoch": 0.2965696917064698, "grad_norm": 0.032677456736564636, "learning_rate": 0.00023670376009916595, "loss": 0.5001, "step": 683 }, { "epoch": 0.29700390794615716, "grad_norm": 0.03544562682509422, "learning_rate": 0.00023534250103810628, "loss": 0.4796, "step": 684 }, { "epoch": 0.29743812418584453, "grad_norm": 0.03991761803627014, "learning_rate": 0.00023398396174233177, "loss": 0.5307, "step": 685 }, { "epoch": 0.2978723404255319, "grad_norm": 0.037654612213373184, "learning_rate": 0.00023262815617293515, "loss": 0.4727, "step": 686 }, { "epoch": 0.29830655666521927, "grad_norm": 0.03447496145963669, "learning_rate": 0.00023127509826291698, "loss": 0.5037, "step": 687 }, { "epoch": 0.29874077290490664, "grad_norm": 0.03626209869980812, "learning_rate": 0.00022992480191704002, "loss": 0.4758, "step": 688 }, { "epoch": 0.299174989144594, "grad_norm": 0.03519487380981445, "learning_rate": 0.000228577281011689, "loss": 0.5046, "step": 689 }, { "epoch": 0.2996092053842814, "grad_norm": 0.036514561623334885, "learning_rate": 0.0002272325493947257, "loss": 0.5014, "step": 690 }, { "epoch": 0.30004342162396874, "grad_norm": 0.036256395280361176, "learning_rate": 0.00022589062088534834, "loss": 0.5311, "step": 691 }, { "epoch": 0.3004776378636561, "grad_norm": 0.036770399659872055, "learning_rate": 0.0002245515092739488, "loss": 0.4922, "step": 692 }, { "epoch": 0.3009118541033435, "grad_norm": 0.037295542657375336, "learning_rate": 0.00022321522832197032, "loss": 0.4817, "step": 693 }, { "epoch": 0.30134607034303085, "grad_norm": 0.03913348540663719, "learning_rate": 0.00022188179176176764, "loss": 0.4565, "step": 694 }, { "epoch": 0.3017802865827182, "grad_norm": 0.04223249852657318, "learning_rate": 0.00022055121329646416, "loss": 0.5297, "step": 695 }, { "epoch": 0.3022145028224056, "grad_norm": 0.03583509102463722, "learning_rate": 0.0002192235065998126, "loss": 0.4808, "step": 696 }, { "epoch": 0.3026487190620929, "grad_norm": 0.04437342286109924, "learning_rate": 0.0002178986853160535, "loss": 0.4794, "step": 697 }, { "epoch": 0.30308293530178027, "grad_norm": 0.043945055454969406, "learning_rate": 0.0002165767630597752, "loss": 0.4959, "step": 698 }, { "epoch": 0.30351715154146763, "grad_norm": 0.04786218702793121, "learning_rate": 0.00021525775341577403, "loss": 0.5047, "step": 699 }, { "epoch": 0.303951367781155, "grad_norm": 0.041139792650938034, "learning_rate": 0.0002139416699389153, "loss": 0.5134, "step": 700 }, { "epoch": 0.30438558402084237, "grad_norm": 0.034917134791612625, "learning_rate": 0.00021262852615399258, "loss": 0.4924, "step": 701 }, { "epoch": 0.30481980026052974, "grad_norm": 0.03301383554935455, "learning_rate": 0.0002113183355555904, "loss": 0.4862, "step": 702 }, { "epoch": 0.3052540165002171, "grad_norm": 0.0419352687895298, "learning_rate": 0.00021001111160794383, "loss": 0.5147, "step": 703 }, { "epoch": 0.3056882327399045, "grad_norm": 0.03364208713173866, "learning_rate": 0.00020870686774480197, "loss": 0.5236, "step": 704 }, { "epoch": 0.30612244897959184, "grad_norm": 0.03912827745079994, "learning_rate": 0.0002074056173692881, "loss": 0.5042, "step": 705 }, { "epoch": 0.3065566652192792, "grad_norm": 0.03657132387161255, "learning_rate": 0.00020610737385376348, "loss": 0.5011, "step": 706 }, { "epoch": 0.3069908814589666, "grad_norm": 0.03518640622496605, "learning_rate": 0.00020481215053968872, "loss": 0.4798, "step": 707 }, { "epoch": 0.30742509769865395, "grad_norm": 0.03507522493600845, "learning_rate": 0.00020351996073748714, "loss": 0.5313, "step": 708 }, { "epoch": 0.3078593139383413, "grad_norm": 0.03609062731266022, "learning_rate": 0.00020223081772640866, "loss": 0.4785, "step": 709 }, { "epoch": 0.3082935301780287, "grad_norm": 0.03479000926017761, "learning_rate": 0.000200944734754392, "loss": 0.4835, "step": 710 }, { "epoch": 0.308727746417716, "grad_norm": 0.044929634779691696, "learning_rate": 0.00019966172503792985, "loss": 0.5122, "step": 711 }, { "epoch": 0.30916196265740337, "grad_norm": 0.03376319631934166, "learning_rate": 0.00019838180176193177, "loss": 0.4876, "step": 712 }, { "epoch": 0.30959617889709073, "grad_norm": 0.03246736153960228, "learning_rate": 0.0001971049780795901, "loss": 0.4804, "step": 713 }, { "epoch": 0.3100303951367781, "grad_norm": 0.031609781086444855, "learning_rate": 0.00019583126711224342, "loss": 0.545, "step": 714 }, { "epoch": 0.31046461137646547, "grad_norm": 0.034195881336927414, "learning_rate": 0.00019456068194924288, "loss": 0.5096, "step": 715 }, { "epoch": 0.31089882761615284, "grad_norm": 0.03344978764653206, "learning_rate": 0.0001932932356478168, "loss": 0.5079, "step": 716 }, { "epoch": 0.3113330438558402, "grad_norm": 0.03219503536820412, "learning_rate": 0.00019202894123293674, "loss": 0.5339, "step": 717 }, { "epoch": 0.3117672600955276, "grad_norm": 0.03737034648656845, "learning_rate": 0.00019076781169718426, "loss": 0.5131, "step": 718 }, { "epoch": 0.31220147633521494, "grad_norm": 0.03394628316164017, "learning_rate": 0.00018950986000061638, "loss": 0.4969, "step": 719 }, { "epoch": 0.3126356925749023, "grad_norm": 0.032426606863737106, "learning_rate": 0.00018825509907063325, "loss": 0.49, "step": 720 }, { "epoch": 0.3130699088145897, "grad_norm": 0.03836345672607422, "learning_rate": 0.00018700354180184464, "loss": 0.4925, "step": 721 }, { "epoch": 0.31350412505427705, "grad_norm": 0.035867128521203995, "learning_rate": 0.0001857552010559382, "loss": 0.5466, "step": 722 }, { "epoch": 0.3139383412939644, "grad_norm": 0.03811186924576759, "learning_rate": 0.0001845100896615462, "loss": 0.4825, "step": 723 }, { "epoch": 0.31437255753365173, "grad_norm": 0.032081685960292816, "learning_rate": 0.00018326822041411523, "loss": 0.4858, "step": 724 }, { "epoch": 0.3148067737733391, "grad_norm": 0.03794045001268387, "learning_rate": 0.00018202960607577247, "loss": 0.5023, "step": 725 }, { "epoch": 0.31524099001302647, "grad_norm": 0.033024583011865616, "learning_rate": 0.00018079425937519728, "loss": 0.4744, "step": 726 }, { "epoch": 0.31567520625271384, "grad_norm": 0.03800756111741066, "learning_rate": 0.00017956219300748795, "loss": 0.4813, "step": 727 }, { "epoch": 0.3161094224924012, "grad_norm": 0.03822485730051994, "learning_rate": 0.0001783334196340331, "loss": 0.4984, "step": 728 }, { "epoch": 0.3165436387320886, "grad_norm": 0.03531699255108833, "learning_rate": 0.00017710795188237988, "loss": 0.5095, "step": 729 }, { "epoch": 0.31697785497177594, "grad_norm": 0.03231954947113991, "learning_rate": 0.0001758858023461059, "loss": 0.4795, "step": 730 }, { "epoch": 0.3174120712114633, "grad_norm": 0.04151546210050583, "learning_rate": 0.00017466698358468825, "loss": 0.4852, "step": 731 }, { "epoch": 0.3178462874511507, "grad_norm": 0.03277713060379028, "learning_rate": 0.00017345150812337563, "loss": 0.4736, "step": 732 }, { "epoch": 0.31828050369083805, "grad_norm": 0.03749159723520279, "learning_rate": 0.0001722393884530593, "loss": 0.5042, "step": 733 }, { "epoch": 0.3187147199305254, "grad_norm": 0.029695043340325356, "learning_rate": 0.00017103063703014372, "loss": 0.511, "step": 734 }, { "epoch": 0.3191489361702128, "grad_norm": 0.033079251646995544, "learning_rate": 0.00016982526627642042, "loss": 0.4871, "step": 735 }, { "epoch": 0.31958315240990015, "grad_norm": 0.03277754411101341, "learning_rate": 0.00016862328857893854, "loss": 0.4805, "step": 736 }, { "epoch": 0.3200173686495875, "grad_norm": 0.031276337802410126, "learning_rate": 0.00016742471628987892, "loss": 0.5195, "step": 737 }, { "epoch": 0.32045158488927483, "grad_norm": 0.031646616756916046, "learning_rate": 0.000166229561726426, "loss": 0.4899, "step": 738 }, { "epoch": 0.3208858011289622, "grad_norm": 0.03364792838692665, "learning_rate": 0.00016503783717064247, "loss": 0.5011, "step": 739 }, { "epoch": 0.32132001736864957, "grad_norm": 0.032123491168022156, "learning_rate": 0.00016384955486934156, "loss": 0.5083, "step": 740 }, { "epoch": 0.32175423360833694, "grad_norm": 0.03388531878590584, "learning_rate": 0.00016266472703396284, "loss": 0.5127, "step": 741 }, { "epoch": 0.3221884498480243, "grad_norm": 0.03697797656059265, "learning_rate": 0.00016148336584044537, "loss": 0.5167, "step": 742 }, { "epoch": 0.3226226660877117, "grad_norm": 0.03302115574479103, "learning_rate": 0.000160305483429103, "loss": 0.483, "step": 743 }, { "epoch": 0.32305688232739904, "grad_norm": 0.03180578723549843, "learning_rate": 0.0001591310919045003, "loss": 0.5144, "step": 744 }, { "epoch": 0.3234910985670864, "grad_norm": 0.033885449171066284, "learning_rate": 0.00015796020333532697, "loss": 0.484, "step": 745 }, { "epoch": 0.3239253148067738, "grad_norm": 0.03529027849435806, "learning_rate": 0.0001567928297542749, "loss": 0.4967, "step": 746 }, { "epoch": 0.32435953104646115, "grad_norm": 0.030900994315743446, "learning_rate": 0.00015562898315791353, "loss": 0.5177, "step": 747 }, { "epoch": 0.3247937472861485, "grad_norm": 0.03337928652763367, "learning_rate": 0.00015446867550656767, "loss": 0.4902, "step": 748 }, { "epoch": 0.3252279635258359, "grad_norm": 0.028887873515486717, "learning_rate": 0.00015331191872419348, "loss": 0.4841, "step": 749 }, { "epoch": 0.32566217976552325, "grad_norm": 0.03361840918660164, "learning_rate": 0.0001521587246982568, "loss": 0.5171, "step": 750 }, { "epoch": 0.3260963960052106, "grad_norm": 0.030776534229516983, "learning_rate": 0.00015100910527961049, "loss": 0.4592, "step": 751 }, { "epoch": 0.32653061224489793, "grad_norm": 0.028537072241306305, "learning_rate": 0.00014986307228237266, "loss": 0.4894, "step": 752 }, { "epoch": 0.3269648284845853, "grad_norm": 0.02929234690964222, "learning_rate": 0.00014872063748380543, "loss": 0.4832, "step": 753 }, { "epoch": 0.32739904472427267, "grad_norm": 0.02816353552043438, "learning_rate": 0.00014758181262419424, "loss": 0.4788, "step": 754 }, { "epoch": 0.32783326096396004, "grad_norm": 0.03165418654680252, "learning_rate": 0.00014644660940672628, "loss": 0.5128, "step": 755 }, { "epoch": 0.3282674772036474, "grad_norm": 0.0279870443046093, "learning_rate": 0.00014531503949737106, "loss": 0.4966, "step": 756 }, { "epoch": 0.3287016934433348, "grad_norm": 0.029440978541970253, "learning_rate": 0.00014418711452476046, "loss": 0.4785, "step": 757 }, { "epoch": 0.32913590968302214, "grad_norm": 0.032016150653362274, "learning_rate": 0.00014306284608006837, "loss": 0.4813, "step": 758 }, { "epoch": 0.3295701259227095, "grad_norm": 0.029074901714920998, "learning_rate": 0.00014194224571689284, "loss": 0.5165, "step": 759 }, { "epoch": 0.3300043421623969, "grad_norm": 0.030839256942272186, "learning_rate": 0.00014082532495113625, "loss": 0.4615, "step": 760 }, { "epoch": 0.33043855840208425, "grad_norm": 0.03017420321702957, "learning_rate": 0.00013971209526088762, "loss": 0.473, "step": 761 }, { "epoch": 0.3308727746417716, "grad_norm": 0.03192548081278801, "learning_rate": 0.00013860256808630427, "loss": 0.5315, "step": 762 }, { "epoch": 0.331306990881459, "grad_norm": 0.029673421755433083, "learning_rate": 0.00013749675482949486, "loss": 0.4978, "step": 763 }, { "epoch": 0.33174120712114635, "grad_norm": 0.035233914852142334, "learning_rate": 0.00013639466685440134, "loss": 0.5174, "step": 764 }, { "epoch": 0.3321754233608337, "grad_norm": 0.03196287900209427, "learning_rate": 0.00013529631548668298, "loss": 0.5121, "step": 765 }, { "epoch": 0.33260963960052103, "grad_norm": 0.028510358184576035, "learning_rate": 0.0001342017120135993, "loss": 0.4629, "step": 766 }, { "epoch": 0.3330438558402084, "grad_norm": 0.032123226672410965, "learning_rate": 0.0001331108676838948, "loss": 0.5538, "step": 767 }, { "epoch": 0.33347807207989577, "grad_norm": 0.0301981084048748, "learning_rate": 0.00013202379370768252, "loss": 0.5278, "step": 768 }, { "epoch": 0.33391228831958314, "grad_norm": 0.029721124097704887, "learning_rate": 0.00013094050125632973, "loss": 0.5353, "step": 769 }, { "epoch": 0.3343465045592705, "grad_norm": 0.029476812109351158, "learning_rate": 0.00012986100146234231, "loss": 0.488, "step": 770 }, { "epoch": 0.3347807207989579, "grad_norm": 0.029208241030573845, "learning_rate": 0.00012878530541925076, "loss": 0.5384, "step": 771 }, { "epoch": 0.33521493703864524, "grad_norm": 0.029901737347245216, "learning_rate": 0.00012771342418149657, "loss": 0.4675, "step": 772 }, { "epoch": 0.3356491532783326, "grad_norm": 0.030133001506328583, "learning_rate": 0.00012664536876431755, "loss": 0.5175, "step": 773 }, { "epoch": 0.33608336951802, "grad_norm": 0.028815865516662598, "learning_rate": 0.0001255811501436359, "loss": 0.5012, "step": 774 }, { "epoch": 0.33651758575770735, "grad_norm": 0.02868812158703804, "learning_rate": 0.00012452077925594434, "loss": 0.521, "step": 775 }, { "epoch": 0.3369518019973947, "grad_norm": 0.035325054079294205, "learning_rate": 0.00012346426699819457, "loss": 0.5075, "step": 776 }, { "epoch": 0.3373860182370821, "grad_norm": 0.03588160499930382, "learning_rate": 0.00012241162422768444, "loss": 0.5091, "step": 777 }, { "epoch": 0.33782023447676945, "grad_norm": 0.03347684070467949, "learning_rate": 0.00012136286176194744, "loss": 0.5039, "step": 778 }, { "epoch": 0.3382544507164568, "grad_norm": 0.03358198329806328, "learning_rate": 0.00012031799037864011, "loss": 0.5014, "step": 779 }, { "epoch": 0.33868866695614414, "grad_norm": 0.02983028069138527, "learning_rate": 0.0001192770208154328, "loss": 0.4823, "step": 780 }, { "epoch": 0.3391228831958315, "grad_norm": 0.028506051748991013, "learning_rate": 0.00011823996376989849, "loss": 0.5135, "step": 781 }, { "epoch": 0.33955709943551887, "grad_norm": 0.031358085572719574, "learning_rate": 0.00011720682989940262, "loss": 0.4677, "step": 782 }, { "epoch": 0.33999131567520624, "grad_norm": 0.031006425619125366, "learning_rate": 0.00011617762982099444, "loss": 0.5188, "step": 783 }, { "epoch": 0.3404255319148936, "grad_norm": 0.0309920571744442, "learning_rate": 0.00011515237411129698, "loss": 0.5087, "step": 784 }, { "epoch": 0.340859748154581, "grad_norm": 0.027928480878472328, "learning_rate": 0.0001141310733063991, "loss": 0.5005, "step": 785 }, { "epoch": 0.34129396439426835, "grad_norm": 0.028440937399864197, "learning_rate": 0.00011311373790174655, "loss": 0.5055, "step": 786 }, { "epoch": 0.3417281806339557, "grad_norm": 0.03135620057582855, "learning_rate": 0.00011210037835203507, "loss": 0.4947, "step": 787 }, { "epoch": 0.3421623968736431, "grad_norm": 0.029479345306754112, "learning_rate": 0.00011109100507110132, "loss": 0.5048, "step": 788 }, { "epoch": 0.34259661311333045, "grad_norm": 0.030894558876752853, "learning_rate": 0.00011008562843181796, "loss": 0.4726, "step": 789 }, { "epoch": 0.3430308293530178, "grad_norm": 0.03545878455042839, "learning_rate": 0.0001090842587659851, "loss": 0.5165, "step": 790 }, { "epoch": 0.3434650455927052, "grad_norm": 0.028414282947778702, "learning_rate": 0.00010808690636422585, "loss": 0.5008, "step": 791 }, { "epoch": 0.34389926183239256, "grad_norm": 0.029046185314655304, "learning_rate": 0.00010709358147587883, "loss": 0.5067, "step": 792 }, { "epoch": 0.34433347807207987, "grad_norm": 0.029067449271678925, "learning_rate": 0.00010610429430889452, "loss": 0.5039, "step": 793 }, { "epoch": 0.34476769431176724, "grad_norm": 0.03735330328345299, "learning_rate": 0.00010511905502972886, "loss": 0.4874, "step": 794 }, { "epoch": 0.3452019105514546, "grad_norm": 0.025629056617617607, "learning_rate": 0.00010413787376324018, "loss": 0.4743, "step": 795 }, { "epoch": 0.345636126791142, "grad_norm": 0.03742838650941849, "learning_rate": 0.00010316076059258389, "loss": 0.4706, "step": 796 }, { "epoch": 0.34607034303082934, "grad_norm": 0.028506414964795113, "learning_rate": 0.00010218772555910954, "loss": 0.4677, "step": 797 }, { "epoch": 0.3465045592705167, "grad_norm": 0.028791414573788643, "learning_rate": 0.00010121877866225782, "loss": 0.4786, "step": 798 }, { "epoch": 0.3469387755102041, "grad_norm": 0.027565713971853256, "learning_rate": 0.00010025392985945703, "loss": 0.4756, "step": 799 }, { "epoch": 0.34737299174989145, "grad_norm": 0.037437453866004944, "learning_rate": 9.929318906602175e-05, "loss": 0.4941, "step": 800 }, { "epoch": 0.3478072079895788, "grad_norm": 0.028768004849553108, "learning_rate": 9.833656615504976e-05, "loss": 0.5263, "step": 801 }, { "epoch": 0.3482414242292662, "grad_norm": 0.030494874343276024, "learning_rate": 9.738407095732193e-05, "loss": 0.4933, "step": 802 }, { "epoch": 0.34867564046895355, "grad_norm": 0.02905452810227871, "learning_rate": 9.643571326119983e-05, "loss": 0.4751, "step": 803 }, { "epoch": 0.3491098567086409, "grad_norm": 0.02821405418217182, "learning_rate": 9.549150281252633e-05, "loss": 0.4874, "step": 804 }, { "epoch": 0.3495440729483283, "grad_norm": 0.02889527939260006, "learning_rate": 9.455144931452459e-05, "loss": 0.4732, "step": 805 }, { "epoch": 0.34997828918801566, "grad_norm": 0.02988606132566929, "learning_rate": 9.36155624276987e-05, "loss": 0.5044, "step": 806 }, { "epoch": 0.35041250542770297, "grad_norm": 0.03306758031249046, "learning_rate": 9.26838517697346e-05, "loss": 0.492, "step": 807 }, { "epoch": 0.35084672166739034, "grad_norm": 0.028606578707695007, "learning_rate": 9.175632691540064e-05, "loss": 0.4696, "step": 808 }, { "epoch": 0.3512809379070777, "grad_norm": 0.027804942801594734, "learning_rate": 9.083299739645007e-05, "loss": 0.504, "step": 809 }, { "epoch": 0.3517151541467651, "grad_norm": 0.02756122685968876, "learning_rate": 8.991387270152201e-05, "loss": 0.5129, "step": 810 }, { "epoch": 0.35214937038645244, "grad_norm": 0.03051147237420082, "learning_rate": 8.899896227604509e-05, "loss": 0.5054, "step": 811 }, { "epoch": 0.3525835866261398, "grad_norm": 0.032840099185705185, "learning_rate": 8.808827552213916e-05, "loss": 0.4926, "step": 812 }, { "epoch": 0.3530178028658272, "grad_norm": 0.026586757972836494, "learning_rate": 8.718182179851997e-05, "loss": 0.4992, "step": 813 }, { "epoch": 0.35345201910551455, "grad_norm": 0.030486930161714554, "learning_rate": 8.627961042040184e-05, "loss": 0.5221, "step": 814 }, { "epoch": 0.3538862353452019, "grad_norm": 0.03037872351706028, "learning_rate": 8.538165065940262e-05, "loss": 0.4801, "step": 815 }, { "epoch": 0.3543204515848893, "grad_norm": 0.027922067791223526, "learning_rate": 8.448795174344803e-05, "loss": 0.5011, "step": 816 }, { "epoch": 0.35475466782457665, "grad_norm": 0.030396727845072746, "learning_rate": 8.359852285667752e-05, "loss": 0.4937, "step": 817 }, { "epoch": 0.355188884064264, "grad_norm": 0.028359906747937202, "learning_rate": 8.271337313934868e-05, "loss": 0.4905, "step": 818 }, { "epoch": 0.3556231003039514, "grad_norm": 0.041622817516326904, "learning_rate": 8.183251168774475e-05, "loss": 0.4829, "step": 819 }, { "epoch": 0.35605731654363876, "grad_norm": 0.030735066160559654, "learning_rate": 8.09559475540797e-05, "loss": 0.5038, "step": 820 }, { "epoch": 0.35649153278332607, "grad_norm": 0.030377600342035294, "learning_rate": 8.008368974640634e-05, "loss": 0.523, "step": 821 }, { "epoch": 0.35692574902301344, "grad_norm": 0.027815300971269608, "learning_rate": 7.921574722852342e-05, "loss": 0.4758, "step": 822 }, { "epoch": 0.3573599652627008, "grad_norm": 0.026980141177773476, "learning_rate": 7.835212891988292e-05, "loss": 0.5151, "step": 823 }, { "epoch": 0.3577941815023882, "grad_norm": 0.028305258601903915, "learning_rate": 7.749284369549953e-05, "loss": 0.4968, "step": 824 }, { "epoch": 0.35822839774207554, "grad_norm": 0.028862129896879196, "learning_rate": 7.663790038585794e-05, "loss": 0.5093, "step": 825 }, { "epoch": 0.3586626139817629, "grad_norm": 0.031168105080723763, "learning_rate": 7.578730777682385e-05, "loss": 0.507, "step": 826 }, { "epoch": 0.3590968302214503, "grad_norm": 0.02467159368097782, "learning_rate": 7.494107460955207e-05, "loss": 0.5031, "step": 827 }, { "epoch": 0.35953104646113765, "grad_norm": 0.028088459745049477, "learning_rate": 7.409920958039795e-05, "loss": 0.5031, "step": 828 }, { "epoch": 0.359965262700825, "grad_norm": 0.027428364381194115, "learning_rate": 7.326172134082704e-05, "loss": 0.5262, "step": 829 }, { "epoch": 0.3603994789405124, "grad_norm": 0.033019062131643295, "learning_rate": 7.242861849732695e-05, "loss": 0.4722, "step": 830 }, { "epoch": 0.36083369518019975, "grad_norm": 0.028655540198087692, "learning_rate": 7.159990961131818e-05, "loss": 0.4899, "step": 831 }, { "epoch": 0.3612679114198871, "grad_norm": 0.02457267791032791, "learning_rate": 7.077560319906695e-05, "loss": 0.4989, "step": 832 }, { "epoch": 0.3617021276595745, "grad_norm": 0.027305081486701965, "learning_rate": 6.995570773159692e-05, "loss": 0.4885, "step": 833 }, { "epoch": 0.36213634389926186, "grad_norm": 0.026189139112830162, "learning_rate": 6.914023163460248e-05, "loss": 0.4727, "step": 834 }, { "epoch": 0.36257056013894917, "grad_norm": 0.028178861364722252, "learning_rate": 6.832918328836246e-05, "loss": 0.5538, "step": 835 }, { "epoch": 0.36300477637863654, "grad_norm": 0.02753199078142643, "learning_rate": 6.752257102765324e-05, "loss": 0.4991, "step": 836 }, { "epoch": 0.3634389926183239, "grad_norm": 0.02777719311416149, "learning_rate": 6.6720403141664e-05, "loss": 0.4607, "step": 837 }, { "epoch": 0.3638732088580113, "grad_norm": 0.026312079280614853, "learning_rate": 6.592268787391076e-05, "loss": 0.5031, "step": 838 }, { "epoch": 0.36430742509769865, "grad_norm": 0.028296776115894318, "learning_rate": 6.512943342215233e-05, "loss": 0.4825, "step": 839 }, { "epoch": 0.364741641337386, "grad_norm": 0.026355035603046417, "learning_rate": 6.43406479383053e-05, "loss": 0.4871, "step": 840 }, { "epoch": 0.3651758575770734, "grad_norm": 0.025023045018315315, "learning_rate": 6.355633952836115e-05, "loss": 0.4797, "step": 841 }, { "epoch": 0.36561007381676075, "grad_norm": 0.027957437559962273, "learning_rate": 6.277651625230219e-05, "loss": 0.5078, "step": 842 }, { "epoch": 0.3660442900564481, "grad_norm": 0.028170911595225334, "learning_rate": 6.200118612401917e-05, "loss": 0.4781, "step": 843 }, { "epoch": 0.3664785062961355, "grad_norm": 0.032493725419044495, "learning_rate": 6.123035711122859e-05, "loss": 0.5078, "step": 844 }, { "epoch": 0.36691272253582286, "grad_norm": 0.029985696077346802, "learning_rate": 6.046403713539139e-05, "loss": 0.478, "step": 845 }, { "epoch": 0.3673469387755102, "grad_norm": 0.03212130442261696, "learning_rate": 5.9702234071631e-05, "loss": 0.5146, "step": 846 }, { "epoch": 0.3677811550151976, "grad_norm": 0.024975216016173363, "learning_rate": 5.89449557486525e-05, "loss": 0.4564, "step": 847 }, { "epoch": 0.36821537125488496, "grad_norm": 0.026881275698542595, "learning_rate": 5.8192209948662365e-05, "loss": 0.518, "step": 848 }, { "epoch": 0.3686495874945723, "grad_norm": 0.04243510961532593, "learning_rate": 5.7444004407288256e-05, "loss": 0.5238, "step": 849 }, { "epoch": 0.36908380373425964, "grad_norm": 0.02634822018444538, "learning_rate": 5.6700346813499945e-05, "loss": 0.5126, "step": 850 }, { "epoch": 0.369518019973947, "grad_norm": 0.029794232919812202, "learning_rate": 5.5961244809529746e-05, "loss": 0.5017, "step": 851 }, { "epoch": 0.3699522362136344, "grad_norm": 0.028596797958016396, "learning_rate": 5.5226705990794155e-05, "loss": 0.508, "step": 852 }, { "epoch": 0.37038645245332175, "grad_norm": 0.04081004112958908, "learning_rate": 5.449673790581611e-05, "loss": 0.4946, "step": 853 }, { "epoch": 0.3708206686930091, "grad_norm": 0.02494877576828003, "learning_rate": 5.377134805614714e-05, "loss": 0.5008, "step": 854 }, { "epoch": 0.3712548849326965, "grad_norm": 0.029261523857712746, "learning_rate": 5.305054389629022e-05, "loss": 0.5155, "step": 855 }, { "epoch": 0.37168910117238385, "grad_norm": 0.026129230856895447, "learning_rate": 5.2334332833623486e-05, "loss": 0.4979, "step": 856 }, { "epoch": 0.3721233174120712, "grad_norm": 0.03483575955033302, "learning_rate": 5.162272222832348e-05, "loss": 0.4956, "step": 857 }, { "epoch": 0.3725575336517586, "grad_norm": 0.028554193675518036, "learning_rate": 5.0915719393290485e-05, "loss": 0.4816, "step": 858 }, { "epoch": 0.37299174989144596, "grad_norm": 0.028302082791924477, "learning_rate": 5.021333159407232e-05, "loss": 0.4734, "step": 859 }, { "epoch": 0.3734259661311333, "grad_norm": 0.03633783757686615, "learning_rate": 4.9515566048790485e-05, "loss": 0.5152, "step": 860 }, { "epoch": 0.3738601823708207, "grad_norm": 0.025067532435059547, "learning_rate": 4.882242992806546e-05, "loss": 0.505, "step": 861 }, { "epoch": 0.374294398610508, "grad_norm": 0.03426215052604675, "learning_rate": 4.813393035494329e-05, "loss": 0.4811, "step": 862 }, { "epoch": 0.3747286148501954, "grad_norm": 0.024589484557509422, "learning_rate": 4.745007440482252e-05, "loss": 0.4881, "step": 863 }, { "epoch": 0.37516283108988274, "grad_norm": 0.032448723912239075, "learning_rate": 4.677086910538092e-05, "loss": 0.5057, "step": 864 }, { "epoch": 0.3755970473295701, "grad_norm": 0.029285568743944168, "learning_rate": 4.6096321436503994e-05, "loss": 0.5035, "step": 865 }, { "epoch": 0.3760312635692575, "grad_norm": 0.02594076097011566, "learning_rate": 4.5426438330212535e-05, "loss": 0.4789, "step": 866 }, { "epoch": 0.37646547980894485, "grad_norm": 0.028019646182656288, "learning_rate": 4.476122667059207e-05, "loss": 0.4841, "step": 867 }, { "epoch": 0.3768996960486322, "grad_norm": 0.022888783365488052, "learning_rate": 4.410069329372152e-05, "loss": 0.4609, "step": 868 }, { "epoch": 0.3773339122883196, "grad_norm": 0.025522053241729736, "learning_rate": 4.3444844987603426e-05, "loss": 0.5148, "step": 869 }, { "epoch": 0.37776812852800695, "grad_norm": 0.027691571041941643, "learning_rate": 4.279368849209381e-05, "loss": 0.4689, "step": 870 }, { "epoch": 0.3782023447676943, "grad_norm": 0.02531832829117775, "learning_rate": 4.214723049883307e-05, "loss": 0.5024, "step": 871 }, { "epoch": 0.3786365610073817, "grad_norm": 0.025689370930194855, "learning_rate": 4.150547765117746e-05, "loss": 0.4839, "step": 872 }, { "epoch": 0.37907077724706906, "grad_norm": 0.031849976629018784, "learning_rate": 4.086843654413031e-05, "loss": 0.4888, "step": 873 }, { "epoch": 0.3795049934867564, "grad_norm": 0.024920670315623283, "learning_rate": 4.0236113724274713e-05, "loss": 0.5088, "step": 874 }, { "epoch": 0.3799392097264438, "grad_norm": 0.03232395276427269, "learning_rate": 3.960851568970586e-05, "loss": 0.5039, "step": 875 }, { "epoch": 0.3803734259661311, "grad_norm": 0.025085026398301125, "learning_rate": 3.898564888996475e-05, "loss": 0.4753, "step": 876 }, { "epoch": 0.3808076422058185, "grad_norm": 0.03306965157389641, "learning_rate": 3.83675197259713e-05, "loss": 0.493, "step": 877 }, { "epoch": 0.38124185844550584, "grad_norm": 0.02512296289205551, "learning_rate": 3.7754134549959293e-05, "loss": 0.5048, "step": 878 }, { "epoch": 0.3816760746851932, "grad_norm": 0.02608814276754856, "learning_rate": 3.714549966541014e-05, "loss": 0.4893, "step": 879 }, { "epoch": 0.3821102909248806, "grad_norm": 0.02658063732087612, "learning_rate": 3.654162132698918e-05, "loss": 0.5069, "step": 880 }, { "epoch": 0.38254450716456795, "grad_norm": 0.027712546288967133, "learning_rate": 3.594250574048058e-05, "loss": 0.4778, "step": 881 }, { "epoch": 0.3829787234042553, "grad_norm": 0.02457358129322529, "learning_rate": 3.534815906272404e-05, "loss": 0.4946, "step": 882 }, { "epoch": 0.3834129396439427, "grad_norm": 0.032855454832315445, "learning_rate": 3.475858740155108e-05, "loss": 0.4796, "step": 883 }, { "epoch": 0.38384715588363005, "grad_norm": 0.02618669532239437, "learning_rate": 3.417379681572297e-05, "loss": 0.4944, "step": 884 }, { "epoch": 0.3842813721233174, "grad_norm": 0.02916988916695118, "learning_rate": 3.3593793314867616e-05, "loss": 0.4865, "step": 885 }, { "epoch": 0.3847155883630048, "grad_norm": 0.02979261800646782, "learning_rate": 3.3018582859418446e-05, "loss": 0.5272, "step": 886 }, { "epoch": 0.38514980460269216, "grad_norm": 0.024905268102884293, "learning_rate": 3.2448171360552835e-05, "loss": 0.4878, "step": 887 }, { "epoch": 0.3855840208423795, "grad_norm": 0.025121575221419334, "learning_rate": 3.18825646801314e-05, "loss": 0.4887, "step": 888 }, { "epoch": 0.3860182370820669, "grad_norm": 0.02593301795423031, "learning_rate": 3.132176863063807e-05, "loss": 0.5255, "step": 889 }, { "epoch": 0.3864524533217542, "grad_norm": 0.02763954922556877, "learning_rate": 3.076578897511978e-05, "loss": 0.4672, "step": 890 }, { "epoch": 0.3868866695614416, "grad_norm": 0.026684967800974846, "learning_rate": 3.0214631427127882e-05, "loss": 0.5277, "step": 891 }, { "epoch": 0.38732088580112894, "grad_norm": 0.030234767124056816, "learning_rate": 2.9668301650658758e-05, "loss": 0.502, "step": 892 }, { "epoch": 0.3877551020408163, "grad_norm": 0.023924171924591064, "learning_rate": 2.9126805260096256e-05, "loss": 0.4813, "step": 893 }, { "epoch": 0.3881893182805037, "grad_norm": 0.03323635831475258, "learning_rate": 2.859014782015351e-05, "loss": 0.5422, "step": 894 }, { "epoch": 0.38862353452019105, "grad_norm": 0.02363520860671997, "learning_rate": 2.8058334845816213e-05, "loss": 0.5091, "step": 895 }, { "epoch": 0.3890577507598784, "grad_norm": 0.024630311876535416, "learning_rate": 2.7531371802285433e-05, "loss": 0.4642, "step": 896 }, { "epoch": 0.3894919669995658, "grad_norm": 0.025726784020662308, "learning_rate": 2.7009264104921606e-05, "loss": 0.511, "step": 897 }, { "epoch": 0.38992618323925315, "grad_norm": 0.02972118742763996, "learning_rate": 2.6492017119189414e-05, "loss": 0.4825, "step": 898 }, { "epoch": 0.3903603994789405, "grad_norm": 0.024715179577469826, "learning_rate": 2.597963616060167e-05, "loss": 0.4879, "step": 899 }, { "epoch": 0.3907946157186279, "grad_norm": 0.02915225736796856, "learning_rate": 2.547212649466568e-05, "loss": 0.4903, "step": 900 }, { "epoch": 0.39122883195831526, "grad_norm": 0.02453678846359253, "learning_rate": 2.4969493336828353e-05, "loss": 0.5024, "step": 901 }, { "epoch": 0.39166304819800263, "grad_norm": 0.025468653067946434, "learning_rate": 2.4471741852423235e-05, "loss": 0.4964, "step": 902 }, { "epoch": 0.39209726443769, "grad_norm": 0.024076983332633972, "learning_rate": 2.3978877156616784e-05, "loss": 0.4872, "step": 903 }, { "epoch": 0.3925314806773773, "grad_norm": 0.02444782853126526, "learning_rate": 2.349090431435641e-05, "loss": 0.482, "step": 904 }, { "epoch": 0.3929656969170647, "grad_norm": 0.025424372404813766, "learning_rate": 2.3007828340318116e-05, "loss": 0.4907, "step": 905 }, { "epoch": 0.39339991315675205, "grad_norm": 0.023807033896446228, "learning_rate": 2.252965419885483e-05, "loss": 0.5007, "step": 906 }, { "epoch": 0.3938341293964394, "grad_norm": 0.022182505577802658, "learning_rate": 2.2056386803945726e-05, "loss": 0.4712, "step": 907 }, { "epoch": 0.3942683456361268, "grad_norm": 0.02266152761876583, "learning_rate": 2.1588031019145637e-05, "loss": 0.479, "step": 908 }, { "epoch": 0.39470256187581415, "grad_norm": 0.024663135409355164, "learning_rate": 2.1124591657534777e-05, "loss": 0.514, "step": 909 }, { "epoch": 0.3951367781155015, "grad_norm": 0.023098768666386604, "learning_rate": 2.0666073481669712e-05, "loss": 0.4857, "step": 910 }, { "epoch": 0.3955709943551889, "grad_norm": 0.025155888870358467, "learning_rate": 2.0212481203534084e-05, "loss": 0.5054, "step": 911 }, { "epoch": 0.39600521059487626, "grad_norm": 0.030762221664190292, "learning_rate": 1.9763819484490355e-05, "loss": 0.4967, "step": 912 }, { "epoch": 0.3964394268345636, "grad_norm": 0.029380332678556442, "learning_rate": 1.932009293523196e-05, "loss": 0.5412, "step": 913 }, { "epoch": 0.396873643074251, "grad_norm": 0.023896733298897743, "learning_rate": 1.888130611573563e-05, "loss": 0.5002, "step": 914 }, { "epoch": 0.39730785931393836, "grad_norm": 0.024988802149891853, "learning_rate": 1.844746353521487e-05, "loss": 0.4888, "step": 915 }, { "epoch": 0.39774207555362573, "grad_norm": 0.02422316186130047, "learning_rate": 1.801856965207338e-05, "loss": 0.4818, "step": 916 }, { "epoch": 0.3981762917933131, "grad_norm": 0.031409528106451035, "learning_rate": 1.7594628873859486e-05, "loss": 0.4923, "step": 917 }, { "epoch": 0.3986105080330004, "grad_norm": 0.023978758603334427, "learning_rate": 1.7175645557220565e-05, "loss": 0.4798, "step": 918 }, { "epoch": 0.3990447242726878, "grad_norm": 0.02715397998690605, "learning_rate": 1.6761624007858522e-05, "loss": 0.4474, "step": 919 }, { "epoch": 0.39947894051237515, "grad_norm": 0.02678792178630829, "learning_rate": 1.6352568480485276e-05, "loss": 0.514, "step": 920 }, { "epoch": 0.3999131567520625, "grad_norm": 0.02234644442796707, "learning_rate": 1.594848317877934e-05, "loss": 0.4869, "step": 921 }, { "epoch": 0.4003473729917499, "grad_norm": 0.029807301238179207, "learning_rate": 1.5549372255342365e-05, "loss": 0.5242, "step": 922 }, { "epoch": 0.40078158923143725, "grad_norm": 0.03037095069885254, "learning_rate": 1.5155239811656562e-05, "loss": 0.5035, "step": 923 }, { "epoch": 0.4012158054711246, "grad_norm": 0.022767795249819756, "learning_rate": 1.4766089898042678e-05, "loss": 0.4903, "step": 924 }, { "epoch": 0.401650021710812, "grad_norm": 0.026207391172647476, "learning_rate": 1.4381926513618138e-05, "loss": 0.4676, "step": 925 }, { "epoch": 0.40208423795049936, "grad_norm": 0.025653038173913956, "learning_rate": 1.400275360625608e-05, "loss": 0.4804, "step": 926 }, { "epoch": 0.4025184541901867, "grad_norm": 0.024481656029820442, "learning_rate": 1.362857507254478e-05, "loss": 0.4871, "step": 927 }, { "epoch": 0.4029526704298741, "grad_norm": 0.02475779317319393, "learning_rate": 1.3259394757747678e-05, "loss": 0.5199, "step": 928 }, { "epoch": 0.40338688666956146, "grad_norm": 0.02371501363813877, "learning_rate": 1.289521645576358e-05, "loss": 0.528, "step": 929 }, { "epoch": 0.40382110290924883, "grad_norm": 0.023029111325740814, "learning_rate": 1.2536043909088191e-05, "loss": 0.4801, "step": 930 }, { "epoch": 0.40425531914893614, "grad_norm": 0.025304608047008514, "learning_rate": 1.2181880808775025e-05, "loss": 0.4793, "step": 931 }, { "epoch": 0.4046895353886235, "grad_norm": 0.02734360471367836, "learning_rate": 1.183273079439795e-05, "loss": 0.5022, "step": 932 }, { "epoch": 0.4051237516283109, "grad_norm": 0.024530354887247086, "learning_rate": 1.1488597454013538e-05, "loss": 0.5032, "step": 933 }, { "epoch": 0.40555796786799825, "grad_norm": 0.02369566634297371, "learning_rate": 1.1149484324124327e-05, "loss": 0.4633, "step": 934 }, { "epoch": 0.4059921841076856, "grad_norm": 0.024855339899659157, "learning_rate": 1.0815394889642338e-05, "loss": 0.5175, "step": 935 }, { "epoch": 0.406426400347373, "grad_norm": 0.023386195302009583, "learning_rate": 1.0486332583853564e-05, "loss": 0.5014, "step": 936 }, { "epoch": 0.40686061658706035, "grad_norm": 0.029485292732715607, "learning_rate": 1.0162300788382261e-05, "loss": 0.4942, "step": 937 }, { "epoch": 0.4072948328267477, "grad_norm": 0.023054329678416252, "learning_rate": 9.843302833156376e-06, "loss": 0.4683, "step": 938 }, { "epoch": 0.4077290490664351, "grad_norm": 0.031031260266900063, "learning_rate": 9.529341996373674e-06, "loss": 0.5173, "step": 939 }, { "epoch": 0.40816326530612246, "grad_norm": 0.023358464241027832, "learning_rate": 9.22042150446728e-06, "loss": 0.4722, "step": 940 }, { "epoch": 0.4085974815458098, "grad_norm": 0.025074800476431847, "learning_rate": 8.916544532073411e-06, "loss": 0.4784, "step": 941 }, { "epoch": 0.4090316977854972, "grad_norm": 0.02519945055246353, "learning_rate": 8.617714201998083e-06, "loss": 0.5022, "step": 942 }, { "epoch": 0.40946591402518456, "grad_norm": 0.027328645810484886, "learning_rate": 8.323933585185184e-06, "loss": 0.4896, "step": 943 }, { "epoch": 0.40990013026487193, "grad_norm": 0.02542622946202755, "learning_rate": 8.035205700685167e-06, "loss": 0.4676, "step": 944 }, { "epoch": 0.41033434650455924, "grad_norm": 0.027195794507861137, "learning_rate": 7.751533515623799e-06, "loss": 0.4903, "step": 945 }, { "epoch": 0.4107685627442466, "grad_norm": 0.02717754803597927, "learning_rate": 7.47291994517163e-06, "loss": 0.4812, "step": 946 }, { "epoch": 0.411202778983934, "grad_norm": 0.02495565079152584, "learning_rate": 7.199367852514238e-06, "loss": 0.5014, "step": 947 }, { "epoch": 0.41163699522362135, "grad_norm": 0.024513866752386093, "learning_rate": 6.93088004882253e-06, "loss": 0.5071, "step": 948 }, { "epoch": 0.4120712114633087, "grad_norm": 0.03526155650615692, "learning_rate": 6.667459293224154e-06, "loss": 0.531, "step": 949 }, { "epoch": 0.4125054277029961, "grad_norm": 0.023175369948148727, "learning_rate": 6.4091082927749125e-06, "loss": 0.4811, "step": 950 }, { "epoch": 0.41293964394268345, "grad_norm": 0.026199523359537125, "learning_rate": 6.15582970243117e-06, "loss": 0.5087, "step": 951 }, { "epoch": 0.4133738601823708, "grad_norm": 0.024454379454255104, "learning_rate": 5.907626125022158e-06, "loss": 0.479, "step": 952 }, { "epoch": 0.4138080764220582, "grad_norm": 0.02492498978972435, "learning_rate": 5.664500111223769e-06, "loss": 0.4977, "step": 953 }, { "epoch": 0.41424229266174556, "grad_norm": 0.023200612515211105, "learning_rate": 5.426454159531913e-06, "loss": 0.4803, "step": 954 }, { "epoch": 0.4146765089014329, "grad_norm": 0.024389084428548813, "learning_rate": 5.193490716237037e-06, "loss": 0.4733, "step": 955 }, { "epoch": 0.4151107251411203, "grad_norm": 0.027189958840608597, "learning_rate": 4.9656121753990924e-06, "loss": 0.4994, "step": 956 }, { "epoch": 0.41554494138080766, "grad_norm": 0.02239627204835415, "learning_rate": 4.7428208788224955e-06, "loss": 0.4978, "step": 957 }, { "epoch": 0.41597915762049503, "grad_norm": 0.021820807829499245, "learning_rate": 4.52511911603265e-06, "loss": 0.4871, "step": 958 }, { "epoch": 0.41641337386018235, "grad_norm": 0.02775476686656475, "learning_rate": 4.312509124251906e-06, "loss": 0.4918, "step": 959 }, { "epoch": 0.4168475900998697, "grad_norm": 0.023040220141410828, "learning_rate": 4.104993088376974e-06, "loss": 0.4497, "step": 960 }, { "epoch": 0.4172818063395571, "grad_norm": 0.025357872247695923, "learning_rate": 3.9025731409561e-06, "loss": 0.4945, "step": 961 }, { "epoch": 0.41771602257924445, "grad_norm": 0.02308405004441738, "learning_rate": 3.7052513621674833e-06, "loss": 0.4945, "step": 962 }, { "epoch": 0.4181502388189318, "grad_norm": 0.02784230373799801, "learning_rate": 3.513029779797783e-06, "loss": 0.4543, "step": 963 }, { "epoch": 0.4185844550586192, "grad_norm": 0.026198869571089745, "learning_rate": 3.325910369220975e-06, "loss": 0.4975, "step": 964 }, { "epoch": 0.41901867129830656, "grad_norm": 0.023575875908136368, "learning_rate": 3.143895053378698e-06, "loss": 0.4987, "step": 965 }, { "epoch": 0.4194528875379939, "grad_norm": 0.02861775830388069, "learning_rate": 2.966985702759828e-06, "loss": 0.4857, "step": 966 }, { "epoch": 0.4198871037776813, "grad_norm": 0.033127009868621826, "learning_rate": 2.7951841353817676e-06, "loss": 0.465, "step": 967 }, { "epoch": 0.42032132001736866, "grad_norm": 0.02451484464108944, "learning_rate": 2.6284921167712973e-06, "loss": 0.4863, "step": 968 }, { "epoch": 0.42075553625705603, "grad_norm": 0.02445485256612301, "learning_rate": 2.466911359946977e-06, "loss": 0.4896, "step": 969 }, { "epoch": 0.4211897524967434, "grad_norm": 0.02243492379784584, "learning_rate": 2.3104435254008848e-06, "loss": 0.4839, "step": 970 }, { "epoch": 0.42162396873643077, "grad_norm": 0.026183119043707848, "learning_rate": 2.1590902210822937e-06, "loss": 0.4594, "step": 971 }, { "epoch": 0.42205818497611813, "grad_norm": 0.025734947994351387, "learning_rate": 2.012853002380466e-06, "loss": 0.4684, "step": 972 }, { "epoch": 0.42249240121580545, "grad_norm": 0.027873944491147995, "learning_rate": 1.8717333721091634e-06, "loss": 0.5056, "step": 973 }, { "epoch": 0.4229266174554928, "grad_norm": 0.02484060265123844, "learning_rate": 1.735732780490884e-06, "loss": 0.5153, "step": 974 }, { "epoch": 0.4233608336951802, "grad_norm": 0.022599438205361366, "learning_rate": 1.6048526251421502e-06, "loss": 0.4936, "step": 975 }, { "epoch": 0.42379504993486755, "grad_norm": 0.020465485751628876, "learning_rate": 1.4790942510590765e-06, "loss": 0.4848, "step": 976 }, { "epoch": 0.4242292661745549, "grad_norm": 0.027462849393486977, "learning_rate": 1.3584589506034362e-06, "loss": 0.4666, "step": 977 }, { "epoch": 0.4246634824142423, "grad_norm": 0.02553846500813961, "learning_rate": 1.2429479634897267e-06, "loss": 0.4899, "step": 978 }, { "epoch": 0.42509769865392966, "grad_norm": 0.024199191480875015, "learning_rate": 1.132562476771959e-06, "loss": 0.4792, "step": 979 }, { "epoch": 0.425531914893617, "grad_norm": 0.023328760638833046, "learning_rate": 1.0273036248318324e-06, "loss": 0.4973, "step": 980 }, { "epoch": 0.4259661311333044, "grad_norm": 0.02446364425122738, "learning_rate": 9.271724893669121e-07, "loss": 0.4654, "step": 981 }, { "epoch": 0.42640034737299176, "grad_norm": 0.02676616981625557, "learning_rate": 8.321700993795811e-07, "loss": 0.4881, "step": 982 }, { "epoch": 0.42683456361267913, "grad_norm": 0.02478119358420372, "learning_rate": 7.422974311662722e-07, "loss": 0.5055, "step": 983 }, { "epoch": 0.4272687798523665, "grad_norm": 0.03248545154929161, "learning_rate": 6.575554083078084e-07, "loss": 0.4949, "step": 984 }, { "epoch": 0.42770299609205387, "grad_norm": 0.024504756554961205, "learning_rate": 5.779449016595773e-07, "loss": 0.4986, "step": 985 }, { "epoch": 0.42813721233174123, "grad_norm": 0.02531745657324791, "learning_rate": 5.034667293427053e-07, "loss": 0.4851, "step": 986 }, { "epoch": 0.42857142857142855, "grad_norm": 0.024369893595576286, "learning_rate": 4.3412165673556436e-07, "loss": 0.5044, "step": 987 }, { "epoch": 0.4290056448111159, "grad_norm": 0.02750435471534729, "learning_rate": 3.6991039646616656e-07, "loss": 0.468, "step": 988 }, { "epoch": 0.4294398610508033, "grad_norm": 0.02262257970869541, "learning_rate": 3.1083360840455934e-07, "loss": 0.498, "step": 989 }, { "epoch": 0.42987407729049065, "grad_norm": 0.024247439578175545, "learning_rate": 2.568918996560532e-07, "loss": 0.4885, "step": 990 }, { "epoch": 0.430308293530178, "grad_norm": 0.024190355092287064, "learning_rate": 2.0808582455528192e-07, "loss": 0.541, "step": 991 }, { "epoch": 0.4307425097698654, "grad_norm": 0.023794230073690414, "learning_rate": 1.6441588466009628e-07, "loss": 0.4616, "step": 992 }, { "epoch": 0.43117672600955276, "grad_norm": 0.023178860545158386, "learning_rate": 1.2588252874673466e-07, "loss": 0.5077, "step": 993 }, { "epoch": 0.4316109422492401, "grad_norm": 0.0218358151614666, "learning_rate": 9.248615280499361e-08, "loss": 0.517, "step": 994 }, { "epoch": 0.4320451584889275, "grad_norm": 0.021379638463258743, "learning_rate": 6.422710003439747e-08, "loss": 0.4589, "step": 995 }, { "epoch": 0.43247937472861486, "grad_norm": 0.024375783279538155, "learning_rate": 4.1105660840368156e-08, "loss": 0.4963, "step": 996 }, { "epoch": 0.43291359096830223, "grad_norm": 0.02347472496330738, "learning_rate": 2.3122072831505136e-08, "loss": 0.4855, "step": 997 }, { "epoch": 0.4333478072079896, "grad_norm": 0.02579214982688427, "learning_rate": 1.0276520816976388e-08, "loss": 0.4786, "step": 998 }, { "epoch": 0.43378202344767697, "grad_norm": 0.023950695991516113, "learning_rate": 2.5691368046865116e-09, "loss": 0.5068, "step": 999 }, { "epoch": 0.4342162396873643, "grad_norm": 0.022179216146469116, "learning_rate": 0.0, "loss": 0.4841, "step": 1000 }, { "epoch": 0.4342162396873643, "step": 1000, "total_flos": 1.0100654119636173e+20, "train_loss": 0.710578792899847, "train_runtime": 20977.6715, "train_samples_per_second": 0.763, "train_steps_per_second": 0.048 } ], "logging_steps": 1.0, "max_steps": 1000, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": false, "should_training_stop": false }, "attributes": {} } }, "total_flos": 1.0100654119636173e+20, "train_batch_size": 1, "trial_name": null, "trial_params": null }