{ "best_metric": 0.012959838844835758, "best_model_checkpoint": "/home/paperspace/Data/models/rommel_importgenius_4b8/llm3br256/checkpoint-350", "epoch": 4.97907949790795, "eval_steps": 5, "global_step": 595, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.008368200836820083, "grad_norm": 0.2135079950094223, "learning_rate": 1.6666666666666667e-06, "loss": 0.0776, "step": 1 }, { "epoch": 0.016736401673640166, "grad_norm": 0.18678432703018188, "learning_rate": 3.3333333333333333e-06, "loss": 0.0663, "step": 2 }, { "epoch": 0.02510460251046025, "grad_norm": 0.18199113011360168, "learning_rate": 5e-06, "loss": 0.0666, "step": 3 }, { "epoch": 0.03347280334728033, "grad_norm": 0.13908497989177704, "learning_rate": 6.666666666666667e-06, "loss": 0.0596, "step": 4 }, { "epoch": 0.04184100418410042, "grad_norm": 0.1845937818288803, "learning_rate": 8.333333333333334e-06, "loss": 0.0672, "step": 5 }, { "epoch": 0.04184100418410042, "eval_loss": 0.07553808391094208, "eval_runtime": 10.1109, "eval_samples_per_second": 4.945, "eval_steps_per_second": 1.286, "step": 5 }, { "epoch": 0.0502092050209205, "grad_norm": 0.1750287413597107, "learning_rate": 1e-05, "loss": 0.0618, "step": 6 }, { "epoch": 0.058577405857740586, "grad_norm": 0.1204298809170723, "learning_rate": 1.1666666666666668e-05, "loss": 0.0529, "step": 7 }, { "epoch": 0.06694560669456066, "grad_norm": 0.12102019786834717, "learning_rate": 1.3333333333333333e-05, "loss": 0.0519, "step": 8 }, { "epoch": 0.07531380753138076, "grad_norm": 0.1123175099492073, "learning_rate": 1.5e-05, "loss": 0.0494, "step": 9 }, { "epoch": 0.08368200836820083, "grad_norm": 0.08419068902730942, "learning_rate": 1.6666666666666667e-05, "loss": 0.0476, "step": 10 }, { "epoch": 0.08368200836820083, "eval_loss": 0.04517042636871338, "eval_runtime": 8.1834, "eval_samples_per_second": 6.11, "eval_steps_per_second": 1.589, "step": 10 }, { "epoch": 0.09205020920502092, "grad_norm": 0.0977497473359108, "learning_rate": 1.8333333333333333e-05, "loss": 0.0382, "step": 11 }, { "epoch": 0.100418410041841, "grad_norm": 0.09294793009757996, "learning_rate": 2e-05, "loss": 0.0374, "step": 12 }, { "epoch": 0.1087866108786611, "grad_norm": 0.0773313045501709, "learning_rate": 2.1666666666666667e-05, "loss": 0.0365, "step": 13 }, { "epoch": 0.11715481171548117, "grad_norm": 0.07890919595956802, "learning_rate": 2.3333333333333336e-05, "loss": 0.0323, "step": 14 }, { "epoch": 0.12552301255230125, "grad_norm": 0.0590483732521534, "learning_rate": 2.5e-05, "loss": 0.0337, "step": 15 }, { "epoch": 0.12552301255230125, "eval_loss": 0.03555193170905113, "eval_runtime": 8.2187, "eval_samples_per_second": 6.084, "eval_steps_per_second": 1.582, "step": 15 }, { "epoch": 0.13389121338912133, "grad_norm": 0.05162455514073372, "learning_rate": 2.6666666666666667e-05, "loss": 0.0336, "step": 16 }, { "epoch": 0.14225941422594143, "grad_norm": 0.04780131205916405, "learning_rate": 2.8333333333333335e-05, "loss": 0.0303, "step": 17 }, { "epoch": 0.1506276150627615, "grad_norm": 0.05417422577738762, "learning_rate": 3e-05, "loss": 0.0303, "step": 18 }, { "epoch": 0.1589958158995816, "grad_norm": 0.05216076225042343, "learning_rate": 3.1666666666666666e-05, "loss": 0.0329, "step": 19 }, { "epoch": 0.16736401673640167, "grad_norm": 0.04367639496922493, "learning_rate": 3.3333333333333335e-05, "loss": 0.0333, "step": 20 }, { "epoch": 0.16736401673640167, "eval_loss": 0.030791474506258965, "eval_runtime": 8.1829, "eval_samples_per_second": 6.11, "eval_steps_per_second": 1.589, "step": 20 }, { "epoch": 0.17573221757322174, "grad_norm": 0.0510227307677269, "learning_rate": 3.5e-05, "loss": 0.0302, "step": 21 }, { "epoch": 0.18410041841004185, "grad_norm": 0.03975889831781387, "learning_rate": 3.6666666666666666e-05, "loss": 0.0293, "step": 22 }, { "epoch": 0.19246861924686193, "grad_norm": 0.033559419214725494, "learning_rate": 3.8333333333333334e-05, "loss": 0.0276, "step": 23 }, { "epoch": 0.200836820083682, "grad_norm": 0.03443528711795807, "learning_rate": 4e-05, "loss": 0.0255, "step": 24 }, { "epoch": 0.20920502092050208, "grad_norm": 0.04452834278345108, "learning_rate": 4.166666666666667e-05, "loss": 0.0258, "step": 25 }, { "epoch": 0.20920502092050208, "eval_loss": 0.027234511449933052, "eval_runtime": 8.2238, "eval_samples_per_second": 6.08, "eval_steps_per_second": 1.581, "step": 25 }, { "epoch": 0.2175732217573222, "grad_norm": 0.04225548356771469, "learning_rate": 4.3333333333333334e-05, "loss": 0.0265, "step": 26 }, { "epoch": 0.22594142259414227, "grad_norm": 0.035125982016325, "learning_rate": 4.5e-05, "loss": 0.0248, "step": 27 }, { "epoch": 0.23430962343096234, "grad_norm": 0.0293523371219635, "learning_rate": 4.666666666666667e-05, "loss": 0.0233, "step": 28 }, { "epoch": 0.24267782426778242, "grad_norm": 0.046518724411726, "learning_rate": 4.8333333333333334e-05, "loss": 0.025, "step": 29 }, { "epoch": 0.2510460251046025, "grad_norm": 0.031807683408260345, "learning_rate": 5e-05, "loss": 0.023, "step": 30 }, { "epoch": 0.2510460251046025, "eval_loss": 0.025461601093411446, "eval_runtime": 8.1735, "eval_samples_per_second": 6.117, "eval_steps_per_second": 1.591, "step": 30 }, { "epoch": 0.2594142259414226, "grad_norm": 0.03979400172829628, "learning_rate": 5.166666666666667e-05, "loss": 0.0225, "step": 31 }, { "epoch": 0.26778242677824265, "grad_norm": 0.030477149412035942, "learning_rate": 5.333333333333333e-05, "loss": 0.0212, "step": 32 }, { "epoch": 0.27615062761506276, "grad_norm": 0.029403740540146828, "learning_rate": 5.500000000000001e-05, "loss": 0.0223, "step": 33 }, { "epoch": 0.28451882845188287, "grad_norm": 0.03240867331624031, "learning_rate": 5.666666666666667e-05, "loss": 0.024, "step": 34 }, { "epoch": 0.2928870292887029, "grad_norm": 0.032563403248786926, "learning_rate": 5.833333333333334e-05, "loss": 0.0202, "step": 35 }, { "epoch": 0.2928870292887029, "eval_loss": 0.023448586463928223, "eval_runtime": 8.2019, "eval_samples_per_second": 6.096, "eval_steps_per_second": 1.585, "step": 35 }, { "epoch": 0.301255230125523, "grad_norm": 0.029798880219459534, "learning_rate": 6e-05, "loss": 0.0214, "step": 36 }, { "epoch": 0.30962343096234307, "grad_norm": 0.02481868863105774, "learning_rate": 6.166666666666667e-05, "loss": 0.0175, "step": 37 }, { "epoch": 0.3179916317991632, "grad_norm": 0.029312577098608017, "learning_rate": 6.333333333333333e-05, "loss": 0.0207, "step": 38 }, { "epoch": 0.3263598326359833, "grad_norm": 0.035519231110811234, "learning_rate": 6.500000000000001e-05, "loss": 0.019, "step": 39 }, { "epoch": 0.33472803347280333, "grad_norm": 0.019138505682349205, "learning_rate": 6.666666666666667e-05, "loss": 0.0188, "step": 40 }, { "epoch": 0.33472803347280333, "eval_loss": 0.02175285294651985, "eval_runtime": 8.1801, "eval_samples_per_second": 6.112, "eval_steps_per_second": 1.589, "step": 40 }, { "epoch": 0.34309623430962344, "grad_norm": 0.028007732704281807, "learning_rate": 6.833333333333333e-05, "loss": 0.0198, "step": 41 }, { "epoch": 0.3514644351464435, "grad_norm": 0.02277914620935917, "learning_rate": 7e-05, "loss": 0.0214, "step": 42 }, { "epoch": 0.3598326359832636, "grad_norm": 0.03379930928349495, "learning_rate": 7.166666666666667e-05, "loss": 0.0228, "step": 43 }, { "epoch": 0.3682008368200837, "grad_norm": 0.02658325619995594, "learning_rate": 7.333333333333333e-05, "loss": 0.0201, "step": 44 }, { "epoch": 0.37656903765690375, "grad_norm": 0.0228717178106308, "learning_rate": 7.500000000000001e-05, "loss": 0.0185, "step": 45 }, { "epoch": 0.37656903765690375, "eval_loss": 0.02076469548046589, "eval_runtime": 8.1924, "eval_samples_per_second": 6.103, "eval_steps_per_second": 1.587, "step": 45 }, { "epoch": 0.38493723849372385, "grad_norm": 0.022528983652591705, "learning_rate": 7.666666666666667e-05, "loss": 0.0175, "step": 46 }, { "epoch": 0.39330543933054396, "grad_norm": 0.02600259892642498, "learning_rate": 7.833333333333333e-05, "loss": 0.0199, "step": 47 }, { "epoch": 0.401673640167364, "grad_norm": 0.022605430334806442, "learning_rate": 8e-05, "loss": 0.0178, "step": 48 }, { "epoch": 0.4100418410041841, "grad_norm": 0.020774831995368004, "learning_rate": 8.166666666666667e-05, "loss": 0.0168, "step": 49 }, { "epoch": 0.41841004184100417, "grad_norm": 0.02565855346620083, "learning_rate": 8.333333333333334e-05, "loss": 0.0199, "step": 50 }, { "epoch": 0.41841004184100417, "eval_loss": 0.020035894587635994, "eval_runtime": 8.1839, "eval_samples_per_second": 6.11, "eval_steps_per_second": 1.588, "step": 50 }, { "epoch": 0.42677824267782427, "grad_norm": 0.027278423309326172, "learning_rate": 8.5e-05, "loss": 0.0208, "step": 51 }, { "epoch": 0.4351464435146444, "grad_norm": 0.022894414141774178, "learning_rate": 8.666666666666667e-05, "loss": 0.0166, "step": 52 }, { "epoch": 0.4435146443514644, "grad_norm": 0.02473956160247326, "learning_rate": 8.833333333333333e-05, "loss": 0.021, "step": 53 }, { "epoch": 0.45188284518828453, "grad_norm": 0.02041654661297798, "learning_rate": 9e-05, "loss": 0.0189, "step": 54 }, { "epoch": 0.4602510460251046, "grad_norm": 0.022939356043934822, "learning_rate": 9.166666666666667e-05, "loss": 0.0198, "step": 55 }, { "epoch": 0.4602510460251046, "eval_loss": 0.019482074305415154, "eval_runtime": 8.206, "eval_samples_per_second": 6.093, "eval_steps_per_second": 1.584, "step": 55 }, { "epoch": 0.4686192468619247, "grad_norm": 0.018536586314439774, "learning_rate": 9.333333333333334e-05, "loss": 0.0176, "step": 56 }, { "epoch": 0.4769874476987448, "grad_norm": 0.02208622545003891, "learning_rate": 9.5e-05, "loss": 0.0189, "step": 57 }, { "epoch": 0.48535564853556484, "grad_norm": 0.021715424954891205, "learning_rate": 9.666666666666667e-05, "loss": 0.0176, "step": 58 }, { "epoch": 0.49372384937238495, "grad_norm": 0.022242875769734383, "learning_rate": 9.833333333333333e-05, "loss": 0.0208, "step": 59 }, { "epoch": 0.502092050209205, "grad_norm": 0.022756585851311684, "learning_rate": 0.0001, "loss": 0.0179, "step": 60 }, { "epoch": 0.502092050209205, "eval_loss": 0.018894275650382042, "eval_runtime": 8.1933, "eval_samples_per_second": 6.103, "eval_steps_per_second": 1.587, "step": 60 }, { "epoch": 0.5104602510460251, "grad_norm": 0.021328022703528404, "learning_rate": 9.999913795300544e-05, "loss": 0.0183, "step": 61 }, { "epoch": 0.5188284518828452, "grad_norm": 0.018391713500022888, "learning_rate": 9.999655184174672e-05, "loss": 0.0172, "step": 62 }, { "epoch": 0.5271966527196653, "grad_norm": 0.019118035212159157, "learning_rate": 9.999224175539785e-05, "loss": 0.0172, "step": 63 }, { "epoch": 0.5355648535564853, "grad_norm": 0.024857187643647194, "learning_rate": 9.99862078425787e-05, "loss": 0.0192, "step": 64 }, { "epoch": 0.5439330543933054, "grad_norm": 0.025595078244805336, "learning_rate": 9.997845031134992e-05, "loss": 0.0185, "step": 65 }, { "epoch": 0.5439330543933054, "eval_loss": 0.018585730344057083, "eval_runtime": 8.172, "eval_samples_per_second": 6.118, "eval_steps_per_second": 1.591, "step": 65 }, { "epoch": 0.5523012552301255, "grad_norm": 0.02300618588924408, "learning_rate": 9.996896942920578e-05, "loss": 0.0174, "step": 66 }, { "epoch": 0.5606694560669456, "grad_norm": 0.022543633356690407, "learning_rate": 9.99577655230649e-05, "loss": 0.0155, "step": 67 }, { "epoch": 0.5690376569037657, "grad_norm": 0.028140394017100334, "learning_rate": 9.994483897925905e-05, "loss": 0.0167, "step": 68 }, { "epoch": 0.5774058577405857, "grad_norm": 0.027847809717059135, "learning_rate": 9.993019024351974e-05, "loss": 0.0184, "step": 69 }, { "epoch": 0.5857740585774058, "grad_norm": 0.02762455679476261, "learning_rate": 9.991381982096292e-05, "loss": 0.0174, "step": 70 }, { "epoch": 0.5857740585774058, "eval_loss": 0.018603580072522163, "eval_runtime": 8.1911, "eval_samples_per_second": 6.104, "eval_steps_per_second": 1.587, "step": 70 }, { "epoch": 0.5941422594142259, "grad_norm": 0.02379084751009941, "learning_rate": 9.989572827607153e-05, "loss": 0.0151, "step": 71 }, { "epoch": 0.602510460251046, "grad_norm": 0.028387486934661865, "learning_rate": 9.987591623267606e-05, "loss": 0.0192, "step": 72 }, { "epoch": 0.6108786610878661, "grad_norm": 0.029841607436537743, "learning_rate": 9.9854384373933e-05, "loss": 0.0166, "step": 73 }, { "epoch": 0.6192468619246861, "grad_norm": 0.024572154507040977, "learning_rate": 9.983113344230129e-05, "loss": 0.0181, "step": 74 }, { "epoch": 0.6276150627615062, "grad_norm": 0.024675268679857254, "learning_rate": 9.98061642395168e-05, "loss": 0.0157, "step": 75 }, { "epoch": 0.6276150627615062, "eval_loss": 0.0182618610560894, "eval_runtime": 8.1895, "eval_samples_per_second": 6.105, "eval_steps_per_second": 1.587, "step": 75 }, { "epoch": 0.6359832635983264, "grad_norm": 0.019830523058772087, "learning_rate": 9.977947762656455e-05, "loss": 0.0183, "step": 76 }, { "epoch": 0.6443514644351465, "grad_norm": 0.017791615799069405, "learning_rate": 9.975107452364913e-05, "loss": 0.0174, "step": 77 }, { "epoch": 0.6527196652719666, "grad_norm": 0.01926375925540924, "learning_rate": 9.972095591016294e-05, "loss": 0.0158, "step": 78 }, { "epoch": 0.6610878661087866, "grad_norm": 0.022102929651737213, "learning_rate": 9.968912282465236e-05, "loss": 0.0155, "step": 79 }, { "epoch": 0.6694560669456067, "grad_norm": 0.024682268500328064, "learning_rate": 9.965557636478203e-05, "loss": 0.0175, "step": 80 }, { "epoch": 0.6694560669456067, "eval_loss": 0.017632750794291496, "eval_runtime": 8.184, "eval_samples_per_second": 6.11, "eval_steps_per_second": 1.588, "step": 80 }, { "epoch": 0.6778242677824268, "grad_norm": 0.022384189069271088, "learning_rate": 9.962031768729693e-05, "loss": 0.0184, "step": 81 }, { "epoch": 0.6861924686192469, "grad_norm": 0.024247104302048683, "learning_rate": 9.958334800798256e-05, "loss": 0.0162, "step": 82 }, { "epoch": 0.694560669456067, "grad_norm": 0.02073194645345211, "learning_rate": 9.954466860162295e-05, "loss": 0.0146, "step": 83 }, { "epoch": 0.702928870292887, "grad_norm": 0.021671872586011887, "learning_rate": 9.950428080195675e-05, "loss": 0.0183, "step": 84 }, { "epoch": 0.7112970711297071, "grad_norm": 0.022358493879437447, "learning_rate": 9.94621860016312e-05, "loss": 0.0175, "step": 85 }, { "epoch": 0.7112970711297071, "eval_loss": 0.01763150468468666, "eval_runtime": 8.1866, "eval_samples_per_second": 6.108, "eval_steps_per_second": 1.588, "step": 85 }, { "epoch": 0.7196652719665272, "grad_norm": 0.027475032955408096, "learning_rate": 9.941838565215413e-05, "loss": 0.018, "step": 86 }, { "epoch": 0.7280334728033473, "grad_norm": 0.022491537034511566, "learning_rate": 9.937288126384396e-05, "loss": 0.0189, "step": 87 }, { "epoch": 0.7364016736401674, "grad_norm": 0.017695261165499687, "learning_rate": 9.932567440577751e-05, "loss": 0.0169, "step": 88 }, { "epoch": 0.7447698744769874, "grad_norm": 0.021199015900492668, "learning_rate": 9.9276766705736e-05, "loss": 0.016, "step": 89 }, { "epoch": 0.7531380753138075, "grad_norm": 0.02306659147143364, "learning_rate": 9.922615985014887e-05, "loss": 0.0164, "step": 90 }, { "epoch": 0.7531380753138075, "eval_loss": 0.017131321132183075, "eval_runtime": 8.1843, "eval_samples_per_second": 6.109, "eval_steps_per_second": 1.588, "step": 90 }, { "epoch": 0.7615062761506276, "grad_norm": 0.023978663608431816, "learning_rate": 9.91738555840356e-05, "loss": 0.0176, "step": 91 }, { "epoch": 0.7698744769874477, "grad_norm": 0.020384812727570534, "learning_rate": 9.911985571094564e-05, "loss": 0.0178, "step": 92 }, { "epoch": 0.7782426778242678, "grad_norm": 0.02201078273355961, "learning_rate": 9.906416209289608e-05, "loss": 0.0161, "step": 93 }, { "epoch": 0.7866108786610879, "grad_norm": 0.021811528131365776, "learning_rate": 9.900677665030762e-05, "loss": 0.0161, "step": 94 }, { "epoch": 0.7949790794979079, "grad_norm": 0.0257060918956995, "learning_rate": 9.894770136193814e-05, "loss": 0.0182, "step": 95 }, { "epoch": 0.7949790794979079, "eval_loss": 0.016829807311296463, "eval_runtime": 8.1842, "eval_samples_per_second": 6.109, "eval_steps_per_second": 1.588, "step": 95 }, { "epoch": 0.803347280334728, "grad_norm": 0.018415937200188637, "learning_rate": 9.888693826481467e-05, "loss": 0.0154, "step": 96 }, { "epoch": 0.8117154811715481, "grad_norm": 0.02051800675690174, "learning_rate": 9.882448945416298e-05, "loss": 0.0168, "step": 97 }, { "epoch": 0.8200836820083682, "grad_norm": 0.01812230981886387, "learning_rate": 9.87603570833355e-05, "loss": 0.0142, "step": 98 }, { "epoch": 0.8284518828451883, "grad_norm": 0.02094101719558239, "learning_rate": 9.869454336373689e-05, "loss": 0.0163, "step": 99 }, { "epoch": 0.8368200836820083, "grad_norm": 0.02226194366812706, "learning_rate": 9.862705056474795e-05, "loss": 0.019, "step": 100 }, { "epoch": 0.8368200836820083, "eval_loss": 0.016707511618733406, "eval_runtime": 8.1876, "eval_samples_per_second": 6.107, "eval_steps_per_second": 1.588, "step": 100 }, { "epoch": 0.8451882845188284, "grad_norm": 0.018342219293117523, "learning_rate": 9.855788101364722e-05, "loss": 0.0139, "step": 101 }, { "epoch": 0.8535564853556485, "grad_norm": 0.024607762694358826, "learning_rate": 9.848703709553089e-05, "loss": 0.017, "step": 102 }, { "epoch": 0.8619246861924686, "grad_norm": 0.021850740537047386, "learning_rate": 9.841452125323041e-05, "loss": 0.0171, "step": 103 }, { "epoch": 0.8702928870292888, "grad_norm": 0.022089634090662003, "learning_rate": 9.834033598722831e-05, "loss": 0.0151, "step": 104 }, { "epoch": 0.8786610878661087, "grad_norm": 0.01841079257428646, "learning_rate": 9.826448385557207e-05, "loss": 0.0163, "step": 105 }, { "epoch": 0.8786610878661087, "eval_loss": 0.01583385095000267, "eval_runtime": 8.2041, "eval_samples_per_second": 6.094, "eval_steps_per_second": 1.585, "step": 105 }, { "epoch": 0.8870292887029289, "grad_norm": 0.020294038578867912, "learning_rate": 9.818696747378573e-05, "loss": 0.0177, "step": 106 }, { "epoch": 0.895397489539749, "grad_norm": 0.015671249479055405, "learning_rate": 9.810778951477986e-05, "loss": 0.0144, "step": 107 }, { "epoch": 0.9037656903765691, "grad_norm": 0.018139947205781937, "learning_rate": 9.802695270875932e-05, "loss": 0.0154, "step": 108 }, { "epoch": 0.9121338912133892, "grad_norm": 0.018227294087409973, "learning_rate": 9.794445984312915e-05, "loss": 0.0149, "step": 109 }, { "epoch": 0.9205020920502092, "grad_norm": 0.015631545335054398, "learning_rate": 9.786031376239842e-05, "loss": 0.0145, "step": 110 }, { "epoch": 0.9205020920502092, "eval_loss": 0.0157511904835701, "eval_runtime": 8.1802, "eval_samples_per_second": 6.112, "eval_steps_per_second": 1.589, "step": 110 }, { "epoch": 0.9288702928870293, "grad_norm": 0.019501695409417152, "learning_rate": 9.777451736808216e-05, "loss": 0.0138, "step": 111 }, { "epoch": 0.9372384937238494, "grad_norm": 0.018380407243967056, "learning_rate": 9.768707361860134e-05, "loss": 0.0155, "step": 112 }, { "epoch": 0.9456066945606695, "grad_norm": 0.016666430979967117, "learning_rate": 9.75979855291808e-05, "loss": 0.0153, "step": 113 }, { "epoch": 0.9539748953974896, "grad_norm": 0.019959956407546997, "learning_rate": 9.750725617174534e-05, "loss": 0.0162, "step": 114 }, { "epoch": 0.9623430962343096, "grad_norm": 0.019904401153326035, "learning_rate": 9.741488867481376e-05, "loss": 0.0165, "step": 115 }, { "epoch": 0.9623430962343096, "eval_loss": 0.015494490042328835, "eval_runtime": 8.1797, "eval_samples_per_second": 6.113, "eval_steps_per_second": 1.589, "step": 115 }, { "epoch": 0.9707112970711297, "grad_norm": 0.02116929367184639, "learning_rate": 9.7320886223391e-05, "loss": 0.0153, "step": 116 }, { "epoch": 0.9790794979079498, "grad_norm": 0.019681984558701515, "learning_rate": 9.722525205885825e-05, "loss": 0.0149, "step": 117 }, { "epoch": 0.9874476987447699, "grad_norm": 0.02032177522778511, "learning_rate": 9.712798947886128e-05, "loss": 0.0148, "step": 118 }, { "epoch": 0.99581589958159, "grad_norm": 0.021595895290374756, "learning_rate": 9.702910183719671e-05, "loss": 0.0199, "step": 119 }, { "epoch": 1.00418410041841, "grad_norm": 0.027728823944926262, "learning_rate": 9.692859254369631e-05, "loss": 0.0205, "step": 120 }, { "epoch": 1.00418410041841, "eval_loss": 0.015181933529675007, "eval_runtime": 8.1872, "eval_samples_per_second": 6.107, "eval_steps_per_second": 1.588, "step": 120 }, { "epoch": 1.0125523012552302, "grad_norm": 0.017031384631991386, "learning_rate": 9.682646506410943e-05, "loss": 0.0152, "step": 121 }, { "epoch": 1.0209205020920502, "grad_norm": 0.01707589253783226, "learning_rate": 9.672272291998355e-05, "loss": 0.0127, "step": 122 }, { "epoch": 1.0292887029288702, "grad_norm": 0.02119339071214199, "learning_rate": 9.661736968854283e-05, "loss": 0.0143, "step": 123 }, { "epoch": 1.0376569037656904, "grad_norm": 0.01858825981616974, "learning_rate": 9.651040900256471e-05, "loss": 0.0145, "step": 124 }, { "epoch": 1.0460251046025104, "grad_norm": 0.013188479468226433, "learning_rate": 9.640184455025471e-05, "loss": 0.0105, "step": 125 }, { "epoch": 1.0460251046025104, "eval_loss": 0.015529554337263107, "eval_runtime": 8.1918, "eval_samples_per_second": 6.104, "eval_steps_per_second": 1.587, "step": 125 }, { "epoch": 1.0543933054393306, "grad_norm": 0.01512962393462658, "learning_rate": 9.629168007511924e-05, "loss": 0.0121, "step": 126 }, { "epoch": 1.0627615062761506, "grad_norm": 0.018388476222753525, "learning_rate": 9.617991937583648e-05, "loss": 0.0154, "step": 127 }, { "epoch": 1.0711297071129706, "grad_norm": 0.018059156835079193, "learning_rate": 9.60665663061254e-05, "loss": 0.0163, "step": 128 }, { "epoch": 1.0794979079497908, "grad_norm": 0.015875034034252167, "learning_rate": 9.595162477461298e-05, "loss": 0.0126, "step": 129 }, { "epoch": 1.0878661087866108, "grad_norm": 0.01882511004805565, "learning_rate": 9.583509874469923e-05, "loss": 0.0147, "step": 130 }, { "epoch": 1.0878661087866108, "eval_loss": 0.015730947256088257, "eval_runtime": 8.1997, "eval_samples_per_second": 6.098, "eval_steps_per_second": 1.585, "step": 130 }, { "epoch": 1.096234309623431, "grad_norm": 0.01482024509459734, "learning_rate": 9.571699223442074e-05, "loss": 0.0129, "step": 131 }, { "epoch": 1.104602510460251, "grad_norm": 0.017418276518583298, "learning_rate": 9.559730931631198e-05, "loss": 0.0137, "step": 132 }, { "epoch": 1.112970711297071, "grad_norm": 0.017272206023335457, "learning_rate": 9.5476054117265e-05, "loss": 0.0119, "step": 133 }, { "epoch": 1.1213389121338913, "grad_norm": 0.02229718305170536, "learning_rate": 9.535323081838691e-05, "loss": 0.0153, "step": 134 }, { "epoch": 1.1297071129707112, "grad_norm": 0.021549543365836143, "learning_rate": 9.522884365485598e-05, "loss": 0.0148, "step": 135 }, { "epoch": 1.1297071129707112, "eval_loss": 0.015986260026693344, "eval_runtime": 8.1904, "eval_samples_per_second": 6.105, "eval_steps_per_second": 1.587, "step": 135 }, { "epoch": 1.1380753138075315, "grad_norm": 0.01844065822660923, "learning_rate": 9.510289691577544e-05, "loss": 0.0138, "step": 136 }, { "epoch": 1.1464435146443515, "grad_norm": 0.022018112242221832, "learning_rate": 9.49753949440256e-05, "loss": 0.017, "step": 137 }, { "epoch": 1.1548117154811715, "grad_norm": 0.015306939370930195, "learning_rate": 9.484634213611411e-05, "loss": 0.0114, "step": 138 }, { "epoch": 1.1631799163179917, "grad_norm": 0.017746033146977425, "learning_rate": 9.471574294202439e-05, "loss": 0.0132, "step": 139 }, { "epoch": 1.1715481171548117, "grad_norm": 0.01575053483247757, "learning_rate": 9.458360186506213e-05, "loss": 0.0115, "step": 140 }, { "epoch": 1.1715481171548117, "eval_loss": 0.015280088409781456, "eval_runtime": 8.1915, "eval_samples_per_second": 6.104, "eval_steps_per_second": 1.587, "step": 140 }, { "epoch": 1.1799163179916319, "grad_norm": 0.019290335476398468, "learning_rate": 9.444992346170007e-05, "loss": 0.0131, "step": 141 }, { "epoch": 1.1882845188284519, "grad_norm": 0.0183683130890131, "learning_rate": 9.431471234142087e-05, "loss": 0.0136, "step": 142 }, { "epoch": 1.196652719665272, "grad_norm": 0.020060362294316292, "learning_rate": 9.41779731665581e-05, "loss": 0.0149, "step": 143 }, { "epoch": 1.205020920502092, "grad_norm": 0.017927493900060654, "learning_rate": 9.403971065213555e-05, "loss": 0.0128, "step": 144 }, { "epoch": 1.213389121338912, "grad_norm": 0.022669749334454536, "learning_rate": 9.389992956570462e-05, "loss": 0.0166, "step": 145 }, { "epoch": 1.213389121338912, "eval_loss": 0.015301547013223171, "eval_runtime": 8.1792, "eval_samples_per_second": 6.113, "eval_steps_per_second": 1.589, "step": 145 }, { "epoch": 1.2217573221757323, "grad_norm": 0.022278638556599617, "learning_rate": 9.375863472717993e-05, "loss": 0.0171, "step": 146 }, { "epoch": 1.2301255230125523, "grad_norm": 0.019489524886012077, "learning_rate": 9.361583100867313e-05, "loss": 0.013, "step": 147 }, { "epoch": 1.2384937238493725, "grad_norm": 0.018918950110673904, "learning_rate": 9.347152333432485e-05, "loss": 0.0139, "step": 148 }, { "epoch": 1.2468619246861925, "grad_norm": 0.01829618029296398, "learning_rate": 9.332571668013499e-05, "loss": 0.0147, "step": 149 }, { "epoch": 1.2552301255230125, "grad_norm": 0.018324939534068108, "learning_rate": 9.317841607379107e-05, "loss": 0.015, "step": 150 }, { "epoch": 1.2552301255230125, "eval_loss": 0.015605509281158447, "eval_runtime": 8.195, "eval_samples_per_second": 6.101, "eval_steps_per_second": 1.586, "step": 150 }, { "epoch": 1.2635983263598327, "grad_norm": 0.015184380114078522, "learning_rate": 9.302962659449486e-05, "loss": 0.0125, "step": 151 }, { "epoch": 1.2719665271966527, "grad_norm": 0.02084585465490818, "learning_rate": 9.287935337278733e-05, "loss": 0.0139, "step": 152 }, { "epoch": 1.280334728033473, "grad_norm": 0.022940685972571373, "learning_rate": 9.272760159037164e-05, "loss": 0.0168, "step": 153 }, { "epoch": 1.288702928870293, "grad_norm": 0.014851168729364872, "learning_rate": 9.25743764799345e-05, "loss": 0.0107, "step": 154 }, { "epoch": 1.297071129707113, "grad_norm": 0.01901235058903694, "learning_rate": 9.241968332496575e-05, "loss": 0.0148, "step": 155 }, { "epoch": 1.297071129707113, "eval_loss": 0.015689246356487274, "eval_runtime": 8.1747, "eval_samples_per_second": 6.116, "eval_steps_per_second": 1.59, "step": 155 }, { "epoch": 1.3054393305439331, "grad_norm": 0.019857851788401604, "learning_rate": 9.226352745957616e-05, "loss": 0.0121, "step": 156 }, { "epoch": 1.3138075313807531, "grad_norm": 0.020181361585855484, "learning_rate": 9.210591426831352e-05, "loss": 0.0138, "step": 157 }, { "epoch": 1.3221757322175733, "grad_norm": 0.0159543976187706, "learning_rate": 9.194684918597694e-05, "loss": 0.0119, "step": 158 }, { "epoch": 1.3305439330543933, "grad_norm": 0.015190091915428638, "learning_rate": 9.178633769742945e-05, "loss": 0.0122, "step": 159 }, { "epoch": 1.3389121338912133, "grad_norm": 0.01558399386703968, "learning_rate": 9.162438533740892e-05, "loss": 0.0112, "step": 160 }, { "epoch": 1.3389121338912133, "eval_loss": 0.01591303013265133, "eval_runtime": 8.203, "eval_samples_per_second": 6.095, "eval_steps_per_second": 1.585, "step": 160 }, { "epoch": 1.3472803347280335, "grad_norm": 0.020452240481972694, "learning_rate": 9.146099769033715e-05, "loss": 0.0141, "step": 161 }, { "epoch": 1.3556485355648535, "grad_norm": 0.01579088531434536, "learning_rate": 9.129618039012736e-05, "loss": 0.0123, "step": 162 }, { "epoch": 1.3640167364016738, "grad_norm": 0.01603122055530548, "learning_rate": 9.112993911998984e-05, "loss": 0.0142, "step": 163 }, { "epoch": 1.3723849372384938, "grad_norm": 0.018658515065908432, "learning_rate": 9.096227961223613e-05, "loss": 0.0149, "step": 164 }, { "epoch": 1.3807531380753137, "grad_norm": 0.01548719685524702, "learning_rate": 9.07932076480812e-05, "loss": 0.0128, "step": 165 }, { "epoch": 1.3807531380753137, "eval_loss": 0.015296611934900284, "eval_runtime": 8.1994, "eval_samples_per_second": 6.098, "eval_steps_per_second": 1.585, "step": 165 }, { "epoch": 1.389121338912134, "grad_norm": 0.019366098567843437, "learning_rate": 9.06227290574442e-05, "loss": 0.0136, "step": 166 }, { "epoch": 1.397489539748954, "grad_norm": 0.015139399096369743, "learning_rate": 9.045084971874738e-05, "loss": 0.0111, "step": 167 }, { "epoch": 1.4058577405857742, "grad_norm": 0.02105526067316532, "learning_rate": 9.027757555871344e-05, "loss": 0.0146, "step": 168 }, { "epoch": 1.4142259414225942, "grad_norm": 0.020205028355121613, "learning_rate": 9.010291255216117e-05, "loss": 0.0142, "step": 169 }, { "epoch": 1.4225941422594142, "grad_norm": 0.018409285694360733, "learning_rate": 8.99268667217993e-05, "loss": 0.0125, "step": 170 }, { "epoch": 1.4225941422594142, "eval_loss": 0.015141828916966915, "eval_runtime": 8.1926, "eval_samples_per_second": 6.103, "eval_steps_per_second": 1.587, "step": 170 }, { "epoch": 1.4309623430962344, "grad_norm": 0.0186452716588974, "learning_rate": 8.974944413801906e-05, "loss": 0.0158, "step": 171 }, { "epoch": 1.4393305439330544, "grad_norm": 0.017251404002308846, "learning_rate": 8.957065091868459e-05, "loss": 0.0138, "step": 172 }, { "epoch": 1.4476987447698746, "grad_norm": 0.015452460385859013, "learning_rate": 8.939049322892223e-05, "loss": 0.0104, "step": 173 }, { "epoch": 1.4560669456066946, "grad_norm": 0.018598852679133415, "learning_rate": 8.920897728090777e-05, "loss": 0.0142, "step": 174 }, { "epoch": 1.4644351464435146, "grad_norm": 0.017056932672858238, "learning_rate": 8.902610933365229e-05, "loss": 0.0137, "step": 175 }, { "epoch": 1.4644351464435146, "eval_loss": 0.015005652792751789, "eval_runtime": 8.1766, "eval_samples_per_second": 6.115, "eval_steps_per_second": 1.59, "step": 175 }, { "epoch": 1.4728033472803348, "grad_norm": 0.017095167189836502, "learning_rate": 8.88418956927864e-05, "loss": 0.0143, "step": 176 }, { "epoch": 1.4811715481171548, "grad_norm": 0.016802068799734116, "learning_rate": 8.865634271034268e-05, "loss": 0.0144, "step": 177 }, { "epoch": 1.489539748953975, "grad_norm": 0.015921777114272118, "learning_rate": 8.846945678453677e-05, "loss": 0.0138, "step": 178 }, { "epoch": 1.497907949790795, "grad_norm": 0.016825497150421143, "learning_rate": 8.828124435954672e-05, "loss": 0.0147, "step": 179 }, { "epoch": 1.506276150627615, "grad_norm": 0.01446506567299366, "learning_rate": 8.809171192529073e-05, "loss": 0.0131, "step": 180 }, { "epoch": 1.506276150627615, "eval_loss": 0.014517219737172127, "eval_runtime": 8.201, "eval_samples_per_second": 6.097, "eval_steps_per_second": 1.585, "step": 180 }, { "epoch": 1.514644351464435, "grad_norm": 0.016049686819314957, "learning_rate": 8.790086601720339e-05, "loss": 0.0137, "step": 181 }, { "epoch": 1.5230125523012552, "grad_norm": 0.015877775847911835, "learning_rate": 8.77087132160104e-05, "loss": 0.0126, "step": 182 }, { "epoch": 1.5313807531380754, "grad_norm": 0.01960177719593048, "learning_rate": 8.751526014750153e-05, "loss": 0.0149, "step": 183 }, { "epoch": 1.5397489539748954, "grad_norm": 0.0134076913818717, "learning_rate": 8.732051348230223e-05, "loss": 0.0114, "step": 184 }, { "epoch": 1.5481171548117154, "grad_norm": 0.012596129439771175, "learning_rate": 8.712447993564361e-05, "loss": 0.0105, "step": 185 }, { "epoch": 1.5481171548117154, "eval_loss": 0.014493227005004883, "eval_runtime": 8.1918, "eval_samples_per_second": 6.104, "eval_steps_per_second": 1.587, "step": 185 }, { "epoch": 1.5564853556485354, "grad_norm": 0.016738103702664375, "learning_rate": 8.692716626713084e-05, "loss": 0.0133, "step": 186 }, { "epoch": 1.5648535564853556, "grad_norm": 0.02182360552251339, "learning_rate": 8.672857928051013e-05, "loss": 0.014, "step": 187 }, { "epoch": 1.5732217573221758, "grad_norm": 0.017175616696476936, "learning_rate": 8.652872582343408e-05, "loss": 0.011, "step": 188 }, { "epoch": 1.5815899581589958, "grad_norm": 0.01974942907691002, "learning_rate": 8.632761278722556e-05, "loss": 0.0147, "step": 189 }, { "epoch": 1.5899581589958158, "grad_norm": 0.016432059928774834, "learning_rate": 8.612524710664012e-05, "loss": 0.0126, "step": 190 }, { "epoch": 1.5899581589958158, "eval_loss": 0.014403114095330238, "eval_runtime": 8.1978, "eval_samples_per_second": 6.099, "eval_steps_per_second": 1.586, "step": 190 }, { "epoch": 1.5983263598326358, "grad_norm": 0.01811247691512108, "learning_rate": 8.592163575962682e-05, "loss": 0.0128, "step": 191 }, { "epoch": 1.606694560669456, "grad_norm": 0.017857994884252548, "learning_rate": 8.571678576708766e-05, "loss": 0.0117, "step": 192 }, { "epoch": 1.6150627615062763, "grad_norm": 0.02201293222606182, "learning_rate": 8.551070419263545e-05, "loss": 0.017, "step": 193 }, { "epoch": 1.6234309623430963, "grad_norm": 0.018607452511787415, "learning_rate": 8.530339814235027e-05, "loss": 0.0123, "step": 194 }, { "epoch": 1.6317991631799162, "grad_norm": 0.015770630910992622, "learning_rate": 8.509487476453442e-05, "loss": 0.0119, "step": 195 }, { "epoch": 1.6317991631799162, "eval_loss": 0.014461501501500607, "eval_runtime": 8.1846, "eval_samples_per_second": 6.109, "eval_steps_per_second": 1.588, "step": 195 }, { "epoch": 1.6401673640167362, "grad_norm": 0.015161648392677307, "learning_rate": 8.488514124946594e-05, "loss": 0.0137, "step": 196 }, { "epoch": 1.6485355648535565, "grad_norm": 0.015444417484104633, "learning_rate": 8.46742048291507e-05, "loss": 0.0116, "step": 197 }, { "epoch": 1.6569037656903767, "grad_norm": 0.018221288919448853, "learning_rate": 8.446207277707296e-05, "loss": 0.0135, "step": 198 }, { "epoch": 1.6652719665271967, "grad_norm": 0.017886707559227943, "learning_rate": 8.424875240794467e-05, "loss": 0.0117, "step": 199 }, { "epoch": 1.6736401673640167, "grad_norm": 0.021269524469971657, "learning_rate": 8.403425107745316e-05, "loss": 0.016, "step": 200 }, { "epoch": 1.6736401673640167, "eval_loss": 0.014672570861876011, "eval_runtime": 8.2192, "eval_samples_per_second": 6.083, "eval_steps_per_second": 1.582, "step": 200 }, { "epoch": 1.6820083682008367, "grad_norm": 0.014137690886855125, "learning_rate": 8.381857618200746e-05, "loss": 0.0107, "step": 201 }, { "epoch": 1.6903765690376569, "grad_norm": 0.017183968797326088, "learning_rate": 8.360173515848343e-05, "loss": 0.0127, "step": 202 }, { "epoch": 1.698744769874477, "grad_norm": 0.018049761652946472, "learning_rate": 8.33837354839672e-05, "loss": 0.0099, "step": 203 }, { "epoch": 1.707112970711297, "grad_norm": 0.020697997882962227, "learning_rate": 8.316458467549729e-05, "loss": 0.0131, "step": 204 }, { "epoch": 1.715481171548117, "grad_norm": 0.022819431498646736, "learning_rate": 8.294429028980556e-05, "loss": 0.0143, "step": 205 }, { "epoch": 1.715481171548117, "eval_loss": 0.014957732520997524, "eval_runtime": 8.2053, "eval_samples_per_second": 6.094, "eval_steps_per_second": 1.584, "step": 205 }, { "epoch": 1.723849372384937, "grad_norm": 0.020023057237267494, "learning_rate": 8.272285992305653e-05, "loss": 0.014, "step": 206 }, { "epoch": 1.7322175732217573, "grad_norm": 0.01683293841779232, "learning_rate": 8.250030121058547e-05, "loss": 0.0112, "step": 207 }, { "epoch": 1.7405857740585775, "grad_norm": 0.01992541179060936, "learning_rate": 8.227662182663518e-05, "loss": 0.0129, "step": 208 }, { "epoch": 1.7489539748953975, "grad_norm": 0.01862913742661476, "learning_rate": 8.205182948409125e-05, "loss": 0.0131, "step": 209 }, { "epoch": 1.7573221757322175, "grad_norm": 0.01816224865615368, "learning_rate": 8.182593193421625e-05, "loss": 0.0139, "step": 210 }, { "epoch": 1.7573221757322175, "eval_loss": 0.014965851791203022, "eval_runtime": 8.2076, "eval_samples_per_second": 6.092, "eval_steps_per_second": 1.584, "step": 210 }, { "epoch": 1.7656903765690377, "grad_norm": 0.015293323434889317, "learning_rate": 8.159893696638232e-05, "loss": 0.0122, "step": 211 }, { "epoch": 1.7740585774058577, "grad_norm": 0.013619404286146164, "learning_rate": 8.137085240780262e-05, "loss": 0.0107, "step": 212 }, { "epoch": 1.782426778242678, "grad_norm": 0.015562871471047401, "learning_rate": 8.114168612326153e-05, "loss": 0.0106, "step": 213 }, { "epoch": 1.790794979079498, "grad_norm": 0.017086800187826157, "learning_rate": 8.091144601484332e-05, "loss": 0.0131, "step": 214 }, { "epoch": 1.799163179916318, "grad_norm": 0.020527249202132225, "learning_rate": 8.06801400216597e-05, "loss": 0.0139, "step": 215 }, { "epoch": 1.799163179916318, "eval_loss": 0.01450226828455925, "eval_runtime": 8.1927, "eval_samples_per_second": 6.103, "eval_steps_per_second": 1.587, "step": 215 }, { "epoch": 1.8075313807531381, "grad_norm": 0.015211045742034912, "learning_rate": 8.044777611957613e-05, "loss": 0.0114, "step": 216 }, { "epoch": 1.8158995815899581, "grad_norm": 0.01953182741999626, "learning_rate": 8.021436232093675e-05, "loss": 0.0136, "step": 217 }, { "epoch": 1.8242677824267783, "grad_norm": 0.018059900030493736, "learning_rate": 7.997990667428811e-05, "loss": 0.013, "step": 218 }, { "epoch": 1.8326359832635983, "grad_norm": 0.019347647204995155, "learning_rate": 7.974441726410164e-05, "loss": 0.0142, "step": 219 }, { "epoch": 1.8410041841004183, "grad_norm": 0.016715556383132935, "learning_rate": 7.950790221049484e-05, "loss": 0.0161, "step": 220 }, { "epoch": 1.8410041841004183, "eval_loss": 0.014255787245929241, "eval_runtime": 8.1832, "eval_samples_per_second": 6.11, "eval_steps_per_second": 1.589, "step": 220 }, { "epoch": 1.8493723849372385, "grad_norm": 0.015245197340846062, "learning_rate": 7.927036966895138e-05, "loss": 0.0117, "step": 221 }, { "epoch": 1.8577405857740585, "grad_norm": 0.016146808862686157, "learning_rate": 7.90318278300398e-05, "loss": 0.0132, "step": 222 }, { "epoch": 1.8661087866108788, "grad_norm": 0.01640206389129162, "learning_rate": 7.879228491913111e-05, "loss": 0.012, "step": 223 }, { "epoch": 1.8744769874476988, "grad_norm": 0.01924493908882141, "learning_rate": 7.855174919611517e-05, "loss": 0.0135, "step": 224 }, { "epoch": 1.8828451882845187, "grad_norm": 0.01706533320248127, "learning_rate": 7.831022895511587e-05, "loss": 0.0098, "step": 225 }, { "epoch": 1.8828451882845187, "eval_loss": 0.013838106766343117, "eval_runtime": 8.1961, "eval_samples_per_second": 6.1, "eval_steps_per_second": 1.586, "step": 225 }, { "epoch": 1.891213389121339, "grad_norm": 0.01932067796587944, "learning_rate": 7.80677325242051e-05, "loss": 0.0144, "step": 226 }, { "epoch": 1.899581589958159, "grad_norm": 0.018316643312573433, "learning_rate": 7.782426826511564e-05, "loss": 0.0124, "step": 227 }, { "epoch": 1.9079497907949792, "grad_norm": 0.016697222366929054, "learning_rate": 7.757984457295284e-05, "loss": 0.0125, "step": 228 }, { "epoch": 1.9163179916317992, "grad_norm": 0.014732821844518185, "learning_rate": 7.733446987590505e-05, "loss": 0.0115, "step": 229 }, { "epoch": 1.9246861924686192, "grad_norm": 0.01892097480595112, "learning_rate": 7.708815263495308e-05, "loss": 0.0108, "step": 230 }, { "epoch": 1.9246861924686192, "eval_loss": 0.014023682102560997, "eval_runtime": 8.1828, "eval_samples_per_second": 6.11, "eval_steps_per_second": 1.589, "step": 230 }, { "epoch": 1.9330543933054394, "grad_norm": 0.018761448562145233, "learning_rate": 7.684090134357839e-05, "loss": 0.014, "step": 231 }, { "epoch": 1.9414225941422594, "grad_norm": 0.01908651366829872, "learning_rate": 7.65927245274703e-05, "loss": 0.0102, "step": 232 }, { "epoch": 1.9497907949790796, "grad_norm": 0.018542446196079254, "learning_rate": 7.634363074423199e-05, "loss": 0.0133, "step": 233 }, { "epoch": 1.9581589958158996, "grad_norm": 0.018298540264368057, "learning_rate": 7.609362858308531e-05, "loss": 0.0139, "step": 234 }, { "epoch": 1.9665271966527196, "grad_norm": 0.01675599068403244, "learning_rate": 7.58427266645747e-05, "loss": 0.0117, "step": 235 }, { "epoch": 1.9665271966527196, "eval_loss": 0.014116828329861164, "eval_runtime": 8.1955, "eval_samples_per_second": 6.101, "eval_steps_per_second": 1.586, "step": 235 }, { "epoch": 1.9748953974895398, "grad_norm": 0.020569469779729843, "learning_rate": 7.559093364027001e-05, "loss": 0.0141, "step": 236 }, { "epoch": 1.9832635983263598, "grad_norm": 0.016174696385860443, "learning_rate": 7.5338258192468e-05, "loss": 0.0105, "step": 237 }, { "epoch": 1.99163179916318, "grad_norm": 0.02071259915828705, "learning_rate": 7.508470903389311e-05, "loss": 0.0138, "step": 238 }, { "epoch": 2.0, "grad_norm": 0.022858325392007828, "learning_rate": 7.48302949073969e-05, "loss": 0.0166, "step": 239 }, { "epoch": 2.00836820083682, "grad_norm": 0.01647881604731083, "learning_rate": 7.457502458565672e-05, "loss": 0.0109, "step": 240 }, { "epoch": 2.00836820083682, "eval_loss": 0.013845333829522133, "eval_runtime": 8.1876, "eval_samples_per_second": 6.107, "eval_steps_per_second": 1.588, "step": 240 }, { "epoch": 2.01673640167364, "grad_norm": 0.01600758358836174, "learning_rate": 7.431890687087313e-05, "loss": 0.0102, "step": 241 }, { "epoch": 2.0251046025104604, "grad_norm": 0.013506948947906494, "learning_rate": 7.406195059446636e-05, "loss": 0.0103, "step": 242 }, { "epoch": 2.0334728033472804, "grad_norm": 0.019151534885168076, "learning_rate": 7.380416461677187e-05, "loss": 0.0119, "step": 243 }, { "epoch": 2.0418410041841004, "grad_norm": 0.01597500592470169, "learning_rate": 7.354555782673473e-05, "loss": 0.0107, "step": 244 }, { "epoch": 2.0502092050209204, "grad_norm": 0.014115000143647194, "learning_rate": 7.328613914160318e-05, "loss": 0.0093, "step": 245 }, { "epoch": 2.0502092050209204, "eval_loss": 0.014546514488756657, "eval_runtime": 8.1884, "eval_samples_per_second": 6.106, "eval_steps_per_second": 1.588, "step": 245 }, { "epoch": 2.0585774058577404, "grad_norm": 0.01916162669658661, "learning_rate": 7.302591750662116e-05, "loss": 0.0117, "step": 246 }, { "epoch": 2.066945606694561, "grad_norm": 0.019530242308974266, "learning_rate": 7.276490189471977e-05, "loss": 0.01, "step": 247 }, { "epoch": 2.075313807531381, "grad_norm": 0.02205420657992363, "learning_rate": 7.250310130620799e-05, "loss": 0.0111, "step": 248 }, { "epoch": 2.083682008368201, "grad_norm": 0.025347299873828888, "learning_rate": 7.224052476846222e-05, "loss": 0.0115, "step": 249 }, { "epoch": 2.092050209205021, "grad_norm": 0.018653597682714462, "learning_rate": 7.197718133561508e-05, "loss": 0.0102, "step": 250 }, { "epoch": 2.092050209205021, "eval_loss": 0.014336950145661831, "eval_runtime": 8.2008, "eval_samples_per_second": 6.097, "eval_steps_per_second": 1.585, "step": 250 }, { "epoch": 2.100418410041841, "grad_norm": 0.01541041024029255, "learning_rate": 7.171308008824316e-05, "loss": 0.0105, "step": 251 }, { "epoch": 2.1087866108786613, "grad_norm": 0.014182745479047298, "learning_rate": 7.144823013305394e-05, "loss": 0.0085, "step": 252 }, { "epoch": 2.1171548117154813, "grad_norm": 0.022554123774170876, "learning_rate": 7.118264060257169e-05, "loss": 0.0116, "step": 253 }, { "epoch": 2.1255230125523012, "grad_norm": 0.02267191931605339, "learning_rate": 7.091632065482272e-05, "loss": 0.0115, "step": 254 }, { "epoch": 2.1338912133891212, "grad_norm": 0.01669103093445301, "learning_rate": 7.064927947301943e-05, "loss": 0.0104, "step": 255 }, { "epoch": 2.1338912133891212, "eval_loss": 0.014133421704173088, "eval_runtime": 8.1833, "eval_samples_per_second": 6.11, "eval_steps_per_second": 1.589, "step": 255 }, { "epoch": 2.1422594142259412, "grad_norm": 0.01666867360472679, "learning_rate": 7.038152626524373e-05, "loss": 0.0097, "step": 256 }, { "epoch": 2.1506276150627617, "grad_norm": 0.0202019102871418, "learning_rate": 7.011307026412958e-05, "loss": 0.0106, "step": 257 }, { "epoch": 2.1589958158995817, "grad_norm": 0.017287522554397583, "learning_rate": 6.984392072654453e-05, "loss": 0.0105, "step": 258 }, { "epoch": 2.1673640167364017, "grad_norm": 0.014798184856772423, "learning_rate": 6.957408693327055e-05, "loss": 0.0096, "step": 259 }, { "epoch": 2.1757322175732217, "grad_norm": 0.016632311046123505, "learning_rate": 6.930357818868409e-05, "loss": 0.0108, "step": 260 }, { "epoch": 2.1757322175732217, "eval_loss": 0.0147059615701437, "eval_runtime": 8.1958, "eval_samples_per_second": 6.101, "eval_steps_per_second": 1.586, "step": 260 }, { "epoch": 2.1841004184100417, "grad_norm": 0.016555817797780037, "learning_rate": 6.903240382043514e-05, "loss": 0.011, "step": 261 }, { "epoch": 2.192468619246862, "grad_norm": 0.015979807823896408, "learning_rate": 6.876057317912569e-05, "loss": 0.0108, "step": 262 }, { "epoch": 2.200836820083682, "grad_norm": 0.014860883355140686, "learning_rate": 6.848809563798721e-05, "loss": 0.0106, "step": 263 }, { "epoch": 2.209205020920502, "grad_norm": 0.016682744026184082, "learning_rate": 6.821498059255752e-05, "loss": 0.0101, "step": 264 }, { "epoch": 2.217573221757322, "grad_norm": 0.017817750573158264, "learning_rate": 6.79412374603568e-05, "loss": 0.0104, "step": 265 }, { "epoch": 2.217573221757322, "eval_loss": 0.014249371364712715, "eval_runtime": 8.1965, "eval_samples_per_second": 6.1, "eval_steps_per_second": 1.586, "step": 265 }, { "epoch": 2.225941422594142, "grad_norm": 0.014692910015583038, "learning_rate": 6.766687568056281e-05, "loss": 0.0093, "step": 266 }, { "epoch": 2.2343096234309625, "grad_norm": 0.01461927779018879, "learning_rate": 6.739190471368548e-05, "loss": 0.0099, "step": 267 }, { "epoch": 2.2426778242677825, "grad_norm": 0.01550222933292389, "learning_rate": 6.711633404124062e-05, "loss": 0.0105, "step": 268 }, { "epoch": 2.2510460251046025, "grad_norm": 0.01469459943473339, "learning_rate": 6.684017316542302e-05, "loss": 0.01, "step": 269 }, { "epoch": 2.2594142259414225, "grad_norm": 0.018585605546832085, "learning_rate": 6.656343160877881e-05, "loss": 0.0103, "step": 270 }, { "epoch": 2.2594142259414225, "eval_loss": 0.014431845396757126, "eval_runtime": 8.1882, "eval_samples_per_second": 6.106, "eval_steps_per_second": 1.588, "step": 270 }, { "epoch": 2.2677824267782425, "grad_norm": 0.01749185100197792, "learning_rate": 6.628611891387706e-05, "loss": 0.0098, "step": 271 }, { "epoch": 2.276150627615063, "grad_norm": 0.019719472154974937, "learning_rate": 6.600824464298082e-05, "loss": 0.0103, "step": 272 }, { "epoch": 2.284518828451883, "grad_norm": 0.019530480727553368, "learning_rate": 6.572981837771726e-05, "loss": 0.0106, "step": 273 }, { "epoch": 2.292887029288703, "grad_norm": 0.020081467926502228, "learning_rate": 6.545084971874738e-05, "loss": 0.0105, "step": 274 }, { "epoch": 2.301255230125523, "grad_norm": 0.016815155744552612, "learning_rate": 6.517134828543496e-05, "loss": 0.0107, "step": 275 }, { "epoch": 2.301255230125523, "eval_loss": 0.01435365341603756, "eval_runtime": 8.1841, "eval_samples_per_second": 6.109, "eval_steps_per_second": 1.588, "step": 275 }, { "epoch": 2.309623430962343, "grad_norm": 0.02080889418721199, "learning_rate": 6.489132371551482e-05, "loss": 0.0107, "step": 276 }, { "epoch": 2.3179916317991633, "grad_norm": 0.019815733656287193, "learning_rate": 6.461078566476054e-05, "loss": 0.0108, "step": 277 }, { "epoch": 2.3263598326359833, "grad_norm": 0.020922286435961723, "learning_rate": 6.43297438066514e-05, "loss": 0.0115, "step": 278 }, { "epoch": 2.3347280334728033, "grad_norm": 0.01711447723209858, "learning_rate": 6.404820783203901e-05, "loss": 0.0101, "step": 279 }, { "epoch": 2.3430962343096233, "grad_norm": 0.016599150374531746, "learning_rate": 6.3766187448813e-05, "loss": 0.0104, "step": 280 }, { "epoch": 2.3430962343096233, "eval_loss": 0.01405297126621008, "eval_runtime": 8.1827, "eval_samples_per_second": 6.11, "eval_steps_per_second": 1.589, "step": 280 }, { "epoch": 2.3514644351464433, "grad_norm": 0.01762828603386879, "learning_rate": 6.348369238156631e-05, "loss": 0.0099, "step": 281 }, { "epoch": 2.3598326359832638, "grad_norm": 0.014734679833054543, "learning_rate": 6.32007323712599e-05, "loss": 0.0094, "step": 282 }, { "epoch": 2.3682008368200838, "grad_norm": 0.01771124079823494, "learning_rate": 6.29173171748868e-05, "loss": 0.0106, "step": 283 }, { "epoch": 2.3765690376569037, "grad_norm": 0.01341992523521185, "learning_rate": 6.263345656513576e-05, "loss": 0.0096, "step": 284 }, { "epoch": 2.3849372384937237, "grad_norm": 0.014316578395664692, "learning_rate": 6.234916033005421e-05, "loss": 0.0092, "step": 285 }, { "epoch": 2.3849372384937237, "eval_loss": 0.014250432141125202, "eval_runtime": 8.1935, "eval_samples_per_second": 6.102, "eval_steps_per_second": 1.587, "step": 285 }, { "epoch": 2.393305439330544, "grad_norm": 0.017734510824084282, "learning_rate": 6.206443827271073e-05, "loss": 0.011, "step": 286 }, { "epoch": 2.401673640167364, "grad_norm": 0.016910862177610397, "learning_rate": 6.17793002108571e-05, "loss": 0.0101, "step": 287 }, { "epoch": 2.410041841004184, "grad_norm": 0.015546726994216442, "learning_rate": 6.149375597658967e-05, "loss": 0.0092, "step": 288 }, { "epoch": 2.418410041841004, "grad_norm": 0.01636183261871338, "learning_rate": 6.12078154160104e-05, "loss": 0.0105, "step": 289 }, { "epoch": 2.426778242677824, "grad_norm": 0.01692620851099491, "learning_rate": 6.092148838888732e-05, "loss": 0.0107, "step": 290 }, { "epoch": 2.426778242677824, "eval_loss": 0.01402355171740055, "eval_runtime": 8.183, "eval_samples_per_second": 6.11, "eval_steps_per_second": 1.589, "step": 290 }, { "epoch": 2.435146443514644, "grad_norm": 0.01607348956167698, "learning_rate": 6.0634784768314564e-05, "loss": 0.0107, "step": 291 }, { "epoch": 2.4435146443514646, "grad_norm": 0.019470971077680588, "learning_rate": 6.0347714440371924e-05, "loss": 0.0095, "step": 292 }, { "epoch": 2.4518828451882846, "grad_norm": 0.014622722752392292, "learning_rate": 6.0060287303783924e-05, "loss": 0.011, "step": 293 }, { "epoch": 2.4602510460251046, "grad_norm": 0.02121814712882042, "learning_rate": 5.977251326957852e-05, "loss": 0.0099, "step": 294 }, { "epoch": 2.4686192468619246, "grad_norm": 0.018752707168459892, "learning_rate": 5.9484402260745386e-05, "loss": 0.0112, "step": 295 }, { "epoch": 2.4686192468619246, "eval_loss": 0.014348245225846767, "eval_runtime": 8.19, "eval_samples_per_second": 6.105, "eval_steps_per_second": 1.587, "step": 295 }, { "epoch": 2.476987447698745, "grad_norm": 0.016945000737905502, "learning_rate": 5.919596421189368e-05, "loss": 0.0096, "step": 296 }, { "epoch": 2.485355648535565, "grad_norm": 0.0150624830275774, "learning_rate": 5.8907209068909554e-05, "loss": 0.0085, "step": 297 }, { "epoch": 2.493723849372385, "grad_norm": 0.01675923727452755, "learning_rate": 5.8618146788613105e-05, "loss": 0.0101, "step": 298 }, { "epoch": 2.502092050209205, "grad_norm": 0.01698216423392296, "learning_rate": 5.832878733841513e-05, "loss": 0.0102, "step": 299 }, { "epoch": 2.510460251046025, "grad_norm": 0.01541865710169077, "learning_rate": 5.803914069597342e-05, "loss": 0.01, "step": 300 }, { "epoch": 2.510460251046025, "eval_loss": 0.014326265081763268, "eval_runtime": 8.1855, "eval_samples_per_second": 6.108, "eval_steps_per_second": 1.588, "step": 300 }, { "epoch": 2.518828451882845, "grad_norm": 0.015604431740939617, "learning_rate": 5.7749216848848664e-05, "loss": 0.0092, "step": 301 }, { "epoch": 2.5271966527196654, "grad_norm": 0.019720997661352158, "learning_rate": 5.7459025794160136e-05, "loss": 0.0109, "step": 302 }, { "epoch": 2.5355648535564854, "grad_norm": 0.014376216568052769, "learning_rate": 5.716857753824086e-05, "loss": 0.0089, "step": 303 }, { "epoch": 2.5439330543933054, "grad_norm": 0.015419202856719494, "learning_rate": 5.6877882096292715e-05, "loss": 0.0095, "step": 304 }, { "epoch": 2.5523012552301254, "grad_norm": 0.018142806366086006, "learning_rate": 5.6586949492040944e-05, "loss": 0.0096, "step": 305 }, { "epoch": 2.5523012552301254, "eval_loss": 0.01382070779800415, "eval_runtime": 8.2064, "eval_samples_per_second": 6.093, "eval_steps_per_second": 1.584, "step": 305 }, { "epoch": 2.560669456066946, "grad_norm": 0.016867833212018013, "learning_rate": 5.629578975738865e-05, "loss": 0.0099, "step": 306 }, { "epoch": 2.569037656903766, "grad_norm": 0.01875397376716137, "learning_rate": 5.6004412932070835e-05, "loss": 0.0104, "step": 307 }, { "epoch": 2.577405857740586, "grad_norm": 0.018324118107557297, "learning_rate": 5.57128290633081e-05, "loss": 0.0096, "step": 308 }, { "epoch": 2.585774058577406, "grad_norm": 0.014931274577975273, "learning_rate": 5.542104820546041e-05, "loss": 0.0095, "step": 309 }, { "epoch": 2.594142259414226, "grad_norm": 0.020742323249578476, "learning_rate": 5.512908041968018e-05, "loss": 0.0096, "step": 310 }, { "epoch": 2.594142259414226, "eval_loss": 0.013671735301613808, "eval_runtime": 8.1946, "eval_samples_per_second": 6.102, "eval_steps_per_second": 1.586, "step": 310 }, { "epoch": 2.602510460251046, "grad_norm": 0.016727490350604057, "learning_rate": 5.483693577356552e-05, "loss": 0.01, "step": 311 }, { "epoch": 2.6108786610878663, "grad_norm": 0.015919320285320282, "learning_rate": 5.454462434081302e-05, "loss": 0.0099, "step": 312 }, { "epoch": 2.6192468619246863, "grad_norm": 0.01895716041326523, "learning_rate": 5.425215620087033e-05, "loss": 0.0097, "step": 313 }, { "epoch": 2.6276150627615062, "grad_norm": 0.0199336726218462, "learning_rate": 5.395954143858871e-05, "loss": 0.0108, "step": 314 }, { "epoch": 2.6359832635983262, "grad_norm": 0.016174420714378357, "learning_rate": 5.36667901438752e-05, "loss": 0.0099, "step": 315 }, { "epoch": 2.6359832635983262, "eval_loss": 0.013724744319915771, "eval_runtime": 8.2179, "eval_samples_per_second": 6.084, "eval_steps_per_second": 1.582, "step": 315 }, { "epoch": 2.6443514644351467, "grad_norm": 0.0191593486815691, "learning_rate": 5.337391241134476e-05, "loss": 0.0104, "step": 316 }, { "epoch": 2.6527196652719667, "grad_norm": 0.017469942569732666, "learning_rate": 5.3080918339972175e-05, "loss": 0.0101, "step": 317 }, { "epoch": 2.6610878661087867, "grad_norm": 0.024253906682133675, "learning_rate": 5.278781803274375e-05, "loss": 0.0109, "step": 318 }, { "epoch": 2.6694560669456067, "grad_norm": 0.020112276077270508, "learning_rate": 5.249462159630908e-05, "loss": 0.0105, "step": 319 }, { "epoch": 2.6778242677824267, "grad_norm": 0.017300646752119064, "learning_rate": 5.220133914063239e-05, "loss": 0.009, "step": 320 }, { "epoch": 2.6778242677824267, "eval_loss": 0.013836627826094627, "eval_runtime": 8.1921, "eval_samples_per_second": 6.103, "eval_steps_per_second": 1.587, "step": 320 }, { "epoch": 2.6861924686192467, "grad_norm": 0.01683647558093071, "learning_rate": 5.1907980778644114e-05, "loss": 0.0097, "step": 321 }, { "epoch": 2.694560669456067, "grad_norm": 0.02241523563861847, "learning_rate": 5.161455662589201e-05, "loss": 0.0112, "step": 322 }, { "epoch": 2.702928870292887, "grad_norm": 0.017464108765125275, "learning_rate": 5.132107680019241e-05, "loss": 0.0106, "step": 323 }, { "epoch": 2.711297071129707, "grad_norm": 0.014207069762051105, "learning_rate": 5.102755142128143e-05, "loss": 0.0092, "step": 324 }, { "epoch": 2.719665271966527, "grad_norm": 0.01582857221364975, "learning_rate": 5.073399061046583e-05, "loss": 0.0097, "step": 325 }, { "epoch": 2.719665271966527, "eval_loss": 0.01370433159172535, "eval_runtime": 8.1877, "eval_samples_per_second": 6.107, "eval_steps_per_second": 1.588, "step": 325 }, { "epoch": 2.7280334728033475, "grad_norm": 0.015251656994223595, "learning_rate": 5.044040449027426e-05, "loss": 0.0091, "step": 326 }, { "epoch": 2.7364016736401675, "grad_norm": 0.01971331425011158, "learning_rate": 5.0146803184108006e-05, "loss": 0.0097, "step": 327 }, { "epoch": 2.7447698744769875, "grad_norm": 0.017107218503952026, "learning_rate": 4.9853196815892e-05, "loss": 0.0105, "step": 328 }, { "epoch": 2.7531380753138075, "grad_norm": 0.016256961971521378, "learning_rate": 4.9559595509725744e-05, "loss": 0.0091, "step": 329 }, { "epoch": 2.7615062761506275, "grad_norm": 0.016966141760349274, "learning_rate": 4.926600938953418e-05, "loss": 0.0097, "step": 330 }, { "epoch": 2.7615062761506275, "eval_loss": 0.01360410638153553, "eval_runtime": 8.2179, "eval_samples_per_second": 6.084, "eval_steps_per_second": 1.582, "step": 330 }, { "epoch": 2.7698744769874475, "grad_norm": 0.016130153089761734, "learning_rate": 4.89724485787186e-05, "loss": 0.0094, "step": 331 }, { "epoch": 2.778242677824268, "grad_norm": 0.016650136560201645, "learning_rate": 4.86789231998076e-05, "loss": 0.0092, "step": 332 }, { "epoch": 2.786610878661088, "grad_norm": 0.015621655620634556, "learning_rate": 4.8385443374108e-05, "loss": 0.0093, "step": 333 }, { "epoch": 2.794979079497908, "grad_norm": 0.024522442370653152, "learning_rate": 4.809201922135589e-05, "loss": 0.0116, "step": 334 }, { "epoch": 2.803347280334728, "grad_norm": 0.016756724566221237, "learning_rate": 4.7798660859367615e-05, "loss": 0.0108, "step": 335 }, { "epoch": 2.803347280334728, "eval_loss": 0.013578813523054123, "eval_runtime": 8.195, "eval_samples_per_second": 6.101, "eval_steps_per_second": 1.586, "step": 335 }, { "epoch": 2.8117154811715483, "grad_norm": 0.01872294582426548, "learning_rate": 4.750537840369095e-05, "loss": 0.0097, "step": 336 }, { "epoch": 2.8200836820083683, "grad_norm": 0.01649293303489685, "learning_rate": 4.721218196725627e-05, "loss": 0.0092, "step": 337 }, { "epoch": 2.8284518828451883, "grad_norm": 0.017666228115558624, "learning_rate": 4.691908166002784e-05, "loss": 0.0114, "step": 338 }, { "epoch": 2.8368200836820083, "grad_norm": 0.014805554412305355, "learning_rate": 4.662608758865524e-05, "loss": 0.0102, "step": 339 }, { "epoch": 2.8451882845188283, "grad_norm": 0.015506643801927567, "learning_rate": 4.633320985612481e-05, "loss": 0.0092, "step": 340 }, { "epoch": 2.8451882845188283, "eval_loss": 0.01316875871270895, "eval_runtime": 8.1877, "eval_samples_per_second": 6.107, "eval_steps_per_second": 1.588, "step": 340 }, { "epoch": 2.8535564853556483, "grad_norm": 0.016277095302939415, "learning_rate": 4.60404585614113e-05, "loss": 0.0093, "step": 341 }, { "epoch": 2.8619246861924688, "grad_norm": 0.015483858995139599, "learning_rate": 4.5747843799129684e-05, "loss": 0.0101, "step": 342 }, { "epoch": 2.8702928870292888, "grad_norm": 0.014321762137115002, "learning_rate": 4.545537565918699e-05, "loss": 0.0098, "step": 343 }, { "epoch": 2.8786610878661087, "grad_norm": 0.014206680469214916, "learning_rate": 4.5163064226434475e-05, "loss": 0.009, "step": 344 }, { "epoch": 2.8870292887029287, "grad_norm": 0.013214725069701672, "learning_rate": 4.487091958031984e-05, "loss": 0.0092, "step": 345 }, { "epoch": 2.8870292887029287, "eval_loss": 0.013155767694115639, "eval_runtime": 8.1971, "eval_samples_per_second": 6.1, "eval_steps_per_second": 1.586, "step": 345 }, { "epoch": 2.895397489539749, "grad_norm": 0.01570652797818184, "learning_rate": 4.457895179453961e-05, "loss": 0.0101, "step": 346 }, { "epoch": 2.903765690376569, "grad_norm": 0.013275294564664364, "learning_rate": 4.42871709366919e-05, "loss": 0.0096, "step": 347 }, { "epoch": 2.912133891213389, "grad_norm": 0.017738599330186844, "learning_rate": 4.3995587067929177e-05, "loss": 0.0106, "step": 348 }, { "epoch": 2.920502092050209, "grad_norm": 0.02087695151567459, "learning_rate": 4.3704210242611346e-05, "loss": 0.0105, "step": 349 }, { "epoch": 2.928870292887029, "grad_norm": 0.01522638276219368, "learning_rate": 4.341305050795907e-05, "loss": 0.0095, "step": 350 }, { "epoch": 2.928870292887029, "eval_loss": 0.012959838844835758, "eval_runtime": 8.1869, "eval_samples_per_second": 6.107, "eval_steps_per_second": 1.588, "step": 350 }, { "epoch": 2.937238493723849, "grad_norm": 0.015606192871928215, "learning_rate": 4.31221179037073e-05, "loss": 0.01, "step": 351 }, { "epoch": 2.9456066945606696, "grad_norm": 0.019583778455853462, "learning_rate": 4.283142246175915e-05, "loss": 0.0096, "step": 352 }, { "epoch": 2.9539748953974896, "grad_norm": 0.01509869284927845, "learning_rate": 4.2540974205839876e-05, "loss": 0.0094, "step": 353 }, { "epoch": 2.9623430962343096, "grad_norm": 0.018710950389504433, "learning_rate": 4.2250783151151334e-05, "loss": 0.011, "step": 354 }, { "epoch": 2.9707112970711296, "grad_norm": 0.015060734003782272, "learning_rate": 4.1960859304026594e-05, "loss": 0.0094, "step": 355 }, { "epoch": 2.9707112970711296, "eval_loss": 0.012675998732447624, "eval_runtime": 8.1889, "eval_samples_per_second": 6.106, "eval_steps_per_second": 1.588, "step": 355 }, { "epoch": 2.97907949790795, "grad_norm": 0.02269388735294342, "learning_rate": 4.167121266158488e-05, "loss": 0.0102, "step": 356 }, { "epoch": 2.98744769874477, "grad_norm": 0.014178493991494179, "learning_rate": 4.1381853211386914e-05, "loss": 0.0085, "step": 357 }, { "epoch": 2.99581589958159, "grad_norm": 0.017629047855734825, "learning_rate": 4.109279093109045e-05, "loss": 0.0098, "step": 358 }, { "epoch": 3.00418410041841, "grad_norm": 0.02921808324754238, "learning_rate": 4.080403578810631e-05, "loss": 0.0133, "step": 359 }, { "epoch": 3.01255230125523, "grad_norm": 0.016827603802084923, "learning_rate": 4.051559773925462e-05, "loss": 0.0088, "step": 360 }, { "epoch": 3.01255230125523, "eval_loss": 0.012736815959215164, "eval_runtime": 8.1835, "eval_samples_per_second": 6.11, "eval_steps_per_second": 1.589, "step": 360 }, { "epoch": 3.02092050209205, "grad_norm": 0.015738021582365036, "learning_rate": 4.022748673042149e-05, "loss": 0.0093, "step": 361 }, { "epoch": 3.0292887029288704, "grad_norm": 0.013376789167523384, "learning_rate": 3.9939712696216094e-05, "loss": 0.007, "step": 362 }, { "epoch": 3.0376569037656904, "grad_norm": 0.015625160187482834, "learning_rate": 3.965228555962808e-05, "loss": 0.0082, "step": 363 }, { "epoch": 3.0460251046025104, "grad_norm": 0.01764129102230072, "learning_rate": 3.9365215231685434e-05, "loss": 0.0074, "step": 364 }, { "epoch": 3.0543933054393304, "grad_norm": 0.015254415571689606, "learning_rate": 3.907851161111269e-05, "loss": 0.0086, "step": 365 }, { "epoch": 3.0543933054393304, "eval_loss": 0.013066377490758896, "eval_runtime": 8.1923, "eval_samples_per_second": 6.103, "eval_steps_per_second": 1.587, "step": 365 }, { "epoch": 3.062761506276151, "grad_norm": 0.016880949959158897, "learning_rate": 3.879218458398962e-05, "loss": 0.0095, "step": 366 }, { "epoch": 3.071129707112971, "grad_norm": 0.012788892723619938, "learning_rate": 3.8506244023410345e-05, "loss": 0.0076, "step": 367 }, { "epoch": 3.079497907949791, "grad_norm": 0.017358217388391495, "learning_rate": 3.82206997891429e-05, "loss": 0.0073, "step": 368 }, { "epoch": 3.087866108786611, "grad_norm": 0.017496395856142044, "learning_rate": 3.7935561727289265e-05, "loss": 0.0079, "step": 369 }, { "epoch": 3.096234309623431, "grad_norm": 0.016393601894378662, "learning_rate": 3.7650839669945804e-05, "loss": 0.0094, "step": 370 }, { "epoch": 3.096234309623431, "eval_loss": 0.01338866539299488, "eval_runtime": 8.1844, "eval_samples_per_second": 6.109, "eval_steps_per_second": 1.588, "step": 370 }, { "epoch": 3.104602510460251, "grad_norm": 0.019458502531051636, "learning_rate": 3.736654343486425e-05, "loss": 0.0069, "step": 371 }, { "epoch": 3.1129707112970713, "grad_norm": 0.01921907253563404, "learning_rate": 3.708268282511321e-05, "loss": 0.0075, "step": 372 }, { "epoch": 3.1213389121338913, "grad_norm": 0.021092800423502922, "learning_rate": 3.679926762874012e-05, "loss": 0.0086, "step": 373 }, { "epoch": 3.1297071129707112, "grad_norm": 0.025811336934566498, "learning_rate": 3.65163076184337e-05, "loss": 0.0088, "step": 374 }, { "epoch": 3.1380753138075312, "grad_norm": 0.017477277666330338, "learning_rate": 3.623381255118702e-05, "loss": 0.0075, "step": 375 }, { "epoch": 3.1380753138075312, "eval_loss": 0.013682931661605835, "eval_runtime": 8.1808, "eval_samples_per_second": 6.112, "eval_steps_per_second": 1.589, "step": 375 }, { "epoch": 3.1464435146443517, "grad_norm": 0.017589591443538666, "learning_rate": 3.5951792167961e-05, "loss": 0.0075, "step": 376 }, { "epoch": 3.1548117154811717, "grad_norm": 0.020391661673784256, "learning_rate": 3.567025619334861e-05, "loss": 0.0091, "step": 377 }, { "epoch": 3.1631799163179917, "grad_norm": 0.016537083312869072, "learning_rate": 3.538921433523949e-05, "loss": 0.0062, "step": 378 }, { "epoch": 3.1715481171548117, "grad_norm": 0.01650119759142399, "learning_rate": 3.5108676284485176e-05, "loss": 0.0076, "step": 379 }, { "epoch": 3.1799163179916317, "grad_norm": 0.01610580086708069, "learning_rate": 3.482865171456505e-05, "loss": 0.0068, "step": 380 }, { "epoch": 3.1799163179916317, "eval_loss": 0.013625282794237137, "eval_runtime": 8.2399, "eval_samples_per_second": 6.068, "eval_steps_per_second": 1.578, "step": 380 }, { "epoch": 3.1882845188284517, "grad_norm": 0.0224465299397707, "learning_rate": 3.4549150281252636e-05, "loss": 0.0098, "step": 381 }, { "epoch": 3.196652719665272, "grad_norm": 0.02336002141237259, "learning_rate": 3.427018162228276e-05, "loss": 0.0079, "step": 382 }, { "epoch": 3.205020920502092, "grad_norm": 0.01989358477294445, "learning_rate": 3.3991755357019196e-05, "loss": 0.0082, "step": 383 }, { "epoch": 3.213389121338912, "grad_norm": 0.016892483457922935, "learning_rate": 3.371388108612293e-05, "loss": 0.0068, "step": 384 }, { "epoch": 3.221757322175732, "grad_norm": 0.018852969631552696, "learning_rate": 3.343656839122121e-05, "loss": 0.0096, "step": 385 }, { "epoch": 3.221757322175732, "eval_loss": 0.013560273684561253, "eval_runtime": 8.2379, "eval_samples_per_second": 6.07, "eval_steps_per_second": 1.578, "step": 385 }, { "epoch": 3.2301255230125525, "grad_norm": 0.012373587116599083, "learning_rate": 3.3159826834576994e-05, "loss": 0.0059, "step": 386 }, { "epoch": 3.2384937238493725, "grad_norm": 0.016591370105743408, "learning_rate": 3.2883665958759394e-05, "loss": 0.0082, "step": 387 }, { "epoch": 3.2468619246861925, "grad_norm": 0.01938340999186039, "learning_rate": 3.260809528631452e-05, "loss": 0.008, "step": 388 }, { "epoch": 3.2552301255230125, "grad_norm": 0.015290241688489914, "learning_rate": 3.233312431943719e-05, "loss": 0.0067, "step": 389 }, { "epoch": 3.2635983263598325, "grad_norm": 0.017185404896736145, "learning_rate": 3.205876253964321e-05, "loss": 0.0088, "step": 390 }, { "epoch": 3.2635983263598325, "eval_loss": 0.013662769459187984, "eval_runtime": 8.2288, "eval_samples_per_second": 6.076, "eval_steps_per_second": 1.58, "step": 390 }, { "epoch": 3.2719665271966525, "grad_norm": 0.019198745489120483, "learning_rate": 3.178501940744249e-05, "loss": 0.0066, "step": 391 }, { "epoch": 3.280334728033473, "grad_norm": 0.01907048560678959, "learning_rate": 3.1511904362012805e-05, "loss": 0.0086, "step": 392 }, { "epoch": 3.288702928870293, "grad_norm": 0.01648963801562786, "learning_rate": 3.123942682087433e-05, "loss": 0.0066, "step": 393 }, { "epoch": 3.297071129707113, "grad_norm": 0.015369053930044174, "learning_rate": 3.096759617956486e-05, "loss": 0.0082, "step": 394 }, { "epoch": 3.305439330543933, "grad_norm": 0.015744755044579506, "learning_rate": 3.069642181131592e-05, "loss": 0.008, "step": 395 }, { "epoch": 3.305439330543933, "eval_loss": 0.01378766167908907, "eval_runtime": 8.1991, "eval_samples_per_second": 6.098, "eval_steps_per_second": 1.586, "step": 395 }, { "epoch": 3.3138075313807533, "grad_norm": 0.028769735246896744, "learning_rate": 3.042591306672946e-05, "loss": 0.0089, "step": 396 }, { "epoch": 3.3221757322175733, "grad_norm": 0.01585283689200878, "learning_rate": 3.0156079273455483e-05, "loss": 0.0071, "step": 397 }, { "epoch": 3.3305439330543933, "grad_norm": 0.016848471015691757, "learning_rate": 2.9886929735870416e-05, "loss": 0.0079, "step": 398 }, { "epoch": 3.3389121338912133, "grad_norm": 0.015105370432138443, "learning_rate": 2.9618473734756264e-05, "loss": 0.0079, "step": 399 }, { "epoch": 3.3472803347280333, "grad_norm": 0.01963132992386818, "learning_rate": 2.935072052698059e-05, "loss": 0.0085, "step": 400 }, { "epoch": 3.3472803347280333, "eval_loss": 0.013703197240829468, "eval_runtime": 8.1849, "eval_samples_per_second": 6.109, "eval_steps_per_second": 1.588, "step": 400 }, { "epoch": 3.3556485355648533, "grad_norm": 0.015131447464227676, "learning_rate": 2.908367934517729e-05, "loss": 0.0079, "step": 401 }, { "epoch": 3.3640167364016738, "grad_norm": 0.01849834993481636, "learning_rate": 2.881735939742831e-05, "loss": 0.0077, "step": 402 }, { "epoch": 3.3723849372384938, "grad_norm": 0.01487911120057106, "learning_rate": 2.8551769866946077e-05, "loss": 0.0075, "step": 403 }, { "epoch": 3.3807531380753137, "grad_norm": 0.020853998139500618, "learning_rate": 2.8286919911756836e-05, "loss": 0.0079, "step": 404 }, { "epoch": 3.3891213389121337, "grad_norm": 0.026485485956072807, "learning_rate": 2.8022818664384944e-05, "loss": 0.0091, "step": 405 }, { "epoch": 3.3891213389121337, "eval_loss": 0.013571958057582378, "eval_runtime": 8.2051, "eval_samples_per_second": 6.094, "eval_steps_per_second": 1.584, "step": 405 }, { "epoch": 3.397489539748954, "grad_norm": 0.015749875456094742, "learning_rate": 2.77594752315378e-05, "loss": 0.007, "step": 406 }, { "epoch": 3.405857740585774, "grad_norm": 0.01729268953204155, "learning_rate": 2.7496898693792028e-05, "loss": 0.0094, "step": 407 }, { "epoch": 3.414225941422594, "grad_norm": 0.01651543565094471, "learning_rate": 2.7235098105280243e-05, "loss": 0.0079, "step": 408 }, { "epoch": 3.422594142259414, "grad_norm": 0.026226142421364784, "learning_rate": 2.697408249337886e-05, "loss": 0.0084, "step": 409 }, { "epoch": 3.430962343096234, "grad_norm": 0.01325182057917118, "learning_rate": 2.671386085839682e-05, "loss": 0.0049, "step": 410 }, { "epoch": 3.430962343096234, "eval_loss": 0.013367210514843464, "eval_runtime": 8.1901, "eval_samples_per_second": 6.105, "eval_steps_per_second": 1.587, "step": 410 }, { "epoch": 3.439330543933054, "grad_norm": 0.01933259144425392, "learning_rate": 2.6454442173265282e-05, "loss": 0.0089, "step": 411 }, { "epoch": 3.4476987447698746, "grad_norm": 0.017439622431993484, "learning_rate": 2.6195835383228127e-05, "loss": 0.0084, "step": 412 }, { "epoch": 3.4560669456066946, "grad_norm": 0.01493872795253992, "learning_rate": 2.5938049405533627e-05, "loss": 0.0071, "step": 413 }, { "epoch": 3.4644351464435146, "grad_norm": 0.01432398147881031, "learning_rate": 2.5681093129126864e-05, "loss": 0.0072, "step": 414 }, { "epoch": 3.4728033472803346, "grad_norm": 0.017259176820516586, "learning_rate": 2.542497541434329e-05, "loss": 0.0072, "step": 415 }, { "epoch": 3.4728033472803346, "eval_loss": 0.013123424723744392, "eval_runtime": 8.2011, "eval_samples_per_second": 6.097, "eval_steps_per_second": 1.585, "step": 415 }, { "epoch": 3.481171548117155, "grad_norm": 0.0195622555911541, "learning_rate": 2.5169705092603125e-05, "loss": 0.0082, "step": 416 }, { "epoch": 3.489539748953975, "grad_norm": 0.020763585343956947, "learning_rate": 2.491529096610692e-05, "loss": 0.0079, "step": 417 }, { "epoch": 3.497907949790795, "grad_norm": 0.02019021287560463, "learning_rate": 2.4661741807532006e-05, "loss": 0.0088, "step": 418 }, { "epoch": 3.506276150627615, "grad_norm": 0.017682604491710663, "learning_rate": 2.4409066359729997e-05, "loss": 0.0089, "step": 419 }, { "epoch": 3.514644351464435, "grad_norm": 0.015951137989759445, "learning_rate": 2.4157273335425297e-05, "loss": 0.0063, "step": 420 }, { "epoch": 3.514644351464435, "eval_loss": 0.013349597342312336, "eval_runtime": 8.1856, "eval_samples_per_second": 6.108, "eval_steps_per_second": 1.588, "step": 420 }, { "epoch": 3.523012552301255, "grad_norm": 0.02157527022063732, "learning_rate": 2.390637141691471e-05, "loss": 0.0078, "step": 421 }, { "epoch": 3.5313807531380754, "grad_norm": 0.015403731726109982, "learning_rate": 2.3656369255768017e-05, "loss": 0.0065, "step": 422 }, { "epoch": 3.5397489539748954, "grad_norm": 0.016804836690425873, "learning_rate": 2.340727547252969e-05, "loss": 0.0081, "step": 423 }, { "epoch": 3.5481171548117154, "grad_norm": 0.016539255157113075, "learning_rate": 2.3159098656421614e-05, "loss": 0.0092, "step": 424 }, { "epoch": 3.5564853556485354, "grad_norm": 0.016467688605189323, "learning_rate": 2.291184736504695e-05, "loss": 0.0076, "step": 425 }, { "epoch": 3.5564853556485354, "eval_loss": 0.013108824379742146, "eval_runtime": 8.1932, "eval_samples_per_second": 6.103, "eval_steps_per_second": 1.587, "step": 425 }, { "epoch": 3.564853556485356, "grad_norm": 0.017600081861019135, "learning_rate": 2.2665530124094962e-05, "loss": 0.0084, "step": 426 }, { "epoch": 3.573221757322176, "grad_norm": 0.02098635584115982, "learning_rate": 2.2420155427047167e-05, "loss": 0.0086, "step": 427 }, { "epoch": 3.581589958158996, "grad_norm": 0.017134375870227814, "learning_rate": 2.2175731734884365e-05, "loss": 0.0075, "step": 428 }, { "epoch": 3.589958158995816, "grad_norm": 0.017826305702328682, "learning_rate": 2.1932267475794917e-05, "loss": 0.0071, "step": 429 }, { "epoch": 3.598326359832636, "grad_norm": 0.016442451626062393, "learning_rate": 2.1689771044884148e-05, "loss": 0.0076, "step": 430 }, { "epoch": 3.598326359832636, "eval_loss": 0.012915154919028282, "eval_runtime": 8.1885, "eval_samples_per_second": 6.106, "eval_steps_per_second": 1.588, "step": 430 }, { "epoch": 3.606694560669456, "grad_norm": 0.026688870042562485, "learning_rate": 2.1448250803884834e-05, "loss": 0.0087, "step": 431 }, { "epoch": 3.6150627615062763, "grad_norm": 0.016501056030392647, "learning_rate": 2.120771508086889e-05, "loss": 0.0066, "step": 432 }, { "epoch": 3.6234309623430963, "grad_norm": 0.013682221062481403, "learning_rate": 2.0968172169960197e-05, "loss": 0.0067, "step": 433 }, { "epoch": 3.6317991631799162, "grad_norm": 0.018159842118620872, "learning_rate": 2.0729630331048615e-05, "loss": 0.0088, "step": 434 }, { "epoch": 3.6401673640167362, "grad_norm": 0.01990005373954773, "learning_rate": 2.0492097789505178e-05, "loss": 0.0074, "step": 435 }, { "epoch": 3.6401673640167362, "eval_loss": 0.013006238266825676, "eval_runtime": 8.2323, "eval_samples_per_second": 6.074, "eval_steps_per_second": 1.579, "step": 435 }, { "epoch": 3.6485355648535567, "grad_norm": 0.01765499822795391, "learning_rate": 2.0255582735898383e-05, "loss": 0.0063, "step": 436 }, { "epoch": 3.6569037656903767, "grad_norm": 0.019951600581407547, "learning_rate": 2.0020093325711896e-05, "loss": 0.0084, "step": 437 }, { "epoch": 3.6652719665271967, "grad_norm": 0.019513117149472237, "learning_rate": 1.9785637679063258e-05, "loss": 0.0064, "step": 438 }, { "epoch": 3.6736401673640167, "grad_norm": 0.027216019108891487, "learning_rate": 1.9552223880423882e-05, "loss": 0.0092, "step": 439 }, { "epoch": 3.6820083682008367, "grad_norm": 0.018059376627206802, "learning_rate": 1.931985997834031e-05, "loss": 0.0074, "step": 440 }, { "epoch": 3.6820083682008367, "eval_loss": 0.01317664049565792, "eval_runtime": 8.1934, "eval_samples_per_second": 6.102, "eval_steps_per_second": 1.587, "step": 440 }, { "epoch": 3.6903765690376567, "grad_norm": 0.020217331126332283, "learning_rate": 1.908855398515669e-05, "loss": 0.0079, "step": 441 }, { "epoch": 3.698744769874477, "grad_norm": 0.019336042925715446, "learning_rate": 1.8858313876738458e-05, "loss": 0.0069, "step": 442 }, { "epoch": 3.707112970711297, "grad_norm": 0.020630570128560066, "learning_rate": 1.8629147592197372e-05, "loss": 0.0083, "step": 443 }, { "epoch": 3.715481171548117, "grad_norm": 0.01856216788291931, "learning_rate": 1.840106303361771e-05, "loss": 0.0075, "step": 444 }, { "epoch": 3.723849372384937, "grad_norm": 0.01747502014040947, "learning_rate": 1.8174068065783765e-05, "loss": 0.0067, "step": 445 }, { "epoch": 3.723849372384937, "eval_loss": 0.013197257183492184, "eval_runtime": 8.2112, "eval_samples_per_second": 6.089, "eval_steps_per_second": 1.583, "step": 445 }, { "epoch": 3.7322175732217575, "grad_norm": 0.018087679520249367, "learning_rate": 1.7948170515908757e-05, "loss": 0.0077, "step": 446 }, { "epoch": 3.7405857740585775, "grad_norm": 0.01658761128783226, "learning_rate": 1.7723378173364836e-05, "loss": 0.0081, "step": 447 }, { "epoch": 3.7489539748953975, "grad_norm": 0.017800970003008842, "learning_rate": 1.7499698789414536e-05, "loss": 0.0076, "step": 448 }, { "epoch": 3.7573221757322175, "grad_norm": 0.018941445276141167, "learning_rate": 1.727714007694348e-05, "loss": 0.0092, "step": 449 }, { "epoch": 3.7656903765690375, "grad_norm": 0.014227015897631645, "learning_rate": 1.705570971019445e-05, "loss": 0.0064, "step": 450 }, { "epoch": 3.7656903765690375, "eval_loss": 0.013023738749325275, "eval_runtime": 8.1848, "eval_samples_per_second": 6.109, "eval_steps_per_second": 1.588, "step": 450 }, { "epoch": 3.7740585774058575, "grad_norm": 0.01811639405786991, "learning_rate": 1.683541532450272e-05, "loss": 0.0077, "step": 451 }, { "epoch": 3.782426778242678, "grad_norm": 0.01967625319957733, "learning_rate": 1.6616264516032807e-05, "loss": 0.0083, "step": 452 }, { "epoch": 3.790794979079498, "grad_norm": 0.0179898664355278, "learning_rate": 1.6398264841516564e-05, "loss": 0.0076, "step": 453 }, { "epoch": 3.799163179916318, "grad_norm": 0.015487701632082462, "learning_rate": 1.618142381799256e-05, "loss": 0.0069, "step": 454 }, { "epoch": 3.8075313807531384, "grad_norm": 0.017041724175214767, "learning_rate": 1.5965748922546876e-05, "loss": 0.0091, "step": 455 }, { "epoch": 3.8075313807531384, "eval_loss": 0.013014528900384903, "eval_runtime": 8.1813, "eval_samples_per_second": 6.112, "eval_steps_per_second": 1.589, "step": 455 }, { "epoch": 3.8158995815899583, "grad_norm": 0.01787850260734558, "learning_rate": 1.5751247592055333e-05, "loss": 0.0069, "step": 456 }, { "epoch": 3.8242677824267783, "grad_norm": 0.01862199790775776, "learning_rate": 1.5537927222927047e-05, "loss": 0.0086, "step": 457 }, { "epoch": 3.8326359832635983, "grad_norm": 0.016767242923378944, "learning_rate": 1.5325795170849315e-05, "loss": 0.0068, "step": 458 }, { "epoch": 3.8410041841004183, "grad_norm": 0.01677015796303749, "learning_rate": 1.5114858750534067e-05, "loss": 0.0078, "step": 459 }, { "epoch": 3.8493723849372383, "grad_norm": 0.017780860885977745, "learning_rate": 1.4905125235465589e-05, "loss": 0.0074, "step": 460 }, { "epoch": 3.8493723849372383, "eval_loss": 0.013112700544297695, "eval_runtime": 8.184, "eval_samples_per_second": 6.109, "eval_steps_per_second": 1.588, "step": 460 }, { "epoch": 3.8577405857740583, "grad_norm": 0.02164328470826149, "learning_rate": 1.4696601857649738e-05, "loss": 0.0075, "step": 461 }, { "epoch": 3.8661087866108788, "grad_norm": 0.01868240348994732, "learning_rate": 1.4489295807364556e-05, "loss": 0.0082, "step": 462 }, { "epoch": 3.8744769874476988, "grad_norm": 0.018218357115983963, "learning_rate": 1.4283214232912345e-05, "loss": 0.0073, "step": 463 }, { "epoch": 3.8828451882845187, "grad_norm": 0.02058660052716732, "learning_rate": 1.4078364240373192e-05, "loss": 0.0063, "step": 464 }, { "epoch": 3.891213389121339, "grad_norm": 0.03692828491330147, "learning_rate": 1.38747528933599e-05, "loss": 0.0076, "step": 465 }, { "epoch": 3.891213389121339, "eval_loss": 0.01321639958769083, "eval_runtime": 8.1882, "eval_samples_per_second": 6.106, "eval_steps_per_second": 1.588, "step": 465 }, { "epoch": 3.899581589958159, "grad_norm": 0.015658119693398476, "learning_rate": 1.3672387212774457e-05, "loss": 0.0093, "step": 466 }, { "epoch": 3.907949790794979, "grad_norm": 0.014811305329203606, "learning_rate": 1.3471274176565935e-05, "loss": 0.0068, "step": 467 }, { "epoch": 3.916317991631799, "grad_norm": 0.01890689507126808, "learning_rate": 1.327142071948988e-05, "loss": 0.0076, "step": 468 }, { "epoch": 3.924686192468619, "grad_norm": 0.016337791457772255, "learning_rate": 1.3072833732869167e-05, "loss": 0.0079, "step": 469 }, { "epoch": 3.933054393305439, "grad_norm": 0.02025439403951168, "learning_rate": 1.28755200643564e-05, "loss": 0.007, "step": 470 }, { "epoch": 3.933054393305439, "eval_loss": 0.013240823522210121, "eval_runtime": 8.1806, "eval_samples_per_second": 6.112, "eval_steps_per_second": 1.589, "step": 470 }, { "epoch": 3.941422594142259, "grad_norm": 0.01858791708946228, "learning_rate": 1.267948651769777e-05, "loss": 0.0082, "step": 471 }, { "epoch": 3.9497907949790796, "grad_norm": 0.015129518695175648, "learning_rate": 1.2484739852498467e-05, "loss": 0.0067, "step": 472 }, { "epoch": 3.9581589958158996, "grad_norm": 0.01916913501918316, "learning_rate": 1.2291286783989597e-05, "loss": 0.0083, "step": 473 }, { "epoch": 3.9665271966527196, "grad_norm": 0.020614786073565483, "learning_rate": 1.2099133982796612e-05, "loss": 0.0083, "step": 474 }, { "epoch": 3.97489539748954, "grad_norm": 0.016557354480028152, "learning_rate": 1.190828807470929e-05, "loss": 0.0082, "step": 475 }, { "epoch": 3.97489539748954, "eval_loss": 0.013233880512416363, "eval_runtime": 8.1955, "eval_samples_per_second": 6.101, "eval_steps_per_second": 1.586, "step": 475 }, { "epoch": 3.98326359832636, "grad_norm": 0.01762431487441063, "learning_rate": 1.1718755640453288e-05, "loss": 0.0059, "step": 476 }, { "epoch": 3.99163179916318, "grad_norm": 0.01950531266629696, "learning_rate": 1.1530543215463235e-05, "loss": 0.0085, "step": 477 }, { "epoch": 4.0, "grad_norm": 0.02553708292543888, "learning_rate": 1.1343657289657333e-05, "loss": 0.0115, "step": 478 }, { "epoch": 4.00836820083682, "grad_norm": 0.011678546667098999, "learning_rate": 1.1158104307213612e-05, "loss": 0.0066, "step": 479 }, { "epoch": 4.01673640167364, "grad_norm": 0.016237011179327965, "learning_rate": 1.0973890666347702e-05, "loss": 0.0059, "step": 480 }, { "epoch": 4.01673640167364, "eval_loss": 0.013261526823043823, "eval_runtime": 8.1877, "eval_samples_per_second": 6.107, "eval_steps_per_second": 1.588, "step": 480 }, { "epoch": 4.02510460251046, "grad_norm": 0.015487432479858398, "learning_rate": 1.0791022719092231e-05, "loss": 0.0064, "step": 481 }, { "epoch": 4.03347280334728, "grad_norm": 0.015571733005344868, "learning_rate": 1.0609506771077765e-05, "loss": 0.0067, "step": 482 }, { "epoch": 4.0418410041841, "grad_norm": 0.013560911640524864, "learning_rate": 1.0429349081315404e-05, "loss": 0.0058, "step": 483 }, { "epoch": 4.050209205020921, "grad_norm": 0.014713162556290627, "learning_rate": 1.025055586198096e-05, "loss": 0.0064, "step": 484 }, { "epoch": 4.058577405857741, "grad_norm": 0.011582295410335064, "learning_rate": 1.0073133278200703e-05, "loss": 0.0066, "step": 485 }, { "epoch": 4.058577405857741, "eval_loss": 0.013517594896256924, "eval_runtime": 8.1969, "eval_samples_per_second": 6.1, "eval_steps_per_second": 1.586, "step": 485 }, { "epoch": 4.066945606694561, "grad_norm": 0.016617678105831146, "learning_rate": 9.897087447838848e-06, "loss": 0.0068, "step": 486 }, { "epoch": 4.075313807531381, "grad_norm": 0.013949776999652386, "learning_rate": 9.72242444128656e-06, "loss": 0.0068, "step": 487 }, { "epoch": 4.083682008368201, "grad_norm": 0.0159038994461298, "learning_rate": 9.549150281252633e-06, "loss": 0.0059, "step": 488 }, { "epoch": 4.092050209205021, "grad_norm": 0.023309912532567978, "learning_rate": 9.377270942555816e-06, "loss": 0.0065, "step": 489 }, { "epoch": 4.100418410041841, "grad_norm": 0.014613127335906029, "learning_rate": 9.206792351918808e-06, "loss": 0.0063, "step": 490 }, { "epoch": 4.100418410041841, "eval_loss": 0.013989124447107315, "eval_runtime": 8.2202, "eval_samples_per_second": 6.083, "eval_steps_per_second": 1.581, "step": 490 }, { "epoch": 4.108786610878661, "grad_norm": 0.03452659025788307, "learning_rate": 9.037720387763877e-06, "loss": 0.0062, "step": 491 }, { "epoch": 4.117154811715481, "grad_norm": 0.014380170032382011, "learning_rate": 8.870060880010161e-06, "loss": 0.0072, "step": 492 }, { "epoch": 4.125523012552302, "grad_norm": 0.014284541830420494, "learning_rate": 8.703819609872655e-06, "loss": 0.0065, "step": 493 }, { "epoch": 4.133891213389122, "grad_norm": 0.028412694111466408, "learning_rate": 8.539002309662863e-06, "loss": 0.0068, "step": 494 }, { "epoch": 4.142259414225942, "grad_norm": 0.01495958399027586, "learning_rate": 8.375614662591098e-06, "loss": 0.0059, "step": 495 }, { "epoch": 4.142259414225942, "eval_loss": 0.01436224952340126, "eval_runtime": 8.1984, "eval_samples_per_second": 6.099, "eval_steps_per_second": 1.586, "step": 495 }, { "epoch": 4.150627615062762, "grad_norm": 0.01771622896194458, "learning_rate": 8.213662302570564e-06, "loss": 0.0072, "step": 496 }, { "epoch": 4.158995815899582, "grad_norm": 0.02444896660745144, "learning_rate": 8.053150814023075e-06, "loss": 0.006, "step": 497 }, { "epoch": 4.167364016736402, "grad_norm": 0.016917061060667038, "learning_rate": 7.894085731686484e-06, "loss": 0.0068, "step": 498 }, { "epoch": 4.175732217573222, "grad_norm": 0.020339468494057655, "learning_rate": 7.736472540423839e-06, "loss": 0.0062, "step": 499 }, { "epoch": 4.184100418410042, "grad_norm": 0.01869555190205574, "learning_rate": 7.5803166750342545e-06, "loss": 0.0066, "step": 500 }, { "epoch": 4.184100418410042, "eval_loss": 0.014202076941728592, "eval_runtime": 8.1932, "eval_samples_per_second": 6.103, "eval_steps_per_second": 1.587, "step": 500 }, { "epoch": 4.192468619246862, "grad_norm": 0.017789356410503387, "learning_rate": 7.425623520065506e-06, "loss": 0.0068, "step": 501 }, { "epoch": 4.200836820083682, "grad_norm": 0.018326247110962868, "learning_rate": 7.272398409628362e-06, "loss": 0.0058, "step": 502 }, { "epoch": 4.209205020920502, "grad_norm": 0.01593657210469246, "learning_rate": 7.120646627212668e-06, "loss": 0.0065, "step": 503 }, { "epoch": 4.2175732217573225, "grad_norm": 0.01636459119617939, "learning_rate": 6.970373405505149e-06, "loss": 0.0059, "step": 504 }, { "epoch": 4.2259414225941425, "grad_norm": 0.014171932823956013, "learning_rate": 6.8215839262089465e-06, "loss": 0.0055, "step": 505 }, { "epoch": 4.2259414225941425, "eval_loss": 0.01415976881980896, "eval_runtime": 8.2304, "eval_samples_per_second": 6.075, "eval_steps_per_second": 1.58, "step": 505 }, { "epoch": 4.2343096234309625, "grad_norm": 0.013969158753752708, "learning_rate": 6.674283319865015e-06, "loss": 0.0065, "step": 506 }, { "epoch": 4.2426778242677825, "grad_norm": 0.020747659727931023, "learning_rate": 6.5284766656751486e-06, "loss": 0.0067, "step": 507 }, { "epoch": 4.2510460251046025, "grad_norm": 0.015698375180363655, "learning_rate": 6.384168991326872e-06, "loss": 0.0058, "step": 508 }, { "epoch": 4.2594142259414225, "grad_norm": 0.018806830048561096, "learning_rate": 6.241365272820065e-06, "loss": 0.0065, "step": 509 }, { "epoch": 4.2677824267782425, "grad_norm": 0.01827147603034973, "learning_rate": 6.100070434295379e-06, "loss": 0.0067, "step": 510 }, { "epoch": 4.2677824267782425, "eval_loss": 0.014154573902487755, "eval_runtime": 8.2041, "eval_samples_per_second": 6.095, "eval_steps_per_second": 1.585, "step": 510 }, { "epoch": 4.2761506276150625, "grad_norm": 0.016821326687932014, "learning_rate": 5.96028934786445e-06, "loss": 0.0057, "step": 511 }, { "epoch": 4.2845188284518825, "grad_norm": 0.01477895863354206, "learning_rate": 5.822026833441901e-06, "loss": 0.0061, "step": 512 }, { "epoch": 4.292887029288703, "grad_norm": 0.02095729485154152, "learning_rate": 5.685287658579124e-06, "loss": 0.0067, "step": 513 }, { "epoch": 4.301255230125523, "grad_norm": 0.014905648306012154, "learning_rate": 5.550076538299931e-06, "loss": 0.0064, "step": 514 }, { "epoch": 4.309623430962343, "grad_norm": 0.01747983880341053, "learning_rate": 5.416398134937878e-06, "loss": 0.0065, "step": 515 }, { "epoch": 4.309623430962343, "eval_loss": 0.014257834292948246, "eval_runtime": 8.1812, "eval_samples_per_second": 6.112, "eval_steps_per_second": 1.589, "step": 515 }, { "epoch": 4.317991631799163, "grad_norm": 0.01598619669675827, "learning_rate": 5.284257057975628e-06, "loss": 0.0063, "step": 516 }, { "epoch": 4.326359832635983, "grad_norm": 0.016550792381167412, "learning_rate": 5.153657863885902e-06, "loss": 0.0056, "step": 517 }, { "epoch": 4.334728033472803, "grad_norm": 0.02542988397181034, "learning_rate": 5.024605055974408e-06, "loss": 0.0065, "step": 518 }, { "epoch": 4.343096234309623, "grad_norm": 0.01975710317492485, "learning_rate": 4.8971030842245635e-06, "loss": 0.0066, "step": 519 }, { "epoch": 4.351464435146443, "grad_norm": 0.01818425953388214, "learning_rate": 4.771156345144018e-06, "loss": 0.0062, "step": 520 }, { "epoch": 4.351464435146443, "eval_loss": 0.014165216125547886, "eval_runtime": 8.1986, "eval_samples_per_second": 6.099, "eval_steps_per_second": 1.586, "step": 520 }, { "epoch": 4.359832635983263, "grad_norm": 0.01841098628938198, "learning_rate": 4.646769181613098e-06, "loss": 0.0063, "step": 521 }, { "epoch": 4.368200836820083, "grad_norm": 0.01820644736289978, "learning_rate": 4.52394588273502e-06, "loss": 0.0065, "step": 522 }, { "epoch": 4.376569037656903, "grad_norm": 0.01679069548845291, "learning_rate": 4.40269068368801e-06, "loss": 0.0065, "step": 523 }, { "epoch": 4.384937238493724, "grad_norm": 0.017241882160305977, "learning_rate": 4.283007765579267e-06, "loss": 0.0062, "step": 524 }, { "epoch": 4.393305439330544, "grad_norm": 0.02121458761394024, "learning_rate": 4.164901255300779e-06, "loss": 0.0065, "step": 525 }, { "epoch": 4.393305439330544, "eval_loss": 0.01406473945826292, "eval_runtime": 8.2004, "eval_samples_per_second": 6.097, "eval_steps_per_second": 1.585, "step": 525 }, { "epoch": 4.401673640167364, "grad_norm": 0.01692904345691204, "learning_rate": 4.048375225387036e-06, "loss": 0.0057, "step": 526 }, { "epoch": 4.410041841004184, "grad_norm": 0.017979208379983902, "learning_rate": 3.9334336938746e-06, "loss": 0.0059, "step": 527 }, { "epoch": 4.418410041841004, "grad_norm": 0.015546801500022411, "learning_rate": 3.82008062416353e-06, "loss": 0.0064, "step": 528 }, { "epoch": 4.426778242677824, "grad_norm": 0.014850549399852753, "learning_rate": 3.7083199248807656e-06, "loss": 0.0058, "step": 529 }, { "epoch": 4.435146443514644, "grad_norm": 0.02002605050802231, "learning_rate": 3.5981554497452884e-06, "loss": 0.007, "step": 530 }, { "epoch": 4.435146443514644, "eval_loss": 0.013939818367362022, "eval_runtime": 8.1895, "eval_samples_per_second": 6.105, "eval_steps_per_second": 1.587, "step": 530 }, { "epoch": 4.443514644351464, "grad_norm": 0.015825215727090836, "learning_rate": 3.4895909974352935e-06, "loss": 0.0066, "step": 531 }, { "epoch": 4.451882845188284, "grad_norm": 0.017673788592219353, "learning_rate": 3.38263031145718e-06, "loss": 0.007, "step": 532 }, { "epoch": 4.460251046025105, "grad_norm": 0.014746863394975662, "learning_rate": 3.27727708001645e-06, "loss": 0.0055, "step": 533 }, { "epoch": 4.468619246861925, "grad_norm": 0.017045453190803528, "learning_rate": 3.173534935890582e-06, "loss": 0.0058, "step": 534 }, { "epoch": 4.476987447698745, "grad_norm": 0.0173475444316864, "learning_rate": 3.0714074563037044e-06, "loss": 0.0058, "step": 535 }, { "epoch": 4.476987447698745, "eval_loss": 0.013857071287930012, "eval_runtime": 8.227, "eval_samples_per_second": 6.078, "eval_steps_per_second": 1.58, "step": 535 }, { "epoch": 4.485355648535565, "grad_norm": 0.014112849719822407, "learning_rate": 2.970898162803287e-06, "loss": 0.0062, "step": 536 }, { "epoch": 4.493723849372385, "grad_norm": 0.018640313297510147, "learning_rate": 2.8720105211387183e-06, "loss": 0.006, "step": 537 }, { "epoch": 4.502092050209205, "grad_norm": 0.018944760784506798, "learning_rate": 2.774747941141764e-06, "loss": 0.0065, "step": 538 }, { "epoch": 4.510460251046025, "grad_norm": 0.017405930906534195, "learning_rate": 2.6791137766090157e-06, "loss": 0.0068, "step": 539 }, { "epoch": 4.518828451882845, "grad_norm": 0.019485093653202057, "learning_rate": 2.5851113251862345e-06, "loss": 0.0056, "step": 540 }, { "epoch": 4.518828451882845, "eval_loss": 0.01391106192022562, "eval_runtime": 8.1885, "eval_samples_per_second": 6.106, "eval_steps_per_second": 1.588, "step": 540 }, { "epoch": 4.527196652719665, "grad_norm": 0.01877741701900959, "learning_rate": 2.4927438282546567e-06, "loss": 0.0068, "step": 541 }, { "epoch": 4.535564853556485, "grad_norm": 0.0198155976831913, "learning_rate": 2.4020144708192026e-06, "loss": 0.006, "step": 542 }, { "epoch": 4.543933054393305, "grad_norm": 0.01476176455616951, "learning_rate": 2.3129263813986724e-06, "loss": 0.0062, "step": 543 }, { "epoch": 4.552301255230126, "grad_norm": 0.01914817839860916, "learning_rate": 2.225482631917847e-06, "loss": 0.0067, "step": 544 }, { "epoch": 4.560669456066946, "grad_norm": 0.0210384763777256, "learning_rate": 2.1396862376015903e-06, "loss": 0.0062, "step": 545 }, { "epoch": 4.560669456066946, "eval_loss": 0.013888537883758545, "eval_runtime": 8.1978, "eval_samples_per_second": 6.099, "eval_steps_per_second": 1.586, "step": 545 }, { "epoch": 4.569037656903766, "grad_norm": 0.021067198365926743, "learning_rate": 2.0555401568708554e-06, "loss": 0.0065, "step": 546 }, { "epoch": 4.577405857740586, "grad_norm": 0.017656171694397926, "learning_rate": 1.9730472912406816e-06, "loss": 0.0071, "step": 547 }, { "epoch": 4.585774058577406, "grad_norm": 0.017184708267450333, "learning_rate": 1.8922104852201528e-06, "loss": 0.0064, "step": 548 }, { "epoch": 4.594142259414226, "grad_norm": 0.02077127993106842, "learning_rate": 1.8130325262142755e-06, "loss": 0.0066, "step": 549 }, { "epoch": 4.602510460251046, "grad_norm": 0.013661603443324566, "learning_rate": 1.7355161444279344e-06, "loss": 0.0061, "step": 550 }, { "epoch": 4.602510460251046, "eval_loss": 0.013935999944806099, "eval_runtime": 8.2068, "eval_samples_per_second": 6.093, "eval_steps_per_second": 1.584, "step": 550 }, { "epoch": 4.610878661087866, "grad_norm": 0.033053044229745865, "learning_rate": 1.6596640127716778e-06, "loss": 0.0074, "step": 551 }, { "epoch": 4.619246861924686, "grad_norm": 0.019544832408428192, "learning_rate": 1.5854787467695975e-06, "loss": 0.0066, "step": 552 }, { "epoch": 4.627615062761507, "grad_norm": 0.017127815634012222, "learning_rate": 1.512962904469112e-06, "loss": 0.0057, "step": 553 }, { "epoch": 4.635983263598327, "grad_norm": 0.024544665589928627, "learning_rate": 1.442118986352775e-06, "loss": 0.0064, "step": 554 }, { "epoch": 4.644351464435147, "grad_norm": 0.016828225925564766, "learning_rate": 1.3729494352520578e-06, "loss": 0.0061, "step": 555 }, { "epoch": 4.644351464435147, "eval_loss": 0.013867921195924282, "eval_runtime": 8.1969, "eval_samples_per_second": 6.1, "eval_steps_per_second": 1.586, "step": 555 }, { "epoch": 4.652719665271967, "grad_norm": 0.016187148168683052, "learning_rate": 1.3054566362631082e-06, "loss": 0.0068, "step": 556 }, { "epoch": 4.661087866108787, "grad_norm": 0.018801085650920868, "learning_rate": 1.2396429166645073e-06, "loss": 0.0064, "step": 557 }, { "epoch": 4.669456066945607, "grad_norm": 0.01601874642074108, "learning_rate": 1.175510545837022e-06, "loss": 0.0062, "step": 558 }, { "epoch": 4.677824267782427, "grad_norm": 0.02034141682088375, "learning_rate": 1.1130617351853446e-06, "loss": 0.0057, "step": 559 }, { "epoch": 4.686192468619247, "grad_norm": 0.017557624727487564, "learning_rate": 1.0522986380618605e-06, "loss": 0.0068, "step": 560 }, { "epoch": 4.686192468619247, "eval_loss": 0.01384110189974308, "eval_runtime": 8.1896, "eval_samples_per_second": 6.105, "eval_steps_per_second": 1.587, "step": 560 }, { "epoch": 4.694560669456067, "grad_norm": 0.015541495755314827, "learning_rate": 9.932233496923903e-07, "loss": 0.0063, "step": 561 }, { "epoch": 4.702928870292887, "grad_norm": 0.01691366545855999, "learning_rate": 9.358379071039147e-07, "loss": 0.0069, "step": 562 }, { "epoch": 4.711297071129707, "grad_norm": 0.017451630905270576, "learning_rate": 8.801442890543743e-07, "loss": 0.0057, "step": 563 }, { "epoch": 4.7196652719665275, "grad_norm": 0.013958767987787724, "learning_rate": 8.261444159644083e-07, "loss": 0.0059, "step": 564 }, { "epoch": 4.7280334728033475, "grad_norm": 0.015628065913915634, "learning_rate": 7.738401498511405e-07, "loss": 0.0069, "step": 565 }, { "epoch": 4.7280334728033475, "eval_loss": 0.013867964036762714, "eval_runtime": 8.201, "eval_samples_per_second": 6.097, "eval_steps_per_second": 1.585, "step": 565 }, { "epoch": 4.7364016736401675, "grad_norm": 0.017734922468662262, "learning_rate": 7.232332942639974e-07, "loss": 0.0066, "step": 566 }, { "epoch": 4.7447698744769875, "grad_norm": 0.01975751295685768, "learning_rate": 6.743255942224913e-07, "loss": 0.0063, "step": 567 }, { "epoch": 4.7531380753138075, "grad_norm": 0.01823136769235134, "learning_rate": 6.27118736156046e-07, "loss": 0.0063, "step": 568 }, { "epoch": 4.7615062761506275, "grad_norm": 0.01851455122232437, "learning_rate": 5.816143478458714e-07, "loss": 0.0059, "step": 569 }, { "epoch": 4.7698744769874475, "grad_norm": 0.019744612276554108, "learning_rate": 5.378139983688135e-07, "loss": 0.0063, "step": 570 }, { "epoch": 4.7698744769874475, "eval_loss": 0.013921505771577358, "eval_runtime": 8.2248, "eval_samples_per_second": 6.079, "eval_steps_per_second": 1.581, "step": 570 }, { "epoch": 4.7782426778242675, "grad_norm": 0.017613310366868973, "learning_rate": 4.957191980432541e-07, "loss": 0.0061, "step": 571 }, { "epoch": 4.786610878661088, "grad_norm": 0.01842423528432846, "learning_rate": 4.5533139837704574e-07, "loss": 0.0065, "step": 572 }, { "epoch": 4.794979079497908, "grad_norm": 0.015000440180301666, "learning_rate": 4.1665199201744167e-07, "loss": 0.0065, "step": 573 }, { "epoch": 4.803347280334728, "grad_norm": 0.01420409232378006, "learning_rate": 3.7968231270307244e-07, "loss": 0.006, "step": 574 }, { "epoch": 4.811715481171548, "grad_norm": 0.014016808941960335, "learning_rate": 3.4442363521798304e-07, "loss": 0.0065, "step": 575 }, { "epoch": 4.811715481171548, "eval_loss": 0.01387822162359953, "eval_runtime": 8.1941, "eval_samples_per_second": 6.102, "eval_steps_per_second": 1.587, "step": 575 }, { "epoch": 4.820083682008368, "grad_norm": 0.025155773386359215, "learning_rate": 3.1087717534764584e-07, "loss": 0.0065, "step": 576 }, { "epoch": 4.828451882845188, "grad_norm": 0.017688661813735962, "learning_rate": 2.7904408983706607e-07, "loss": 0.0062, "step": 577 }, { "epoch": 4.836820083682008, "grad_norm": 0.014572778716683388, "learning_rate": 2.489254763508642e-07, "loss": 0.0063, "step": 578 }, { "epoch": 4.845188284518828, "grad_norm": 0.015580548904836178, "learning_rate": 2.2052237343545e-07, "loss": 0.0058, "step": 579 }, { "epoch": 4.853556485355648, "grad_norm": 0.01487270649522543, "learning_rate": 1.938357604832075e-07, "loss": 0.0064, "step": 580 }, { "epoch": 4.853556485355648, "eval_loss": 0.013862421736121178, "eval_runtime": 8.192, "eval_samples_per_second": 6.104, "eval_steps_per_second": 1.587, "step": 580 }, { "epoch": 4.861924686192468, "grad_norm": 0.020253285765647888, "learning_rate": 1.6886655769871029e-07, "loss": 0.0058, "step": 581 }, { "epoch": 4.870292887029288, "grad_norm": 0.01656149886548519, "learning_rate": 1.4561562606701385e-07, "loss": 0.0062, "step": 582 }, { "epoch": 4.878661087866108, "grad_norm": 0.016599709168076515, "learning_rate": 1.240837673239459e-07, "loss": 0.0065, "step": 583 }, { "epoch": 4.887029288702929, "grad_norm": 0.01712622120976448, "learning_rate": 1.0427172392847295e-07, "loss": 0.0062, "step": 584 }, { "epoch": 4.895397489539749, "grad_norm": 0.01667969673871994, "learning_rate": 8.618017903708197e-08, "loss": 0.0062, "step": 585 }, { "epoch": 4.895397489539749, "eval_loss": 0.013880819082260132, "eval_runtime": 8.2034, "eval_samples_per_second": 6.095, "eval_steps_per_second": 1.585, "step": 585 }, { "epoch": 4.903765690376569, "grad_norm": 0.019158054143190384, "learning_rate": 6.980975648026578e-08, "loss": 0.0061, "step": 586 }, { "epoch": 4.912133891213389, "grad_norm": 0.013462690636515617, "learning_rate": 5.516102074095697e-08, "loss": 0.0062, "step": 587 }, { "epoch": 4.920502092050209, "grad_norm": 0.014567787759006023, "learning_rate": 4.2234476935099074e-08, "loss": 0.0062, "step": 588 }, { "epoch": 4.928870292887029, "grad_norm": 0.014899435453116894, "learning_rate": 3.103057079422711e-08, "loss": 0.0057, "step": 589 }, { "epoch": 4.937238493723849, "grad_norm": 0.013964378274977207, "learning_rate": 2.154968865007989e-08, "loss": 0.0065, "step": 590 }, { "epoch": 4.937238493723849, "eval_loss": 0.013855883851647377, "eval_runtime": 8.2021, "eval_samples_per_second": 6.096, "eval_steps_per_second": 1.585, "step": 590 }, { "epoch": 4.945606694560669, "grad_norm": 0.013118981383740902, "learning_rate": 1.3792157421299579e-08, "loss": 0.006, "step": 591 }, { "epoch": 4.95397489539749, "grad_norm": 0.01663290336728096, "learning_rate": 7.758244602151799e-09, "loss": 0.0057, "step": 592 }, { "epoch": 4.96234309623431, "grad_norm": 0.01918199472129345, "learning_rate": 3.448158253277489e-09, "loss": 0.0059, "step": 593 }, { "epoch": 4.97071129707113, "grad_norm": 0.016706887632608414, "learning_rate": 8.620469945708198e-10, "loss": 0.0062, "step": 594 }, { "epoch": 4.97907949790795, "grad_norm": 0.016998644918203354, "learning_rate": 0.0, "loss": 0.0055, "step": 595 }, { "epoch": 4.97907949790795, "eval_loss": 0.01388928946107626, "eval_runtime": 8.1966, "eval_samples_per_second": 6.1, "eval_steps_per_second": 1.586, "step": 595 }, { "epoch": 4.97907949790795, "step": 595, "total_flos": 1.199536935920124e+18, "train_loss": 0.012196063881880846, "train_runtime": 9595.7495, "train_samples_per_second": 1.991, "train_steps_per_second": 0.062 } ], "logging_steps": 1, "max_steps": 595, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 50, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.199536935920124e+18, "train_batch_size": 4, "trial_name": null, "trial_params": null }