{ "best_metric": null, "best_model_checkpoint": null, "epoch": 3.0, "eval_steps": 500, "global_step": 375, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.008, "grad_norm": 14.56950855255127, "learning_rate": 2.6315789473684213e-07, "loss": 0.8339, "step": 1 }, { "epoch": 0.016, "grad_norm": 13.73933219909668, "learning_rate": 5.263157894736843e-07, "loss": 0.8732, "step": 2 }, { "epoch": 0.024, "grad_norm": 14.081198692321777, "learning_rate": 7.894736842105263e-07, "loss": 0.894, "step": 3 }, { "epoch": 0.032, "grad_norm": 14.026392936706543, "learning_rate": 1.0526315789473685e-06, "loss": 0.8509, "step": 4 }, { "epoch": 0.04, "grad_norm": 14.380779266357422, "learning_rate": 1.3157894736842106e-06, "loss": 0.8362, "step": 5 }, { "epoch": 0.048, "grad_norm": 12.984580993652344, "learning_rate": 1.5789473684210526e-06, "loss": 0.8653, "step": 6 }, { "epoch": 0.056, "grad_norm": 10.433991432189941, "learning_rate": 1.8421052631578948e-06, "loss": 0.8321, "step": 7 }, { "epoch": 0.064, "grad_norm": 9.493083000183105, "learning_rate": 2.105263157894737e-06, "loss": 0.8143, "step": 8 }, { "epoch": 0.072, "grad_norm": 7.640753746032715, "learning_rate": 2.368421052631579e-06, "loss": 0.7905, "step": 9 }, { "epoch": 0.08, "grad_norm": 4.019260883331299, "learning_rate": 2.631578947368421e-06, "loss": 0.7129, "step": 10 }, { "epoch": 0.088, "grad_norm": 3.2670819759368896, "learning_rate": 2.8947368421052634e-06, "loss": 0.7227, "step": 11 }, { "epoch": 0.096, "grad_norm": 2.992112398147583, "learning_rate": 3.157894736842105e-06, "loss": 0.6972, "step": 12 }, { "epoch": 0.104, "grad_norm": 2.052677869796753, "learning_rate": 3.421052631578948e-06, "loss": 0.678, "step": 13 }, { "epoch": 0.112, "grad_norm": 1.920199990272522, "learning_rate": 3.6842105263157896e-06, "loss": 0.667, "step": 14 }, { "epoch": 0.12, "grad_norm": 1.706879734992981, "learning_rate": 3.947368421052632e-06, "loss": 0.6192, "step": 15 }, { "epoch": 0.128, "grad_norm": 1.6659389734268188, "learning_rate": 4.210526315789474e-06, "loss": 0.6118, "step": 16 }, { "epoch": 0.136, "grad_norm": 1.3653850555419922, "learning_rate": 4.473684210526316e-06, "loss": 0.5681, "step": 17 }, { "epoch": 0.144, "grad_norm": 1.4679189920425415, "learning_rate": 4.736842105263158e-06, "loss": 0.607, "step": 18 }, { "epoch": 0.152, "grad_norm": 1.2237054109573364, "learning_rate": 5e-06, "loss": 0.5273, "step": 19 }, { "epoch": 0.16, "grad_norm": 0.9686553478240967, "learning_rate": 5.263157894736842e-06, "loss": 0.5578, "step": 20 }, { "epoch": 0.168, "grad_norm": 0.7895737290382385, "learning_rate": 5.526315789473685e-06, "loss": 0.5794, "step": 21 }, { "epoch": 0.176, "grad_norm": 0.8087520599365234, "learning_rate": 5.789473684210527e-06, "loss": 0.5229, "step": 22 }, { "epoch": 0.184, "grad_norm": 0.7652890086174011, "learning_rate": 6.0526315789473685e-06, "loss": 0.5292, "step": 23 }, { "epoch": 0.192, "grad_norm": 0.9132710695266724, "learning_rate": 6.31578947368421e-06, "loss": 0.5271, "step": 24 }, { "epoch": 0.2, "grad_norm": 0.8860102295875549, "learning_rate": 6.578947368421054e-06, "loss": 0.5155, "step": 25 }, { "epoch": 0.208, "grad_norm": 0.8902170062065125, "learning_rate": 6.842105263157896e-06, "loss": 0.5287, "step": 26 }, { "epoch": 0.216, "grad_norm": 0.8824247121810913, "learning_rate": 7.1052631578947375e-06, "loss": 0.4791, "step": 27 }, { "epoch": 0.224, "grad_norm": 1.2683454751968384, "learning_rate": 7.368421052631579e-06, "loss": 0.5329, "step": 28 }, { "epoch": 0.232, "grad_norm": 0.7248997092247009, "learning_rate": 7.631578947368423e-06, "loss": 0.5294, "step": 29 }, { "epoch": 0.24, "grad_norm": 0.7209534645080566, "learning_rate": 7.894736842105265e-06, "loss": 0.4842, "step": 30 }, { "epoch": 0.248, "grad_norm": 0.7555510401725769, "learning_rate": 8.157894736842106e-06, "loss": 0.5348, "step": 31 }, { "epoch": 0.256, "grad_norm": 0.6169366240501404, "learning_rate": 8.421052631578948e-06, "loss": 0.5382, "step": 32 }, { "epoch": 0.264, "grad_norm": 0.6152195930480957, "learning_rate": 8.68421052631579e-06, "loss": 0.5039, "step": 33 }, { "epoch": 0.272, "grad_norm": 0.631288468837738, "learning_rate": 8.947368421052632e-06, "loss": 0.5037, "step": 34 }, { "epoch": 0.28, "grad_norm": 0.5909512042999268, "learning_rate": 9.210526315789474e-06, "loss": 0.5017, "step": 35 }, { "epoch": 0.288, "grad_norm": 0.5489953756332397, "learning_rate": 9.473684210526315e-06, "loss": 0.4469, "step": 36 }, { "epoch": 0.296, "grad_norm": 0.565674901008606, "learning_rate": 9.736842105263159e-06, "loss": 0.478, "step": 37 }, { "epoch": 0.304, "grad_norm": 0.680474579334259, "learning_rate": 1e-05, "loss": 0.5392, "step": 38 }, { "epoch": 0.312, "grad_norm": 0.607657253742218, "learning_rate": 9.99978274148479e-06, "loss": 0.5022, "step": 39 }, { "epoch": 0.32, "grad_norm": 0.5720808506011963, "learning_rate": 9.999130984819662e-06, "loss": 0.5057, "step": 40 }, { "epoch": 0.328, "grad_norm": 0.5648075938224792, "learning_rate": 9.998044786644492e-06, "loss": 0.4718, "step": 41 }, { "epoch": 0.336, "grad_norm": 0.5617280006408691, "learning_rate": 9.9965242413536e-06, "loss": 0.5049, "step": 42 }, { "epoch": 0.344, "grad_norm": 0.6486408114433289, "learning_rate": 9.994569481087552e-06, "loss": 0.5203, "step": 43 }, { "epoch": 0.352, "grad_norm": 0.5352612733840942, "learning_rate": 9.992180675721671e-06, "loss": 0.4967, "step": 44 }, { "epoch": 0.36, "grad_norm": 0.5438513159751892, "learning_rate": 9.989358032851283e-06, "loss": 0.5188, "step": 45 }, { "epoch": 0.368, "grad_norm": 0.5454742908477783, "learning_rate": 9.986101797773667e-06, "loss": 0.4864, "step": 46 }, { "epoch": 0.376, "grad_norm": 0.536598265171051, "learning_rate": 9.98241225346674e-06, "loss": 0.4883, "step": 47 }, { "epoch": 0.384, "grad_norm": 0.4659676253795624, "learning_rate": 9.978289720564471e-06, "loss": 0.5121, "step": 48 }, { "epoch": 0.392, "grad_norm": 0.5836896300315857, "learning_rate": 9.97373455732901e-06, "loss": 0.4975, "step": 49 }, { "epoch": 0.4, "grad_norm": 0.5226072669029236, "learning_rate": 9.968747159619556e-06, "loss": 0.4944, "step": 50 }, { "epoch": 0.408, "grad_norm": 0.5444321632385254, "learning_rate": 9.963327960857962e-06, "loss": 0.5209, "step": 51 }, { "epoch": 0.416, "grad_norm": 0.5670180320739746, "learning_rate": 9.957477431991053e-06, "loss": 0.514, "step": 52 }, { "epoch": 0.424, "grad_norm": 0.511341392993927, "learning_rate": 9.95119608144972e-06, "loss": 0.526, "step": 53 }, { "epoch": 0.432, "grad_norm": 0.5060410499572754, "learning_rate": 9.944484455104716e-06, "loss": 0.4806, "step": 54 }, { "epoch": 0.44, "grad_norm": 0.487128347158432, "learning_rate": 9.937343136219234e-06, "loss": 0.5273, "step": 55 }, { "epoch": 0.448, "grad_norm": 0.5598118305206299, "learning_rate": 9.929772745398207e-06, "loss": 0.5745, "step": 56 }, { "epoch": 0.456, "grad_norm": 0.5080751180648804, "learning_rate": 9.921773940534382e-06, "loss": 0.5082, "step": 57 }, { "epoch": 0.464, "grad_norm": 0.5492854118347168, "learning_rate": 9.913347416751148e-06, "loss": 0.5244, "step": 58 }, { "epoch": 0.472, "grad_norm": 0.5565740466117859, "learning_rate": 9.904493906342124e-06, "loss": 0.5117, "step": 59 }, { "epoch": 0.48, "grad_norm": 0.5090964436531067, "learning_rate": 9.895214178707516e-06, "loss": 0.4942, "step": 60 }, { "epoch": 0.488, "grad_norm": 0.5151892304420471, "learning_rate": 9.885509040287267e-06, "loss": 0.5165, "step": 61 }, { "epoch": 0.496, "grad_norm": 0.5298195481300354, "learning_rate": 9.875379334490962e-06, "loss": 0.5218, "step": 62 }, { "epoch": 0.504, "grad_norm": 0.5165826082229614, "learning_rate": 9.864825941624538e-06, "loss": 0.4986, "step": 63 }, { "epoch": 0.512, "grad_norm": 0.506293773651123, "learning_rate": 9.853849778813777e-06, "loss": 0.5127, "step": 64 }, { "epoch": 0.52, "grad_norm": 0.5140017867088318, "learning_rate": 9.842451799924616e-06, "loss": 0.5158, "step": 65 }, { "epoch": 0.528, "grad_norm": 0.5674154162406921, "learning_rate": 9.830632995480243e-06, "loss": 0.5006, "step": 66 }, { "epoch": 0.536, "grad_norm": 0.5346920490264893, "learning_rate": 9.818394392575018e-06, "loss": 0.5126, "step": 67 }, { "epoch": 0.544, "grad_norm": 0.5248146653175354, "learning_rate": 9.805737054785223e-06, "loss": 0.5394, "step": 68 }, { "epoch": 0.552, "grad_norm": 0.5473419427871704, "learning_rate": 9.792662082076618e-06, "loss": 0.5224, "step": 69 }, { "epoch": 0.56, "grad_norm": 0.5517472624778748, "learning_rate": 9.779170610708872e-06, "loss": 0.5125, "step": 70 }, { "epoch": 0.568, "grad_norm": 0.4769006669521332, "learning_rate": 9.765263813136796e-06, "loss": 0.5081, "step": 71 }, { "epoch": 0.576, "grad_norm": 0.5075954794883728, "learning_rate": 9.750942897908468e-06, "loss": 0.4958, "step": 72 }, { "epoch": 0.584, "grad_norm": 0.5006887316703796, "learning_rate": 9.736209109560201e-06, "loss": 0.5279, "step": 73 }, { "epoch": 0.592, "grad_norm": 0.5326476693153381, "learning_rate": 9.721063728508384e-06, "loss": 0.5136, "step": 74 }, { "epoch": 0.6, "grad_norm": 0.5512110590934753, "learning_rate": 9.705508070938219e-06, "loss": 0.474, "step": 75 }, { "epoch": 0.608, "grad_norm": 0.5133175253868103, "learning_rate": 9.689543488689332e-06, "loss": 0.4994, "step": 76 }, { "epoch": 0.616, "grad_norm": 0.5255376696586609, "learning_rate": 9.673171369138297e-06, "loss": 0.5183, "step": 77 }, { "epoch": 0.624, "grad_norm": 0.5707758665084839, "learning_rate": 9.656393135078067e-06, "loss": 0.493, "step": 78 }, { "epoch": 0.632, "grad_norm": 0.5201893448829651, "learning_rate": 9.639210244594335e-06, "loss": 0.4635, "step": 79 }, { "epoch": 0.64, "grad_norm": 0.5058737397193909, "learning_rate": 9.621624190938802e-06, "loss": 0.5312, "step": 80 }, { "epoch": 0.648, "grad_norm": 0.43761327862739563, "learning_rate": 9.603636502399436e-06, "loss": 0.4524, "step": 81 }, { "epoch": 0.656, "grad_norm": 0.5335831642150879, "learning_rate": 9.585248742167638e-06, "loss": 0.4648, "step": 82 }, { "epoch": 0.664, "grad_norm": 0.5373964309692383, "learning_rate": 9.566462508202403e-06, "loss": 0.4911, "step": 83 }, { "epoch": 0.672, "grad_norm": 0.5256152153015137, "learning_rate": 9.547279433091446e-06, "loss": 0.4968, "step": 84 }, { "epoch": 0.68, "grad_norm": 0.5257714986801147, "learning_rate": 9.527701183909336e-06, "loss": 0.4879, "step": 85 }, { "epoch": 0.688, "grad_norm": 0.5462167859077454, "learning_rate": 9.507729462072615e-06, "loss": 0.4577, "step": 86 }, { "epoch": 0.696, "grad_norm": 0.5356786847114563, "learning_rate": 9.48736600319193e-06, "loss": 0.4922, "step": 87 }, { "epoch": 0.704, "grad_norm": 0.5588122010231018, "learning_rate": 9.466612576921223e-06, "loss": 0.4915, "step": 88 }, { "epoch": 0.712, "grad_norm": 0.5467613935470581, "learning_rate": 9.445470986803922e-06, "loss": 0.4711, "step": 89 }, { "epoch": 0.72, "grad_norm": 0.5490957498550415, "learning_rate": 9.423943070116219e-06, "loss": 0.5108, "step": 90 }, { "epoch": 0.728, "grad_norm": 0.4839731752872467, "learning_rate": 9.402030697707398e-06, "loss": 0.5128, "step": 91 }, { "epoch": 0.736, "grad_norm": 0.5173898935317993, "learning_rate": 9.37973577383726e-06, "loss": 0.4823, "step": 92 }, { "epoch": 0.744, "grad_norm": 0.563199520111084, "learning_rate": 9.357060236010626e-06, "loss": 0.4985, "step": 93 }, { "epoch": 0.752, "grad_norm": 0.513721227645874, "learning_rate": 9.334006054808966e-06, "loss": 0.4484, "step": 94 }, { "epoch": 0.76, "grad_norm": 0.5133054256439209, "learning_rate": 9.310575233719155e-06, "loss": 0.5047, "step": 95 }, { "epoch": 0.768, "grad_norm": 0.5787802338600159, "learning_rate": 9.28676980895935e-06, "loss": 0.5064, "step": 96 }, { "epoch": 0.776, "grad_norm": 0.4748121500015259, "learning_rate": 9.262591849302049e-06, "loss": 0.4604, "step": 97 }, { "epoch": 0.784, "grad_norm": 0.556131899356842, "learning_rate": 9.238043455894294e-06, "loss": 0.5295, "step": 98 }, { "epoch": 0.792, "grad_norm": 0.5900002717971802, "learning_rate": 9.213126762075088e-06, "loss": 0.4977, "step": 99 }, { "epoch": 0.8, "grad_norm": 0.5788969397544861, "learning_rate": 9.187843933189994e-06, "loss": 0.4732, "step": 100 }, { "epoch": 0.808, "grad_norm": 0.5433998107910156, "learning_rate": 9.162197166402957e-06, "loss": 0.5175, "step": 101 }, { "epoch": 0.816, "grad_norm": 0.5408159494400024, "learning_rate": 9.136188690505363e-06, "loss": 0.5087, "step": 102 }, { "epoch": 0.824, "grad_norm": 0.5552597045898438, "learning_rate": 9.109820765722357e-06, "loss": 0.4835, "step": 103 }, { "epoch": 0.832, "grad_norm": 0.5368766188621521, "learning_rate": 9.083095683516414e-06, "loss": 0.5349, "step": 104 }, { "epoch": 0.84, "grad_norm": 0.46206405758857727, "learning_rate": 9.056015766388205e-06, "loss": 0.4567, "step": 105 }, { "epoch": 0.848, "grad_norm": 0.5309408903121948, "learning_rate": 9.028583367674767e-06, "loss": 0.5166, "step": 106 }, { "epoch": 0.856, "grad_norm": 0.528049886226654, "learning_rate": 9.00080087134498e-06, "loss": 0.4526, "step": 107 }, { "epoch": 0.864, "grad_norm": 0.5379208326339722, "learning_rate": 8.972670691792409e-06, "loss": 0.4893, "step": 108 }, { "epoch": 0.872, "grad_norm": 0.5074729919433594, "learning_rate": 8.944195273625472e-06, "loss": 0.4924, "step": 109 }, { "epoch": 0.88, "grad_norm": 0.5556723475456238, "learning_rate": 8.915377091454992e-06, "loss": 0.4981, "step": 110 }, { "epoch": 0.888, "grad_norm": 0.5345364212989807, "learning_rate": 8.886218649679162e-06, "loss": 0.5134, "step": 111 }, { "epoch": 0.896, "grad_norm": 0.511635959148407, "learning_rate": 8.856722482265886e-06, "loss": 0.461, "step": 112 }, { "epoch": 0.904, "grad_norm": 0.49792084097862244, "learning_rate": 8.826891152532579e-06, "loss": 0.505, "step": 113 }, { "epoch": 0.912, "grad_norm": 0.4829058051109314, "learning_rate": 8.796727252923403e-06, "loss": 0.4716, "step": 114 }, { "epoch": 0.92, "grad_norm": 0.5941436886787415, "learning_rate": 8.766233404783975e-06, "loss": 0.5296, "step": 115 }, { "epoch": 0.928, "grad_norm": 0.5134578943252563, "learning_rate": 8.735412258133562e-06, "loss": 0.5045, "step": 116 }, { "epoch": 0.936, "grad_norm": 0.5325246453285217, "learning_rate": 8.704266491434787e-06, "loss": 0.472, "step": 117 }, { "epoch": 0.944, "grad_norm": 0.538865327835083, "learning_rate": 8.672798811360863e-06, "loss": 0.5026, "step": 118 }, { "epoch": 0.952, "grad_norm": 0.48347458243370056, "learning_rate": 8.641011952560372e-06, "loss": 0.484, "step": 119 }, { "epoch": 0.96, "grad_norm": 0.5306034684181213, "learning_rate": 8.608908677419606e-06, "loss": 0.4997, "step": 120 }, { "epoch": 0.968, "grad_norm": 0.5399216413497925, "learning_rate": 8.576491775822527e-06, "loss": 0.4804, "step": 121 }, { "epoch": 0.976, "grad_norm": 0.5183812975883484, "learning_rate": 8.543764064908295e-06, "loss": 0.4887, "step": 122 }, { "epoch": 0.984, "grad_norm": 0.4991128444671631, "learning_rate": 8.510728388826464e-06, "loss": 0.4453, "step": 123 }, { "epoch": 0.992, "grad_norm": 0.5225016474723816, "learning_rate": 8.477387618489808e-06, "loss": 0.5123, "step": 124 }, { "epoch": 1.0, "grad_norm": 0.46465304493904114, "learning_rate": 8.443744651324828e-06, "loss": 0.4633, "step": 125 }, { "epoch": 1.008, "grad_norm": 0.5113005638122559, "learning_rate": 8.409802411019962e-06, "loss": 0.427, "step": 126 }, { "epoch": 1.016, "grad_norm": 0.5328509211540222, "learning_rate": 8.375563847271506e-06, "loss": 0.4324, "step": 127 }, { "epoch": 1.024, "grad_norm": 0.519523024559021, "learning_rate": 8.341031935527267e-06, "loss": 0.4387, "step": 128 }, { "epoch": 1.032, "grad_norm": 0.506823718547821, "learning_rate": 8.306209676727994e-06, "loss": 0.4469, "step": 129 }, { "epoch": 1.04, "grad_norm": 0.4721229374408722, "learning_rate": 8.271100097046585e-06, "loss": 0.4226, "step": 130 }, { "epoch": 1.048, "grad_norm": 0.4656391739845276, "learning_rate": 8.235706247625098e-06, "loss": 0.3729, "step": 131 }, { "epoch": 1.056, "grad_norm": 0.4901474118232727, "learning_rate": 8.200031204309604e-06, "loss": 0.4627, "step": 132 }, { "epoch": 1.064, "grad_norm": 0.606958270072937, "learning_rate": 8.16407806738288e-06, "loss": 0.4347, "step": 133 }, { "epoch": 1.072, "grad_norm": 0.4573584496974945, "learning_rate": 8.127849961294984e-06, "loss": 0.402, "step": 134 }, { "epoch": 1.08, "grad_norm": 0.4766193926334381, "learning_rate": 8.091350034391732e-06, "loss": 0.4203, "step": 135 }, { "epoch": 1.088, "grad_norm": 0.5167376399040222, "learning_rate": 8.05458145864109e-06, "loss": 0.4521, "step": 136 }, { "epoch": 1.096, "grad_norm": 0.52271968126297, "learning_rate": 8.017547429357532e-06, "loss": 0.4321, "step": 137 }, { "epoch": 1.104, "grad_norm": 0.49505582451820374, "learning_rate": 7.980251164924342e-06, "loss": 0.4591, "step": 138 }, { "epoch": 1.112, "grad_norm": 0.6967505216598511, "learning_rate": 7.94269590651393e-06, "loss": 0.4322, "step": 139 }, { "epoch": 1.12, "grad_norm": 0.46629568934440613, "learning_rate": 7.904884917806174e-06, "loss": 0.4612, "step": 140 }, { "epoch": 1.1280000000000001, "grad_norm": 0.49921801686286926, "learning_rate": 7.866821484704777e-06, "loss": 0.4353, "step": 141 }, { "epoch": 1.1360000000000001, "grad_norm": 0.5447902083396912, "learning_rate": 7.828508915051724e-06, "loss": 0.4323, "step": 142 }, { "epoch": 1.144, "grad_norm": 0.5017483830451965, "learning_rate": 7.789950538339813e-06, "loss": 0.4352, "step": 143 }, { "epoch": 1.152, "grad_norm": 0.49647992849349976, "learning_rate": 7.751149705423313e-06, "loss": 0.3971, "step": 144 }, { "epoch": 1.16, "grad_norm": 0.5001079440116882, "learning_rate": 7.712109788226763e-06, "loss": 0.4566, "step": 145 }, { "epoch": 1.168, "grad_norm": 0.4786428213119507, "learning_rate": 7.672834179451943e-06, "loss": 0.446, "step": 146 }, { "epoch": 1.176, "grad_norm": 0.5244361758232117, "learning_rate": 7.633326292283028e-06, "loss": 0.4759, "step": 147 }, { "epoch": 1.184, "grad_norm": 0.4763104319572449, "learning_rate": 7.593589560089984e-06, "loss": 0.4239, "step": 148 }, { "epoch": 1.192, "grad_norm": 0.4974442422389984, "learning_rate": 7.553627436130183e-06, "loss": 0.4053, "step": 149 }, { "epoch": 1.2, "grad_norm": 0.5250166654586792, "learning_rate": 7.513443393248312e-06, "loss": 0.4521, "step": 150 }, { "epoch": 1.208, "grad_norm": 0.4822094440460205, "learning_rate": 7.473040923574567e-06, "loss": 0.3883, "step": 151 }, { "epoch": 1.216, "grad_norm": 0.4862370491027832, "learning_rate": 7.432423538221179e-06, "loss": 0.4322, "step": 152 }, { "epoch": 1.224, "grad_norm": 0.482036828994751, "learning_rate": 7.391594766977277e-06, "loss": 0.4105, "step": 153 }, { "epoch": 1.232, "grad_norm": 0.5110259056091309, "learning_rate": 7.350558158002154e-06, "loss": 0.4103, "step": 154 }, { "epoch": 1.24, "grad_norm": 0.5026330351829529, "learning_rate": 7.3093172775169e-06, "loss": 0.4312, "step": 155 }, { "epoch": 1.248, "grad_norm": 0.4886966943740845, "learning_rate": 7.2678757094945e-06, "loss": 0.4325, "step": 156 }, { "epoch": 1.256, "grad_norm": 0.4985504150390625, "learning_rate": 7.226237055348369e-06, "loss": 0.4232, "step": 157 }, { "epoch": 1.264, "grad_norm": 0.47912245988845825, "learning_rate": 7.184404933619377e-06, "loss": 0.4359, "step": 158 }, { "epoch": 1.272, "grad_norm": 0.48068177700042725, "learning_rate": 7.142382979661386e-06, "loss": 0.4158, "step": 159 }, { "epoch": 1.28, "grad_norm": 0.4839794933795929, "learning_rate": 7.100174845325327e-06, "loss": 0.4035, "step": 160 }, { "epoch": 1.288, "grad_norm": 0.5554186105728149, "learning_rate": 7.057784198641835e-06, "loss": 0.3862, "step": 161 }, { "epoch": 1.296, "grad_norm": 0.5477203726768494, "learning_rate": 7.015214723502496e-06, "loss": 0.4347, "step": 162 }, { "epoch": 1.304, "grad_norm": 0.48494377732276917, "learning_rate": 6.972470119339692e-06, "loss": 0.4183, "step": 163 }, { "epoch": 1.312, "grad_norm": 0.4876761734485626, "learning_rate": 6.929554100805118e-06, "loss": 0.4459, "step": 164 }, { "epoch": 1.32, "grad_norm": 0.4854036867618561, "learning_rate": 6.886470397446958e-06, "loss": 0.4249, "step": 165 }, { "epoch": 1.328, "grad_norm": 0.4709065556526184, "learning_rate": 6.843222753385785e-06, "loss": 0.4376, "step": 166 }, { "epoch": 1.336, "grad_norm": 0.4834536910057068, "learning_rate": 6.799814926989171e-06, "loss": 0.4017, "step": 167 }, { "epoch": 1.3439999999999999, "grad_norm": 0.4833991527557373, "learning_rate": 6.756250690545079e-06, "loss": 0.4145, "step": 168 }, { "epoch": 1.3519999999999999, "grad_norm": 0.5171043872833252, "learning_rate": 6.712533829934042e-06, "loss": 0.4373, "step": 169 }, { "epoch": 1.3599999999999999, "grad_norm": 0.5456111431121826, "learning_rate": 6.6686681443001485e-06, "loss": 0.4272, "step": 170 }, { "epoch": 1.3679999999999999, "grad_norm": 0.5275494456291199, "learning_rate": 6.62465744572089e-06, "loss": 0.4628, "step": 171 }, { "epoch": 1.376, "grad_norm": 0.46399417519569397, "learning_rate": 6.580505558875878e-06, "loss": 0.481, "step": 172 }, { "epoch": 1.384, "grad_norm": 0.4880594313144684, "learning_rate": 6.536216320714466e-06, "loss": 0.4285, "step": 173 }, { "epoch": 1.392, "grad_norm": 0.48990365862846375, "learning_rate": 6.491793580122301e-06, "loss": 0.3997, "step": 174 }, { "epoch": 1.4, "grad_norm": 0.49995189905166626, "learning_rate": 6.447241197586847e-06, "loss": 0.4332, "step": 175 }, { "epoch": 1.408, "grad_norm": 0.4638576805591583, "learning_rate": 6.402563044861899e-06, "loss": 0.4236, "step": 176 }, { "epoch": 1.416, "grad_norm": 0.49920183420181274, "learning_rate": 6.357763004631104e-06, "loss": 0.4349, "step": 177 }, { "epoch": 1.424, "grad_norm": 0.5193583965301514, "learning_rate": 6.312844970170551e-06, "loss": 0.4089, "step": 178 }, { "epoch": 1.432, "grad_norm": 0.47679874300956726, "learning_rate": 6.267812845010431e-06, "loss": 0.4298, "step": 179 }, { "epoch": 1.44, "grad_norm": 0.49884992837905884, "learning_rate": 6.2226705425958e-06, "loss": 0.4517, "step": 180 }, { "epoch": 1.448, "grad_norm": 0.48136106133461, "learning_rate": 6.177421985946499e-06, "loss": 0.4393, "step": 181 }, { "epoch": 1.456, "grad_norm": 0.5779682397842407, "learning_rate": 6.132071107316221e-06, "loss": 0.4225, "step": 182 }, { "epoch": 1.464, "grad_norm": 0.5509202480316162, "learning_rate": 6.0866218478507875e-06, "loss": 0.4579, "step": 183 }, { "epoch": 1.472, "grad_norm": 0.5101506114006042, "learning_rate": 6.041078157245649e-06, "loss": 0.4036, "step": 184 }, { "epoch": 1.48, "grad_norm": 0.623524010181427, "learning_rate": 5.995443993402647e-06, "loss": 0.451, "step": 185 }, { "epoch": 1.488, "grad_norm": 0.4368286728858948, "learning_rate": 5.949723322086053e-06, "loss": 0.3881, "step": 186 }, { "epoch": 1.496, "grad_norm": 0.4305906295776367, "learning_rate": 5.9039201165779315e-06, "loss": 0.4036, "step": 187 }, { "epoch": 1.504, "grad_norm": 0.4749639630317688, "learning_rate": 5.858038357332851e-06, "loss": 0.4048, "step": 188 }, { "epoch": 1.512, "grad_norm": 0.48530155420303345, "learning_rate": 5.812082031631966e-06, "loss": 0.4464, "step": 189 }, { "epoch": 1.52, "grad_norm": 0.555038571357727, "learning_rate": 5.766055133236513e-06, "loss": 0.4485, "step": 190 }, { "epoch": 1.528, "grad_norm": 0.4892346262931824, "learning_rate": 5.7199616620407325e-06, "loss": 0.4522, "step": 191 }, { "epoch": 1.536, "grad_norm": 0.5365411639213562, "learning_rate": 5.673805623724272e-06, "loss": 0.4752, "step": 192 }, { "epoch": 1.544, "grad_norm": 0.5186147689819336, "learning_rate": 5.627591029404072e-06, "loss": 0.4021, "step": 193 }, { "epoch": 1.552, "grad_norm": 0.48625391721725464, "learning_rate": 5.581321895285787e-06, "loss": 0.4396, "step": 194 }, { "epoch": 1.56, "grad_norm": 0.5071629285812378, "learning_rate": 5.535002242314772e-06, "loss": 0.4227, "step": 195 }, { "epoch": 1.568, "grad_norm": 0.5189520120620728, "learning_rate": 5.488636095826636e-06, "loss": 0.4512, "step": 196 }, { "epoch": 1.576, "grad_norm": 0.5435962677001953, "learning_rate": 5.4422274851974356e-06, "loss": 0.4767, "step": 197 }, { "epoch": 1.584, "grad_norm": 0.4914834797382355, "learning_rate": 5.395780443493508e-06, "loss": 0.4035, "step": 198 }, { "epoch": 1.592, "grad_norm": 0.5276004672050476, "learning_rate": 5.34929900712098e-06, "loss": 0.4329, "step": 199 }, { "epoch": 1.6, "grad_norm": 0.4546476900577545, "learning_rate": 5.302787215474992e-06, "loss": 0.3965, "step": 200 }, { "epoch": 1.608, "grad_norm": 0.48372510075569153, "learning_rate": 5.256249110588659e-06, "loss": 0.4272, "step": 201 }, { "epoch": 1.616, "grad_norm": 0.49213626980781555, "learning_rate": 5.209688736781811e-06, "loss": 0.4409, "step": 202 }, { "epoch": 1.624, "grad_norm": 0.4945046603679657, "learning_rate": 5.163110140309518e-06, "loss": 0.4419, "step": 203 }, { "epoch": 1.6320000000000001, "grad_norm": 0.4468039274215698, "learning_rate": 5.116517369010467e-06, "loss": 0.398, "step": 204 }, { "epoch": 1.6400000000000001, "grad_norm": 0.4727765917778015, "learning_rate": 5.069914471955179e-06, "loss": 0.4229, "step": 205 }, { "epoch": 1.6480000000000001, "grad_norm": 0.5005367398262024, "learning_rate": 5.023305499094145e-06, "loss": 0.4153, "step": 206 }, { "epoch": 1.6560000000000001, "grad_norm": 0.4818647801876068, "learning_rate": 4.976694500905858e-06, "loss": 0.443, "step": 207 }, { "epoch": 1.6640000000000001, "grad_norm": 0.47024303674697876, "learning_rate": 4.930085528044823e-06, "loss": 0.4541, "step": 208 }, { "epoch": 1.6720000000000002, "grad_norm": 0.47553378343582153, "learning_rate": 4.883482630989536e-06, "loss": 0.4069, "step": 209 }, { "epoch": 1.6800000000000002, "grad_norm": 0.44939181208610535, "learning_rate": 4.8368898596904834e-06, "loss": 0.3964, "step": 210 }, { "epoch": 1.688, "grad_norm": 0.46793460845947266, "learning_rate": 4.790311263218191e-06, "loss": 0.4339, "step": 211 }, { "epoch": 1.696, "grad_norm": 0.45191749930381775, "learning_rate": 4.743750889411342e-06, "loss": 0.426, "step": 212 }, { "epoch": 1.704, "grad_norm": 0.45308157801628113, "learning_rate": 4.697212784525009e-06, "loss": 0.4338, "step": 213 }, { "epoch": 1.712, "grad_norm": 0.4543735384941101, "learning_rate": 4.65070099287902e-06, "loss": 0.4083, "step": 214 }, { "epoch": 1.72, "grad_norm": 0.49145424365997314, "learning_rate": 4.604219556506492e-06, "loss": 0.4658, "step": 215 }, { "epoch": 1.728, "grad_norm": 0.4785975515842438, "learning_rate": 4.557772514802564e-06, "loss": 0.4248, "step": 216 }, { "epoch": 1.736, "grad_norm": 0.5019366145133972, "learning_rate": 4.511363904173366e-06, "loss": 0.4482, "step": 217 }, { "epoch": 1.744, "grad_norm": 0.4578736126422882, "learning_rate": 4.46499775768523e-06, "loss": 0.4391, "step": 218 }, { "epoch": 1.752, "grad_norm": 0.46869972348213196, "learning_rate": 4.418678104714214e-06, "loss": 0.4087, "step": 219 }, { "epoch": 1.76, "grad_norm": 0.480589359998703, "learning_rate": 4.372408970595931e-06, "loss": 0.418, "step": 220 }, { "epoch": 1.768, "grad_norm": 0.5114148855209351, "learning_rate": 4.326194376275729e-06, "loss": 0.4009, "step": 221 }, { "epoch": 1.776, "grad_norm": 0.4714803099632263, "learning_rate": 4.280038337959268e-06, "loss": 0.4119, "step": 222 }, { "epoch": 1.784, "grad_norm": 0.4820714592933655, "learning_rate": 4.2339448667634885e-06, "loss": 0.4364, "step": 223 }, { "epoch": 1.792, "grad_norm": 0.45575451850891113, "learning_rate": 4.187917968368036e-06, "loss": 0.3943, "step": 224 }, { "epoch": 1.8, "grad_norm": 0.4869477152824402, "learning_rate": 4.141961642667152e-06, "loss": 0.4447, "step": 225 }, { "epoch": 1.808, "grad_norm": 0.4546818435192108, "learning_rate": 4.09607988342207e-06, "loss": 0.4397, "step": 226 }, { "epoch": 1.8159999999999998, "grad_norm": 0.5653818249702454, "learning_rate": 4.0502766779139485e-06, "loss": 0.3881, "step": 227 }, { "epoch": 1.8239999999999998, "grad_norm": 0.484779953956604, "learning_rate": 4.0045560065973535e-06, "loss": 0.4673, "step": 228 }, { "epoch": 1.8319999999999999, "grad_norm": 0.48289933800697327, "learning_rate": 3.958921842754351e-06, "loss": 0.4156, "step": 229 }, { "epoch": 1.8399999999999999, "grad_norm": 0.43402060866355896, "learning_rate": 3.913378152149214e-06, "loss": 0.4044, "step": 230 }, { "epoch": 1.8479999999999999, "grad_norm": 0.47271519899368286, "learning_rate": 3.86792889268378e-06, "loss": 0.4387, "step": 231 }, { "epoch": 1.8559999999999999, "grad_norm": 0.5104248523712158, "learning_rate": 3.8225780140535025e-06, "loss": 0.3976, "step": 232 }, { "epoch": 1.8639999999999999, "grad_norm": 0.44085070490837097, "learning_rate": 3.777329457404202e-06, "loss": 0.4062, "step": 233 }, { "epoch": 1.8719999999999999, "grad_norm": 0.5098981857299805, "learning_rate": 3.7321871549895715e-06, "loss": 0.4458, "step": 234 }, { "epoch": 1.88, "grad_norm": 0.42704570293426514, "learning_rate": 3.68715502982945e-06, "loss": 0.3947, "step": 235 }, { "epoch": 1.888, "grad_norm": 0.46332859992980957, "learning_rate": 3.6422369953688973e-06, "loss": 0.4294, "step": 236 }, { "epoch": 1.896, "grad_norm": 0.4884559214115143, "learning_rate": 3.5974369551381023e-06, "loss": 0.4498, "step": 237 }, { "epoch": 1.904, "grad_norm": 0.5290793180465698, "learning_rate": 3.5527588024131542e-06, "loss": 0.4312, "step": 238 }, { "epoch": 1.912, "grad_norm": 0.48744434118270874, "learning_rate": 3.5082064198777e-06, "loss": 0.4772, "step": 239 }, { "epoch": 1.92, "grad_norm": 0.4615407884120941, "learning_rate": 3.463783679285535e-06, "loss": 0.4101, "step": 240 }, { "epoch": 1.928, "grad_norm": 0.49090084433555603, "learning_rate": 3.4194944411241213e-06, "loss": 0.4222, "step": 241 }, { "epoch": 1.936, "grad_norm": 0.4603979289531708, "learning_rate": 3.3753425542791106e-06, "loss": 0.4322, "step": 242 }, { "epoch": 1.944, "grad_norm": 0.4630926847457886, "learning_rate": 3.3313318556998523e-06, "loss": 0.4553, "step": 243 }, { "epoch": 1.952, "grad_norm": 0.5014007091522217, "learning_rate": 3.2874661700659586e-06, "loss": 0.4541, "step": 244 }, { "epoch": 1.96, "grad_norm": 0.5217999815940857, "learning_rate": 3.2437493094549223e-06, "loss": 0.4164, "step": 245 }, { "epoch": 1.968, "grad_norm": 0.5072699785232544, "learning_rate": 3.200185073010831e-06, "loss": 0.4151, "step": 246 }, { "epoch": 1.976, "grad_norm": 0.4602278172969818, "learning_rate": 3.1567772466142156e-06, "loss": 0.3739, "step": 247 }, { "epoch": 1.984, "grad_norm": 0.5061410069465637, "learning_rate": 3.1135296025530426e-06, "loss": 0.4892, "step": 248 }, { "epoch": 1.992, "grad_norm": 0.5533497333526611, "learning_rate": 3.070445899194885e-06, "loss": 0.4132, "step": 249 }, { "epoch": 2.0, "grad_norm": 0.52688068151474, "learning_rate": 3.0275298806603102e-06, "loss": 0.4463, "step": 250 }, { "epoch": 2.008, "grad_norm": 0.5126904249191284, "learning_rate": 2.984785276497507e-06, "loss": 0.3639, "step": 251 }, { "epoch": 2.016, "grad_norm": 0.5104655623435974, "learning_rate": 2.9422158013581658e-06, "loss": 0.3535, "step": 252 }, { "epoch": 2.024, "grad_norm": 0.5136793255805969, "learning_rate": 2.899825154674674e-06, "loss": 0.3835, "step": 253 }, { "epoch": 2.032, "grad_norm": 0.45347532629966736, "learning_rate": 2.8576170203386144e-06, "loss": 0.3412, "step": 254 }, { "epoch": 2.04, "grad_norm": 0.4745730757713318, "learning_rate": 2.8155950663806234e-06, "loss": 0.3863, "step": 255 }, { "epoch": 2.048, "grad_norm": 0.5067371129989624, "learning_rate": 2.7737629446516325e-06, "loss": 0.3514, "step": 256 }, { "epoch": 2.056, "grad_norm": 0.45462149381637573, "learning_rate": 2.732124290505501e-06, "loss": 0.3528, "step": 257 }, { "epoch": 2.064, "grad_norm": 0.46529221534729004, "learning_rate": 2.6906827224831024e-06, "loss": 0.3296, "step": 258 }, { "epoch": 2.072, "grad_norm": 0.5160210728645325, "learning_rate": 2.6494418419978485e-06, "loss": 0.3416, "step": 259 }, { "epoch": 2.08, "grad_norm": 0.505672037601471, "learning_rate": 2.608405233022724e-06, "loss": 0.3793, "step": 260 }, { "epoch": 2.088, "grad_norm": 0.44298309087753296, "learning_rate": 2.5675764617788233e-06, "loss": 0.3298, "step": 261 }, { "epoch": 2.096, "grad_norm": 0.4680778682231903, "learning_rate": 2.526959076425434e-06, "loss": 0.3427, "step": 262 }, { "epoch": 2.104, "grad_norm": 0.4816036820411682, "learning_rate": 2.4865566067516896e-06, "loss": 0.3709, "step": 263 }, { "epoch": 2.112, "grad_norm": 0.47969356179237366, "learning_rate": 2.4463725638698182e-06, "loss": 0.3255, "step": 264 }, { "epoch": 2.12, "grad_norm": 0.5166645646095276, "learning_rate": 2.406410439910017e-06, "loss": 0.3928, "step": 265 }, { "epoch": 2.128, "grad_norm": 0.523476243019104, "learning_rate": 2.366673707716973e-06, "loss": 0.3497, "step": 266 }, { "epoch": 2.136, "grad_norm": 0.5414285063743591, "learning_rate": 2.327165820548059e-06, "loss": 0.3653, "step": 267 }, { "epoch": 2.144, "grad_norm": 0.4763210713863373, "learning_rate": 2.287890211773238e-06, "loss": 0.3628, "step": 268 }, { "epoch": 2.152, "grad_norm": 0.5000659227371216, "learning_rate": 2.2488502945766893e-06, "loss": 0.3681, "step": 269 }, { "epoch": 2.16, "grad_norm": 0.48179784417152405, "learning_rate": 2.210049461660189e-06, "loss": 0.3542, "step": 270 }, { "epoch": 2.168, "grad_norm": 0.4680255651473999, "learning_rate": 2.1714910849482777e-06, "loss": 0.3618, "step": 271 }, { "epoch": 2.176, "grad_norm": 0.4781685173511505, "learning_rate": 2.1331785152952243e-06, "loss": 0.3533, "step": 272 }, { "epoch": 2.184, "grad_norm": 0.508620023727417, "learning_rate": 2.0951150821938278e-06, "loss": 0.3761, "step": 273 }, { "epoch": 2.192, "grad_norm": 0.48846498131752014, "learning_rate": 2.0573040934860717e-06, "loss": 0.3776, "step": 274 }, { "epoch": 2.2, "grad_norm": 0.48690298199653625, "learning_rate": 2.0197488350756618e-06, "loss": 0.326, "step": 275 }, { "epoch": 2.208, "grad_norm": 0.5058321356773376, "learning_rate": 1.98245257064247e-06, "loss": 0.3596, "step": 276 }, { "epoch": 2.216, "grad_norm": 0.5168304443359375, "learning_rate": 1.945418541358911e-06, "loss": 0.3443, "step": 277 }, { "epoch": 2.224, "grad_norm": 0.4696882367134094, "learning_rate": 1.9086499656082685e-06, "loss": 0.3469, "step": 278 }, { "epoch": 2.232, "grad_norm": 0.46734338998794556, "learning_rate": 1.872150038705015e-06, "loss": 0.3494, "step": 279 }, { "epoch": 2.24, "grad_norm": 0.4970659017562866, "learning_rate": 1.835921932617119e-06, "loss": 0.3614, "step": 280 }, { "epoch": 2.248, "grad_norm": 0.4725131094455719, "learning_rate": 1.7999687956903955e-06, "loss": 0.3678, "step": 281 }, { "epoch": 2.2560000000000002, "grad_norm": 0.48716068267822266, "learning_rate": 1.7642937523749038e-06, "loss": 0.3514, "step": 282 }, { "epoch": 2.2640000000000002, "grad_norm": 0.4228149950504303, "learning_rate": 1.7288999029534177e-06, "loss": 0.3349, "step": 283 }, { "epoch": 2.2720000000000002, "grad_norm": 0.47141751646995544, "learning_rate": 1.6937903232720076e-06, "loss": 0.3655, "step": 284 }, { "epoch": 2.2800000000000002, "grad_norm": 0.44692301750183105, "learning_rate": 1.6589680644727347e-06, "loss": 0.3446, "step": 285 }, { "epoch": 2.288, "grad_norm": 0.46537840366363525, "learning_rate": 1.6244361527284953e-06, "loss": 0.3473, "step": 286 }, { "epoch": 2.296, "grad_norm": 0.4578072428703308, "learning_rate": 1.5901975889800387e-06, "loss": 0.3614, "step": 287 }, { "epoch": 2.304, "grad_norm": 0.4702107012271881, "learning_rate": 1.556255348675174e-06, "loss": 0.3417, "step": 288 }, { "epoch": 2.312, "grad_norm": 0.49962151050567627, "learning_rate": 1.522612381510195e-06, "loss": 0.348, "step": 289 }, { "epoch": 2.32, "grad_norm": 0.514380156993866, "learning_rate": 1.489271611173538e-06, "loss": 0.3587, "step": 290 }, { "epoch": 2.328, "grad_norm": 0.4479122459888458, "learning_rate": 1.4562359350917054e-06, "loss": 0.3479, "step": 291 }, { "epoch": 2.336, "grad_norm": 0.48699259757995605, "learning_rate": 1.423508224177474e-06, "loss": 0.3678, "step": 292 }, { "epoch": 2.344, "grad_norm": 0.46162307262420654, "learning_rate": 1.3910913225803946e-06, "loss": 0.3321, "step": 293 }, { "epoch": 2.352, "grad_norm": 0.4586437940597534, "learning_rate": 1.35898804743963e-06, "loss": 0.3748, "step": 294 }, { "epoch": 2.36, "grad_norm": 0.479818195104599, "learning_rate": 1.3272011886391368e-06, "loss": 0.3489, "step": 295 }, { "epoch": 2.368, "grad_norm": 0.44310837984085083, "learning_rate": 1.295733508565213e-06, "loss": 0.3427, "step": 296 }, { "epoch": 2.376, "grad_norm": 0.4900333881378174, "learning_rate": 1.2645877418664394e-06, "loss": 0.3484, "step": 297 }, { "epoch": 2.384, "grad_norm": 0.6010957360267639, "learning_rate": 1.2337665952160266e-06, "loss": 0.3486, "step": 298 }, { "epoch": 2.392, "grad_norm": 0.44970938563346863, "learning_rate": 1.2032727470765982e-06, "loss": 0.3345, "step": 299 }, { "epoch": 2.4, "grad_norm": 0.46644240617752075, "learning_rate": 1.1731088474674235e-06, "loss": 0.3695, "step": 300 }, { "epoch": 2.408, "grad_norm": 0.4730045199394226, "learning_rate": 1.1432775177341165e-06, "loss": 0.3605, "step": 301 }, { "epoch": 2.416, "grad_norm": 0.4608900845050812, "learning_rate": 1.11378135032084e-06, "loss": 0.3328, "step": 302 }, { "epoch": 2.424, "grad_norm": 0.47910159826278687, "learning_rate": 1.08462290854501e-06, "loss": 0.3592, "step": 303 }, { "epoch": 2.432, "grad_norm": 0.5008491277694702, "learning_rate": 1.0558047263745297e-06, "loss": 0.3653, "step": 304 }, { "epoch": 2.44, "grad_norm": 0.45120254158973694, "learning_rate": 1.0273293082075914e-06, "loss": 0.3317, "step": 305 }, { "epoch": 2.448, "grad_norm": 0.44771674275398254, "learning_rate": 9.991991286550207e-07, "loss": 0.3417, "step": 306 }, { "epoch": 2.456, "grad_norm": 0.47719165682792664, "learning_rate": 9.71416632325235e-07, "loss": 0.3634, "step": 307 }, { "epoch": 2.464, "grad_norm": 0.5096871256828308, "learning_rate": 9.439842336117954e-07, "loss": 0.3428, "step": 308 }, { "epoch": 2.472, "grad_norm": 0.43439558148384094, "learning_rate": 9.169043164835867e-07, "loss": 0.3336, "step": 309 }, { "epoch": 2.48, "grad_norm": 0.4950801730155945, "learning_rate": 8.901792342776439e-07, "loss": 0.4171, "step": 310 }, { "epoch": 2.488, "grad_norm": 0.4548616409301758, "learning_rate": 8.638113094946382e-07, "loss": 0.3123, "step": 311 }, { "epoch": 2.496, "grad_norm": 0.5122199058532715, "learning_rate": 8.378028335970451e-07, "loss": 0.3784, "step": 312 }, { "epoch": 2.504, "grad_norm": 0.452438622713089, "learning_rate": 8.121560668100065e-07, "loss": 0.3631, "step": 313 }, { "epoch": 2.512, "grad_norm": 0.43378862738609314, "learning_rate": 7.868732379249122e-07, "loss": 0.3319, "step": 314 }, { "epoch": 2.52, "grad_norm": 0.4786088466644287, "learning_rate": 7.619565441057075e-07, "loss": 0.3517, "step": 315 }, { "epoch": 2.528, "grad_norm": 0.5133551955223083, "learning_rate": 7.37408150697953e-07, "loss": 0.3702, "step": 316 }, { "epoch": 2.536, "grad_norm": 0.4935527741909027, "learning_rate": 7.132301910406503e-07, "loss": 0.3759, "step": 317 }, { "epoch": 2.544, "grad_norm": 0.4293310344219208, "learning_rate": 6.894247662808456e-07, "loss": 0.322, "step": 318 }, { "epoch": 2.552, "grad_norm": 0.46054044365882874, "learning_rate": 6.659939451910341e-07, "loss": 0.3343, "step": 319 }, { "epoch": 2.56, "grad_norm": 0.5166534185409546, "learning_rate": 6.429397639893758e-07, "loss": 0.3813, "step": 320 }, { "epoch": 2.568, "grad_norm": 0.4581873118877411, "learning_rate": 6.202642261627411e-07, "loss": 0.368, "step": 321 }, { "epoch": 2.576, "grad_norm": 0.5063138008117676, "learning_rate": 5.979693022926025e-07, "loss": 0.4142, "step": 322 }, { "epoch": 2.584, "grad_norm": 0.47579020261764526, "learning_rate": 5.760569298837825e-07, "loss": 0.3717, "step": 323 }, { "epoch": 2.592, "grad_norm": 0.4639912545681, "learning_rate": 5.54529013196079e-07, "loss": 0.352, "step": 324 }, { "epoch": 2.6, "grad_norm": 0.4613383710384369, "learning_rate": 5.333874230787772e-07, "loss": 0.3661, "step": 325 }, { "epoch": 2.608, "grad_norm": 0.42720070481300354, "learning_rate": 5.126339968080696e-07, "loss": 0.3616, "step": 326 }, { "epoch": 2.616, "grad_norm": 0.4943825304508209, "learning_rate": 4.922705379273862e-07, "loss": 0.3645, "step": 327 }, { "epoch": 2.624, "grad_norm": 0.4619237780570984, "learning_rate": 4.7229881609066387e-07, "loss": 0.3307, "step": 328 }, { "epoch": 2.632, "grad_norm": 0.47587689757347107, "learning_rate": 4.5272056690855494e-07, "loss": 0.3624, "step": 329 }, { "epoch": 2.64, "grad_norm": 0.45437976717948914, "learning_rate": 4.335374917975982e-07, "loss": 0.3242, "step": 330 }, { "epoch": 2.648, "grad_norm": 0.5222877264022827, "learning_rate": 4.147512578323615e-07, "loss": 0.3808, "step": 331 }, { "epoch": 2.656, "grad_norm": 0.46543389558792114, "learning_rate": 3.9636349760056427e-07, "loss": 0.3389, "step": 332 }, { "epoch": 2.664, "grad_norm": 0.4624124765396118, "learning_rate": 3.783758090611983e-07, "loss": 0.3336, "step": 333 }, { "epoch": 2.672, "grad_norm": 0.45994681119918823, "learning_rate": 3.6078975540566716e-07, "loss": 0.3482, "step": 334 }, { "epoch": 2.68, "grad_norm": 0.45068469643592834, "learning_rate": 3.4360686492193263e-07, "loss": 0.3077, "step": 335 }, { "epoch": 2.6879999999999997, "grad_norm": 0.5126630663871765, "learning_rate": 3.268286308617041e-07, "loss": 0.3906, "step": 336 }, { "epoch": 2.6959999999999997, "grad_norm": 0.4749516546726227, "learning_rate": 3.104565113106689e-07, "loss": 0.342, "step": 337 }, { "epoch": 2.7039999999999997, "grad_norm": 0.43455156683921814, "learning_rate": 2.9449192906178205e-07, "loss": 0.3491, "step": 338 }, { "epoch": 2.7119999999999997, "grad_norm": 0.46082979440689087, "learning_rate": 2.789362714916172e-07, "loss": 0.3875, "step": 339 }, { "epoch": 2.7199999999999998, "grad_norm": 0.44106754660606384, "learning_rate": 2.6379089043980064e-07, "loss": 0.32, "step": 340 }, { "epoch": 2.7279999999999998, "grad_norm": 0.46814629435539246, "learning_rate": 2.4905710209153224e-07, "loss": 0.3451, "step": 341 }, { "epoch": 2.7359999999999998, "grad_norm": 0.5077950358390808, "learning_rate": 2.3473618686320477e-07, "loss": 0.3861, "step": 342 }, { "epoch": 2.7439999999999998, "grad_norm": 0.49315568804740906, "learning_rate": 2.208293892911284e-07, "loss": 0.3469, "step": 343 }, { "epoch": 2.752, "grad_norm": 0.4634096920490265, "learning_rate": 2.0733791792338197e-07, "loss": 0.3761, "step": 344 }, { "epoch": 2.76, "grad_norm": 0.4556006193161011, "learning_rate": 1.9426294521477874e-07, "loss": 0.3504, "step": 345 }, { "epoch": 2.768, "grad_norm": 0.48851218819618225, "learning_rate": 1.8160560742498223e-07, "loss": 0.3652, "step": 346 }, { "epoch": 2.776, "grad_norm": 0.45232683420181274, "learning_rate": 1.6936700451975818e-07, "loss": 0.3259, "step": 347 }, { "epoch": 2.784, "grad_norm": 0.4734852910041809, "learning_rate": 1.5754820007538473e-07, "loss": 0.3555, "step": 348 }, { "epoch": 2.792, "grad_norm": 0.46094757318496704, "learning_rate": 1.461502211862237e-07, "loss": 0.3451, "step": 349 }, { "epoch": 2.8, "grad_norm": 0.45597511529922485, "learning_rate": 1.3517405837546404e-07, "loss": 0.3481, "step": 350 }, { "epoch": 2.808, "grad_norm": 0.44162362813949585, "learning_rate": 1.2462066550903818e-07, "loss": 0.3324, "step": 351 }, { "epoch": 2.816, "grad_norm": 0.4328787326812744, "learning_rate": 1.1449095971273305e-07, "loss": 0.3304, "step": 352 }, { "epoch": 2.824, "grad_norm": 0.472141832113266, "learning_rate": 1.0478582129248516e-07, "loss": 0.3655, "step": 353 }, { "epoch": 2.832, "grad_norm": 0.4979706406593323, "learning_rate": 9.550609365787888e-08, "loss": 0.3815, "step": 354 }, { "epoch": 2.84, "grad_norm": 0.49750280380249023, "learning_rate": 8.66525832488535e-08, "loss": 0.3656, "step": 355 }, { "epoch": 2.848, "grad_norm": 0.46096086502075195, "learning_rate": 7.822605946561923e-08, "loss": 0.3437, "step": 356 }, { "epoch": 2.856, "grad_norm": 0.5005020499229431, "learning_rate": 7.022725460179459e-08, "loss": 0.3697, "step": 357 }, { "epoch": 2.864, "grad_norm": 0.4813578426837921, "learning_rate": 6.265686378076729e-08, "loss": 0.3595, "step": 358 }, { "epoch": 2.872, "grad_norm": 0.4750673472881317, "learning_rate": 5.5515544895284324e-08, "loss": 0.3706, "step": 359 }, { "epoch": 2.88, "grad_norm": 0.48565149307250977, "learning_rate": 4.880391855028088e-08, "loss": 0.3524, "step": 360 }, { "epoch": 2.888, "grad_norm": 0.4671759307384491, "learning_rate": 4.252256800894694e-08, "loss": 0.3407, "step": 361 }, { "epoch": 2.896, "grad_norm": 0.4376312494277954, "learning_rate": 3.6672039142039426e-08, "loss": 0.3292, "step": 362 }, { "epoch": 2.904, "grad_norm": 0.4908435344696045, "learning_rate": 3.125284038044407e-08, "loss": 0.3429, "step": 363 }, { "epoch": 2.912, "grad_norm": 0.44629916548728943, "learning_rate": 2.6265442670991293e-08, "loss": 0.3428, "step": 364 }, { "epoch": 2.92, "grad_norm": 0.465819388628006, "learning_rate": 2.1710279435530058e-08, "loss": 0.3448, "step": 365 }, { "epoch": 2.928, "grad_norm": 0.4819665253162384, "learning_rate": 1.7587746533260786e-08, "loss": 0.3519, "step": 366 }, { "epoch": 2.936, "grad_norm": 0.4784250855445862, "learning_rate": 1.3898202226333424e-08, "loss": 0.3739, "step": 367 }, { "epoch": 2.944, "grad_norm": 0.495050847530365, "learning_rate": 1.0641967148716236e-08, "loss": 0.3628, "step": 368 }, { "epoch": 2.952, "grad_norm": 0.4823502004146576, "learning_rate": 7.819324278328099e-09, "loss": 0.3411, "step": 369 }, { "epoch": 2.96, "grad_norm": 0.41492488980293274, "learning_rate": 5.430518912448169e-09, "loss": 0.3326, "step": 370 }, { "epoch": 2.968, "grad_norm": 0.47112229466438293, "learning_rate": 3.4757586464001513e-09, "loss": 0.3696, "step": 371 }, { "epoch": 2.976, "grad_norm": 0.46793392300605774, "learning_rate": 1.9552133555084117e-09, "loss": 0.3661, "step": 372 }, { "epoch": 2.984, "grad_norm": 0.4385625720024109, "learning_rate": 8.690151803386615e-10, "loss": 0.3421, "step": 373 }, { "epoch": 2.992, "grad_norm": 0.43801286816596985, "learning_rate": 2.1725851521103847e-10, "loss": 0.3395, "step": 374 }, { "epoch": 3.0, "grad_norm": 0.44443294405937195, "learning_rate": 0.0, "loss": 0.3399, "step": 375 }, { "epoch": 3.0, "step": 375, "total_flos": 223704105418752.0, "train_loss": 0.4401542440255483, "train_runtime": 59266.2263, "train_samples_per_second": 0.607, "train_steps_per_second": 0.006 } ], "logging_steps": 1, "max_steps": 375, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 50, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 223704105418752.0, "train_batch_size": 1, "trial_name": null, "trial_params": null }