|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 3.0, |
|
"eval_steps": 500, |
|
"global_step": 375, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.008, |
|
"grad_norm": 14.56950855255127, |
|
"learning_rate": 2.6315789473684213e-07, |
|
"loss": 0.8339, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.016, |
|
"grad_norm": 13.73933219909668, |
|
"learning_rate": 5.263157894736843e-07, |
|
"loss": 0.8732, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.024, |
|
"grad_norm": 14.081198692321777, |
|
"learning_rate": 7.894736842105263e-07, |
|
"loss": 0.894, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.032, |
|
"grad_norm": 14.026392936706543, |
|
"learning_rate": 1.0526315789473685e-06, |
|
"loss": 0.8509, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 14.380779266357422, |
|
"learning_rate": 1.3157894736842106e-06, |
|
"loss": 0.8362, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.048, |
|
"grad_norm": 12.984580993652344, |
|
"learning_rate": 1.5789473684210526e-06, |
|
"loss": 0.8653, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.056, |
|
"grad_norm": 10.433991432189941, |
|
"learning_rate": 1.8421052631578948e-06, |
|
"loss": 0.8321, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.064, |
|
"grad_norm": 9.493083000183105, |
|
"learning_rate": 2.105263157894737e-06, |
|
"loss": 0.8143, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.072, |
|
"grad_norm": 7.640753746032715, |
|
"learning_rate": 2.368421052631579e-06, |
|
"loss": 0.7905, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 4.019260883331299, |
|
"learning_rate": 2.631578947368421e-06, |
|
"loss": 0.7129, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.088, |
|
"grad_norm": 3.2670819759368896, |
|
"learning_rate": 2.8947368421052634e-06, |
|
"loss": 0.7227, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.096, |
|
"grad_norm": 2.992112398147583, |
|
"learning_rate": 3.157894736842105e-06, |
|
"loss": 0.6972, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.104, |
|
"grad_norm": 2.052677869796753, |
|
"learning_rate": 3.421052631578948e-06, |
|
"loss": 0.678, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.112, |
|
"grad_norm": 1.920199990272522, |
|
"learning_rate": 3.6842105263157896e-06, |
|
"loss": 0.667, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 1.706879734992981, |
|
"learning_rate": 3.947368421052632e-06, |
|
"loss": 0.6192, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.128, |
|
"grad_norm": 1.6659389734268188, |
|
"learning_rate": 4.210526315789474e-06, |
|
"loss": 0.6118, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.136, |
|
"grad_norm": 1.3653850555419922, |
|
"learning_rate": 4.473684210526316e-06, |
|
"loss": 0.5681, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.144, |
|
"grad_norm": 1.4679189920425415, |
|
"learning_rate": 4.736842105263158e-06, |
|
"loss": 0.607, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.152, |
|
"grad_norm": 1.2237054109573364, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5273, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.9686553478240967, |
|
"learning_rate": 5.263157894736842e-06, |
|
"loss": 0.5578, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.168, |
|
"grad_norm": 0.7895737290382385, |
|
"learning_rate": 5.526315789473685e-06, |
|
"loss": 0.5794, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.176, |
|
"grad_norm": 0.8087520599365234, |
|
"learning_rate": 5.789473684210527e-06, |
|
"loss": 0.5229, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.184, |
|
"grad_norm": 0.7652890086174011, |
|
"learning_rate": 6.0526315789473685e-06, |
|
"loss": 0.5292, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.192, |
|
"grad_norm": 0.9132710695266724, |
|
"learning_rate": 6.31578947368421e-06, |
|
"loss": 0.5271, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.8860102295875549, |
|
"learning_rate": 6.578947368421054e-06, |
|
"loss": 0.5155, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.208, |
|
"grad_norm": 0.8902170062065125, |
|
"learning_rate": 6.842105263157896e-06, |
|
"loss": 0.5287, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.216, |
|
"grad_norm": 0.8824247121810913, |
|
"learning_rate": 7.1052631578947375e-06, |
|
"loss": 0.4791, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.224, |
|
"grad_norm": 1.2683454751968384, |
|
"learning_rate": 7.368421052631579e-06, |
|
"loss": 0.5329, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.232, |
|
"grad_norm": 0.7248997092247009, |
|
"learning_rate": 7.631578947368423e-06, |
|
"loss": 0.5294, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.7209534645080566, |
|
"learning_rate": 7.894736842105265e-06, |
|
"loss": 0.4842, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.248, |
|
"grad_norm": 0.7555510401725769, |
|
"learning_rate": 8.157894736842106e-06, |
|
"loss": 0.5348, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.256, |
|
"grad_norm": 0.6169366240501404, |
|
"learning_rate": 8.421052631578948e-06, |
|
"loss": 0.5382, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.264, |
|
"grad_norm": 0.6152195930480957, |
|
"learning_rate": 8.68421052631579e-06, |
|
"loss": 0.5039, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.272, |
|
"grad_norm": 0.631288468837738, |
|
"learning_rate": 8.947368421052632e-06, |
|
"loss": 0.5037, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.5909512042999268, |
|
"learning_rate": 9.210526315789474e-06, |
|
"loss": 0.5017, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.288, |
|
"grad_norm": 0.5489953756332397, |
|
"learning_rate": 9.473684210526315e-06, |
|
"loss": 0.4469, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.296, |
|
"grad_norm": 0.565674901008606, |
|
"learning_rate": 9.736842105263159e-06, |
|
"loss": 0.478, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.304, |
|
"grad_norm": 0.680474579334259, |
|
"learning_rate": 1e-05, |
|
"loss": 0.5392, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.312, |
|
"grad_norm": 0.607657253742218, |
|
"learning_rate": 9.99978274148479e-06, |
|
"loss": 0.5022, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.5720808506011963, |
|
"learning_rate": 9.999130984819662e-06, |
|
"loss": 0.5057, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.328, |
|
"grad_norm": 0.5648075938224792, |
|
"learning_rate": 9.998044786644492e-06, |
|
"loss": 0.4718, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.336, |
|
"grad_norm": 0.5617280006408691, |
|
"learning_rate": 9.9965242413536e-06, |
|
"loss": 0.5049, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.344, |
|
"grad_norm": 0.6486408114433289, |
|
"learning_rate": 9.994569481087552e-06, |
|
"loss": 0.5203, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.352, |
|
"grad_norm": 0.5352612733840942, |
|
"learning_rate": 9.992180675721671e-06, |
|
"loss": 0.4967, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.5438513159751892, |
|
"learning_rate": 9.989358032851283e-06, |
|
"loss": 0.5188, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.368, |
|
"grad_norm": 0.5454742908477783, |
|
"learning_rate": 9.986101797773667e-06, |
|
"loss": 0.4864, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.376, |
|
"grad_norm": 0.536598265171051, |
|
"learning_rate": 9.98241225346674e-06, |
|
"loss": 0.4883, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.384, |
|
"grad_norm": 0.4659676253795624, |
|
"learning_rate": 9.978289720564471e-06, |
|
"loss": 0.5121, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.392, |
|
"grad_norm": 0.5836896300315857, |
|
"learning_rate": 9.97373455732901e-06, |
|
"loss": 0.4975, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.5226072669029236, |
|
"learning_rate": 9.968747159619556e-06, |
|
"loss": 0.4944, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.408, |
|
"grad_norm": 0.5444321632385254, |
|
"learning_rate": 9.963327960857962e-06, |
|
"loss": 0.5209, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.416, |
|
"grad_norm": 0.5670180320739746, |
|
"learning_rate": 9.957477431991053e-06, |
|
"loss": 0.514, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.424, |
|
"grad_norm": 0.511341392993927, |
|
"learning_rate": 9.95119608144972e-06, |
|
"loss": 0.526, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.432, |
|
"grad_norm": 0.5060410499572754, |
|
"learning_rate": 9.944484455104716e-06, |
|
"loss": 0.4806, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.487128347158432, |
|
"learning_rate": 9.937343136219234e-06, |
|
"loss": 0.5273, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.448, |
|
"grad_norm": 0.5598118305206299, |
|
"learning_rate": 9.929772745398207e-06, |
|
"loss": 0.5745, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.456, |
|
"grad_norm": 0.5080751180648804, |
|
"learning_rate": 9.921773940534382e-06, |
|
"loss": 0.5082, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.464, |
|
"grad_norm": 0.5492854118347168, |
|
"learning_rate": 9.913347416751148e-06, |
|
"loss": 0.5244, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.472, |
|
"grad_norm": 0.5565740466117859, |
|
"learning_rate": 9.904493906342124e-06, |
|
"loss": 0.5117, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.5090964436531067, |
|
"learning_rate": 9.895214178707516e-06, |
|
"loss": 0.4942, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.488, |
|
"grad_norm": 0.5151892304420471, |
|
"learning_rate": 9.885509040287267e-06, |
|
"loss": 0.5165, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.496, |
|
"grad_norm": 0.5298195481300354, |
|
"learning_rate": 9.875379334490962e-06, |
|
"loss": 0.5218, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.504, |
|
"grad_norm": 0.5165826082229614, |
|
"learning_rate": 9.864825941624538e-06, |
|
"loss": 0.4986, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.512, |
|
"grad_norm": 0.506293773651123, |
|
"learning_rate": 9.853849778813777e-06, |
|
"loss": 0.5127, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 0.5140017867088318, |
|
"learning_rate": 9.842451799924616e-06, |
|
"loss": 0.5158, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.528, |
|
"grad_norm": 0.5674154162406921, |
|
"learning_rate": 9.830632995480243e-06, |
|
"loss": 0.5006, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.536, |
|
"grad_norm": 0.5346920490264893, |
|
"learning_rate": 9.818394392575018e-06, |
|
"loss": 0.5126, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.544, |
|
"grad_norm": 0.5248146653175354, |
|
"learning_rate": 9.805737054785223e-06, |
|
"loss": 0.5394, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.552, |
|
"grad_norm": 0.5473419427871704, |
|
"learning_rate": 9.792662082076618e-06, |
|
"loss": 0.5224, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 0.5517472624778748, |
|
"learning_rate": 9.779170610708872e-06, |
|
"loss": 0.5125, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.568, |
|
"grad_norm": 0.4769006669521332, |
|
"learning_rate": 9.765263813136796e-06, |
|
"loss": 0.5081, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.576, |
|
"grad_norm": 0.5075954794883728, |
|
"learning_rate": 9.750942897908468e-06, |
|
"loss": 0.4958, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.584, |
|
"grad_norm": 0.5006887316703796, |
|
"learning_rate": 9.736209109560201e-06, |
|
"loss": 0.5279, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.592, |
|
"grad_norm": 0.5326476693153381, |
|
"learning_rate": 9.721063728508384e-06, |
|
"loss": 0.5136, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 0.5512110590934753, |
|
"learning_rate": 9.705508070938219e-06, |
|
"loss": 0.474, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.608, |
|
"grad_norm": 0.5133175253868103, |
|
"learning_rate": 9.689543488689332e-06, |
|
"loss": 0.4994, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.616, |
|
"grad_norm": 0.5255376696586609, |
|
"learning_rate": 9.673171369138297e-06, |
|
"loss": 0.5183, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.624, |
|
"grad_norm": 0.5707758665084839, |
|
"learning_rate": 9.656393135078067e-06, |
|
"loss": 0.493, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.632, |
|
"grad_norm": 0.5201893448829651, |
|
"learning_rate": 9.639210244594335e-06, |
|
"loss": 0.4635, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 0.5058737397193909, |
|
"learning_rate": 9.621624190938802e-06, |
|
"loss": 0.5312, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.648, |
|
"grad_norm": 0.43761327862739563, |
|
"learning_rate": 9.603636502399436e-06, |
|
"loss": 0.4524, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.656, |
|
"grad_norm": 0.5335831642150879, |
|
"learning_rate": 9.585248742167638e-06, |
|
"loss": 0.4648, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.664, |
|
"grad_norm": 0.5373964309692383, |
|
"learning_rate": 9.566462508202403e-06, |
|
"loss": 0.4911, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.672, |
|
"grad_norm": 0.5256152153015137, |
|
"learning_rate": 9.547279433091446e-06, |
|
"loss": 0.4968, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 0.5257714986801147, |
|
"learning_rate": 9.527701183909336e-06, |
|
"loss": 0.4879, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.688, |
|
"grad_norm": 0.5462167859077454, |
|
"learning_rate": 9.507729462072615e-06, |
|
"loss": 0.4577, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.696, |
|
"grad_norm": 0.5356786847114563, |
|
"learning_rate": 9.48736600319193e-06, |
|
"loss": 0.4922, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.704, |
|
"grad_norm": 0.5588122010231018, |
|
"learning_rate": 9.466612576921223e-06, |
|
"loss": 0.4915, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.712, |
|
"grad_norm": 0.5467613935470581, |
|
"learning_rate": 9.445470986803922e-06, |
|
"loss": 0.4711, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 0.5490957498550415, |
|
"learning_rate": 9.423943070116219e-06, |
|
"loss": 0.5108, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.728, |
|
"grad_norm": 0.4839731752872467, |
|
"learning_rate": 9.402030697707398e-06, |
|
"loss": 0.5128, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.736, |
|
"grad_norm": 0.5173898935317993, |
|
"learning_rate": 9.37973577383726e-06, |
|
"loss": 0.4823, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.744, |
|
"grad_norm": 0.563199520111084, |
|
"learning_rate": 9.357060236010626e-06, |
|
"loss": 0.4985, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.752, |
|
"grad_norm": 0.513721227645874, |
|
"learning_rate": 9.334006054808966e-06, |
|
"loss": 0.4484, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 0.5133054256439209, |
|
"learning_rate": 9.310575233719155e-06, |
|
"loss": 0.5047, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.768, |
|
"grad_norm": 0.5787802338600159, |
|
"learning_rate": 9.28676980895935e-06, |
|
"loss": 0.5064, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.776, |
|
"grad_norm": 0.4748121500015259, |
|
"learning_rate": 9.262591849302049e-06, |
|
"loss": 0.4604, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.784, |
|
"grad_norm": 0.556131899356842, |
|
"learning_rate": 9.238043455894294e-06, |
|
"loss": 0.5295, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.792, |
|
"grad_norm": 0.5900002717971802, |
|
"learning_rate": 9.213126762075088e-06, |
|
"loss": 0.4977, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 0.5788969397544861, |
|
"learning_rate": 9.187843933189994e-06, |
|
"loss": 0.4732, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.808, |
|
"grad_norm": 0.5433998107910156, |
|
"learning_rate": 9.162197166402957e-06, |
|
"loss": 0.5175, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.816, |
|
"grad_norm": 0.5408159494400024, |
|
"learning_rate": 9.136188690505363e-06, |
|
"loss": 0.5087, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.824, |
|
"grad_norm": 0.5552597045898438, |
|
"learning_rate": 9.109820765722357e-06, |
|
"loss": 0.4835, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.832, |
|
"grad_norm": 0.5368766188621521, |
|
"learning_rate": 9.083095683516414e-06, |
|
"loss": 0.5349, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 0.46206405758857727, |
|
"learning_rate": 9.056015766388205e-06, |
|
"loss": 0.4567, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.848, |
|
"grad_norm": 0.5309408903121948, |
|
"learning_rate": 9.028583367674767e-06, |
|
"loss": 0.5166, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.856, |
|
"grad_norm": 0.528049886226654, |
|
"learning_rate": 9.00080087134498e-06, |
|
"loss": 0.4526, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.864, |
|
"grad_norm": 0.5379208326339722, |
|
"learning_rate": 8.972670691792409e-06, |
|
"loss": 0.4893, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.872, |
|
"grad_norm": 0.5074729919433594, |
|
"learning_rate": 8.944195273625472e-06, |
|
"loss": 0.4924, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 0.5556723475456238, |
|
"learning_rate": 8.915377091454992e-06, |
|
"loss": 0.4981, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.888, |
|
"grad_norm": 0.5345364212989807, |
|
"learning_rate": 8.886218649679162e-06, |
|
"loss": 0.5134, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.896, |
|
"grad_norm": 0.511635959148407, |
|
"learning_rate": 8.856722482265886e-06, |
|
"loss": 0.461, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.904, |
|
"grad_norm": 0.49792084097862244, |
|
"learning_rate": 8.826891152532579e-06, |
|
"loss": 0.505, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.912, |
|
"grad_norm": 0.4829058051109314, |
|
"learning_rate": 8.796727252923403e-06, |
|
"loss": 0.4716, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 0.5941436886787415, |
|
"learning_rate": 8.766233404783975e-06, |
|
"loss": 0.5296, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.928, |
|
"grad_norm": 0.5134578943252563, |
|
"learning_rate": 8.735412258133562e-06, |
|
"loss": 0.5045, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.936, |
|
"grad_norm": 0.5325246453285217, |
|
"learning_rate": 8.704266491434787e-06, |
|
"loss": 0.472, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.944, |
|
"grad_norm": 0.538865327835083, |
|
"learning_rate": 8.672798811360863e-06, |
|
"loss": 0.5026, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.952, |
|
"grad_norm": 0.48347458243370056, |
|
"learning_rate": 8.641011952560372e-06, |
|
"loss": 0.484, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 0.5306034684181213, |
|
"learning_rate": 8.608908677419606e-06, |
|
"loss": 0.4997, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.968, |
|
"grad_norm": 0.5399216413497925, |
|
"learning_rate": 8.576491775822527e-06, |
|
"loss": 0.4804, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.976, |
|
"grad_norm": 0.5183812975883484, |
|
"learning_rate": 8.543764064908295e-06, |
|
"loss": 0.4887, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.984, |
|
"grad_norm": 0.4991128444671631, |
|
"learning_rate": 8.510728388826464e-06, |
|
"loss": 0.4453, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.992, |
|
"grad_norm": 0.5225016474723816, |
|
"learning_rate": 8.477387618489808e-06, |
|
"loss": 0.5123, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.46465304493904114, |
|
"learning_rate": 8.443744651324828e-06, |
|
"loss": 0.4633, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 1.008, |
|
"grad_norm": 0.5113005638122559, |
|
"learning_rate": 8.409802411019962e-06, |
|
"loss": 0.427, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 1.016, |
|
"grad_norm": 0.5328509211540222, |
|
"learning_rate": 8.375563847271506e-06, |
|
"loss": 0.4324, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 1.024, |
|
"grad_norm": 0.519523024559021, |
|
"learning_rate": 8.341031935527267e-06, |
|
"loss": 0.4387, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 1.032, |
|
"grad_norm": 0.506823718547821, |
|
"learning_rate": 8.306209676727994e-06, |
|
"loss": 0.4469, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"grad_norm": 0.4721229374408722, |
|
"learning_rate": 8.271100097046585e-06, |
|
"loss": 0.4226, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 1.048, |
|
"grad_norm": 0.4656391739845276, |
|
"learning_rate": 8.235706247625098e-06, |
|
"loss": 0.3729, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 1.056, |
|
"grad_norm": 0.4901474118232727, |
|
"learning_rate": 8.200031204309604e-06, |
|
"loss": 0.4627, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 1.064, |
|
"grad_norm": 0.606958270072937, |
|
"learning_rate": 8.16407806738288e-06, |
|
"loss": 0.4347, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 1.072, |
|
"grad_norm": 0.4573584496974945, |
|
"learning_rate": 8.127849961294984e-06, |
|
"loss": 0.402, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"grad_norm": 0.4766193926334381, |
|
"learning_rate": 8.091350034391732e-06, |
|
"loss": 0.4203, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 1.088, |
|
"grad_norm": 0.5167376399040222, |
|
"learning_rate": 8.05458145864109e-06, |
|
"loss": 0.4521, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 1.096, |
|
"grad_norm": 0.52271968126297, |
|
"learning_rate": 8.017547429357532e-06, |
|
"loss": 0.4321, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 1.104, |
|
"grad_norm": 0.49505582451820374, |
|
"learning_rate": 7.980251164924342e-06, |
|
"loss": 0.4591, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 1.112, |
|
"grad_norm": 0.6967505216598511, |
|
"learning_rate": 7.94269590651393e-06, |
|
"loss": 0.4322, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"grad_norm": 0.46629568934440613, |
|
"learning_rate": 7.904884917806174e-06, |
|
"loss": 0.4612, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 1.1280000000000001, |
|
"grad_norm": 0.49921801686286926, |
|
"learning_rate": 7.866821484704777e-06, |
|
"loss": 0.4353, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 1.1360000000000001, |
|
"grad_norm": 0.5447902083396912, |
|
"learning_rate": 7.828508915051724e-06, |
|
"loss": 0.4323, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 1.144, |
|
"grad_norm": 0.5017483830451965, |
|
"learning_rate": 7.789950538339813e-06, |
|
"loss": 0.4352, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 1.152, |
|
"grad_norm": 0.49647992849349976, |
|
"learning_rate": 7.751149705423313e-06, |
|
"loss": 0.3971, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"grad_norm": 0.5001079440116882, |
|
"learning_rate": 7.712109788226763e-06, |
|
"loss": 0.4566, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 1.168, |
|
"grad_norm": 0.4786428213119507, |
|
"learning_rate": 7.672834179451943e-06, |
|
"loss": 0.446, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 1.176, |
|
"grad_norm": 0.5244361758232117, |
|
"learning_rate": 7.633326292283028e-06, |
|
"loss": 0.4759, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 1.184, |
|
"grad_norm": 0.4763104319572449, |
|
"learning_rate": 7.593589560089984e-06, |
|
"loss": 0.4239, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 1.192, |
|
"grad_norm": 0.4974442422389984, |
|
"learning_rate": 7.553627436130183e-06, |
|
"loss": 0.4053, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"grad_norm": 0.5250166654586792, |
|
"learning_rate": 7.513443393248312e-06, |
|
"loss": 0.4521, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 1.208, |
|
"grad_norm": 0.4822094440460205, |
|
"learning_rate": 7.473040923574567e-06, |
|
"loss": 0.3883, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 1.216, |
|
"grad_norm": 0.4862370491027832, |
|
"learning_rate": 7.432423538221179e-06, |
|
"loss": 0.4322, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 1.224, |
|
"grad_norm": 0.482036828994751, |
|
"learning_rate": 7.391594766977277e-06, |
|
"loss": 0.4105, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 1.232, |
|
"grad_norm": 0.5110259056091309, |
|
"learning_rate": 7.350558158002154e-06, |
|
"loss": 0.4103, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"grad_norm": 0.5026330351829529, |
|
"learning_rate": 7.3093172775169e-06, |
|
"loss": 0.4312, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 1.248, |
|
"grad_norm": 0.4886966943740845, |
|
"learning_rate": 7.2678757094945e-06, |
|
"loss": 0.4325, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 1.256, |
|
"grad_norm": 0.4985504150390625, |
|
"learning_rate": 7.226237055348369e-06, |
|
"loss": 0.4232, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 1.264, |
|
"grad_norm": 0.47912245988845825, |
|
"learning_rate": 7.184404933619377e-06, |
|
"loss": 0.4359, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 1.272, |
|
"grad_norm": 0.48068177700042725, |
|
"learning_rate": 7.142382979661386e-06, |
|
"loss": 0.4158, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"grad_norm": 0.4839794933795929, |
|
"learning_rate": 7.100174845325327e-06, |
|
"loss": 0.4035, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 1.288, |
|
"grad_norm": 0.5554186105728149, |
|
"learning_rate": 7.057784198641835e-06, |
|
"loss": 0.3862, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 1.296, |
|
"grad_norm": 0.5477203726768494, |
|
"learning_rate": 7.015214723502496e-06, |
|
"loss": 0.4347, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 1.304, |
|
"grad_norm": 0.48494377732276917, |
|
"learning_rate": 6.972470119339692e-06, |
|
"loss": 0.4183, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 1.312, |
|
"grad_norm": 0.4876761734485626, |
|
"learning_rate": 6.929554100805118e-06, |
|
"loss": 0.4459, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"grad_norm": 0.4854036867618561, |
|
"learning_rate": 6.886470397446958e-06, |
|
"loss": 0.4249, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 1.328, |
|
"grad_norm": 0.4709065556526184, |
|
"learning_rate": 6.843222753385785e-06, |
|
"loss": 0.4376, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 1.336, |
|
"grad_norm": 0.4834536910057068, |
|
"learning_rate": 6.799814926989171e-06, |
|
"loss": 0.4017, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 1.3439999999999999, |
|
"grad_norm": 0.4833991527557373, |
|
"learning_rate": 6.756250690545079e-06, |
|
"loss": 0.4145, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 1.3519999999999999, |
|
"grad_norm": 0.5171043872833252, |
|
"learning_rate": 6.712533829934042e-06, |
|
"loss": 0.4373, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 1.3599999999999999, |
|
"grad_norm": 0.5456111431121826, |
|
"learning_rate": 6.6686681443001485e-06, |
|
"loss": 0.4272, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 1.3679999999999999, |
|
"grad_norm": 0.5275494456291199, |
|
"learning_rate": 6.62465744572089e-06, |
|
"loss": 0.4628, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 1.376, |
|
"grad_norm": 0.46399417519569397, |
|
"learning_rate": 6.580505558875878e-06, |
|
"loss": 0.481, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 1.384, |
|
"grad_norm": 0.4880594313144684, |
|
"learning_rate": 6.536216320714466e-06, |
|
"loss": 0.4285, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 1.392, |
|
"grad_norm": 0.48990365862846375, |
|
"learning_rate": 6.491793580122301e-06, |
|
"loss": 0.3997, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"grad_norm": 0.49995189905166626, |
|
"learning_rate": 6.447241197586847e-06, |
|
"loss": 0.4332, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 1.408, |
|
"grad_norm": 0.4638576805591583, |
|
"learning_rate": 6.402563044861899e-06, |
|
"loss": 0.4236, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 1.416, |
|
"grad_norm": 0.49920183420181274, |
|
"learning_rate": 6.357763004631104e-06, |
|
"loss": 0.4349, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 1.424, |
|
"grad_norm": 0.5193583965301514, |
|
"learning_rate": 6.312844970170551e-06, |
|
"loss": 0.4089, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 1.432, |
|
"grad_norm": 0.47679874300956726, |
|
"learning_rate": 6.267812845010431e-06, |
|
"loss": 0.4298, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"grad_norm": 0.49884992837905884, |
|
"learning_rate": 6.2226705425958e-06, |
|
"loss": 0.4517, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 1.448, |
|
"grad_norm": 0.48136106133461, |
|
"learning_rate": 6.177421985946499e-06, |
|
"loss": 0.4393, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 1.456, |
|
"grad_norm": 0.5779682397842407, |
|
"learning_rate": 6.132071107316221e-06, |
|
"loss": 0.4225, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 1.464, |
|
"grad_norm": 0.5509202480316162, |
|
"learning_rate": 6.0866218478507875e-06, |
|
"loss": 0.4579, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 1.472, |
|
"grad_norm": 0.5101506114006042, |
|
"learning_rate": 6.041078157245649e-06, |
|
"loss": 0.4036, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"grad_norm": 0.623524010181427, |
|
"learning_rate": 5.995443993402647e-06, |
|
"loss": 0.451, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 1.488, |
|
"grad_norm": 0.4368286728858948, |
|
"learning_rate": 5.949723322086053e-06, |
|
"loss": 0.3881, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 1.496, |
|
"grad_norm": 0.4305906295776367, |
|
"learning_rate": 5.9039201165779315e-06, |
|
"loss": 0.4036, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 1.504, |
|
"grad_norm": 0.4749639630317688, |
|
"learning_rate": 5.858038357332851e-06, |
|
"loss": 0.4048, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 1.512, |
|
"grad_norm": 0.48530155420303345, |
|
"learning_rate": 5.812082031631966e-06, |
|
"loss": 0.4464, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"grad_norm": 0.555038571357727, |
|
"learning_rate": 5.766055133236513e-06, |
|
"loss": 0.4485, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 1.528, |
|
"grad_norm": 0.4892346262931824, |
|
"learning_rate": 5.7199616620407325e-06, |
|
"loss": 0.4522, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 1.536, |
|
"grad_norm": 0.5365411639213562, |
|
"learning_rate": 5.673805623724272e-06, |
|
"loss": 0.4752, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 1.544, |
|
"grad_norm": 0.5186147689819336, |
|
"learning_rate": 5.627591029404072e-06, |
|
"loss": 0.4021, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 1.552, |
|
"grad_norm": 0.48625391721725464, |
|
"learning_rate": 5.581321895285787e-06, |
|
"loss": 0.4396, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"grad_norm": 0.5071629285812378, |
|
"learning_rate": 5.535002242314772e-06, |
|
"loss": 0.4227, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 1.568, |
|
"grad_norm": 0.5189520120620728, |
|
"learning_rate": 5.488636095826636e-06, |
|
"loss": 0.4512, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 1.576, |
|
"grad_norm": 0.5435962677001953, |
|
"learning_rate": 5.4422274851974356e-06, |
|
"loss": 0.4767, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 1.584, |
|
"grad_norm": 0.4914834797382355, |
|
"learning_rate": 5.395780443493508e-06, |
|
"loss": 0.4035, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 1.592, |
|
"grad_norm": 0.5276004672050476, |
|
"learning_rate": 5.34929900712098e-06, |
|
"loss": 0.4329, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"grad_norm": 0.4546476900577545, |
|
"learning_rate": 5.302787215474992e-06, |
|
"loss": 0.3965, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 1.608, |
|
"grad_norm": 0.48372510075569153, |
|
"learning_rate": 5.256249110588659e-06, |
|
"loss": 0.4272, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 1.616, |
|
"grad_norm": 0.49213626980781555, |
|
"learning_rate": 5.209688736781811e-06, |
|
"loss": 0.4409, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 1.624, |
|
"grad_norm": 0.4945046603679657, |
|
"learning_rate": 5.163110140309518e-06, |
|
"loss": 0.4419, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 1.6320000000000001, |
|
"grad_norm": 0.4468039274215698, |
|
"learning_rate": 5.116517369010467e-06, |
|
"loss": 0.398, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 1.6400000000000001, |
|
"grad_norm": 0.4727765917778015, |
|
"learning_rate": 5.069914471955179e-06, |
|
"loss": 0.4229, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 1.6480000000000001, |
|
"grad_norm": 0.5005367398262024, |
|
"learning_rate": 5.023305499094145e-06, |
|
"loss": 0.4153, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 1.6560000000000001, |
|
"grad_norm": 0.4818647801876068, |
|
"learning_rate": 4.976694500905858e-06, |
|
"loss": 0.443, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 1.6640000000000001, |
|
"grad_norm": 0.47024303674697876, |
|
"learning_rate": 4.930085528044823e-06, |
|
"loss": 0.4541, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 1.6720000000000002, |
|
"grad_norm": 0.47553378343582153, |
|
"learning_rate": 4.883482630989536e-06, |
|
"loss": 0.4069, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 1.6800000000000002, |
|
"grad_norm": 0.44939181208610535, |
|
"learning_rate": 4.8368898596904834e-06, |
|
"loss": 0.3964, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 1.688, |
|
"grad_norm": 0.46793460845947266, |
|
"learning_rate": 4.790311263218191e-06, |
|
"loss": 0.4339, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 1.696, |
|
"grad_norm": 0.45191749930381775, |
|
"learning_rate": 4.743750889411342e-06, |
|
"loss": 0.426, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 1.704, |
|
"grad_norm": 0.45308157801628113, |
|
"learning_rate": 4.697212784525009e-06, |
|
"loss": 0.4338, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 1.712, |
|
"grad_norm": 0.4543735384941101, |
|
"learning_rate": 4.65070099287902e-06, |
|
"loss": 0.4083, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"grad_norm": 0.49145424365997314, |
|
"learning_rate": 4.604219556506492e-06, |
|
"loss": 0.4658, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 1.728, |
|
"grad_norm": 0.4785975515842438, |
|
"learning_rate": 4.557772514802564e-06, |
|
"loss": 0.4248, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 1.736, |
|
"grad_norm": 0.5019366145133972, |
|
"learning_rate": 4.511363904173366e-06, |
|
"loss": 0.4482, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 1.744, |
|
"grad_norm": 0.4578736126422882, |
|
"learning_rate": 4.46499775768523e-06, |
|
"loss": 0.4391, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 1.752, |
|
"grad_norm": 0.46869972348213196, |
|
"learning_rate": 4.418678104714214e-06, |
|
"loss": 0.4087, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"grad_norm": 0.480589359998703, |
|
"learning_rate": 4.372408970595931e-06, |
|
"loss": 0.418, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 1.768, |
|
"grad_norm": 0.5114148855209351, |
|
"learning_rate": 4.326194376275729e-06, |
|
"loss": 0.4009, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 1.776, |
|
"grad_norm": 0.4714803099632263, |
|
"learning_rate": 4.280038337959268e-06, |
|
"loss": 0.4119, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 1.784, |
|
"grad_norm": 0.4820714592933655, |
|
"learning_rate": 4.2339448667634885e-06, |
|
"loss": 0.4364, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 1.792, |
|
"grad_norm": 0.45575451850891113, |
|
"learning_rate": 4.187917968368036e-06, |
|
"loss": 0.3943, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"grad_norm": 0.4869477152824402, |
|
"learning_rate": 4.141961642667152e-06, |
|
"loss": 0.4447, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 1.808, |
|
"grad_norm": 0.4546818435192108, |
|
"learning_rate": 4.09607988342207e-06, |
|
"loss": 0.4397, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 1.8159999999999998, |
|
"grad_norm": 0.5653818249702454, |
|
"learning_rate": 4.0502766779139485e-06, |
|
"loss": 0.3881, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 1.8239999999999998, |
|
"grad_norm": 0.484779953956604, |
|
"learning_rate": 4.0045560065973535e-06, |
|
"loss": 0.4673, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 1.8319999999999999, |
|
"grad_norm": 0.48289933800697327, |
|
"learning_rate": 3.958921842754351e-06, |
|
"loss": 0.4156, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 1.8399999999999999, |
|
"grad_norm": 0.43402060866355896, |
|
"learning_rate": 3.913378152149214e-06, |
|
"loss": 0.4044, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 1.8479999999999999, |
|
"grad_norm": 0.47271519899368286, |
|
"learning_rate": 3.86792889268378e-06, |
|
"loss": 0.4387, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 1.8559999999999999, |
|
"grad_norm": 0.5104248523712158, |
|
"learning_rate": 3.8225780140535025e-06, |
|
"loss": 0.3976, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 1.8639999999999999, |
|
"grad_norm": 0.44085070490837097, |
|
"learning_rate": 3.777329457404202e-06, |
|
"loss": 0.4062, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 1.8719999999999999, |
|
"grad_norm": 0.5098981857299805, |
|
"learning_rate": 3.7321871549895715e-06, |
|
"loss": 0.4458, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"grad_norm": 0.42704570293426514, |
|
"learning_rate": 3.68715502982945e-06, |
|
"loss": 0.3947, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 1.888, |
|
"grad_norm": 0.46332859992980957, |
|
"learning_rate": 3.6422369953688973e-06, |
|
"loss": 0.4294, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 1.896, |
|
"grad_norm": 0.4884559214115143, |
|
"learning_rate": 3.5974369551381023e-06, |
|
"loss": 0.4498, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 1.904, |
|
"grad_norm": 0.5290793180465698, |
|
"learning_rate": 3.5527588024131542e-06, |
|
"loss": 0.4312, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 1.912, |
|
"grad_norm": 0.48744434118270874, |
|
"learning_rate": 3.5082064198777e-06, |
|
"loss": 0.4772, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"grad_norm": 0.4615407884120941, |
|
"learning_rate": 3.463783679285535e-06, |
|
"loss": 0.4101, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 1.928, |
|
"grad_norm": 0.49090084433555603, |
|
"learning_rate": 3.4194944411241213e-06, |
|
"loss": 0.4222, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 1.936, |
|
"grad_norm": 0.4603979289531708, |
|
"learning_rate": 3.3753425542791106e-06, |
|
"loss": 0.4322, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 1.944, |
|
"grad_norm": 0.4630926847457886, |
|
"learning_rate": 3.3313318556998523e-06, |
|
"loss": 0.4553, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 1.952, |
|
"grad_norm": 0.5014007091522217, |
|
"learning_rate": 3.2874661700659586e-06, |
|
"loss": 0.4541, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"grad_norm": 0.5217999815940857, |
|
"learning_rate": 3.2437493094549223e-06, |
|
"loss": 0.4164, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 1.968, |
|
"grad_norm": 0.5072699785232544, |
|
"learning_rate": 3.200185073010831e-06, |
|
"loss": 0.4151, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 1.976, |
|
"grad_norm": 0.4602278172969818, |
|
"learning_rate": 3.1567772466142156e-06, |
|
"loss": 0.3739, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 1.984, |
|
"grad_norm": 0.5061410069465637, |
|
"learning_rate": 3.1135296025530426e-06, |
|
"loss": 0.4892, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 1.992, |
|
"grad_norm": 0.5533497333526611, |
|
"learning_rate": 3.070445899194885e-06, |
|
"loss": 0.4132, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 0.52688068151474, |
|
"learning_rate": 3.0275298806603102e-06, |
|
"loss": 0.4463, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 2.008, |
|
"grad_norm": 0.5126904249191284, |
|
"learning_rate": 2.984785276497507e-06, |
|
"loss": 0.3639, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 2.016, |
|
"grad_norm": 0.5104655623435974, |
|
"learning_rate": 2.9422158013581658e-06, |
|
"loss": 0.3535, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 2.024, |
|
"grad_norm": 0.5136793255805969, |
|
"learning_rate": 2.899825154674674e-06, |
|
"loss": 0.3835, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 2.032, |
|
"grad_norm": 0.45347532629966736, |
|
"learning_rate": 2.8576170203386144e-06, |
|
"loss": 0.3412, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"grad_norm": 0.4745730757713318, |
|
"learning_rate": 2.8155950663806234e-06, |
|
"loss": 0.3863, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 2.048, |
|
"grad_norm": 0.5067371129989624, |
|
"learning_rate": 2.7737629446516325e-06, |
|
"loss": 0.3514, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 2.056, |
|
"grad_norm": 0.45462149381637573, |
|
"learning_rate": 2.732124290505501e-06, |
|
"loss": 0.3528, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 2.064, |
|
"grad_norm": 0.46529221534729004, |
|
"learning_rate": 2.6906827224831024e-06, |
|
"loss": 0.3296, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 2.072, |
|
"grad_norm": 0.5160210728645325, |
|
"learning_rate": 2.6494418419978485e-06, |
|
"loss": 0.3416, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"grad_norm": 0.505672037601471, |
|
"learning_rate": 2.608405233022724e-06, |
|
"loss": 0.3793, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 2.088, |
|
"grad_norm": 0.44298309087753296, |
|
"learning_rate": 2.5675764617788233e-06, |
|
"loss": 0.3298, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 2.096, |
|
"grad_norm": 0.4680778682231903, |
|
"learning_rate": 2.526959076425434e-06, |
|
"loss": 0.3427, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 2.104, |
|
"grad_norm": 0.4816036820411682, |
|
"learning_rate": 2.4865566067516896e-06, |
|
"loss": 0.3709, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 2.112, |
|
"grad_norm": 0.47969356179237366, |
|
"learning_rate": 2.4463725638698182e-06, |
|
"loss": 0.3255, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"grad_norm": 0.5166645646095276, |
|
"learning_rate": 2.406410439910017e-06, |
|
"loss": 0.3928, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 2.128, |
|
"grad_norm": 0.523476243019104, |
|
"learning_rate": 2.366673707716973e-06, |
|
"loss": 0.3497, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 2.136, |
|
"grad_norm": 0.5414285063743591, |
|
"learning_rate": 2.327165820548059e-06, |
|
"loss": 0.3653, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 2.144, |
|
"grad_norm": 0.4763210713863373, |
|
"learning_rate": 2.287890211773238e-06, |
|
"loss": 0.3628, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 2.152, |
|
"grad_norm": 0.5000659227371216, |
|
"learning_rate": 2.2488502945766893e-06, |
|
"loss": 0.3681, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"grad_norm": 0.48179784417152405, |
|
"learning_rate": 2.210049461660189e-06, |
|
"loss": 0.3542, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 2.168, |
|
"grad_norm": 0.4680255651473999, |
|
"learning_rate": 2.1714910849482777e-06, |
|
"loss": 0.3618, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 2.176, |
|
"grad_norm": 0.4781685173511505, |
|
"learning_rate": 2.1331785152952243e-06, |
|
"loss": 0.3533, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 2.184, |
|
"grad_norm": 0.508620023727417, |
|
"learning_rate": 2.0951150821938278e-06, |
|
"loss": 0.3761, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 2.192, |
|
"grad_norm": 0.48846498131752014, |
|
"learning_rate": 2.0573040934860717e-06, |
|
"loss": 0.3776, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"grad_norm": 0.48690298199653625, |
|
"learning_rate": 2.0197488350756618e-06, |
|
"loss": 0.326, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 2.208, |
|
"grad_norm": 0.5058321356773376, |
|
"learning_rate": 1.98245257064247e-06, |
|
"loss": 0.3596, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 2.216, |
|
"grad_norm": 0.5168304443359375, |
|
"learning_rate": 1.945418541358911e-06, |
|
"loss": 0.3443, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 2.224, |
|
"grad_norm": 0.4696882367134094, |
|
"learning_rate": 1.9086499656082685e-06, |
|
"loss": 0.3469, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 2.232, |
|
"grad_norm": 0.46734338998794556, |
|
"learning_rate": 1.872150038705015e-06, |
|
"loss": 0.3494, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"grad_norm": 0.4970659017562866, |
|
"learning_rate": 1.835921932617119e-06, |
|
"loss": 0.3614, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 2.248, |
|
"grad_norm": 0.4725131094455719, |
|
"learning_rate": 1.7999687956903955e-06, |
|
"loss": 0.3678, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 2.2560000000000002, |
|
"grad_norm": 0.48716068267822266, |
|
"learning_rate": 1.7642937523749038e-06, |
|
"loss": 0.3514, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 2.2640000000000002, |
|
"grad_norm": 0.4228149950504303, |
|
"learning_rate": 1.7288999029534177e-06, |
|
"loss": 0.3349, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 2.2720000000000002, |
|
"grad_norm": 0.47141751646995544, |
|
"learning_rate": 1.6937903232720076e-06, |
|
"loss": 0.3655, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 2.2800000000000002, |
|
"grad_norm": 0.44692301750183105, |
|
"learning_rate": 1.6589680644727347e-06, |
|
"loss": 0.3446, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 2.288, |
|
"grad_norm": 0.46537840366363525, |
|
"learning_rate": 1.6244361527284953e-06, |
|
"loss": 0.3473, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 2.296, |
|
"grad_norm": 0.4578072428703308, |
|
"learning_rate": 1.5901975889800387e-06, |
|
"loss": 0.3614, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 2.304, |
|
"grad_norm": 0.4702107012271881, |
|
"learning_rate": 1.556255348675174e-06, |
|
"loss": 0.3417, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 2.312, |
|
"grad_norm": 0.49962151050567627, |
|
"learning_rate": 1.522612381510195e-06, |
|
"loss": 0.348, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"grad_norm": 0.514380156993866, |
|
"learning_rate": 1.489271611173538e-06, |
|
"loss": 0.3587, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 2.328, |
|
"grad_norm": 0.4479122459888458, |
|
"learning_rate": 1.4562359350917054e-06, |
|
"loss": 0.3479, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 2.336, |
|
"grad_norm": 0.48699259757995605, |
|
"learning_rate": 1.423508224177474e-06, |
|
"loss": 0.3678, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 2.344, |
|
"grad_norm": 0.46162307262420654, |
|
"learning_rate": 1.3910913225803946e-06, |
|
"loss": 0.3321, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 2.352, |
|
"grad_norm": 0.4586437940597534, |
|
"learning_rate": 1.35898804743963e-06, |
|
"loss": 0.3748, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"grad_norm": 0.479818195104599, |
|
"learning_rate": 1.3272011886391368e-06, |
|
"loss": 0.3489, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 2.368, |
|
"grad_norm": 0.44310837984085083, |
|
"learning_rate": 1.295733508565213e-06, |
|
"loss": 0.3427, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 2.376, |
|
"grad_norm": 0.4900333881378174, |
|
"learning_rate": 1.2645877418664394e-06, |
|
"loss": 0.3484, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 2.384, |
|
"grad_norm": 0.6010957360267639, |
|
"learning_rate": 1.2337665952160266e-06, |
|
"loss": 0.3486, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 2.392, |
|
"grad_norm": 0.44970938563346863, |
|
"learning_rate": 1.2032727470765982e-06, |
|
"loss": 0.3345, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"grad_norm": 0.46644240617752075, |
|
"learning_rate": 1.1731088474674235e-06, |
|
"loss": 0.3695, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 2.408, |
|
"grad_norm": 0.4730045199394226, |
|
"learning_rate": 1.1432775177341165e-06, |
|
"loss": 0.3605, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 2.416, |
|
"grad_norm": 0.4608900845050812, |
|
"learning_rate": 1.11378135032084e-06, |
|
"loss": 0.3328, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 2.424, |
|
"grad_norm": 0.47910159826278687, |
|
"learning_rate": 1.08462290854501e-06, |
|
"loss": 0.3592, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 2.432, |
|
"grad_norm": 0.5008491277694702, |
|
"learning_rate": 1.0558047263745297e-06, |
|
"loss": 0.3653, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"grad_norm": 0.45120254158973694, |
|
"learning_rate": 1.0273293082075914e-06, |
|
"loss": 0.3317, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 2.448, |
|
"grad_norm": 0.44771674275398254, |
|
"learning_rate": 9.991991286550207e-07, |
|
"loss": 0.3417, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 2.456, |
|
"grad_norm": 0.47719165682792664, |
|
"learning_rate": 9.71416632325235e-07, |
|
"loss": 0.3634, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 2.464, |
|
"grad_norm": 0.5096871256828308, |
|
"learning_rate": 9.439842336117954e-07, |
|
"loss": 0.3428, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 2.472, |
|
"grad_norm": 0.43439558148384094, |
|
"learning_rate": 9.169043164835867e-07, |
|
"loss": 0.3336, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"grad_norm": 0.4950801730155945, |
|
"learning_rate": 8.901792342776439e-07, |
|
"loss": 0.4171, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 2.488, |
|
"grad_norm": 0.4548616409301758, |
|
"learning_rate": 8.638113094946382e-07, |
|
"loss": 0.3123, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 2.496, |
|
"grad_norm": 0.5122199058532715, |
|
"learning_rate": 8.378028335970451e-07, |
|
"loss": 0.3784, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 2.504, |
|
"grad_norm": 0.452438622713089, |
|
"learning_rate": 8.121560668100065e-07, |
|
"loss": 0.3631, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 2.512, |
|
"grad_norm": 0.43378862738609314, |
|
"learning_rate": 7.868732379249122e-07, |
|
"loss": 0.3319, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"grad_norm": 0.4786088466644287, |
|
"learning_rate": 7.619565441057075e-07, |
|
"loss": 0.3517, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 2.528, |
|
"grad_norm": 0.5133551955223083, |
|
"learning_rate": 7.37408150697953e-07, |
|
"loss": 0.3702, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 2.536, |
|
"grad_norm": 0.4935527741909027, |
|
"learning_rate": 7.132301910406503e-07, |
|
"loss": 0.3759, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 2.544, |
|
"grad_norm": 0.4293310344219208, |
|
"learning_rate": 6.894247662808456e-07, |
|
"loss": 0.322, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 2.552, |
|
"grad_norm": 0.46054044365882874, |
|
"learning_rate": 6.659939451910341e-07, |
|
"loss": 0.3343, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"grad_norm": 0.5166534185409546, |
|
"learning_rate": 6.429397639893758e-07, |
|
"loss": 0.3813, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 2.568, |
|
"grad_norm": 0.4581873118877411, |
|
"learning_rate": 6.202642261627411e-07, |
|
"loss": 0.368, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 2.576, |
|
"grad_norm": 0.5063138008117676, |
|
"learning_rate": 5.979693022926025e-07, |
|
"loss": 0.4142, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 2.584, |
|
"grad_norm": 0.47579020261764526, |
|
"learning_rate": 5.760569298837825e-07, |
|
"loss": 0.3717, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 2.592, |
|
"grad_norm": 0.4639912545681, |
|
"learning_rate": 5.54529013196079e-07, |
|
"loss": 0.352, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"grad_norm": 0.4613383710384369, |
|
"learning_rate": 5.333874230787772e-07, |
|
"loss": 0.3661, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 2.608, |
|
"grad_norm": 0.42720070481300354, |
|
"learning_rate": 5.126339968080696e-07, |
|
"loss": 0.3616, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 2.616, |
|
"grad_norm": 0.4943825304508209, |
|
"learning_rate": 4.922705379273862e-07, |
|
"loss": 0.3645, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 2.624, |
|
"grad_norm": 0.4619237780570984, |
|
"learning_rate": 4.7229881609066387e-07, |
|
"loss": 0.3307, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 2.632, |
|
"grad_norm": 0.47587689757347107, |
|
"learning_rate": 4.5272056690855494e-07, |
|
"loss": 0.3624, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"grad_norm": 0.45437976717948914, |
|
"learning_rate": 4.335374917975982e-07, |
|
"loss": 0.3242, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 2.648, |
|
"grad_norm": 0.5222877264022827, |
|
"learning_rate": 4.147512578323615e-07, |
|
"loss": 0.3808, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 2.656, |
|
"grad_norm": 0.46543389558792114, |
|
"learning_rate": 3.9636349760056427e-07, |
|
"loss": 0.3389, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 2.664, |
|
"grad_norm": 0.4624124765396118, |
|
"learning_rate": 3.783758090611983e-07, |
|
"loss": 0.3336, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 2.672, |
|
"grad_norm": 0.45994681119918823, |
|
"learning_rate": 3.6078975540566716e-07, |
|
"loss": 0.3482, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"grad_norm": 0.45068469643592834, |
|
"learning_rate": 3.4360686492193263e-07, |
|
"loss": 0.3077, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 2.6879999999999997, |
|
"grad_norm": 0.5126630663871765, |
|
"learning_rate": 3.268286308617041e-07, |
|
"loss": 0.3906, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 2.6959999999999997, |
|
"grad_norm": 0.4749516546726227, |
|
"learning_rate": 3.104565113106689e-07, |
|
"loss": 0.342, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 2.7039999999999997, |
|
"grad_norm": 0.43455156683921814, |
|
"learning_rate": 2.9449192906178205e-07, |
|
"loss": 0.3491, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 2.7119999999999997, |
|
"grad_norm": 0.46082979440689087, |
|
"learning_rate": 2.789362714916172e-07, |
|
"loss": 0.3875, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 2.7199999999999998, |
|
"grad_norm": 0.44106754660606384, |
|
"learning_rate": 2.6379089043980064e-07, |
|
"loss": 0.32, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 2.7279999999999998, |
|
"grad_norm": 0.46814629435539246, |
|
"learning_rate": 2.4905710209153224e-07, |
|
"loss": 0.3451, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 2.7359999999999998, |
|
"grad_norm": 0.5077950358390808, |
|
"learning_rate": 2.3473618686320477e-07, |
|
"loss": 0.3861, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 2.7439999999999998, |
|
"grad_norm": 0.49315568804740906, |
|
"learning_rate": 2.208293892911284e-07, |
|
"loss": 0.3469, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 2.752, |
|
"grad_norm": 0.4634096920490265, |
|
"learning_rate": 2.0733791792338197e-07, |
|
"loss": 0.3761, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"grad_norm": 0.4556006193161011, |
|
"learning_rate": 1.9426294521477874e-07, |
|
"loss": 0.3504, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 2.768, |
|
"grad_norm": 0.48851218819618225, |
|
"learning_rate": 1.8160560742498223e-07, |
|
"loss": 0.3652, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 2.776, |
|
"grad_norm": 0.45232683420181274, |
|
"learning_rate": 1.6936700451975818e-07, |
|
"loss": 0.3259, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 2.784, |
|
"grad_norm": 0.4734852910041809, |
|
"learning_rate": 1.5754820007538473e-07, |
|
"loss": 0.3555, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 2.792, |
|
"grad_norm": 0.46094757318496704, |
|
"learning_rate": 1.461502211862237e-07, |
|
"loss": 0.3451, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"grad_norm": 0.45597511529922485, |
|
"learning_rate": 1.3517405837546404e-07, |
|
"loss": 0.3481, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 2.808, |
|
"grad_norm": 0.44162362813949585, |
|
"learning_rate": 1.2462066550903818e-07, |
|
"loss": 0.3324, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 2.816, |
|
"grad_norm": 0.4328787326812744, |
|
"learning_rate": 1.1449095971273305e-07, |
|
"loss": 0.3304, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 2.824, |
|
"grad_norm": 0.472141832113266, |
|
"learning_rate": 1.0478582129248516e-07, |
|
"loss": 0.3655, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 2.832, |
|
"grad_norm": 0.4979706406593323, |
|
"learning_rate": 9.550609365787888e-08, |
|
"loss": 0.3815, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"grad_norm": 0.49750280380249023, |
|
"learning_rate": 8.66525832488535e-08, |
|
"loss": 0.3656, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 2.848, |
|
"grad_norm": 0.46096086502075195, |
|
"learning_rate": 7.822605946561923e-08, |
|
"loss": 0.3437, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 2.856, |
|
"grad_norm": 0.5005020499229431, |
|
"learning_rate": 7.022725460179459e-08, |
|
"loss": 0.3697, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 2.864, |
|
"grad_norm": 0.4813578426837921, |
|
"learning_rate": 6.265686378076729e-08, |
|
"loss": 0.3595, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 2.872, |
|
"grad_norm": 0.4750673472881317, |
|
"learning_rate": 5.5515544895284324e-08, |
|
"loss": 0.3706, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"grad_norm": 0.48565149307250977, |
|
"learning_rate": 4.880391855028088e-08, |
|
"loss": 0.3524, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 2.888, |
|
"grad_norm": 0.4671759307384491, |
|
"learning_rate": 4.252256800894694e-08, |
|
"loss": 0.3407, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 2.896, |
|
"grad_norm": 0.4376312494277954, |
|
"learning_rate": 3.6672039142039426e-08, |
|
"loss": 0.3292, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 2.904, |
|
"grad_norm": 0.4908435344696045, |
|
"learning_rate": 3.125284038044407e-08, |
|
"loss": 0.3429, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 2.912, |
|
"grad_norm": 0.44629916548728943, |
|
"learning_rate": 2.6265442670991293e-08, |
|
"loss": 0.3428, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"grad_norm": 0.465819388628006, |
|
"learning_rate": 2.1710279435530058e-08, |
|
"loss": 0.3448, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 2.928, |
|
"grad_norm": 0.4819665253162384, |
|
"learning_rate": 1.7587746533260786e-08, |
|
"loss": 0.3519, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 2.936, |
|
"grad_norm": 0.4784250855445862, |
|
"learning_rate": 1.3898202226333424e-08, |
|
"loss": 0.3739, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 2.944, |
|
"grad_norm": 0.495050847530365, |
|
"learning_rate": 1.0641967148716236e-08, |
|
"loss": 0.3628, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 2.952, |
|
"grad_norm": 0.4823502004146576, |
|
"learning_rate": 7.819324278328099e-09, |
|
"loss": 0.3411, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"grad_norm": 0.41492488980293274, |
|
"learning_rate": 5.430518912448169e-09, |
|
"loss": 0.3326, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 2.968, |
|
"grad_norm": 0.47112229466438293, |
|
"learning_rate": 3.4757586464001513e-09, |
|
"loss": 0.3696, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 2.976, |
|
"grad_norm": 0.46793392300605774, |
|
"learning_rate": 1.9552133555084117e-09, |
|
"loss": 0.3661, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 2.984, |
|
"grad_norm": 0.4385625720024109, |
|
"learning_rate": 8.690151803386615e-10, |
|
"loss": 0.3421, |
|
"step": 373 |
|
}, |
|
{ |
|
"epoch": 2.992, |
|
"grad_norm": 0.43801286816596985, |
|
"learning_rate": 2.1725851521103847e-10, |
|
"loss": 0.3395, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"grad_norm": 0.44443294405937195, |
|
"learning_rate": 0.0, |
|
"loss": 0.3399, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"step": 375, |
|
"total_flos": 223704105418752.0, |
|
"train_loss": 0.4401542440255483, |
|
"train_runtime": 59266.2263, |
|
"train_samples_per_second": 0.607, |
|
"train_steps_per_second": 0.006 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 375, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 50, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 223704105418752.0, |
|
"train_batch_size": 1, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|