lillian039's picture
Model save
b749358 verified
raw
history blame
82.1 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 3.0,
"eval_steps": 500,
"global_step": 465,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0064516129032258064,
"grad_norm": 0.15169227525554954,
"learning_rate": 4.255319148936171e-06,
"loss": 0.1803,
"step": 1
},
{
"epoch": 0.012903225806451613,
"grad_norm": 0.1108587617034759,
"learning_rate": 8.510638297872341e-06,
"loss": 0.1441,
"step": 2
},
{
"epoch": 0.01935483870967742,
"grad_norm": 0.08345105115839851,
"learning_rate": 1.2765957446808511e-05,
"loss": 0.1118,
"step": 3
},
{
"epoch": 0.025806451612903226,
"grad_norm": 0.1426234633379846,
"learning_rate": 1.7021276595744682e-05,
"loss": 0.1313,
"step": 4
},
{
"epoch": 0.03225806451612903,
"grad_norm": 0.14217530777409554,
"learning_rate": 2.1276595744680852e-05,
"loss": 0.1195,
"step": 5
},
{
"epoch": 0.03870967741935484,
"grad_norm": 0.1459141077043439,
"learning_rate": 2.5531914893617022e-05,
"loss": 0.1453,
"step": 6
},
{
"epoch": 0.04516129032258064,
"grad_norm": 0.10195797727709278,
"learning_rate": 2.9787234042553192e-05,
"loss": 0.1119,
"step": 7
},
{
"epoch": 0.05161290322580645,
"grad_norm": 0.13553076939195616,
"learning_rate": 3.4042553191489365e-05,
"loss": 0.1791,
"step": 8
},
{
"epoch": 0.05806451612903226,
"grad_norm": 0.15732711391586654,
"learning_rate": 3.829787234042553e-05,
"loss": 0.1282,
"step": 9
},
{
"epoch": 0.06451612903225806,
"grad_norm": 0.1192108991071351,
"learning_rate": 4.2553191489361704e-05,
"loss": 0.1288,
"step": 10
},
{
"epoch": 0.07096774193548387,
"grad_norm": 0.1534053229135478,
"learning_rate": 4.680851063829788e-05,
"loss": 0.15,
"step": 11
},
{
"epoch": 0.07741935483870968,
"grad_norm": 0.11808236646609899,
"learning_rate": 5.1063829787234044e-05,
"loss": 0.1098,
"step": 12
},
{
"epoch": 0.08387096774193549,
"grad_norm": 0.1127350100020763,
"learning_rate": 5.531914893617022e-05,
"loss": 0.1214,
"step": 13
},
{
"epoch": 0.09032258064516129,
"grad_norm": 0.13802666747672132,
"learning_rate": 5.9574468085106384e-05,
"loss": 0.1555,
"step": 14
},
{
"epoch": 0.0967741935483871,
"grad_norm": 0.19187355993638458,
"learning_rate": 6.382978723404256e-05,
"loss": 0.1474,
"step": 15
},
{
"epoch": 0.1032258064516129,
"grad_norm": 0.16430813087649648,
"learning_rate": 6.808510638297873e-05,
"loss": 0.1461,
"step": 16
},
{
"epoch": 0.10967741935483871,
"grad_norm": 0.13789502094335715,
"learning_rate": 7.23404255319149e-05,
"loss": 0.1226,
"step": 17
},
{
"epoch": 0.11612903225806452,
"grad_norm": 0.173980091579583,
"learning_rate": 7.659574468085106e-05,
"loss": 0.1558,
"step": 18
},
{
"epoch": 0.12258064516129032,
"grad_norm": 0.1418687450951512,
"learning_rate": 8.085106382978723e-05,
"loss": 0.1379,
"step": 19
},
{
"epoch": 0.12903225806451613,
"grad_norm": 0.1765388069700401,
"learning_rate": 8.510638297872341e-05,
"loss": 0.149,
"step": 20
},
{
"epoch": 0.13548387096774195,
"grad_norm": 0.17659794510341198,
"learning_rate": 8.936170212765958e-05,
"loss": 0.1101,
"step": 21
},
{
"epoch": 0.14193548387096774,
"grad_norm": 0.22633201272355616,
"learning_rate": 9.361702127659576e-05,
"loss": 0.1656,
"step": 22
},
{
"epoch": 0.14838709677419354,
"grad_norm": 0.16004849032165383,
"learning_rate": 9.787234042553192e-05,
"loss": 0.1616,
"step": 23
},
{
"epoch": 0.15483870967741936,
"grad_norm": 0.16981049690586422,
"learning_rate": 0.00010212765957446809,
"loss": 0.1321,
"step": 24
},
{
"epoch": 0.16129032258064516,
"grad_norm": 0.13581353274905067,
"learning_rate": 0.00010638297872340425,
"loss": 0.1122,
"step": 25
},
{
"epoch": 0.16774193548387098,
"grad_norm": 0.2089185781750498,
"learning_rate": 0.00011063829787234043,
"loss": 0.1527,
"step": 26
},
{
"epoch": 0.17419354838709677,
"grad_norm": 0.16125655225492164,
"learning_rate": 0.00011489361702127661,
"loss": 0.1263,
"step": 27
},
{
"epoch": 0.18064516129032257,
"grad_norm": 0.13753328983503424,
"learning_rate": 0.00011914893617021277,
"loss": 0.1224,
"step": 28
},
{
"epoch": 0.1870967741935484,
"grad_norm": 0.19089590529877826,
"learning_rate": 0.00012340425531914893,
"loss": 0.168,
"step": 29
},
{
"epoch": 0.1935483870967742,
"grad_norm": 0.15419681175807715,
"learning_rate": 0.00012765957446808513,
"loss": 0.138,
"step": 30
},
{
"epoch": 0.2,
"grad_norm": 0.18178665024271073,
"learning_rate": 0.00013191489361702127,
"loss": 0.1955,
"step": 31
},
{
"epoch": 0.2064516129032258,
"grad_norm": 0.11760523051627117,
"learning_rate": 0.00013617021276595746,
"loss": 0.1367,
"step": 32
},
{
"epoch": 0.2129032258064516,
"grad_norm": 0.10853469531105876,
"learning_rate": 0.00014042553191489363,
"loss": 0.1208,
"step": 33
},
{
"epoch": 0.21935483870967742,
"grad_norm": 0.12890024443339382,
"learning_rate": 0.0001446808510638298,
"loss": 0.142,
"step": 34
},
{
"epoch": 0.22580645161290322,
"grad_norm": 0.12692879603118554,
"learning_rate": 0.00014893617021276596,
"loss": 0.1268,
"step": 35
},
{
"epoch": 0.23225806451612904,
"grad_norm": 0.1529612097417899,
"learning_rate": 0.00015319148936170213,
"loss": 0.1194,
"step": 36
},
{
"epoch": 0.23870967741935484,
"grad_norm": 0.129432114467059,
"learning_rate": 0.00015744680851063832,
"loss": 0.128,
"step": 37
},
{
"epoch": 0.24516129032258063,
"grad_norm": 0.11343138573627701,
"learning_rate": 0.00016170212765957446,
"loss": 0.102,
"step": 38
},
{
"epoch": 0.25161290322580643,
"grad_norm": 0.16577469112721435,
"learning_rate": 0.00016595744680851065,
"loss": 0.1691,
"step": 39
},
{
"epoch": 0.25806451612903225,
"grad_norm": 0.13614466586953358,
"learning_rate": 0.00017021276595744682,
"loss": 0.1348,
"step": 40
},
{
"epoch": 0.2645161290322581,
"grad_norm": 0.1524812917659128,
"learning_rate": 0.00017446808510638298,
"loss": 0.1281,
"step": 41
},
{
"epoch": 0.2709677419354839,
"grad_norm": 0.14519269708506746,
"learning_rate": 0.00017872340425531915,
"loss": 0.1349,
"step": 42
},
{
"epoch": 0.27741935483870966,
"grad_norm": 0.18300481897670345,
"learning_rate": 0.00018297872340425532,
"loss": 0.1694,
"step": 43
},
{
"epoch": 0.2838709677419355,
"grad_norm": 0.11929331561330575,
"learning_rate": 0.0001872340425531915,
"loss": 0.1331,
"step": 44
},
{
"epoch": 0.2903225806451613,
"grad_norm": 0.12604932986068976,
"learning_rate": 0.00019148936170212768,
"loss": 0.0949,
"step": 45
},
{
"epoch": 0.2967741935483871,
"grad_norm": 0.14226790630565908,
"learning_rate": 0.00019574468085106384,
"loss": 0.1477,
"step": 46
},
{
"epoch": 0.3032258064516129,
"grad_norm": 0.1465980952326029,
"learning_rate": 0.0002,
"loss": 0.1408,
"step": 47
},
{
"epoch": 0.3096774193548387,
"grad_norm": 0.14535685898764863,
"learning_rate": 0.0001999971756719333,
"loss": 0.1474,
"step": 48
},
{
"epoch": 0.3161290322580645,
"grad_norm": 0.1559112597851861,
"learning_rate": 0.00019998870284726968,
"loss": 0.1568,
"step": 49
},
{
"epoch": 0.3225806451612903,
"grad_norm": 0.14392375991483491,
"learning_rate": 0.00019997458200460993,
"loss": 0.1526,
"step": 50
},
{
"epoch": 0.32903225806451614,
"grad_norm": 0.18920444771524633,
"learning_rate": 0.00019995481394159188,
"loss": 0.1384,
"step": 51
},
{
"epoch": 0.33548387096774196,
"grad_norm": 0.17943388616343298,
"learning_rate": 0.0001999293997748454,
"loss": 0.1555,
"step": 52
},
{
"epoch": 0.3419354838709677,
"grad_norm": 0.16752646291327727,
"learning_rate": 0.00019989834093992945,
"loss": 0.1634,
"step": 53
},
{
"epoch": 0.34838709677419355,
"grad_norm": 0.1484740777328073,
"learning_rate": 0.00019986163919125075,
"loss": 0.124,
"step": 54
},
{
"epoch": 0.3548387096774194,
"grad_norm": 0.15268630472434244,
"learning_rate": 0.00019981929660196492,
"loss": 0.1228,
"step": 55
},
{
"epoch": 0.36129032258064514,
"grad_norm": 0.19583236181584973,
"learning_rate": 0.0001997713155638592,
"loss": 0.1227,
"step": 56
},
{
"epoch": 0.36774193548387096,
"grad_norm": 0.1469999186659826,
"learning_rate": 0.00019971769878721743,
"loss": 0.1188,
"step": 57
},
{
"epoch": 0.3741935483870968,
"grad_norm": 0.10845193424043034,
"learning_rate": 0.000199658449300667,
"loss": 0.1177,
"step": 58
},
{
"epoch": 0.38064516129032255,
"grad_norm": 0.1845188187089657,
"learning_rate": 0.00019959357045100764,
"loss": 0.1726,
"step": 59
},
{
"epoch": 0.3870967741935484,
"grad_norm": 0.14769124665651473,
"learning_rate": 0.00019952306590302247,
"loss": 0.1634,
"step": 60
},
{
"epoch": 0.3935483870967742,
"grad_norm": 0.16408468392163889,
"learning_rate": 0.00019944693963927092,
"loss": 0.1584,
"step": 61
},
{
"epoch": 0.4,
"grad_norm": 0.15156206723802879,
"learning_rate": 0.00019936519595986394,
"loss": 0.1454,
"step": 62
},
{
"epoch": 0.4064516129032258,
"grad_norm": 0.15835599927161748,
"learning_rate": 0.00019927783948222084,
"loss": 0.1621,
"step": 63
},
{
"epoch": 0.4129032258064516,
"grad_norm": 0.1333411665662589,
"learning_rate": 0.00019918487514080865,
"loss": 0.1293,
"step": 64
},
{
"epoch": 0.41935483870967744,
"grad_norm": 0.1589874051481937,
"learning_rate": 0.00019908630818686338,
"loss": 0.1391,
"step": 65
},
{
"epoch": 0.4258064516129032,
"grad_norm": 0.12399538770065353,
"learning_rate": 0.0001989821441880933,
"loss": 0.1208,
"step": 66
},
{
"epoch": 0.432258064516129,
"grad_norm": 0.16949673743147955,
"learning_rate": 0.00019887238902836448,
"loss": 0.1139,
"step": 67
},
{
"epoch": 0.43870967741935485,
"grad_norm": 0.1875430265393267,
"learning_rate": 0.00019875704890736853,
"loss": 0.1856,
"step": 68
},
{
"epoch": 0.44516129032258067,
"grad_norm": 0.14927152386929934,
"learning_rate": 0.00019863613034027224,
"loss": 0.1516,
"step": 69
},
{
"epoch": 0.45161290322580644,
"grad_norm": 0.17501653378304205,
"learning_rate": 0.0001985096401573497,
"loss": 0.161,
"step": 70
},
{
"epoch": 0.45806451612903226,
"grad_norm": 0.17582762670350804,
"learning_rate": 0.00019837758550359636,
"loss": 0.1564,
"step": 71
},
{
"epoch": 0.4645161290322581,
"grad_norm": 0.15692049884405931,
"learning_rate": 0.0001982399738383255,
"loss": 0.1282,
"step": 72
},
{
"epoch": 0.47096774193548385,
"grad_norm": 0.14436392088728167,
"learning_rate": 0.00019809681293474693,
"loss": 0.1299,
"step": 73
},
{
"epoch": 0.4774193548387097,
"grad_norm": 0.19048463762976417,
"learning_rate": 0.0001979481108795278,
"loss": 0.1983,
"step": 74
},
{
"epoch": 0.4838709677419355,
"grad_norm": 0.13475205452089994,
"learning_rate": 0.00019779387607233586,
"loss": 0.156,
"step": 75
},
{
"epoch": 0.49032258064516127,
"grad_norm": 0.14145398172929924,
"learning_rate": 0.00019763411722536502,
"loss": 0.1355,
"step": 76
},
{
"epoch": 0.4967741935483871,
"grad_norm": 0.14400838677113517,
"learning_rate": 0.00019746884336284317,
"loss": 0.1371,
"step": 77
},
{
"epoch": 0.5032258064516129,
"grad_norm": 0.13421206031331137,
"learning_rate": 0.00019729806382052248,
"loss": 0.1156,
"step": 78
},
{
"epoch": 0.5096774193548387,
"grad_norm": 0.11613915473665105,
"learning_rate": 0.00019712178824515212,
"loss": 0.1293,
"step": 79
},
{
"epoch": 0.5161290322580645,
"grad_norm": 0.12472786830602234,
"learning_rate": 0.00019694002659393305,
"loss": 0.1189,
"step": 80
},
{
"epoch": 0.5225806451612903,
"grad_norm": 0.11050712097688373,
"learning_rate": 0.00019675278913395606,
"loss": 0.12,
"step": 81
},
{
"epoch": 0.5290322580645161,
"grad_norm": 0.1303276052231771,
"learning_rate": 0.0001965600864416213,
"loss": 0.1428,
"step": 82
},
{
"epoch": 0.535483870967742,
"grad_norm": 0.13774570876008593,
"learning_rate": 0.00019636192940204134,
"loss": 0.1438,
"step": 83
},
{
"epoch": 0.5419354838709678,
"grad_norm": 0.14810394982940484,
"learning_rate": 0.00019615832920842586,
"loss": 0.1404,
"step": 84
},
{
"epoch": 0.5483870967741935,
"grad_norm": 0.144275852100491,
"learning_rate": 0.00019594929736144976,
"loss": 0.1357,
"step": 85
},
{
"epoch": 0.5548387096774193,
"grad_norm": 0.16879600090790034,
"learning_rate": 0.0001957348456686032,
"loss": 0.1578,
"step": 86
},
{
"epoch": 0.5612903225806452,
"grad_norm": 0.1588074767274021,
"learning_rate": 0.00019551498624352496,
"loss": 0.1228,
"step": 87
},
{
"epoch": 0.567741935483871,
"grad_norm": 0.15067916918276408,
"learning_rate": 0.00019528973150531787,
"loss": 0.1323,
"step": 88
},
{
"epoch": 0.5741935483870968,
"grad_norm": 0.16266341913656662,
"learning_rate": 0.00019505909417784754,
"loss": 0.13,
"step": 89
},
{
"epoch": 0.5806451612903226,
"grad_norm": 0.121529537729675,
"learning_rate": 0.00019482308728902356,
"loss": 0.1067,
"step": 90
},
{
"epoch": 0.5870967741935483,
"grad_norm": 0.1740468182649888,
"learning_rate": 0.00019458172417006347,
"loss": 0.1513,
"step": 91
},
{
"epoch": 0.5935483870967742,
"grad_norm": 0.14062538318374462,
"learning_rate": 0.00019433501845473995,
"loss": 0.1438,
"step": 92
},
{
"epoch": 0.6,
"grad_norm": 0.17690034130801896,
"learning_rate": 0.00019408298407861042,
"loss": 0.1356,
"step": 93
},
{
"epoch": 0.6064516129032258,
"grad_norm": 0.18789482750619546,
"learning_rate": 0.00019382563527823026,
"loss": 0.1758,
"step": 94
},
{
"epoch": 0.6129032258064516,
"grad_norm": 0.16993067663839118,
"learning_rate": 0.00019356298659034817,
"loss": 0.1599,
"step": 95
},
{
"epoch": 0.6193548387096774,
"grad_norm": 0.16495058136550117,
"learning_rate": 0.00019329505285108542,
"loss": 0.1283,
"step": 96
},
{
"epoch": 0.6258064516129033,
"grad_norm": 0.15497395377626808,
"learning_rate": 0.00019302184919509755,
"loss": 0.1493,
"step": 97
},
{
"epoch": 0.632258064516129,
"grad_norm": 0.13660921526912856,
"learning_rate": 0.00019274339105471971,
"loss": 0.1307,
"step": 98
},
{
"epoch": 0.6387096774193548,
"grad_norm": 0.18246231884688152,
"learning_rate": 0.00019245969415909465,
"loss": 0.1598,
"step": 99
},
{
"epoch": 0.6451612903225806,
"grad_norm": 0.11344749529118914,
"learning_rate": 0.00019217077453328449,
"loss": 0.1304,
"step": 100
},
{
"epoch": 0.6516129032258065,
"grad_norm": 0.11682283731326468,
"learning_rate": 0.0001918766484973654,
"loss": 0.0977,
"step": 101
},
{
"epoch": 0.6580645161290323,
"grad_norm": 0.14494532290577813,
"learning_rate": 0.00019157733266550575,
"loss": 0.1338,
"step": 102
},
{
"epoch": 0.6645161290322581,
"grad_norm": 0.12095275202026515,
"learning_rate": 0.0001912728439450276,
"loss": 0.1513,
"step": 103
},
{
"epoch": 0.6709677419354839,
"grad_norm": 0.13423702981009097,
"learning_rate": 0.00019096319953545185,
"loss": 0.1335,
"step": 104
},
{
"epoch": 0.6774193548387096,
"grad_norm": 0.1670132860946028,
"learning_rate": 0.0001906484169275263,
"loss": 0.1607,
"step": 105
},
{
"epoch": 0.6838709677419355,
"grad_norm": 0.14053854153152684,
"learning_rate": 0.00019032851390223812,
"loss": 0.1365,
"step": 106
},
{
"epoch": 0.6903225806451613,
"grad_norm": 0.1399807021991922,
"learning_rate": 0.00019000350852980909,
"loss": 0.1589,
"step": 107
},
{
"epoch": 0.6967741935483871,
"grad_norm": 0.15473299551506894,
"learning_rate": 0.00018967341916867518,
"loss": 0.166,
"step": 108
},
{
"epoch": 0.7032258064516129,
"grad_norm": 0.1536872117526864,
"learning_rate": 0.00018933826446444933,
"loss": 0.1657,
"step": 109
},
{
"epoch": 0.7096774193548387,
"grad_norm": 0.1282004975556196,
"learning_rate": 0.0001889980633488683,
"loss": 0.1212,
"step": 110
},
{
"epoch": 0.7161290322580646,
"grad_norm": 0.16458328388975405,
"learning_rate": 0.00018865283503872324,
"loss": 0.1655,
"step": 111
},
{
"epoch": 0.7225806451612903,
"grad_norm": 0.1505113828615181,
"learning_rate": 0.00018830259903477426,
"loss": 0.1571,
"step": 112
},
{
"epoch": 0.7290322580645161,
"grad_norm": 0.16575595088070239,
"learning_rate": 0.0001879473751206489,
"loss": 0.1504,
"step": 113
},
{
"epoch": 0.7354838709677419,
"grad_norm": 0.1484230902451611,
"learning_rate": 0.0001875871833617246,
"loss": 0.1498,
"step": 114
},
{
"epoch": 0.7419354838709677,
"grad_norm": 0.2059907492830938,
"learning_rate": 0.0001872220441039952,
"loss": 0.134,
"step": 115
},
{
"epoch": 0.7483870967741936,
"grad_norm": 0.1491671097000444,
"learning_rate": 0.0001868519779729218,
"loss": 0.1374,
"step": 116
},
{
"epoch": 0.7548387096774194,
"grad_norm": 0.1727675086328308,
"learning_rate": 0.0001864770058722676,
"loss": 0.1624,
"step": 117
},
{
"epoch": 0.7612903225806451,
"grad_norm": 0.16357318423846662,
"learning_rate": 0.00018609714898291718,
"loss": 0.1528,
"step": 118
},
{
"epoch": 0.7677419354838709,
"grad_norm": 0.1584052674932312,
"learning_rate": 0.00018571242876167996,
"loss": 0.1321,
"step": 119
},
{
"epoch": 0.7741935483870968,
"grad_norm": 0.21714396600094343,
"learning_rate": 0.0001853228669400784,
"loss": 0.1748,
"step": 120
},
{
"epoch": 0.7806451612903226,
"grad_norm": 0.13743651841776636,
"learning_rate": 0.00018492848552312014,
"loss": 0.1493,
"step": 121
},
{
"epoch": 0.7870967741935484,
"grad_norm": 0.1541126978032927,
"learning_rate": 0.00018452930678805536,
"loss": 0.1331,
"step": 122
},
{
"epoch": 0.7935483870967742,
"grad_norm": 0.1571822882709535,
"learning_rate": 0.00018412535328311814,
"loss": 0.1427,
"step": 123
},
{
"epoch": 0.8,
"grad_norm": 0.17825934340851243,
"learning_rate": 0.00018371664782625287,
"loss": 0.1871,
"step": 124
},
{
"epoch": 0.8064516129032258,
"grad_norm": 0.1745800846806893,
"learning_rate": 0.00018330321350382544,
"loss": 0.1672,
"step": 125
},
{
"epoch": 0.8129032258064516,
"grad_norm": 0.13970218689990957,
"learning_rate": 0.00018288507366931905,
"loss": 0.1715,
"step": 126
},
{
"epoch": 0.8193548387096774,
"grad_norm": 0.15100292169016535,
"learning_rate": 0.00018246225194201517,
"loss": 0.1411,
"step": 127
},
{
"epoch": 0.8258064516129032,
"grad_norm": 0.1538126122981586,
"learning_rate": 0.00018203477220565912,
"loss": 0.1516,
"step": 128
},
{
"epoch": 0.832258064516129,
"grad_norm": 0.15630735898296536,
"learning_rate": 0.00018160265860711134,
"loss": 0.1636,
"step": 129
},
{
"epoch": 0.8387096774193549,
"grad_norm": 0.1427718560771215,
"learning_rate": 0.00018116593555498307,
"loss": 0.1297,
"step": 130
},
{
"epoch": 0.8451612903225807,
"grad_norm": 0.11911885929622754,
"learning_rate": 0.0001807246277182578,
"loss": 0.1115,
"step": 131
},
{
"epoch": 0.8516129032258064,
"grad_norm": 0.13702745725576418,
"learning_rate": 0.0001802787600248977,
"loss": 0.157,
"step": 132
},
{
"epoch": 0.8580645161290322,
"grad_norm": 0.17220644805792673,
"learning_rate": 0.0001798283576604356,
"loss": 0.1561,
"step": 133
},
{
"epoch": 0.864516129032258,
"grad_norm": 0.196867226472361,
"learning_rate": 0.0001793734460665523,
"loss": 0.1657,
"step": 134
},
{
"epoch": 0.8709677419354839,
"grad_norm": 0.1759359784948508,
"learning_rate": 0.00017891405093963938,
"loss": 0.1909,
"step": 135
},
{
"epoch": 0.8774193548387097,
"grad_norm": 0.16063163300244113,
"learning_rate": 0.0001784501982293479,
"loss": 0.1552,
"step": 136
},
{
"epoch": 0.8838709677419355,
"grad_norm": 0.17034183209183734,
"learning_rate": 0.00017798191413712243,
"loss": 0.1502,
"step": 137
},
{
"epoch": 0.8903225806451613,
"grad_norm": 0.14694529779128243,
"learning_rate": 0.0001775092251147211,
"loss": 0.1277,
"step": 138
},
{
"epoch": 0.896774193548387,
"grad_norm": 0.13174927545002138,
"learning_rate": 0.0001770321578627213,
"loss": 0.1277,
"step": 139
},
{
"epoch": 0.9032258064516129,
"grad_norm": 0.15784079444219237,
"learning_rate": 0.00017655073932901168,
"loss": 0.1534,
"step": 140
},
{
"epoch": 0.9096774193548387,
"grad_norm": 0.19691370909219638,
"learning_rate": 0.0001760649967072697,
"loss": 0.1688,
"step": 141
},
{
"epoch": 0.9161290322580645,
"grad_norm": 0.15697904965484202,
"learning_rate": 0.00017557495743542585,
"loss": 0.1523,
"step": 142
},
{
"epoch": 0.9225806451612903,
"grad_norm": 0.1759980765103477,
"learning_rate": 0.00017508064919411344,
"loss": 0.15,
"step": 143
},
{
"epoch": 0.9290322580645162,
"grad_norm": 0.19254834048997346,
"learning_rate": 0.00017458209990510527,
"loss": 0.1474,
"step": 144
},
{
"epoch": 0.9354838709677419,
"grad_norm": 0.19220369870461818,
"learning_rate": 0.00017407933772973637,
"loss": 0.1678,
"step": 145
},
{
"epoch": 0.9419354838709677,
"grad_norm": 0.14789056250556576,
"learning_rate": 0.00017357239106731317,
"loss": 0.1634,
"step": 146
},
{
"epoch": 0.9483870967741935,
"grad_norm": 0.15823815803270533,
"learning_rate": 0.00017306128855350942,
"loss": 0.1744,
"step": 147
},
{
"epoch": 0.9548387096774194,
"grad_norm": 0.159128130793647,
"learning_rate": 0.0001725460590587486,
"loss": 0.1732,
"step": 148
},
{
"epoch": 0.9612903225806452,
"grad_norm": 0.1420932941022579,
"learning_rate": 0.00017202673168657318,
"loss": 0.1193,
"step": 149
},
{
"epoch": 0.967741935483871,
"grad_norm": 0.1581041276537875,
"learning_rate": 0.0001715033357720006,
"loss": 0.157,
"step": 150
},
{
"epoch": 0.9741935483870968,
"grad_norm": 0.13409040520330398,
"learning_rate": 0.00017097590087986633,
"loss": 0.1187,
"step": 151
},
{
"epoch": 0.9806451612903225,
"grad_norm": 0.15236261967510367,
"learning_rate": 0.00017044445680315372,
"loss": 0.1541,
"step": 152
},
{
"epoch": 0.9870967741935484,
"grad_norm": 0.15028826750982388,
"learning_rate": 0.00016990903356131124,
"loss": 0.1462,
"step": 153
},
{
"epoch": 0.9935483870967742,
"grad_norm": 0.14170487741522195,
"learning_rate": 0.00016936966139855663,
"loss": 0.1275,
"step": 154
},
{
"epoch": 1.0,
"grad_norm": 0.14288119090775778,
"learning_rate": 0.00016882637078216868,
"loss": 0.1316,
"step": 155
},
{
"epoch": 1.0,
"eval_loss": 0.14944089949131012,
"eval_runtime": 27.6083,
"eval_samples_per_second": 4.745,
"eval_steps_per_second": 0.616,
"step": 155
},
{
"epoch": 1.0064516129032257,
"grad_norm": 0.08453811084356862,
"learning_rate": 0.0001682791924007661,
"loss": 0.0732,
"step": 156
},
{
"epoch": 1.0129032258064516,
"grad_norm": 0.1163343033490188,
"learning_rate": 0.00016772815716257412,
"loss": 0.0955,
"step": 157
},
{
"epoch": 1.0193548387096774,
"grad_norm": 0.12774969124655083,
"learning_rate": 0.0001671732961936785,
"loss": 0.1154,
"step": 158
},
{
"epoch": 1.0258064516129033,
"grad_norm": 0.10149504758995384,
"learning_rate": 0.00016661464083626734,
"loss": 0.0834,
"step": 159
},
{
"epoch": 1.032258064516129,
"grad_norm": 0.1558767636437416,
"learning_rate": 0.00016605222264686086,
"loss": 0.0978,
"step": 160
},
{
"epoch": 1.038709677419355,
"grad_norm": 0.14800857594022712,
"learning_rate": 0.00016548607339452853,
"loss": 0.0783,
"step": 161
},
{
"epoch": 1.0451612903225806,
"grad_norm": 0.10427437820954576,
"learning_rate": 0.00016491622505909482,
"loss": 0.0714,
"step": 162
},
{
"epoch": 1.0516129032258064,
"grad_norm": 0.16509881361577539,
"learning_rate": 0.00016434270982933273,
"loss": 0.0971,
"step": 163
},
{
"epoch": 1.0580645161290323,
"grad_norm": 0.15528612822142446,
"learning_rate": 0.0001637655601011454,
"loss": 0.0806,
"step": 164
},
{
"epoch": 1.064516129032258,
"grad_norm": 0.1578175598943513,
"learning_rate": 0.00016318480847573642,
"loss": 0.0962,
"step": 165
},
{
"epoch": 1.070967741935484,
"grad_norm": 0.2591491854389689,
"learning_rate": 0.00016260048775776804,
"loss": 0.1107,
"step": 166
},
{
"epoch": 1.0774193548387097,
"grad_norm": 0.1557337965418426,
"learning_rate": 0.00016201263095350833,
"loss": 0.0707,
"step": 167
},
{
"epoch": 1.0838709677419356,
"grad_norm": 0.24212581528206514,
"learning_rate": 0.0001614212712689668,
"loss": 0.0964,
"step": 168
},
{
"epoch": 1.0903225806451613,
"grad_norm": 0.12600203218602474,
"learning_rate": 0.00016082644210801844,
"loss": 0.0649,
"step": 169
},
{
"epoch": 1.096774193548387,
"grad_norm": 0.14377456483205683,
"learning_rate": 0.00016022817707051724,
"loss": 0.0778,
"step": 170
},
{
"epoch": 1.103225806451613,
"grad_norm": 0.1828183442232092,
"learning_rate": 0.00015962650995039783,
"loss": 0.1068,
"step": 171
},
{
"epoch": 1.1096774193548387,
"grad_norm": 0.1695078607562437,
"learning_rate": 0.00015902147473376694,
"loss": 0.0938,
"step": 172
},
{
"epoch": 1.1161290322580646,
"grad_norm": 0.10525262692642562,
"learning_rate": 0.00015841310559698343,
"loss": 0.0645,
"step": 173
},
{
"epoch": 1.1225806451612903,
"grad_norm": 0.15379971536703851,
"learning_rate": 0.0001578014369047279,
"loss": 0.0752,
"step": 174
},
{
"epoch": 1.129032258064516,
"grad_norm": 0.1352511385337785,
"learning_rate": 0.00015718650320806142,
"loss": 0.0803,
"step": 175
},
{
"epoch": 1.135483870967742,
"grad_norm": 0.1708537982318491,
"learning_rate": 0.00015656833924247398,
"loss": 0.0908,
"step": 176
},
{
"epoch": 1.1419354838709677,
"grad_norm": 0.16847128771716718,
"learning_rate": 0.00015594697992592232,
"loss": 0.0697,
"step": 177
},
{
"epoch": 1.1483870967741936,
"grad_norm": 0.14044376651199897,
"learning_rate": 0.00015532246035685756,
"loss": 0.0707,
"step": 178
},
{
"epoch": 1.1548387096774193,
"grad_norm": 0.13752732543534765,
"learning_rate": 0.00015469481581224272,
"loss": 0.0769,
"step": 179
},
{
"epoch": 1.1612903225806452,
"grad_norm": 0.20084535405957962,
"learning_rate": 0.00015406408174555976,
"loss": 0.1114,
"step": 180
},
{
"epoch": 1.167741935483871,
"grad_norm": 0.16882912932677738,
"learning_rate": 0.0001534302937848073,
"loss": 0.0977,
"step": 181
},
{
"epoch": 1.1741935483870969,
"grad_norm": 0.16088758279122345,
"learning_rate": 0.00015279348773048786,
"loss": 0.0862,
"step": 182
},
{
"epoch": 1.1806451612903226,
"grad_norm": 0.1610420842518173,
"learning_rate": 0.00015215369955358566,
"loss": 0.0909,
"step": 183
},
{
"epoch": 1.1870967741935483,
"grad_norm": 0.1567858759713509,
"learning_rate": 0.0001515109653935348,
"loss": 0.0988,
"step": 184
},
{
"epoch": 1.1935483870967742,
"grad_norm": 0.12400211503217344,
"learning_rate": 0.00015086532155617784,
"loss": 0.0813,
"step": 185
},
{
"epoch": 1.2,
"grad_norm": 0.12785181430507606,
"learning_rate": 0.00015021680451171498,
"loss": 0.0802,
"step": 186
},
{
"epoch": 1.206451612903226,
"grad_norm": 0.14298258948499543,
"learning_rate": 0.00014956545089264407,
"loss": 0.0843,
"step": 187
},
{
"epoch": 1.2129032258064516,
"grad_norm": 0.20148810623427083,
"learning_rate": 0.0001489112974916912,
"loss": 0.0942,
"step": 188
},
{
"epoch": 1.2193548387096773,
"grad_norm": 0.14657915199625932,
"learning_rate": 0.00014825438125973264,
"loss": 0.0829,
"step": 189
},
{
"epoch": 1.2258064516129032,
"grad_norm": 0.16147059815020365,
"learning_rate": 0.00014759473930370736,
"loss": 0.0756,
"step": 190
},
{
"epoch": 1.232258064516129,
"grad_norm": 0.17956881851269735,
"learning_rate": 0.0001469324088845212,
"loss": 0.1007,
"step": 191
},
{
"epoch": 1.238709677419355,
"grad_norm": 0.1578318532174886,
"learning_rate": 0.00014626742741494206,
"loss": 0.0983,
"step": 192
},
{
"epoch": 1.2451612903225806,
"grad_norm": 0.1649792762044239,
"learning_rate": 0.00014559983245748638,
"loss": 0.0905,
"step": 193
},
{
"epoch": 1.2516129032258063,
"grad_norm": 0.1455653603561888,
"learning_rate": 0.00014492966172229777,
"loss": 0.0791,
"step": 194
},
{
"epoch": 1.2580645161290323,
"grad_norm": 0.1312841252501999,
"learning_rate": 0.00014425695306501658,
"loss": 0.0613,
"step": 195
},
{
"epoch": 1.2645161290322582,
"grad_norm": 0.1361347125426188,
"learning_rate": 0.00014358174448464154,
"loss": 0.0773,
"step": 196
},
{
"epoch": 1.270967741935484,
"grad_norm": 0.17551022044504175,
"learning_rate": 0.00014290407412138366,
"loss": 0.0985,
"step": 197
},
{
"epoch": 1.2774193548387096,
"grad_norm": 0.1456899710914213,
"learning_rate": 0.00014222398025451135,
"loss": 0.0827,
"step": 198
},
{
"epoch": 1.2838709677419355,
"grad_norm": 0.19482064326323745,
"learning_rate": 0.00014154150130018866,
"loss": 0.0974,
"step": 199
},
{
"epoch": 1.2903225806451613,
"grad_norm": 0.13544624920378107,
"learning_rate": 0.0001408566758093048,
"loss": 0.0742,
"step": 200
},
{
"epoch": 1.2967741935483872,
"grad_norm": 0.12685823273525554,
"learning_rate": 0.00014016954246529696,
"loss": 0.0677,
"step": 201
},
{
"epoch": 1.303225806451613,
"grad_norm": 0.15457862761329938,
"learning_rate": 0.00013948014008196487,
"loss": 0.0815,
"step": 202
},
{
"epoch": 1.3096774193548386,
"grad_norm": 0.1482256655702007,
"learning_rate": 0.0001387885076012785,
"loss": 0.0961,
"step": 203
},
{
"epoch": 1.3161290322580645,
"grad_norm": 0.20896278322687534,
"learning_rate": 0.00013809468409117846,
"loss": 0.1049,
"step": 204
},
{
"epoch": 1.3225806451612903,
"grad_norm": 0.175019480492322,
"learning_rate": 0.00013739870874336898,
"loss": 0.0884,
"step": 205
},
{
"epoch": 1.3290322580645162,
"grad_norm": 0.14363356864891735,
"learning_rate": 0.00013670062087110422,
"loss": 0.0864,
"step": 206
},
{
"epoch": 1.335483870967742,
"grad_norm": 0.1397081258858762,
"learning_rate": 0.00013600045990696762,
"loss": 0.0797,
"step": 207
},
{
"epoch": 1.3419354838709676,
"grad_norm": 0.1667674455506937,
"learning_rate": 0.0001352982654006444,
"loss": 0.0977,
"step": 208
},
{
"epoch": 1.3483870967741935,
"grad_norm": 0.1409898972781143,
"learning_rate": 0.00013459407701668763,
"loss": 0.0923,
"step": 209
},
{
"epoch": 1.3548387096774195,
"grad_norm": 0.16290520924324786,
"learning_rate": 0.00013388793453227767,
"loss": 0.1033,
"step": 210
},
{
"epoch": 1.3612903225806452,
"grad_norm": 0.14967784709530377,
"learning_rate": 0.0001331798778349752,
"loss": 0.0843,
"step": 211
},
{
"epoch": 1.367741935483871,
"grad_norm": 0.14370370779202904,
"learning_rate": 0.00013246994692046836,
"loss": 0.0845,
"step": 212
},
{
"epoch": 1.3741935483870968,
"grad_norm": 0.15288920332301245,
"learning_rate": 0.00013175818189031327,
"loss": 0.0962,
"step": 213
},
{
"epoch": 1.3806451612903226,
"grad_norm": 0.16907810212754584,
"learning_rate": 0.00013104462294966896,
"loss": 0.1033,
"step": 214
},
{
"epoch": 1.3870967741935485,
"grad_norm": 0.17023880837738012,
"learning_rate": 0.00013032931040502627,
"loss": 0.1155,
"step": 215
},
{
"epoch": 1.3935483870967742,
"grad_norm": 0.12189251069561345,
"learning_rate": 0.00012961228466193116,
"loss": 0.0692,
"step": 216
},
{
"epoch": 1.4,
"grad_norm": 0.18108512316647296,
"learning_rate": 0.00012889358622270223,
"loss": 0.086,
"step": 217
},
{
"epoch": 1.4064516129032258,
"grad_norm": 0.15804455793477432,
"learning_rate": 0.00012817325568414297,
"loss": 0.0751,
"step": 218
},
{
"epoch": 1.4129032258064516,
"grad_norm": 0.13668879316730062,
"learning_rate": 0.00012745133373524853,
"loss": 0.0786,
"step": 219
},
{
"epoch": 1.4193548387096775,
"grad_norm": 0.18610949095399393,
"learning_rate": 0.0001267278611549073,
"loss": 0.0732,
"step": 220
},
{
"epoch": 1.4258064516129032,
"grad_norm": 0.186875270696186,
"learning_rate": 0.00012600287880959763,
"loss": 0.1051,
"step": 221
},
{
"epoch": 1.432258064516129,
"grad_norm": 0.15558801012557805,
"learning_rate": 0.0001252764276510792,
"loss": 0.0879,
"step": 222
},
{
"epoch": 1.4387096774193548,
"grad_norm": 0.16188457902393685,
"learning_rate": 0.00012454854871407994,
"loss": 0.0887,
"step": 223
},
{
"epoch": 1.4451612903225808,
"grad_norm": 0.1332789968563102,
"learning_rate": 0.00012381928311397806,
"loss": 0.0819,
"step": 224
},
{
"epoch": 1.4516129032258065,
"grad_norm": 0.13104439295337186,
"learning_rate": 0.0001230886720444796,
"loss": 0.0992,
"step": 225
},
{
"epoch": 1.4580645161290322,
"grad_norm": 0.1741921210517873,
"learning_rate": 0.00012235675677529158,
"loss": 0.1049,
"step": 226
},
{
"epoch": 1.4645161290322581,
"grad_norm": 0.17686059237450052,
"learning_rate": 0.00012162357864979072,
"loss": 0.0942,
"step": 227
},
{
"epoch": 1.4709677419354839,
"grad_norm": 0.14208255193217198,
"learning_rate": 0.00012088917908268821,
"loss": 0.0932,
"step": 228
},
{
"epoch": 1.4774193548387098,
"grad_norm": 0.16207984639669018,
"learning_rate": 0.00012015359955769021,
"loss": 0.0952,
"step": 229
},
{
"epoch": 1.4838709677419355,
"grad_norm": 0.10567516069189406,
"learning_rate": 0.00011941688162515467,
"loss": 0.0698,
"step": 230
},
{
"epoch": 1.4903225806451612,
"grad_norm": 0.14315458863752037,
"learning_rate": 0.00011867906689974428,
"loss": 0.0706,
"step": 231
},
{
"epoch": 1.4967741935483871,
"grad_norm": 0.14644874259664967,
"learning_rate": 0.00011794019705807584,
"loss": 0.0954,
"step": 232
},
{
"epoch": 1.5032258064516129,
"grad_norm": 0.1522258926853921,
"learning_rate": 0.00011720031383636585,
"loss": 0.1026,
"step": 233
},
{
"epoch": 1.5096774193548388,
"grad_norm": 0.2389589012648026,
"learning_rate": 0.00011645945902807341,
"loss": 0.0981,
"step": 234
},
{
"epoch": 1.5161290322580645,
"grad_norm": 0.1539160607687386,
"learning_rate": 0.00011571767448153901,
"loss": 0.074,
"step": 235
},
{
"epoch": 1.5225806451612902,
"grad_norm": 0.17340052794011998,
"learning_rate": 0.00011497500209762102,
"loss": 0.0943,
"step": 236
},
{
"epoch": 1.5290322580645161,
"grad_norm": 0.11294207936842038,
"learning_rate": 0.00011423148382732853,
"loss": 0.0644,
"step": 237
},
{
"epoch": 1.535483870967742,
"grad_norm": 0.14391081206665055,
"learning_rate": 0.00011348716166945195,
"loss": 0.0774,
"step": 238
},
{
"epoch": 1.5419354838709678,
"grad_norm": 0.16298875474891092,
"learning_rate": 0.0001127420776681905,
"loss": 0.0786,
"step": 239
},
{
"epoch": 1.5483870967741935,
"grad_norm": 0.1325076119120941,
"learning_rate": 0.00011199627391077732,
"loss": 0.0889,
"step": 240
},
{
"epoch": 1.5548387096774192,
"grad_norm": 0.10971482413094012,
"learning_rate": 0.00011124979252510208,
"loss": 0.0737,
"step": 241
},
{
"epoch": 1.5612903225806452,
"grad_norm": 0.14614299703062,
"learning_rate": 0.0001105026756773314,
"loss": 0.0682,
"step": 242
},
{
"epoch": 1.567741935483871,
"grad_norm": 0.22690104820583093,
"learning_rate": 0.00010975496556952682,
"loss": 0.1094,
"step": 243
},
{
"epoch": 1.5741935483870968,
"grad_norm": 0.2561727438038473,
"learning_rate": 0.00010900670443726135,
"loss": 0.0851,
"step": 244
},
{
"epoch": 1.5806451612903225,
"grad_norm": 0.15371305701947427,
"learning_rate": 0.00010825793454723325,
"loss": 0.0923,
"step": 245
},
{
"epoch": 1.5870967741935482,
"grad_norm": 0.16267574566743875,
"learning_rate": 0.00010750869819487883,
"loss": 0.1036,
"step": 246
},
{
"epoch": 1.5935483870967742,
"grad_norm": 0.15591155698404394,
"learning_rate": 0.00010675903770198333,
"loss": 0.0893,
"step": 247
},
{
"epoch": 1.6,
"grad_norm": 0.14338972962339533,
"learning_rate": 0.00010600899541429004,
"loss": 0.0837,
"step": 248
},
{
"epoch": 1.6064516129032258,
"grad_norm": 0.12387607320751257,
"learning_rate": 0.00010525861369910877,
"loss": 0.0755,
"step": 249
},
{
"epoch": 1.6129032258064515,
"grad_norm": 0.16606169294386383,
"learning_rate": 0.00010450793494292224,
"loss": 0.1043,
"step": 250
},
{
"epoch": 1.6193548387096774,
"grad_norm": 0.1795920159350681,
"learning_rate": 0.00010375700154899208,
"loss": 0.1008,
"step": 251
},
{
"epoch": 1.6258064516129034,
"grad_norm": 0.13025097291519463,
"learning_rate": 0.00010300585593496348,
"loss": 0.0851,
"step": 252
},
{
"epoch": 1.632258064516129,
"grad_norm": 0.14349816154023654,
"learning_rate": 0.00010225454053046921,
"loss": 0.0807,
"step": 253
},
{
"epoch": 1.6387096774193548,
"grad_norm": 0.16695448888874226,
"learning_rate": 0.00010150309777473306,
"loss": 0.1117,
"step": 254
},
{
"epoch": 1.6451612903225805,
"grad_norm": 0.15743712528269815,
"learning_rate": 0.0001007515701141722,
"loss": 0.086,
"step": 255
},
{
"epoch": 1.6516129032258065,
"grad_norm": 0.16704335210894908,
"learning_rate": 0.0001,
"loss": 0.0886,
"step": 256
},
{
"epoch": 1.6580645161290324,
"grad_norm": 0.10245531520994122,
"learning_rate": 9.924842988582782e-05,
"loss": 0.0678,
"step": 257
},
{
"epoch": 1.664516129032258,
"grad_norm": 0.1688033836669086,
"learning_rate": 9.849690222526698e-05,
"loss": 0.0958,
"step": 258
},
{
"epoch": 1.6709677419354838,
"grad_norm": 0.1338126992775965,
"learning_rate": 9.77454594695308e-05,
"loss": 0.0782,
"step": 259
},
{
"epoch": 1.6774193548387095,
"grad_norm": 0.1332820451604909,
"learning_rate": 9.699414406503654e-05,
"loss": 0.0856,
"step": 260
},
{
"epoch": 1.6838709677419355,
"grad_norm": 0.1449401720605745,
"learning_rate": 9.624299845100795e-05,
"loss": 0.0882,
"step": 261
},
{
"epoch": 1.6903225806451614,
"grad_norm": 0.19395703151279187,
"learning_rate": 9.549206505707777e-05,
"loss": 0.1148,
"step": 262
},
{
"epoch": 1.696774193548387,
"grad_norm": 0.13482277559178169,
"learning_rate": 9.474138630089124e-05,
"loss": 0.0644,
"step": 263
},
{
"epoch": 1.7032258064516128,
"grad_norm": 0.18530593840863338,
"learning_rate": 9.399100458570997e-05,
"loss": 0.1074,
"step": 264
},
{
"epoch": 1.7096774193548387,
"grad_norm": 0.1897995033582595,
"learning_rate": 9.324096229801674e-05,
"loss": 0.0867,
"step": 265
},
{
"epoch": 1.7161290322580647,
"grad_norm": 0.16472780150681127,
"learning_rate": 9.249130180512118e-05,
"loss": 0.0896,
"step": 266
},
{
"epoch": 1.7225806451612904,
"grad_norm": 0.15525608679066774,
"learning_rate": 9.174206545276677e-05,
"loss": 0.0865,
"step": 267
},
{
"epoch": 1.729032258064516,
"grad_norm": 0.15426999703424252,
"learning_rate": 9.099329556273866e-05,
"loss": 0.0853,
"step": 268
},
{
"epoch": 1.7354838709677418,
"grad_norm": 0.19103115451320254,
"learning_rate": 9.024503443047319e-05,
"loss": 0.0993,
"step": 269
},
{
"epoch": 1.7419354838709677,
"grad_norm": 0.12323460303269068,
"learning_rate": 8.949732432266866e-05,
"loss": 0.0723,
"step": 270
},
{
"epoch": 1.7483870967741937,
"grad_norm": 0.1339668030381976,
"learning_rate": 8.875020747489794e-05,
"loss": 0.0852,
"step": 271
},
{
"epoch": 1.7548387096774194,
"grad_norm": 0.1385020226698345,
"learning_rate": 8.800372608922271e-05,
"loss": 0.0773,
"step": 272
},
{
"epoch": 1.761290322580645,
"grad_norm": 0.21835470061774626,
"learning_rate": 8.72579223318095e-05,
"loss": 0.1167,
"step": 273
},
{
"epoch": 1.7677419354838708,
"grad_norm": 0.1425322245524885,
"learning_rate": 8.651283833054809e-05,
"loss": 0.0801,
"step": 274
},
{
"epoch": 1.7741935483870968,
"grad_norm": 0.1794495686810833,
"learning_rate": 8.57685161726715e-05,
"loss": 0.0784,
"step": 275
},
{
"epoch": 1.7806451612903227,
"grad_norm": 0.15688310999446023,
"learning_rate": 8.5024997902379e-05,
"loss": 0.1001,
"step": 276
},
{
"epoch": 1.7870967741935484,
"grad_norm": 0.15417421931142297,
"learning_rate": 8.428232551846101e-05,
"loss": 0.0898,
"step": 277
},
{
"epoch": 1.793548387096774,
"grad_norm": 0.17958092903510822,
"learning_rate": 8.35405409719266e-05,
"loss": 0.0921,
"step": 278
},
{
"epoch": 1.8,
"grad_norm": 0.165061388870699,
"learning_rate": 8.279968616363418e-05,
"loss": 0.0935,
"step": 279
},
{
"epoch": 1.8064516129032258,
"grad_norm": 0.1116000947584692,
"learning_rate": 8.205980294192421e-05,
"loss": 0.0684,
"step": 280
},
{
"epoch": 1.8129032258064517,
"grad_norm": 0.12370745355237868,
"learning_rate": 8.132093310025571e-05,
"loss": 0.0764,
"step": 281
},
{
"epoch": 1.8193548387096774,
"grad_norm": 0.1724137684896053,
"learning_rate": 8.058311837484535e-05,
"loss": 0.0969,
"step": 282
},
{
"epoch": 1.8258064516129031,
"grad_norm": 0.16125193051943912,
"learning_rate": 7.984640044230983e-05,
"loss": 0.0868,
"step": 283
},
{
"epoch": 1.832258064516129,
"grad_norm": 0.13929471780512187,
"learning_rate": 7.911082091731181e-05,
"loss": 0.0701,
"step": 284
},
{
"epoch": 1.838709677419355,
"grad_norm": 0.1172859356132756,
"learning_rate": 7.837642135020929e-05,
"loss": 0.0705,
"step": 285
},
{
"epoch": 1.8451612903225807,
"grad_norm": 0.1416371336298496,
"learning_rate": 7.764324322470841e-05,
"loss": 0.0683,
"step": 286
},
{
"epoch": 1.8516129032258064,
"grad_norm": 0.19198949443360017,
"learning_rate": 7.691132795552043e-05,
"loss": 0.0894,
"step": 287
},
{
"epoch": 1.8580645161290321,
"grad_norm": 0.19384526534149363,
"learning_rate": 7.618071688602199e-05,
"loss": 0.0954,
"step": 288
},
{
"epoch": 1.864516129032258,
"grad_norm": 0.21508649416468206,
"learning_rate": 7.54514512859201e-05,
"loss": 0.0986,
"step": 289
},
{
"epoch": 1.870967741935484,
"grad_norm": 0.1952670486171591,
"learning_rate": 7.472357234892082e-05,
"loss": 0.1049,
"step": 290
},
{
"epoch": 1.8774193548387097,
"grad_norm": 0.13259304301639457,
"learning_rate": 7.399712119040238e-05,
"loss": 0.0784,
"step": 291
},
{
"epoch": 1.8838709677419354,
"grad_norm": 0.1367442063153438,
"learning_rate": 7.327213884509272e-05,
"loss": 0.0828,
"step": 292
},
{
"epoch": 1.8903225806451613,
"grad_norm": 0.16389743747526248,
"learning_rate": 7.254866626475152e-05,
"loss": 0.0927,
"step": 293
},
{
"epoch": 1.896774193548387,
"grad_norm": 0.22089121957380548,
"learning_rate": 7.182674431585704e-05,
"loss": 0.1077,
"step": 294
},
{
"epoch": 1.903225806451613,
"grad_norm": 0.14882797154919192,
"learning_rate": 7.110641377729778e-05,
"loss": 0.0879,
"step": 295
},
{
"epoch": 1.9096774193548387,
"grad_norm": 0.12794357419440644,
"learning_rate": 7.038771533806884e-05,
"loss": 0.0719,
"step": 296
},
{
"epoch": 1.9161290322580644,
"grad_norm": 0.1846824594193936,
"learning_rate": 6.967068959497376e-05,
"loss": 0.0895,
"step": 297
},
{
"epoch": 1.9225806451612903,
"grad_norm": 0.17646276082900025,
"learning_rate": 6.895537705033108e-05,
"loss": 0.0996,
"step": 298
},
{
"epoch": 1.9290322580645163,
"grad_norm": 0.15604696338212207,
"learning_rate": 6.824181810968675e-05,
"loss": 0.087,
"step": 299
},
{
"epoch": 1.935483870967742,
"grad_norm": 0.15478000466177547,
"learning_rate": 6.753005307953167e-05,
"loss": 0.0831,
"step": 300
},
{
"epoch": 1.9419354838709677,
"grad_norm": 0.19642059718505833,
"learning_rate": 6.682012216502484e-05,
"loss": 0.1073,
"step": 301
},
{
"epoch": 1.9483870967741934,
"grad_norm": 0.17255606883063224,
"learning_rate": 6.611206546772237e-05,
"loss": 0.1018,
"step": 302
},
{
"epoch": 1.9548387096774194,
"grad_norm": 0.1393602083174125,
"learning_rate": 6.54059229833124e-05,
"loss": 0.0859,
"step": 303
},
{
"epoch": 1.9612903225806453,
"grad_norm": 0.16652643955266602,
"learning_rate": 6.47017345993556e-05,
"loss": 0.0899,
"step": 304
},
{
"epoch": 1.967741935483871,
"grad_norm": 0.16878660937501272,
"learning_rate": 6.39995400930324e-05,
"loss": 0.0998,
"step": 305
},
{
"epoch": 1.9741935483870967,
"grad_norm": 0.12969061293334938,
"learning_rate": 6.329937912889582e-05,
"loss": 0.0706,
"step": 306
},
{
"epoch": 1.9806451612903224,
"grad_norm": 0.15739894009886643,
"learning_rate": 6.260129125663106e-05,
"loss": 0.0888,
"step": 307
},
{
"epoch": 1.9870967741935484,
"grad_norm": 0.1752067977727585,
"learning_rate": 6.190531590882159e-05,
"loss": 0.0826,
"step": 308
},
{
"epoch": 1.9935483870967743,
"grad_norm": 0.21002280616145647,
"learning_rate": 6.121149239872151e-05,
"loss": 0.123,
"step": 309
},
{
"epoch": 2.0,
"grad_norm": 0.13850366944373244,
"learning_rate": 6.051985991803517e-05,
"loss": 0.073,
"step": 310
},
{
"epoch": 2.0,
"eval_loss": 0.16690203547477722,
"eval_runtime": 25.4682,
"eval_samples_per_second": 5.144,
"eval_steps_per_second": 0.668,
"step": 310
},
{
"epoch": 2.0064516129032257,
"grad_norm": 0.1124990090834071,
"learning_rate": 5.983045753470308e-05,
"loss": 0.0485,
"step": 311
},
{
"epoch": 2.0129032258064514,
"grad_norm": 0.08246086368829486,
"learning_rate": 5.9143324190695196e-05,
"loss": 0.0411,
"step": 312
},
{
"epoch": 2.0193548387096776,
"grad_norm": 0.11824209839228844,
"learning_rate": 5.845849869981137e-05,
"loss": 0.0578,
"step": 313
},
{
"epoch": 2.0258064516129033,
"grad_norm": 0.09183702414760068,
"learning_rate": 5.777601974548866e-05,
"loss": 0.0405,
"step": 314
},
{
"epoch": 2.032258064516129,
"grad_norm": 0.10138990147374953,
"learning_rate": 5.709592587861637e-05,
"loss": 0.045,
"step": 315
},
{
"epoch": 2.0387096774193547,
"grad_norm": 0.13174518232765414,
"learning_rate": 5.6418255515358486e-05,
"loss": 0.0417,
"step": 316
},
{
"epoch": 2.0451612903225804,
"grad_norm": 0.11134581768568716,
"learning_rate": 5.574304693498346e-05,
"loss": 0.0351,
"step": 317
},
{
"epoch": 2.0516129032258066,
"grad_norm": 0.10372734844639428,
"learning_rate": 5.507033827770225e-05,
"loss": 0.0458,
"step": 318
},
{
"epoch": 2.0580645161290323,
"grad_norm": 0.09893496819685892,
"learning_rate": 5.4400167542513636e-05,
"loss": 0.0421,
"step": 319
},
{
"epoch": 2.064516129032258,
"grad_norm": 0.1544849048513997,
"learning_rate": 5.3732572585057974e-05,
"loss": 0.0449,
"step": 320
},
{
"epoch": 2.0709677419354837,
"grad_norm": 0.12211629106983518,
"learning_rate": 5.306759111547881e-05,
"loss": 0.0341,
"step": 321
},
{
"epoch": 2.07741935483871,
"grad_norm": 0.1519039459170261,
"learning_rate": 5.240526069629265e-05,
"loss": 0.0477,
"step": 322
},
{
"epoch": 2.0838709677419356,
"grad_norm": 0.10882973801143421,
"learning_rate": 5.174561874026741e-05,
"loss": 0.0376,
"step": 323
},
{
"epoch": 2.0903225806451613,
"grad_norm": 0.09785322895617321,
"learning_rate": 5.108870250830882e-05,
"loss": 0.0305,
"step": 324
},
{
"epoch": 2.096774193548387,
"grad_norm": 0.14990301799413153,
"learning_rate": 5.0434549107355944e-05,
"loss": 0.0409,
"step": 325
},
{
"epoch": 2.1032258064516127,
"grad_norm": 0.1915279670852206,
"learning_rate": 4.978319548828504e-05,
"loss": 0.045,
"step": 326
},
{
"epoch": 2.109677419354839,
"grad_norm": 0.15445162243894608,
"learning_rate": 4.9134678443822166e-05,
"loss": 0.0334,
"step": 327
},
{
"epoch": 2.1161290322580646,
"grad_norm": 0.34562522599625123,
"learning_rate": 4.8489034606465225e-05,
"loss": 0.0395,
"step": 328
},
{
"epoch": 2.1225806451612903,
"grad_norm": 0.3183911745576579,
"learning_rate": 4.784630044641435e-05,
"loss": 0.0546,
"step": 329
},
{
"epoch": 2.129032258064516,
"grad_norm": 0.1619196710614726,
"learning_rate": 4.7206512269512124e-05,
"loss": 0.0446,
"step": 330
},
{
"epoch": 2.135483870967742,
"grad_norm": 0.24657637060341897,
"learning_rate": 4.65697062151927e-05,
"loss": 0.0349,
"step": 331
},
{
"epoch": 2.141935483870968,
"grad_norm": 0.11723613581448042,
"learning_rate": 4.593591825444028e-05,
"loss": 0.0337,
"step": 332
},
{
"epoch": 2.1483870967741936,
"grad_norm": 0.12413977056977685,
"learning_rate": 4.530518418775733e-05,
"loss": 0.0355,
"step": 333
},
{
"epoch": 2.1548387096774193,
"grad_norm": 0.1368012687546921,
"learning_rate": 4.4677539643142454e-05,
"loss": 0.0369,
"step": 334
},
{
"epoch": 2.161290322580645,
"grad_norm": 0.132092425970093,
"learning_rate": 4.40530200740777e-05,
"loss": 0.0346,
"step": 335
},
{
"epoch": 2.167741935483871,
"grad_norm": 0.17255652388608525,
"learning_rate": 4.343166075752605e-05,
"loss": 0.0425,
"step": 336
},
{
"epoch": 2.174193548387097,
"grad_norm": 0.10537625462278236,
"learning_rate": 4.281349679193861e-05,
"loss": 0.0363,
"step": 337
},
{
"epoch": 2.1806451612903226,
"grad_norm": 0.08446560264535163,
"learning_rate": 4.2198563095272116e-05,
"loss": 0.0303,
"step": 338
},
{
"epoch": 2.1870967741935483,
"grad_norm": 0.11155624490505092,
"learning_rate": 4.158689440301657e-05,
"loss": 0.0424,
"step": 339
},
{
"epoch": 2.193548387096774,
"grad_norm": 0.10614591478252769,
"learning_rate": 4.097852526623307e-05,
"loss": 0.0374,
"step": 340
},
{
"epoch": 2.2,
"grad_norm": 0.15380955207075353,
"learning_rate": 4.0373490049602204e-05,
"loss": 0.0506,
"step": 341
},
{
"epoch": 2.206451612903226,
"grad_norm": 0.11905414571153279,
"learning_rate": 3.977182292948283e-05,
"loss": 0.0517,
"step": 342
},
{
"epoch": 2.2129032258064516,
"grad_norm": 0.14425425935781702,
"learning_rate": 3.9173557891981573e-05,
"loss": 0.0467,
"step": 343
},
{
"epoch": 2.2193548387096773,
"grad_norm": 0.12453230757386329,
"learning_rate": 3.857872873103322e-05,
"loss": 0.0435,
"step": 344
},
{
"epoch": 2.225806451612903,
"grad_norm": 0.10038267160538758,
"learning_rate": 3.7987369046491684e-05,
"loss": 0.0387,
"step": 345
},
{
"epoch": 2.232258064516129,
"grad_norm": 0.1438025015993235,
"learning_rate": 3.7399512242231995e-05,
"loss": 0.0346,
"step": 346
},
{
"epoch": 2.238709677419355,
"grad_norm": 0.26072179652296745,
"learning_rate": 3.6815191524263624e-05,
"loss": 0.0545,
"step": 347
},
{
"epoch": 2.2451612903225806,
"grad_norm": 0.08770511437142381,
"learning_rate": 3.623443989885462e-05,
"loss": 0.0327,
"step": 348
},
{
"epoch": 2.2516129032258063,
"grad_norm": 0.12301181296016188,
"learning_rate": 3.565729017066729e-05,
"loss": 0.039,
"step": 349
},
{
"epoch": 2.258064516129032,
"grad_norm": 0.13814667341858822,
"learning_rate": 3.508377494090521e-05,
"loss": 0.0356,
"step": 350
},
{
"epoch": 2.264516129032258,
"grad_norm": 0.12413533495298362,
"learning_rate": 3.45139266054715e-05,
"loss": 0.0415,
"step": 351
},
{
"epoch": 2.270967741935484,
"grad_norm": 0.11031185879435731,
"learning_rate": 3.394777735313919e-05,
"loss": 0.0363,
"step": 352
},
{
"epoch": 2.2774193548387096,
"grad_norm": 0.11353520726841541,
"learning_rate": 3.338535916373266e-05,
"loss": 0.0376,
"step": 353
},
{
"epoch": 2.2838709677419353,
"grad_norm": 0.23000368159173218,
"learning_rate": 3.2826703806321525e-05,
"loss": 0.0444,
"step": 354
},
{
"epoch": 2.2903225806451615,
"grad_norm": 0.10390022981327145,
"learning_rate": 3.227184283742591e-05,
"loss": 0.0308,
"step": 355
},
{
"epoch": 2.296774193548387,
"grad_norm": 0.14881153611607173,
"learning_rate": 3.17208075992339e-05,
"loss": 0.0428,
"step": 356
},
{
"epoch": 2.303225806451613,
"grad_norm": 0.22847373170988944,
"learning_rate": 3.117362921783134e-05,
"loss": 0.0449,
"step": 357
},
{
"epoch": 2.3096774193548386,
"grad_norm": 0.1261660214676178,
"learning_rate": 3.063033860144339e-05,
"loss": 0.0354,
"step": 358
},
{
"epoch": 2.3161290322580643,
"grad_norm": 0.16508383182402034,
"learning_rate": 3.0090966438688772e-05,
"loss": 0.0404,
"step": 359
},
{
"epoch": 2.3225806451612905,
"grad_norm": 0.14003607495149023,
"learning_rate": 2.9555543196846292e-05,
"loss": 0.0417,
"step": 360
},
{
"epoch": 2.329032258064516,
"grad_norm": 0.14067522061651452,
"learning_rate": 2.9024099120133673e-05,
"loss": 0.0325,
"step": 361
},
{
"epoch": 2.335483870967742,
"grad_norm": 0.17575236098692995,
"learning_rate": 2.8496664227999415e-05,
"loss": 0.0533,
"step": 362
},
{
"epoch": 2.3419354838709676,
"grad_norm": 0.11116834491408094,
"learning_rate": 2.7973268313426837e-05,
"loss": 0.0347,
"step": 363
},
{
"epoch": 2.3483870967741938,
"grad_norm": 0.16113939577201203,
"learning_rate": 2.745394094125141e-05,
"loss": 0.0472,
"step": 364
},
{
"epoch": 2.3548387096774195,
"grad_norm": 0.10955543396615161,
"learning_rate": 2.6938711446490606e-05,
"loss": 0.0357,
"step": 365
},
{
"epoch": 2.361290322580645,
"grad_norm": 0.17969798691111394,
"learning_rate": 2.6427608932686843e-05,
"loss": 0.0584,
"step": 366
},
{
"epoch": 2.367741935483871,
"grad_norm": 0.08542932983952008,
"learning_rate": 2.5920662270263653e-05,
"loss": 0.032,
"step": 367
},
{
"epoch": 2.3741935483870966,
"grad_norm": 0.14296192462496796,
"learning_rate": 2.5417900094894744e-05,
"loss": 0.0463,
"step": 368
},
{
"epoch": 2.3806451612903228,
"grad_norm": 0.11857064948287055,
"learning_rate": 2.4919350805886577e-05,
"loss": 0.0328,
"step": 369
},
{
"epoch": 2.3870967741935485,
"grad_norm": 0.1614783419773078,
"learning_rate": 2.4425042564574184e-05,
"loss": 0.0398,
"step": 370
},
{
"epoch": 2.393548387096774,
"grad_norm": 0.12389155943230613,
"learning_rate": 2.3935003292730296e-05,
"loss": 0.0389,
"step": 371
},
{
"epoch": 2.4,
"grad_norm": 0.11380344182842446,
"learning_rate": 2.344926067098836e-05,
"loss": 0.0412,
"step": 372
},
{
"epoch": 2.4064516129032256,
"grad_norm": 0.09336252258103142,
"learning_rate": 2.2967842137278706e-05,
"loss": 0.0311,
"step": 373
},
{
"epoch": 2.412903225806452,
"grad_norm": 0.1312469591783424,
"learning_rate": 2.2490774885278908e-05,
"loss": 0.0363,
"step": 374
},
{
"epoch": 2.4193548387096775,
"grad_norm": 0.1928829483525128,
"learning_rate": 2.201808586287757e-05,
"loss": 0.0447,
"step": 375
},
{
"epoch": 2.425806451612903,
"grad_norm": 0.1407497782508695,
"learning_rate": 2.15498017706521e-05,
"loss": 0.0425,
"step": 376
},
{
"epoch": 2.432258064516129,
"grad_norm": 0.12625194439970722,
"learning_rate": 2.1085949060360654e-05,
"loss": 0.0394,
"step": 377
},
{
"epoch": 2.4387096774193546,
"grad_norm": 0.1650711686265322,
"learning_rate": 2.0626553933447734e-05,
"loss": 0.0356,
"step": 378
},
{
"epoch": 2.445161290322581,
"grad_norm": 0.203794329158885,
"learning_rate": 2.01716423395644e-05,
"loss": 0.046,
"step": 379
},
{
"epoch": 2.4516129032258065,
"grad_norm": 0.0983865374506401,
"learning_rate": 1.9721239975102313e-05,
"loss": 0.0383,
"step": 380
},
{
"epoch": 2.458064516129032,
"grad_norm": 0.11475237817508949,
"learning_rate": 1.9275372281742242e-05,
"loss": 0.0341,
"step": 381
},
{
"epoch": 2.464516129032258,
"grad_norm": 0.11696076741124051,
"learning_rate": 1.8834064445016953e-05,
"loss": 0.0351,
"step": 382
},
{
"epoch": 2.4709677419354836,
"grad_norm": 0.11440255447742229,
"learning_rate": 1.839734139288868e-05,
"loss": 0.0344,
"step": 383
},
{
"epoch": 2.47741935483871,
"grad_norm": 0.12722166866147971,
"learning_rate": 1.7965227794340877e-05,
"loss": 0.0327,
"step": 384
},
{
"epoch": 2.4838709677419355,
"grad_norm": 0.11977054711252778,
"learning_rate": 1.753774805798486e-05,
"loss": 0.0374,
"step": 385
},
{
"epoch": 2.490322580645161,
"grad_norm": 0.19041811914083892,
"learning_rate": 1.7114926330680957e-05,
"loss": 0.0448,
"step": 386
},
{
"epoch": 2.496774193548387,
"grad_norm": 0.12908385372034334,
"learning_rate": 1.6696786496174578e-05,
"loss": 0.0421,
"step": 387
},
{
"epoch": 2.5032258064516126,
"grad_norm": 0.11345958947234494,
"learning_rate": 1.6283352173747145e-05,
"loss": 0.0337,
"step": 388
},
{
"epoch": 2.509677419354839,
"grad_norm": 0.13882462080577593,
"learning_rate": 1.587464671688187e-05,
"loss": 0.0399,
"step": 389
},
{
"epoch": 2.5161290322580645,
"grad_norm": 0.15241110627485502,
"learning_rate": 1.5470693211944643e-05,
"loss": 0.0331,
"step": 390
},
{
"epoch": 2.52258064516129,
"grad_norm": 0.10975692065298487,
"learning_rate": 1.5071514476879878e-05,
"loss": 0.0379,
"step": 391
},
{
"epoch": 2.5290322580645164,
"grad_norm": 0.1379096296948562,
"learning_rate": 1.4677133059921632e-05,
"loss": 0.0362,
"step": 392
},
{
"epoch": 2.535483870967742,
"grad_norm": 0.10190107251103887,
"learning_rate": 1.4287571238320053e-05,
"loss": 0.0371,
"step": 393
},
{
"epoch": 2.541935483870968,
"grad_norm": 0.0875699618751271,
"learning_rate": 1.3902851017082864e-05,
"loss": 0.0335,
"step": 394
},
{
"epoch": 2.5483870967741935,
"grad_norm": 0.13655405710076232,
"learning_rate": 1.3522994127732414e-05,
"loss": 0.0388,
"step": 395
},
{
"epoch": 2.554838709677419,
"grad_norm": 0.11073272357568731,
"learning_rate": 1.3148022027078222e-05,
"loss": 0.0366,
"step": 396
},
{
"epoch": 2.5612903225806454,
"grad_norm": 0.11393589351391524,
"learning_rate": 1.2777955896004812e-05,
"loss": 0.0375,
"step": 397
},
{
"epoch": 2.567741935483871,
"grad_norm": 0.1294793091961239,
"learning_rate": 1.2412816638275404e-05,
"loss": 0.0447,
"step": 398
},
{
"epoch": 2.574193548387097,
"grad_norm": 0.12252957295031998,
"learning_rate": 1.2052624879351104e-05,
"loss": 0.036,
"step": 399
},
{
"epoch": 2.5806451612903225,
"grad_norm": 0.10701768631052949,
"learning_rate": 1.1697400965225747e-05,
"loss": 0.0375,
"step": 400
},
{
"epoch": 2.587096774193548,
"grad_norm": 0.1826126683217801,
"learning_rate": 1.134716496127679e-05,
"loss": 0.0398,
"step": 401
},
{
"epoch": 2.5935483870967744,
"grad_norm": 0.12179126729794809,
"learning_rate": 1.1001936651131717e-05,
"loss": 0.0396,
"step": 402
},
{
"epoch": 2.6,
"grad_norm": 0.11972903116479075,
"learning_rate": 1.0661735535550666e-05,
"loss": 0.0336,
"step": 403
},
{
"epoch": 2.606451612903226,
"grad_norm": 0.09797391900877583,
"learning_rate": 1.0326580831324817e-05,
"loss": 0.0291,
"step": 404
},
{
"epoch": 2.6129032258064515,
"grad_norm": 0.08807259670414254,
"learning_rate": 9.996491470190917e-06,
"loss": 0.0318,
"step": 405
},
{
"epoch": 2.6193548387096772,
"grad_norm": 0.12986706205285126,
"learning_rate": 9.671486097761917e-06,
"loss": 0.0408,
"step": 406
},
{
"epoch": 2.6258064516129034,
"grad_norm": 0.1331746828106328,
"learning_rate": 9.351583072473713e-06,
"loss": 0.0427,
"step": 407
},
{
"epoch": 2.632258064516129,
"grad_norm": 0.14912986532560119,
"learning_rate": 9.036800464548157e-06,
"loss": 0.0464,
"step": 408
},
{
"epoch": 2.638709677419355,
"grad_norm": 0.1440292521357826,
"learning_rate": 8.727156054972374e-06,
"loss": 0.0448,
"step": 409
},
{
"epoch": 2.6451612903225805,
"grad_norm": 0.16571651133316453,
"learning_rate": 8.422667334494249e-06,
"loss": 0.0449,
"step": 410
},
{
"epoch": 2.6516129032258062,
"grad_norm": 0.15006924502881824,
"learning_rate": 8.123351502634625e-06,
"loss": 0.0371,
"step": 411
},
{
"epoch": 2.6580645161290324,
"grad_norm": 0.3766926409227814,
"learning_rate": 7.82922546671555e-06,
"loss": 0.0522,
"step": 412
},
{
"epoch": 2.664516129032258,
"grad_norm": 0.1693245533307095,
"learning_rate": 7.54030584090537e-06,
"loss": 0.0523,
"step": 413
},
{
"epoch": 2.670967741935484,
"grad_norm": 0.16933380856734512,
"learning_rate": 7.256608945280319e-06,
"loss": 0.0448,
"step": 414
},
{
"epoch": 2.6774193548387095,
"grad_norm": 0.12372656470631721,
"learning_rate": 6.97815080490245e-06,
"loss": 0.041,
"step": 415
},
{
"epoch": 2.6838709677419352,
"grad_norm": 0.1139014768011305,
"learning_rate": 6.704947148914609e-06,
"loss": 0.034,
"step": 416
},
{
"epoch": 2.6903225806451614,
"grad_norm": 0.14509781428626975,
"learning_rate": 6.437013409651849e-06,
"loss": 0.0289,
"step": 417
},
{
"epoch": 2.696774193548387,
"grad_norm": 0.1403482292705624,
"learning_rate": 6.174364721769743e-06,
"loss": 0.0493,
"step": 418
},
{
"epoch": 2.703225806451613,
"grad_norm": 0.12152856591733673,
"learning_rate": 5.917015921389568e-06,
"loss": 0.0343,
"step": 419
},
{
"epoch": 2.709677419354839,
"grad_norm": 0.10494471511565585,
"learning_rate": 5.664981545260073e-06,
"loss": 0.03,
"step": 420
},
{
"epoch": 2.7161290322580647,
"grad_norm": 0.2176951472197835,
"learning_rate": 5.418275829936537e-06,
"loss": 0.0486,
"step": 421
},
{
"epoch": 2.7225806451612904,
"grad_norm": 0.09677627092271344,
"learning_rate": 5.176912710976467e-06,
"loss": 0.0318,
"step": 422
},
{
"epoch": 2.729032258064516,
"grad_norm": 0.14416553291024004,
"learning_rate": 4.940905822152453e-06,
"loss": 0.0356,
"step": 423
},
{
"epoch": 2.735483870967742,
"grad_norm": 0.10141360651760384,
"learning_rate": 4.710268494682146e-06,
"loss": 0.0346,
"step": 424
},
{
"epoch": 2.741935483870968,
"grad_norm": 0.13463566674635727,
"learning_rate": 4.485013756475076e-06,
"loss": 0.0443,
"step": 425
},
{
"epoch": 2.7483870967741937,
"grad_norm": 0.26842778961550184,
"learning_rate": 4.2651543313968145e-06,
"loss": 0.0445,
"step": 426
},
{
"epoch": 2.7548387096774194,
"grad_norm": 0.16014595507517537,
"learning_rate": 4.050702638550275e-06,
"loss": 0.0469,
"step": 427
},
{
"epoch": 2.761290322580645,
"grad_norm": 0.2690022963645026,
"learning_rate": 3.841670791574137e-06,
"loss": 0.0423,
"step": 428
},
{
"epoch": 2.767741935483871,
"grad_norm": 0.2323526833237967,
"learning_rate": 3.638070597958665e-06,
"loss": 0.035,
"step": 429
},
{
"epoch": 2.774193548387097,
"grad_norm": 0.11643458974488918,
"learning_rate": 3.4399135583787043e-06,
"loss": 0.0401,
"step": 430
},
{
"epoch": 2.7806451612903227,
"grad_norm": 0.12278976372135998,
"learning_rate": 3.2472108660439706e-06,
"loss": 0.04,
"step": 431
},
{
"epoch": 2.7870967741935484,
"grad_norm": 0.11369002878299092,
"learning_rate": 3.059973406066963e-06,
"loss": 0.0335,
"step": 432
},
{
"epoch": 2.793548387096774,
"grad_norm": 0.1426764505452139,
"learning_rate": 2.878211754847926e-06,
"loss": 0.0431,
"step": 433
},
{
"epoch": 2.8,
"grad_norm": 0.17614309627774977,
"learning_rate": 2.7019361794775156e-06,
"loss": 0.0336,
"step": 434
},
{
"epoch": 2.806451612903226,
"grad_norm": 0.13119646329419238,
"learning_rate": 2.5311566371568507e-06,
"loss": 0.0388,
"step": 435
},
{
"epoch": 2.8129032258064517,
"grad_norm": 0.2274358694692015,
"learning_rate": 2.365882774634998e-06,
"loss": 0.0483,
"step": 436
},
{
"epoch": 2.8193548387096774,
"grad_norm": 0.16809646480030987,
"learning_rate": 2.206123927664161e-06,
"loss": 0.0428,
"step": 437
},
{
"epoch": 2.825806451612903,
"grad_norm": 0.08861174696327136,
"learning_rate": 2.0518891204722168e-06,
"loss": 0.0297,
"step": 438
},
{
"epoch": 2.832258064516129,
"grad_norm": 0.1516310249747229,
"learning_rate": 1.903187065253076e-06,
"loss": 0.0486,
"step": 439
},
{
"epoch": 2.838709677419355,
"grad_norm": 0.1154092981393085,
"learning_rate": 1.7600261616745106e-06,
"loss": 0.044,
"step": 440
},
{
"epoch": 2.8451612903225807,
"grad_norm": 0.14626098217303432,
"learning_rate": 1.6224144964036681e-06,
"loss": 0.045,
"step": 441
},
{
"epoch": 2.8516129032258064,
"grad_norm": 0.09272906884038755,
"learning_rate": 1.4903598426503241e-06,
"loss": 0.0305,
"step": 442
},
{
"epoch": 2.858064516129032,
"grad_norm": 0.1688562089326058,
"learning_rate": 1.3638696597277679e-06,
"loss": 0.0402,
"step": 443
},
{
"epoch": 2.864516129032258,
"grad_norm": 0.11628209549950047,
"learning_rate": 1.2429510926314836e-06,
"loss": 0.0309,
"step": 444
},
{
"epoch": 2.870967741935484,
"grad_norm": 0.14018105443781761,
"learning_rate": 1.1276109716355287e-06,
"loss": 0.0466,
"step": 445
},
{
"epoch": 2.8774193548387097,
"grad_norm": 0.12354234788520546,
"learning_rate": 1.0178558119067315e-06,
"loss": 0.0315,
"step": 446
},
{
"epoch": 2.8838709677419354,
"grad_norm": 0.15472531322652747,
"learning_rate": 9.136918131366412e-07,
"loss": 0.0436,
"step": 447
},
{
"epoch": 2.8903225806451616,
"grad_norm": 0.16727260037004013,
"learning_rate": 8.151248591913518e-07,
"loss": 0.0413,
"step": 448
},
{
"epoch": 2.896774193548387,
"grad_norm": 0.15358557638143366,
"learning_rate": 7.221605177791691e-07,
"loss": 0.0477,
"step": 449
},
{
"epoch": 2.903225806451613,
"grad_norm": 0.11899732193116695,
"learning_rate": 6.348040401360833e-07,
"loss": 0.0337,
"step": 450
},
{
"epoch": 2.9096774193548387,
"grad_norm": 0.19561410636845064,
"learning_rate": 5.530603607290851e-07,
"loss": 0.0556,
"step": 451
},
{
"epoch": 2.9161290322580644,
"grad_norm": 0.14981745610583072,
"learning_rate": 4.76934096977566e-07,
"loss": 0.0372,
"step": 452
},
{
"epoch": 2.9225806451612906,
"grad_norm": 0.1039480982798509,
"learning_rate": 4.0642954899238197e-07,
"loss": 0.0343,
"step": 453
},
{
"epoch": 2.9290322580645163,
"grad_norm": 0.11333580055218463,
"learning_rate": 3.415506993330153e-07,
"loss": 0.0388,
"step": 454
},
{
"epoch": 2.935483870967742,
"grad_norm": 0.10481858385197364,
"learning_rate": 2.8230121278257637e-07,
"loss": 0.0313,
"step": 455
},
{
"epoch": 2.9419354838709677,
"grad_norm": 0.15064675470229316,
"learning_rate": 2.2868443614082469e-07,
"loss": 0.0449,
"step": 456
},
{
"epoch": 2.9483870967741934,
"grad_norm": 0.13918413557689208,
"learning_rate": 1.8070339803509807e-07,
"loss": 0.0414,
"step": 457
},
{
"epoch": 2.9548387096774196,
"grad_norm": 0.15721864030858848,
"learning_rate": 1.3836080874926049e-07,
"loss": 0.0342,
"step": 458
},
{
"epoch": 2.9612903225806453,
"grad_norm": 0.15411456561887094,
"learning_rate": 1.0165906007056914e-07,
"loss": 0.0418,
"step": 459
},
{
"epoch": 2.967741935483871,
"grad_norm": 0.1513944041768792,
"learning_rate": 7.060022515460451e-08,
"loss": 0.0364,
"step": 460
},
{
"epoch": 2.9741935483870967,
"grad_norm": 0.22512284396579382,
"learning_rate": 4.518605840815315e-08,
"loss": 0.0447,
"step": 461
},
{
"epoch": 2.9806451612903224,
"grad_norm": 0.1529295717768871,
"learning_rate": 2.5417995390086824e-08,
"loss": 0.0449,
"step": 462
},
{
"epoch": 2.9870967741935486,
"grad_norm": 0.13942712092352366,
"learning_rate": 1.129715273033849e-08,
"loss": 0.0367,
"step": 463
},
{
"epoch": 2.9935483870967743,
"grad_norm": 0.13569899513050301,
"learning_rate": 2.824328066730608e-09,
"loss": 0.0398,
"step": 464
},
{
"epoch": 3.0,
"grad_norm": 0.11004047385534296,
"learning_rate": 0.0,
"loss": 0.0305,
"step": 465
},
{
"epoch": 3.0,
"eval_loss": 0.2107405662536621,
"eval_runtime": 25.5157,
"eval_samples_per_second": 5.134,
"eval_steps_per_second": 0.666,
"step": 465
},
{
"epoch": 3.0,
"step": 465,
"total_flos": 261374226563072.0,
"train_loss": 0.09038178783751304,
"train_runtime": 3542.1658,
"train_samples_per_second": 2.097,
"train_steps_per_second": 0.131
}
],
"logging_steps": 1,
"max_steps": 465,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 261374226563072.0,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}