diff --git "a/sft_full/hyperrouter/trainer_state.json" "b/sft_full/hyperrouter/trainer_state.json" new file mode 100644--- /dev/null +++ "b/sft_full/hyperrouter/trainer_state.json" @@ -0,0 +1,99835 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.9999549082382648, + "eval_steps": 500, + "global_step": 11088, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 9.018352347026198e-05, + "flos": 20416263415680.0, + "grad_norm": 57.00024842325251, + "learning_rate": 0.0, + "loss": 2.0008, + "num_input_tokens_seen": 26480, + "step": 1 + }, + { + "epoch": 0.00018036704694052397, + "flos": 22494810728160.0, + "grad_norm": 54.76518143543624, + "learning_rate": 4.773623799730706e-07, + "loss": 1.9301, + "num_input_tokens_seen": 52215, + "step": 2 + }, + { + "epoch": 0.0002705505704107859, + "flos": 23986593935040.0, + "grad_norm": 34.817962710987985, + "learning_rate": 7.566014715123208e-07, + "loss": 1.6523, + "num_input_tokens_seen": 81590, + "step": 3 + }, + { + "epoch": 0.00036073409388104793, + "flos": 22241235660000.0, + "grad_norm": 55.03509622419361, + "learning_rate": 9.547247599461412e-07, + "loss": 1.9717, + "num_input_tokens_seen": 108995, + "step": 4 + }, + { + "epoch": 0.0004509176173513099, + "flos": 26064784673760.0, + "grad_norm": 40.07419571381783, + "learning_rate": 1.108401121501769e-06, + "loss": 1.5705, + "num_input_tokens_seen": 140930, + "step": 5 + }, + { + "epoch": 0.0005411011408215718, + "flos": 15355606232160.0, + "grad_norm": 59.82754961854425, + "learning_rate": 1.2339638514853914e-06, + "loss": 1.7504, + "num_input_tokens_seen": 168255, + "step": 6 + }, + { + "epoch": 0.0006312846642918339, + "flos": 24900518654880.0, + "grad_norm": 35.35920411193792, + "learning_rate": 1.3401256270225321e-06, + "loss": 1.4796, + "num_input_tokens_seen": 200695, + "step": 7 + }, + { + "epoch": 0.0007214681877620959, + "flos": 15538138421760.0, + "grad_norm": 27.048388037185436, + "learning_rate": 1.4320871399192119e-06, + "loss": 1.4154, + "num_input_tokens_seen": 226405, + "step": 8 + }, + { + "epoch": 0.0008116517112323579, + "flos": 21547055793120.0, + "grad_norm": 16.10454924385207, + "learning_rate": 1.5132029430246416e-06, + "loss": 1.2858, + "num_input_tokens_seen": 253615, + "step": 9 + }, + { + "epoch": 0.0009018352347026198, + "flos": 20742984914880.0, + "grad_norm": 16.46113030855282, + "learning_rate": 1.5857635014748399e-06, + "loss": 1.3977, + "num_input_tokens_seen": 279335, + "step": 10 + }, + { + "epoch": 0.0009920187581728818, + "flos": 11639569365120.0, + "grad_norm": 17.9625583761882, + "learning_rate": 1.6514025108267924e-06, + "loss": 1.4724, + "num_input_tokens_seen": 304145, + "step": 11 + }, + { + "epoch": 0.0010822022816431437, + "flos": 22423693510080.0, + "grad_norm": 14.815109087529212, + "learning_rate": 1.711326231458462e-06, + "loss": 1.353, + "num_input_tokens_seen": 333970, + "step": 12 + }, + { + "epoch": 0.0011723858051134058, + "flos": 16557568121280.0, + "grad_norm": 7.849771931244792, + "learning_rate": 1.7664507107987104e-06, + "loss": 1.1956, + "num_input_tokens_seen": 356980, + "step": 13 + }, + { + "epoch": 0.0012625693285836677, + "flos": 24209498217600.0, + "grad_norm": 5.595695455655739, + "learning_rate": 1.8174880069956024e-06, + "loss": 1.1579, + "num_input_tokens_seen": 386575, + "step": 14 + }, + { + "epoch": 0.0013527528520539298, + "flos": 22747493722080.0, + "grad_norm": 4.876320880060846, + "learning_rate": 1.8650025930140899e-06, + "loss": 1.1153, + "num_input_tokens_seen": 414945, + "step": 15 + }, + { + "epoch": 0.0014429363755241917, + "flos": 15574607689920.0, + "grad_norm": 5.614686454245882, + "learning_rate": 1.9094495198922823e-06, + "loss": 1.2292, + "num_input_tokens_seen": 441905, + "step": 16 + }, + { + "epoch": 0.0015331198989944536, + "flos": 24208903501440.0, + "grad_norm": 4.60576920503978, + "learning_rate": 1.9512009899507514e-06, + "loss": 1.0754, + "num_input_tokens_seen": 470450, + "step": 17 + }, + { + "epoch": 0.0016233034224647158, + "flos": 22606411548480.0, + "grad_norm": 4.33890202534264, + "learning_rate": 1.990565322997712e-06, + "loss": 1.1294, + "num_input_tokens_seen": 499605, + "step": 18 + }, + { + "epoch": 0.0017134869459349777, + "flos": 37577848144800.0, + "grad_norm": 3.492866057624997, + "learning_rate": 2.027800787770518e-06, + "loss": 1.0404, + "num_input_tokens_seen": 532040, + "step": 19 + }, + { + "epoch": 0.0018036704694052396, + "flos": 24391138332960.0, + "grad_norm": 4.874433044330603, + "learning_rate": 2.06312588144791e-06, + "loss": 1.1335, + "num_input_tokens_seen": 560030, + "step": 20 + }, + { + "epoch": 0.0018938539928755017, + "flos": 24860109392160.0, + "grad_norm": 3.326765328921013, + "learning_rate": 2.0967270985348526e-06, + "loss": 1.0392, + "num_input_tokens_seen": 586835, + "step": 21 + }, + { + "epoch": 0.0019840375163457636, + "flos": 28436513706240.0, + "grad_norm": 5.329131179623811, + "learning_rate": 2.128764890799863e-06, + "loss": 0.9558, + "num_input_tokens_seen": 617945, + "step": 22 + }, + { + "epoch": 0.0020742210398160257, + "flos": 36487189717920.0, + "grad_norm": 3.3794795790234655, + "learning_rate": 2.1593783012990145e-06, + "loss": 0.9578, + "num_input_tokens_seen": 649070, + "step": 23 + }, + { + "epoch": 0.0021644045632862874, + "flos": 21622113005760.0, + "grad_norm": 2.782376264921038, + "learning_rate": 2.188688611431533e-06, + "loss": 1.1077, + "num_input_tokens_seen": 675560, + "step": 24 + }, + { + "epoch": 0.0022545880867565495, + "flos": 23735026033920.0, + "grad_norm": 3.2216965604665764, + "learning_rate": 2.216802243003538e-06, + "loss": 1.0862, + "num_input_tokens_seen": 703870, + "step": 25 + }, + { + "epoch": 0.0023447716102268116, + "flos": 17869644040320.0, + "grad_norm": 4.520480703263119, + "learning_rate": 2.243813090771781e-06, + "loss": 0.9281, + "num_input_tokens_seen": 733050, + "step": 26 + }, + { + "epoch": 0.0024349551336970737, + "flos": 21476087254080.0, + "grad_norm": 4.98197740630697, + "learning_rate": 2.269804414536962e-06, + "loss": 1.0944, + "num_input_tokens_seen": 758950, + "step": 27 + }, + { + "epoch": 0.0025251386571673354, + "flos": 31203977903520.0, + "grad_norm": 5.810245856173359, + "learning_rate": 2.2948503869686733e-06, + "loss": 1.0051, + "num_input_tokens_seen": 791220, + "step": 28 + }, + { + "epoch": 0.0026153221806375975, + "flos": 63704303184000.0, + "grad_norm": 2.5243527668755172, + "learning_rate": 2.3190173696980436e-06, + "loss": 0.6974, + "num_input_tokens_seen": 884040, + "step": 29 + }, + { + "epoch": 0.0027055057041078597, + "flos": 23587216133760.0, + "grad_norm": 2.4685266128103778, + "learning_rate": 2.3423649729871604e-06, + "loss": 1.0075, + "num_input_tokens_seen": 910420, + "step": 30 + }, + { + "epoch": 0.0027956892275781214, + "flos": 18817361805600.0, + "grad_norm": 2.143913249481254, + "learning_rate": 2.364946941580084e-06, + "loss": 1.0044, + "num_input_tokens_seen": 937625, + "step": 31 + }, + { + "epoch": 0.0028858727510483835, + "flos": 14983330919520.0, + "grad_norm": 2.6855840728957925, + "learning_rate": 2.3868118998653532e-06, + "loss": 1.0676, + "num_input_tokens_seen": 962050, + "step": 32 + }, + { + "epoch": 0.0029760562745186456, + "flos": 70229663131680.0, + "grad_norm": 2.2868821106930737, + "learning_rate": 2.408003982339113e-06, + "loss": 0.6433, + "num_input_tokens_seen": 1059370, + "step": 33 + }, + { + "epoch": 0.0030662397979889073, + "flos": 24536606538240.0, + "grad_norm": 2.1916661507880275, + "learning_rate": 2.4285633699238223e-06, + "loss": 0.9158, + "num_input_tokens_seen": 1088155, + "step": 34 + }, + { + "epoch": 0.0031564233214591694, + "flos": 48399556650720.0, + "grad_norm": 4.269600493348615, + "learning_rate": 2.4485267485243007e-06, + "loss": 0.9468, + "num_input_tokens_seen": 1120235, + "step": 35 + }, + { + "epoch": 0.0032466068449294315, + "flos": 22492060165920.0, + "grad_norm": 1.8521889701156484, + "learning_rate": 2.467927702970783e-06, + "loss": 0.9835, + "num_input_tokens_seen": 1148730, + "step": 36 + }, + { + "epoch": 0.003336790368399693, + "flos": 25296365328960.0, + "grad_norm": 2.160964014603693, + "learning_rate": 2.4867970569753584e-06, + "loss": 1.0025, + "num_input_tokens_seen": 1176690, + "step": 37 + }, + { + "epoch": 0.0034269738918699553, + "flos": 18118758737280.0, + "grad_norm": 3.2039539288031387, + "learning_rate": 2.5051631677435883e-06, + "loss": 0.8433, + "num_input_tokens_seen": 1202550, + "step": 38 + }, + { + "epoch": 0.0035171574153402174, + "flos": 39290156769600.0, + "grad_norm": 2.8976393266411478, + "learning_rate": 2.523052182311031e-06, + "loss": 0.8118, + "num_input_tokens_seen": 1237825, + "step": 39 + }, + { + "epoch": 0.003607340938810479, + "flos": 23950236176160.0, + "grad_norm": 2.955931606446148, + "learning_rate": 2.540488261420981e-06, + "loss": 1.0162, + "num_input_tokens_seen": 1267345, + "step": 40 + }, + { + "epoch": 0.0036975244622807412, + "flos": 24828249174240.0, + "grad_norm": 2.220288777342096, + "learning_rate": 2.557493775753984e-06, + "loss": 0.8151, + "num_input_tokens_seen": 1298280, + "step": 41 + }, + { + "epoch": 0.0037877079857510034, + "flos": 25155134476320.0, + "grad_norm": 1.83756745402769, + "learning_rate": 2.5740894785079235e-06, + "loss": 0.9309, + "num_input_tokens_seen": 1327675, + "step": 42 + }, + { + "epoch": 0.0038778915092212655, + "flos": 24499802742240.0, + "grad_norm": 2.498780239088275, + "learning_rate": 2.5902946576685834e-06, + "loss": 0.9341, + "num_input_tokens_seen": 1356790, + "step": 43 + }, + { + "epoch": 0.003968075032691527, + "flos": 25156286738880.0, + "grad_norm": 2.4684913596345304, + "learning_rate": 2.606127270772933e-06, + "loss": 0.8622, + "num_input_tokens_seen": 1384125, + "step": 44 + }, + { + "epoch": 0.004058258556161789, + "flos": 28253200951680.0, + "grad_norm": 2.7506361138956374, + "learning_rate": 2.62160406452641e-06, + "loss": 0.8348, + "num_input_tokens_seen": 1414960, + "step": 45 + }, + { + "epoch": 0.004148442079632051, + "flos": 19145213521440.0, + "grad_norm": 2.10890961210875, + "learning_rate": 2.636740681272085e-06, + "loss": 0.9975, + "num_input_tokens_seen": 1442965, + "step": 46 + }, + { + "epoch": 0.004238625603102313, + "flos": 22312464387360.0, + "grad_norm": 2.124738452747591, + "learning_rate": 2.651551754008722e-06, + "loss": 1.0333, + "num_input_tokens_seen": 1471250, + "step": 47 + }, + { + "epoch": 0.004328809126572575, + "flos": 19253469063360.0, + "grad_norm": 2.093272254299793, + "learning_rate": 2.6660509914046035e-06, + "loss": 0.9696, + "num_input_tokens_seen": 1498710, + "step": 48 + }, + { + "epoch": 0.004418992650042837, + "flos": 32801303259840.0, + "grad_norm": 3.2826028124127444, + "learning_rate": 2.6802512540450642e-06, + "loss": 0.9285, + "num_input_tokens_seen": 1531230, + "step": 49 + }, + { + "epoch": 0.004509176173513099, + "flos": 22678681029120.0, + "grad_norm": 2.130306462372658, + "learning_rate": 2.694164622976609e-06, + "loss": 0.9148, + "num_input_tokens_seen": 1558370, + "step": 50 + }, + { + "epoch": 0.0045993596969833616, + "flos": 19181608450080.0, + "grad_norm": 1.830559633598101, + "learning_rate": 2.707802461463072e-06, + "loss": 0.8516, + "num_input_tokens_seen": 1586025, + "step": 51 + }, + { + "epoch": 0.004689543220453623, + "flos": 22349639880960.0, + "grad_norm": 1.9170224716429534, + "learning_rate": 2.7211754707448516e-06, + "loss": 0.8946, + "num_input_tokens_seen": 1612485, + "step": 52 + }, + { + "epoch": 0.004779726743923885, + "flos": 20633019564000.0, + "grad_norm": 1.9349332851175605, + "learning_rate": 2.734293740486721e-06, + "loss": 0.8748, + "num_input_tokens_seen": 1641400, + "step": 53 + }, + { + "epoch": 0.0048699102673941475, + "flos": 24498092933280.0, + "grad_norm": 2.210801838193833, + "learning_rate": 2.747166794510033e-06, + "loss": 0.9196, + "num_input_tokens_seen": 1671910, + "step": 54 + }, + { + "epoch": 0.004960093790864409, + "flos": 28323909302400.0, + "grad_norm": 1.9792441168906416, + "learning_rate": 2.759803632328562e-06, + "loss": 0.8877, + "num_input_tokens_seen": 1701190, + "step": 55 + }, + { + "epoch": 0.005050277314334671, + "flos": 27633223392960.0, + "grad_norm": 2.236156473693491, + "learning_rate": 2.772212766941744e-06, + "loss": 1.0072, + "num_input_tokens_seen": 1729240, + "step": 56 + }, + { + "epoch": 0.005140460837804933, + "flos": 23262077810400.0, + "grad_norm": 2.6314084981629327, + "learning_rate": 2.7844022592828385e-06, + "loss": 0.9949, + "num_input_tokens_seen": 1756890, + "step": 57 + }, + { + "epoch": 0.005230644361275195, + "flos": 26172780027360.0, + "grad_norm": 2.030030926435944, + "learning_rate": 2.7963797496711145e-06, + "loss": 0.8476, + "num_input_tokens_seen": 1785850, + "step": 58 + }, + { + "epoch": 0.005320827884745457, + "flos": 22896121356960.0, + "grad_norm": 2.077043983548429, + "learning_rate": 2.80815248657541e-06, + "loss": 0.9385, + "num_input_tokens_seen": 1813755, + "step": 59 + }, + { + "epoch": 0.005411011408215719, + "flos": 66614856721920.0, + "grad_norm": 1.3836573374528762, + "learning_rate": 2.819727352960231e-06, + "loss": 0.6167, + "num_input_tokens_seen": 1915265, + "step": 60 + }, + { + "epoch": 0.005501194931685981, + "flos": 21583450721760.0, + "grad_norm": 2.2149348164562093, + "learning_rate": 2.8311108904541717e-06, + "loss": 0.8974, + "num_input_tokens_seen": 1942385, + "step": 61 + }, + { + "epoch": 0.005591378455156243, + "flos": 28034236663680.0, + "grad_norm": 1.8451566987316133, + "learning_rate": 2.842309321553155e-06, + "loss": 0.8448, + "num_input_tokens_seen": 1973990, + "step": 62 + }, + { + "epoch": 0.005681561978626505, + "flos": 25262237755680.0, + "grad_norm": 1.6805959684418699, + "learning_rate": 2.8533285700471737e-06, + "loss": 0.831, + "num_input_tokens_seen": 2004565, + "step": 63 + }, + { + "epoch": 0.005771745502096767, + "flos": 25077029343360.0, + "grad_norm": 2.3634519405042367, + "learning_rate": 2.8641742798384237e-06, + "loss": 0.9689, + "num_input_tokens_seen": 2032840, + "step": 64 + }, + { + "epoch": 0.005861929025567029, + "flos": 23261297245440.0, + "grad_norm": 1.9346551005208115, + "learning_rate": 2.874851832300479e-06, + "loss": 0.925, + "num_input_tokens_seen": 2060190, + "step": 65 + }, + { + "epoch": 0.005952112549037291, + "flos": 16704002740320.0, + "grad_norm": 2.1514117374387585, + "learning_rate": 2.8853663623121834e-06, + "loss": 0.9263, + "num_input_tokens_seen": 2084785, + "step": 66 + }, + { + "epoch": 0.006042296072507553, + "flos": 18999262109280.0, + "grad_norm": 3.6662288028638814, + "learning_rate": 2.895722773085839e-06, + "loss": 0.9538, + "num_input_tokens_seen": 2112460, + "step": 67 + }, + { + "epoch": 0.0061324795959778146, + "flos": 28763436178080.0, + "grad_norm": 3.776492570285668, + "learning_rate": 2.905925749896893e-06, + "loss": 0.9218, + "num_input_tokens_seen": 2142930, + "step": 68 + }, + { + "epoch": 0.006222663119448077, + "flos": 30365890961280.0, + "grad_norm": 1.717567661401882, + "learning_rate": 2.915979772811335e-06, + "loss": 0.9116, + "num_input_tokens_seen": 2172770, + "step": 69 + }, + { + "epoch": 0.006312846642918339, + "flos": 25849240003680.0, + "grad_norm": 2.5433373718762944, + "learning_rate": 2.925889128497372e-06, + "loss": 0.8707, + "num_input_tokens_seen": 2202520, + "step": 70 + }, + { + "epoch": 0.0064030301663886005, + "flos": 21695014372320.0, + "grad_norm": 2.1292479074684425, + "learning_rate": 2.9356579211992906e-06, + "loss": 0.9701, + "num_input_tokens_seen": 2230980, + "step": 71 + }, + { + "epoch": 0.006493213689858863, + "flos": 31529413584960.0, + "grad_norm": 2.2386965274867867, + "learning_rate": 2.9452900829438533e-06, + "loss": 0.8412, + "num_input_tokens_seen": 2262440, + "step": 72 + }, + { + "epoch": 0.006583397213329125, + "flos": 27487457829600.0, + "grad_norm": 2.4839060598722367, + "learning_rate": 2.954789383042727e-06, + "loss": 0.8741, + "num_input_tokens_seen": 2288125, + "step": 73 + }, + { + "epoch": 0.006673580736799386, + "flos": 72446334160800.0, + "grad_norm": 1.2170255149229159, + "learning_rate": 2.9641594369484293e-06, + "loss": 0.6854, + "num_input_tokens_seen": 2386595, + "step": 74 + }, + { + "epoch": 0.006763764260269649, + "flos": 22674815374080.0, + "grad_norm": 3.1184061292765213, + "learning_rate": 2.9734037145158586e-06, + "loss": 0.899, + "num_input_tokens_seen": 2414850, + "step": 75 + }, + { + "epoch": 0.006853947783739911, + "flos": 19247744920320.0, + "grad_norm": 2.2209214880685337, + "learning_rate": 2.982525547716659e-06, + "loss": 0.898, + "num_input_tokens_seen": 2441125, + "step": 76 + }, + { + "epoch": 0.006944131307210172, + "flos": 15131252328960.0, + "grad_norm": 2.664736486293126, + "learning_rate": 2.9915281378493246e-06, + "loss": 0.8682, + "num_input_tokens_seen": 2463890, + "step": 77 + }, + { + "epoch": 0.007034314830680435, + "flos": 31528707359520.0, + "grad_norm": 2.1190262020880013, + "learning_rate": 3.000414562284102e-06, + "loss": 0.8927, + "num_input_tokens_seen": 2496130, + "step": 78 + }, + { + "epoch": 0.0071244983541506966, + "flos": 24681442857600.0, + "grad_norm": 3.350384414024863, + "learning_rate": 3.009187780778246e-06, + "loss": 0.8862, + "num_input_tokens_seen": 2526185, + "step": 79 + }, + { + "epoch": 0.007214681877620958, + "flos": 18852790320480.0, + "grad_norm": 2.5517017520391883, + "learning_rate": 3.017850641394051e-06, + "loss": 0.8788, + "num_input_tokens_seen": 2551390, + "step": 80 + }, + { + "epoch": 0.007304865401091221, + "flos": 45485286136800.0, + "grad_norm": 1.9257490608047638, + "learning_rate": 3.0264058860492832e-06, + "loss": 0.7478, + "num_input_tokens_seen": 2583315, + "step": 81 + }, + { + "epoch": 0.0073950489245614825, + "flos": 24317902438560.0, + "grad_norm": 2.166011332171411, + "learning_rate": 3.0348561557270548e-06, + "loss": 0.9362, + "num_input_tokens_seen": 2612770, + "step": 82 + }, + { + "epoch": 0.007485232448031745, + "flos": 64804254050880.0, + "grad_norm": 0.8748958048771477, + "learning_rate": 3.043203995369939e-06, + "loss": 0.6459, + "num_input_tokens_seen": 2701920, + "step": 83 + }, + { + "epoch": 0.007575415971502007, + "flos": 22202833564320.0, + "grad_norm": 3.5726065246204683, + "learning_rate": 3.051451858480994e-06, + "loss": 0.9798, + "num_input_tokens_seen": 2730395, + "step": 84 + }, + { + "epoch": 0.007665599494972268, + "flos": 32038682397600.0, + "grad_norm": 2.618383606837419, + "learning_rate": 3.05960211145252e-06, + "loss": 0.809, + "num_input_tokens_seen": 2761515, + "step": 85 + }, + { + "epoch": 0.007755783018442531, + "flos": 31602352121280.0, + "grad_norm": 1.9907251713100373, + "learning_rate": 3.0676570376416543e-06, + "loss": 0.8102, + "num_input_tokens_seen": 2791690, + "step": 86 + }, + { + "epoch": 0.007845966541912792, + "flos": 20710083943680.0, + "grad_norm": 3.090316254070269, + "learning_rate": 3.0756188412103647e-06, + "loss": 0.9078, + "num_input_tokens_seen": 2818570, + "step": 87 + }, + { + "epoch": 0.007936150065383054, + "flos": 26722941309600.0, + "grad_norm": 1.9036793494937867, + "learning_rate": 3.083489650746004e-06, + "loss": 0.8856, + "num_input_tokens_seen": 2846040, + "step": 88 + }, + { + "epoch": 0.008026333588853317, + "flos": 21039422449920.0, + "grad_norm": 1.8977251011312204, + "learning_rate": 3.0912715226772975e-06, + "loss": 0.9147, + "num_input_tokens_seen": 2873275, + "step": 89 + }, + { + "epoch": 0.008116517112323578, + "flos": 23407546015680.0, + "grad_norm": 1.618538039186167, + "learning_rate": 3.098966444499481e-06, + "loss": 0.8267, + "num_input_tokens_seen": 2905390, + "step": 90 + }, + { + "epoch": 0.00820670063579384, + "flos": 21586647321120.0, + "grad_norm": 2.2874551278804685, + "learning_rate": 3.1065763378212426e-06, + "loss": 0.8945, + "num_input_tokens_seen": 2934495, + "step": 91 + }, + { + "epoch": 0.008296884159264103, + "flos": 23586993115200.0, + "grad_norm": 1.6115771291977532, + "learning_rate": 3.1141030612451554e-06, + "loss": 0.8953, + "num_input_tokens_seen": 2965185, + "step": 92 + }, + { + "epoch": 0.008387067682734364, + "flos": 62278767956640.0, + "grad_norm": 0.8732800950926431, + "learning_rate": 3.1215484130924052e-06, + "loss": 0.6769, + "num_input_tokens_seen": 3051530, + "step": 93 + }, + { + "epoch": 0.008477251206204626, + "flos": 21841077293760.0, + "grad_norm": 1.9394212856541229, + "learning_rate": 3.128914133981793e-06, + "loss": 0.8988, + "num_input_tokens_seen": 3079185, + "step": 94 + }, + { + "epoch": 0.008567434729674889, + "flos": 26904953122560.0, + "grad_norm": 2.7741177436576607, + "learning_rate": 3.136201909272287e-06, + "loss": 0.8931, + "num_input_tokens_seen": 3109260, + "step": 95 + }, + { + "epoch": 0.00865761825314515, + "flos": 25991920476960.0, + "grad_norm": 2.1501188600198677, + "learning_rate": 3.1434133713776735e-06, + "loss": 0.8554, + "num_input_tokens_seen": 3138075, + "step": 96 + }, + { + "epoch": 0.008747801776615412, + "flos": 19983263293920.0, + "grad_norm": 1.8875340187089205, + "learning_rate": 3.15055010196128e-06, + "loss": 0.8819, + "num_input_tokens_seen": 3165025, + "step": 97 + }, + { + "epoch": 0.008837985300085675, + "flos": 41112951121920.0, + "grad_norm": 2.568253547605851, + "learning_rate": 3.157613634018135e-06, + "loss": 0.8753, + "num_input_tokens_seen": 3196850, + "step": 98 + }, + { + "epoch": 0.008928168823555935, + "flos": 23007127461120.0, + "grad_norm": 1.9552268963438804, + "learning_rate": 3.1646054538514336e-06, + "loss": 0.9013, + "num_input_tokens_seen": 3226975, + "step": 99 + }, + { + "epoch": 0.009018352347026198, + "flos": 28868234932320.0, + "grad_norm": 2.3908802968258365, + "learning_rate": 3.1715270029496797e-06, + "loss": 0.9338, + "num_input_tokens_seen": 3255725, + "step": 100 + }, + { + "epoch": 0.00910853587049646, + "flos": 28616072315040.0, + "grad_norm": 1.6604318906053428, + "learning_rate": 3.1783796797704243e-06, + "loss": 0.9132, + "num_input_tokens_seen": 3287410, + "step": 101 + }, + { + "epoch": 0.009198719393966723, + "flos": 19725896910240.0, + "grad_norm": 2.0865551894801007, + "learning_rate": 3.185164841436142e-06, + "loss": 0.8399, + "num_input_tokens_seen": 3314510, + "step": 102 + }, + { + "epoch": 0.009288902917436984, + "flos": 18708400038240.0, + "grad_norm": 2.653160093185487, + "learning_rate": 3.1918838053473723e-06, + "loss": 0.8796, + "num_input_tokens_seen": 3341505, + "step": 103 + }, + { + "epoch": 0.009379086440907246, + "flos": 34191038274720.0, + "grad_norm": 1.6678987796588922, + "learning_rate": 3.198537850717922e-06, + "loss": 0.785, + "num_input_tokens_seen": 3372185, + "step": 104 + }, + { + "epoch": 0.009469269964377509, + "flos": 24974200586400.0, + "grad_norm": 2.0397665043511597, + "learning_rate": 3.205128220036622e-06, + "loss": 0.8388, + "num_input_tokens_seen": 3401880, + "step": 105 + }, + { + "epoch": 0.00955945348784777, + "flos": 21622261684800.0, + "grad_norm": 1.9944135010295607, + "learning_rate": 3.2116561204597917e-06, + "loss": 0.8378, + "num_input_tokens_seen": 3431870, + "step": 106 + }, + { + "epoch": 0.009649637011318032, + "flos": 61086358695840.0, + "grad_norm": 0.8630207756255796, + "learning_rate": 3.218122725138335e-06, + "loss": 0.6675, + "num_input_tokens_seen": 3522085, + "step": 107 + }, + { + "epoch": 0.009739820534788295, + "flos": 24317716589760.0, + "grad_norm": 9.193565390656943, + "learning_rate": 3.224529174483104e-06, + "loss": 0.8717, + "num_input_tokens_seen": 3550490, + "step": 108 + }, + { + "epoch": 0.009830004058258556, + "flos": 29527655340000.0, + "grad_norm": 2.447845219861327, + "learning_rate": 3.2308765773719435e-06, + "loss": 0.8152, + "num_input_tokens_seen": 3580390, + "step": 109 + }, + { + "epoch": 0.009920187581728818, + "flos": 18488878203840.0, + "grad_norm": 1.919473279388928, + "learning_rate": 3.2371660123016323e-06, + "loss": 0.8659, + "num_input_tokens_seen": 3607695, + "step": 110 + }, + { + "epoch": 0.010010371105199081, + "flos": 29630781455040.0, + "grad_norm": 1.9000270138878073, + "learning_rate": 3.2433985284876787e-06, + "loss": 0.7929, + "num_input_tokens_seen": 3638605, + "step": 111 + }, + { + "epoch": 0.010100554628669342, + "flos": 20419779419040.0, + "grad_norm": 1.527120198233759, + "learning_rate": 3.2495751469148143e-06, + "loss": 0.8636, + "num_input_tokens_seen": 3666420, + "step": 112 + }, + { + "epoch": 0.010190738152139604, + "flos": 18124297031520.0, + "grad_norm": 5.1452259935694125, + "learning_rate": 3.2556968613407816e-06, + "loss": 0.8374, + "num_input_tokens_seen": 3693230, + "step": 113 + }, + { + "epoch": 0.010280921675609867, + "flos": 18627693022080.0, + "grad_norm": 1.8847568333870408, + "learning_rate": 3.2617646392559094e-06, + "loss": 0.9047, + "num_input_tokens_seen": 3719965, + "step": 114 + }, + { + "epoch": 0.010371105199080128, + "flos": 19618830800640.0, + "grad_norm": 1.9727746304787384, + "learning_rate": 3.2677794228007836e-06, + "loss": 0.9175, + "num_input_tokens_seen": 3745190, + "step": 115 + }, + { + "epoch": 0.01046128872255039, + "flos": 36122459866560.0, + "grad_norm": 2.0330298035665737, + "learning_rate": 3.273742129644185e-06, + "loss": 0.8506, + "num_input_tokens_seen": 3775615, + "step": 116 + }, + { + "epoch": 0.010551472246020653, + "flos": 15319397152320.0, + "grad_norm": 2.1659498006933435, + "learning_rate": 3.279653653823352e-06, + "loss": 0.8709, + "num_input_tokens_seen": 3802000, + "step": 117 + }, + { + "epoch": 0.010641655769490914, + "flos": 21109796272800.0, + "grad_norm": 2.4132804715003093, + "learning_rate": 3.285514866548481e-06, + "loss": 0.865, + "num_input_tokens_seen": 3830415, + "step": 118 + }, + { + "epoch": 0.010731839292961176, + "flos": 64429679946720.0, + "grad_norm": 2.010715034231121, + "learning_rate": 3.2913266169732838e-06, + "loss": 0.9277, + "num_input_tokens_seen": 3865140, + "step": 119 + }, + { + "epoch": 0.010822022816431439, + "flos": 23443792265280.0, + "grad_norm": 2.759129023577912, + "learning_rate": 3.2970897329333017e-06, + "loss": 0.8463, + "num_input_tokens_seen": 3890865, + "step": 120 + }, + { + "epoch": 0.0109122063399017, + "flos": 24354260197440.0, + "grad_norm": 1.9159953420578437, + "learning_rate": 3.302805021653585e-06, + "loss": 0.8425, + "num_input_tokens_seen": 3919770, + "step": 121 + }, + { + "epoch": 0.011002389863371962, + "flos": 19501989044160.0, + "grad_norm": 2.0504124482371204, + "learning_rate": 3.3084732704272426e-06, + "loss": 0.8548, + "num_input_tokens_seen": 3947505, + "step": 122 + }, + { + "epoch": 0.011092573386842225, + "flos": 24637874165280.0, + "grad_norm": 2.807387558544354, + "learning_rate": 3.314095247266304e-06, + "loss": 0.8565, + "num_input_tokens_seen": 3973630, + "step": 123 + }, + { + "epoch": 0.011182756910312485, + "flos": 24278794117440.0, + "grad_norm": 5.2435946672277876, + "learning_rate": 3.3196717015262255e-06, + "loss": 0.8211, + "num_input_tokens_seen": 4000955, + "step": 124 + }, + { + "epoch": 0.011272940433782748, + "flos": 24241061077440.0, + "grad_norm": 2.6527130813654005, + "learning_rate": 3.325203364505307e-06, + "loss": 0.8469, + "num_input_tokens_seen": 4030425, + "step": 125 + }, + { + "epoch": 0.01136312395725301, + "flos": 17389893750720.0, + "grad_norm": 2.944089114343484, + "learning_rate": 3.3306909500202442e-06, + "loss": 0.8951, + "num_input_tokens_seen": 4055555, + "step": 126 + }, + { + "epoch": 0.011453307480723271, + "flos": 23734579996800.0, + "grad_norm": 3.109355677252304, + "learning_rate": 3.3361351549589145e-06, + "loss": 0.9231, + "num_input_tokens_seen": 4083935, + "step": 127 + }, + { + "epoch": 0.011543491004193534, + "flos": 27267155430240.0, + "grad_norm": 2.14906906555158, + "learning_rate": 3.341536659811494e-06, + "loss": 0.7971, + "num_input_tokens_seen": 4113720, + "step": 128 + }, + { + "epoch": 0.011633674527663796, + "flos": 15319583001120.0, + "grad_norm": 2.3532032866202615, + "learning_rate": 3.346896129180904e-06, + "loss": 0.8822, + "num_input_tokens_seen": 4139575, + "step": 129 + }, + { + "epoch": 0.011723858051134057, + "flos": 27230277294720.0, + "grad_norm": 1.5864040646838096, + "learning_rate": 3.35221421227355e-06, + "loss": 0.8655, + "num_input_tokens_seen": 4170730, + "step": 130 + }, + { + "epoch": 0.01181404157460432, + "flos": 19035211000800.0, + "grad_norm": 5.206887657434952, + "learning_rate": 3.357491543371255e-06, + "loss": 0.8867, + "num_input_tokens_seen": 4197565, + "step": 131 + }, + { + "epoch": 0.011904225098074582, + "flos": 20889568212960.0, + "grad_norm": 1.6384883227661229, + "learning_rate": 3.3627287422852543e-06, + "loss": 0.815, + "num_input_tokens_seen": 4226190, + "step": 132 + }, + { + "epoch": 0.011994408621544843, + "flos": 23404423755840.0, + "grad_norm": 1.7008885389630057, + "learning_rate": 3.3679264147930497e-06, + "loss": 0.8891, + "num_input_tokens_seen": 4255825, + "step": 133 + }, + { + "epoch": 0.012084592145015106, + "flos": 26722160744640.0, + "grad_norm": 3.200313022324056, + "learning_rate": 3.37308515305891e-06, + "loss": 0.8084, + "num_input_tokens_seen": 4287045, + "step": 134 + }, + { + "epoch": 0.012174775668485368, + "flos": 24785981423520.0, + "grad_norm": 1.834870113380589, + "learning_rate": 3.3782055360387313e-06, + "loss": 0.9264, + "num_input_tokens_seen": 4316580, + "step": 135 + }, + { + "epoch": 0.012264959191955629, + "flos": 32989039215840.0, + "grad_norm": 1.89301086454503, + "learning_rate": 3.3832881298699633e-06, + "loss": 0.8214, + "num_input_tokens_seen": 4346600, + "step": 136 + }, + { + "epoch": 0.012355142715425892, + "flos": 19582435872000.0, + "grad_norm": 2.4065919112996617, + "learning_rate": 3.388333488247249e-06, + "loss": 0.8339, + "num_input_tokens_seen": 4375235, + "step": 137 + }, + { + "epoch": 0.012445326238896154, + "flos": 32369321845440.0, + "grad_norm": 1.7134564812384774, + "learning_rate": 3.393342152784406e-06, + "loss": 0.9117, + "num_input_tokens_seen": 4407085, + "step": 138 + }, + { + "epoch": 0.012535509762366415, + "flos": 19655077050240.0, + "grad_norm": 2.0349183044156223, + "learning_rate": 3.3983146533633376e-06, + "loss": 0.8682, + "num_input_tokens_seen": 4433680, + "step": 139 + }, + { + "epoch": 0.012625693285836678, + "flos": 24646088682240.0, + "grad_norm": 2.1374532772557093, + "learning_rate": 3.403251508470442e-06, + "loss": 0.9234, + "num_input_tokens_seen": 4462520, + "step": 140 + }, + { + "epoch": 0.01271587680930694, + "flos": 19436447290080.0, + "grad_norm": 2.033440333297442, + "learning_rate": 3.408153225521043e-06, + "loss": 0.8247, + "num_input_tokens_seen": 4490805, + "step": 141 + }, + { + "epoch": 0.012806060332777201, + "flos": 29163557374560.0, + "grad_norm": 2.2396188304364992, + "learning_rate": 3.413020301172361e-06, + "loss": 0.7541, + "num_input_tokens_seen": 4520415, + "step": 142 + }, + { + "epoch": 0.012896243856247463, + "flos": 21295153364160.0, + "grad_norm": 1.7193660672342985, + "learning_rate": 3.4178532216255024e-06, + "loss": 0.845, + "num_input_tokens_seen": 4548350, + "step": 143 + }, + { + "epoch": 0.012986427379717726, + "flos": 30584297702880.0, + "grad_norm": 4.213414640205477, + "learning_rate": 3.422652462916924e-06, + "loss": 0.9013, + "num_input_tokens_seen": 4578295, + "step": 144 + }, + { + "epoch": 0.013076610903187987, + "flos": 29746582458240.0, + "grad_norm": 1.6384155654966908, + "learning_rate": 3.4274184911998124e-06, + "loss": 0.7635, + "num_input_tokens_seen": 4608405, + "step": 145 + }, + { + "epoch": 0.01316679442665825, + "flos": 19545632076000.0, + "grad_norm": 2.173951178921118, + "learning_rate": 3.4321517630157976e-06, + "loss": 0.9525, + "num_input_tokens_seen": 4635210, + "step": 146 + }, + { + "epoch": 0.013256977950128512, + "flos": 18780334991040.0, + "grad_norm": 3.872806297290507, + "learning_rate": 3.4368527255573845e-06, + "loss": 0.9191, + "num_input_tokens_seen": 4662245, + "step": 147 + }, + { + "epoch": 0.013347161473598773, + "flos": 34228027919520.0, + "grad_norm": 1.9002852376353925, + "learning_rate": 3.4415218169214994e-06, + "loss": 0.8318, + "num_input_tokens_seen": 4694005, + "step": 148 + }, + { + "epoch": 0.013437344997069035, + "flos": 34991429346720.0, + "grad_norm": 1.6108908319974775, + "learning_rate": 3.4461594663544882e-06, + "loss": 0.8371, + "num_input_tokens_seen": 4725580, + "step": 149 + }, + { + "epoch": 0.013527528520539298, + "flos": 33821141826720.0, + "grad_norm": 2.4140942025186245, + "learning_rate": 3.450766094488929e-06, + "loss": 0.8029, + "num_input_tokens_seen": 4755245, + "step": 150 + }, + { + "epoch": 0.013617712044009559, + "flos": 21804496516320.0, + "grad_norm": 1.8533688518267568, + "learning_rate": 3.4553421135725735e-06, + "loss": 0.9033, + "num_input_tokens_seen": 4784355, + "step": 151 + }, + { + "epoch": 0.013707895567479821, + "flos": 27779323484160.0, + "grad_norm": 2.257846079645944, + "learning_rate": 3.45988792768973e-06, + "loss": 0.8719, + "num_input_tokens_seen": 4815590, + "step": 152 + }, + { + "epoch": 0.013798079090950084, + "flos": 21585755246880.0, + "grad_norm": 1.7928255975626175, + "learning_rate": 3.464403932975393e-06, + "loss": 0.9205, + "num_input_tokens_seen": 4843955, + "step": 153 + }, + { + "epoch": 0.013888262614420345, + "flos": 25771580907840.0, + "grad_norm": 1.7537544711566464, + "learning_rate": 3.468890517822395e-06, + "loss": 0.9115, + "num_input_tokens_seen": 4874525, + "step": 154 + }, + { + "epoch": 0.013978446137890607, + "flos": 20419816588800.0, + "grad_norm": 2.2743633259690905, + "learning_rate": 3.473348063081853e-06, + "loss": 0.8482, + "num_input_tokens_seen": 4901030, + "step": 155 + }, + { + "epoch": 0.01406862966136087, + "flos": 22242016224960.0, + "grad_norm": 1.8347056050317745, + "learning_rate": 3.4777769422571727e-06, + "loss": 0.8107, + "num_input_tokens_seen": 4930025, + "step": 156 + }, + { + "epoch": 0.01415881318483113, + "flos": 26649296547840.0, + "grad_norm": 1.8486346180821875, + "learning_rate": 3.4821775216918497e-06, + "loss": 0.8522, + "num_input_tokens_seen": 4963575, + "step": 157 + }, + { + "epoch": 0.014248996708301393, + "flos": 24937136602080.0, + "grad_norm": 1.7933598509438693, + "learning_rate": 3.4865501607513164e-06, + "loss": 0.8031, + "num_input_tokens_seen": 4996240, + "step": 158 + }, + { + "epoch": 0.014339180231771656, + "flos": 30766086497280.0, + "grad_norm": 3.604709929754867, + "learning_rate": 3.4908952119990423e-06, + "loss": 0.9249, + "num_input_tokens_seen": 5024610, + "step": 159 + }, + { + "epoch": 0.014429363755241916, + "flos": 55666822437120.0, + "grad_norm": 0.9140730963198742, + "learning_rate": 3.495213021367122e-06, + "loss": 0.6271, + "num_input_tokens_seen": 5117515, + "step": 160 + }, + { + "epoch": 0.014519547278712179, + "flos": 30219753700320.0, + "grad_norm": 2.712795269624348, + "learning_rate": 3.4995039283215464e-06, + "loss": 0.8763, + "num_input_tokens_seen": 5143085, + "step": 161 + }, + { + "epoch": 0.014609730802182442, + "flos": 23771866999680.0, + "grad_norm": 2.180965472795058, + "learning_rate": 3.5037682660223533e-06, + "loss": 0.8699, + "num_input_tokens_seen": 5171385, + "step": 162 + }, + { + "epoch": 0.014699914325652702, + "flos": 20055755793120.0, + "grad_norm": 2.13299012055861, + "learning_rate": 3.508006361478857e-06, + "loss": 0.8845, + "num_input_tokens_seen": 5199690, + "step": 163 + }, + { + "epoch": 0.014790097849122965, + "flos": 18270285613440.0, + "grad_norm": 2.101397993604937, + "learning_rate": 3.5122185357001253e-06, + "loss": 0.8368, + "num_input_tokens_seen": 5227625, + "step": 164 + }, + { + "epoch": 0.014880281372593228, + "flos": 23004488408160.0, + "grad_norm": 2.6103207821499406, + "learning_rate": 3.5164051038408817e-06, + "loss": 0.9091, + "num_input_tokens_seen": 5254215, + "step": 165 + }, + { + "epoch": 0.01497046489606349, + "flos": 24718767030240.0, + "grad_norm": 1.8755976042483455, + "learning_rate": 3.5205663753430093e-06, + "loss": 0.8946, + "num_input_tokens_seen": 5283430, + "step": 166 + }, + { + "epoch": 0.015060648419533751, + "flos": 25300602681600.0, + "grad_norm": 1.741686693407969, + "learning_rate": 3.5247026540727915e-06, + "loss": 0.9496, + "num_input_tokens_seen": 5313155, + "step": 167 + }, + { + "epoch": 0.015150831943004013, + "flos": 27267415618560.0, + "grad_norm": 2.5163919926322085, + "learning_rate": 3.5288142384540645e-06, + "loss": 0.7994, + "num_input_tokens_seen": 5345475, + "step": 168 + }, + { + "epoch": 0.015241015466474276, + "flos": 26504311549440.0, + "grad_norm": 2.2273125706861667, + "learning_rate": 3.532901421597421e-06, + "loss": 0.8852, + "num_input_tokens_seen": 5376045, + "step": 169 + }, + { + "epoch": 0.015331198989944537, + "flos": 29594498035680.0, + "grad_norm": 2.9439747653688233, + "learning_rate": 3.5369644914255915e-06, + "loss": 0.7978, + "num_input_tokens_seen": 5400890, + "step": 170 + }, + { + "epoch": 0.0154213825134148, + "flos": 25993890474240.0, + "grad_norm": 2.221235790132099, + "learning_rate": 3.5410037307951596e-06, + "loss": 0.8508, + "num_input_tokens_seen": 5431340, + "step": 171 + }, + { + "epoch": 0.015511566036885062, + "flos": 28907045895360.0, + "grad_norm": 1.9215228370621673, + "learning_rate": 3.545019417614725e-06, + "loss": 0.899, + "num_input_tokens_seen": 5464405, + "step": 172 + }, + { + "epoch": 0.015601749560355323, + "flos": 20929865966400.0, + "grad_norm": 2.030276255860237, + "learning_rate": 3.5490118249596387e-06, + "loss": 0.8953, + "num_input_tokens_seen": 5492020, + "step": 173 + }, + { + "epoch": 0.015691933083825584, + "flos": 23513645711520.0, + "grad_norm": 2.35781384285395, + "learning_rate": 3.5529812211834352e-06, + "loss": 0.904, + "num_input_tokens_seen": 5522460, + "step": 174 + }, + { + "epoch": 0.015782116607295848, + "flos": 21620143008480.0, + "grad_norm": 1.9019169573460302, + "learning_rate": 3.5569278700260707e-06, + "loss": 0.8563, + "num_input_tokens_seen": 5551315, + "step": 175 + }, + { + "epoch": 0.01587230013076611, + "flos": 22387447260480.0, + "grad_norm": 2.0622751517989286, + "learning_rate": 3.5608520307190746e-06, + "loss": 0.8384, + "num_input_tokens_seen": 5580890, + "step": 176 + }, + { + "epoch": 0.01596248365423637, + "flos": 33751920266400.0, + "grad_norm": 1.6895564935185357, + "learning_rate": 3.564753958087731e-06, + "loss": 0.7929, + "num_input_tokens_seen": 5612435, + "step": 177 + }, + { + "epoch": 0.016052667177706634, + "flos": 19728312944640.0, + "grad_norm": 1.8725643232474276, + "learning_rate": 3.5686339026503684e-06, + "loss": 0.8645, + "num_input_tokens_seen": 5642500, + "step": 178 + }, + { + "epoch": 0.016142850701176895, + "flos": 26685617136960.0, + "grad_norm": 6.801543778600326, + "learning_rate": 3.5724921107148806e-06, + "loss": 0.9393, + "num_input_tokens_seen": 5672890, + "step": 179 + }, + { + "epoch": 0.016233034224647155, + "flos": 21221285583840.0, + "grad_norm": 2.0876911235734825, + "learning_rate": 3.576328824472552e-06, + "loss": 0.9137, + "num_input_tokens_seen": 5701850, + "step": 180 + }, + { + "epoch": 0.01632321774811742, + "flos": 25075728401760.0, + "grad_norm": 5.447218880054452, + "learning_rate": 3.5801442820892838e-06, + "loss": 0.9471, + "num_input_tokens_seen": 5730545, + "step": 181 + }, + { + "epoch": 0.01641340127158768, + "flos": 26504088530880.0, + "grad_norm": 1.4646115538626225, + "learning_rate": 3.583938717794313e-06, + "loss": 0.8785, + "num_input_tokens_seen": 5761535, + "step": 182 + }, + { + "epoch": 0.01650358479505794, + "flos": 26610188226720.0, + "grad_norm": 2.1552181262725973, + "learning_rate": 3.5877123619664928e-06, + "loss": 0.8982, + "num_input_tokens_seen": 5791780, + "step": 183 + }, + { + "epoch": 0.016593768318528206, + "flos": 27701701558080.0, + "grad_norm": 2.268264811104713, + "learning_rate": 3.5914654412182268e-06, + "loss": 0.8737, + "num_input_tokens_seen": 5821560, + "step": 184 + }, + { + "epoch": 0.016683951841998466, + "flos": 23042481636480.0, + "grad_norm": 2.486006770676395, + "learning_rate": 3.595198178477127e-06, + "loss": 0.8931, + "num_input_tokens_seen": 5851315, + "step": 185 + }, + { + "epoch": 0.016774135365468727, + "flos": 45958717567200.0, + "grad_norm": 4.59175201355159, + "learning_rate": 3.5989107930654757e-06, + "loss": 0.7914, + "num_input_tokens_seen": 5885980, + "step": 186 + }, + { + "epoch": 0.01686431888893899, + "flos": 23002369731840.0, + "grad_norm": 1.906621427777382, + "learning_rate": 3.6026035007775437e-06, + "loss": 0.9214, + "num_input_tokens_seen": 5912870, + "step": 187 + }, + { + "epoch": 0.016954502412409252, + "flos": 17359631832480.0, + "grad_norm": 2.198903794713853, + "learning_rate": 3.6062765139548636e-06, + "loss": 0.826, + "num_input_tokens_seen": 5939170, + "step": 188 + }, + { + "epoch": 0.017044685935879513, + "flos": 29889820477920.0, + "grad_norm": 9.224987131332064, + "learning_rate": 3.6099300415594945e-06, + "loss": 0.7223, + "num_input_tokens_seen": 5971245, + "step": 189 + }, + { + "epoch": 0.017134869459349777, + "flos": 30437156858400.0, + "grad_norm": 1.8110270598995015, + "learning_rate": 3.6135642892453575e-06, + "loss": 0.7385, + "num_input_tokens_seen": 6004155, + "step": 190 + }, + { + "epoch": 0.01722505298282004, + "flos": 25557485858400.0, + "grad_norm": 1.5372220571239295, + "learning_rate": 3.6171794594277004e-06, + "loss": 0.8714, + "num_input_tokens_seen": 6035355, + "step": 191 + }, + { + "epoch": 0.0173152365062903, + "flos": 21039682638240.0, + "grad_norm": 4.11226027697315, + "learning_rate": 3.620775751350745e-06, + "loss": 0.8068, + "num_input_tokens_seen": 6063520, + "step": 192 + }, + { + "epoch": 0.017405420029760563, + "flos": 17430265843680.0, + "grad_norm": 4.162017262736004, + "learning_rate": 3.6243533611535794e-06, + "loss": 0.8954, + "num_input_tokens_seen": 6085410, + "step": 193 + }, + { + "epoch": 0.017495603553230824, + "flos": 25628565906720.0, + "grad_norm": 1.9594690668380716, + "learning_rate": 3.627912481934351e-06, + "loss": 0.848, + "num_input_tokens_seen": 6115620, + "step": 194 + }, + { + "epoch": 0.017585787076701085, + "flos": 22569310394400.0, + "grad_norm": 2.0892465468986727, + "learning_rate": 3.6314533038128e-06, + "loss": 0.831, + "num_input_tokens_seen": 6141860, + "step": 195 + }, + { + "epoch": 0.01767597060017135, + "flos": 65342706858720.0, + "grad_norm": 0.8074622599519082, + "learning_rate": 3.6349760139912048e-06, + "loss": 0.612, + "num_input_tokens_seen": 6225910, + "step": 196 + }, + { + "epoch": 0.01776615412364161, + "flos": 25738902955200.0, + "grad_norm": 1.833591731963898, + "learning_rate": 3.638480796813769e-06, + "loss": 0.9177, + "num_input_tokens_seen": 6251540, + "step": 197 + }, + { + "epoch": 0.01785633764711187, + "flos": 24864049386720.0, + "grad_norm": 3.5871003037937954, + "learning_rate": 3.641967833824504e-06, + "loss": 0.8522, + "num_input_tokens_seen": 6283125, + "step": 198 + }, + { + "epoch": 0.017946521170582135, + "flos": 22312538726880.0, + "grad_norm": 3.639026339962641, + "learning_rate": 3.645437303823663e-06, + "loss": 0.848, + "num_input_tokens_seen": 6312710, + "step": 199 + }, + { + "epoch": 0.018036704694052396, + "flos": 36923780182560.0, + "grad_norm": 2.5436691983673225, + "learning_rate": 3.64888938292275e-06, + "loss": 0.8351, + "num_input_tokens_seen": 6343685, + "step": 200 + }, + { + "epoch": 0.01812688821752266, + "flos": 20383087132320.0, + "grad_norm": 2.20228452410027, + "learning_rate": 3.6523242445981603e-06, + "loss": 0.8593, + "num_input_tokens_seen": 6372390, + "step": 201 + }, + { + "epoch": 0.01821707174099292, + "flos": 21877509392160.0, + "grad_norm": 1.7694094449661861, + "learning_rate": 3.655742059743495e-06, + "loss": 0.8217, + "num_input_tokens_seen": 6403210, + "step": 202 + }, + { + "epoch": 0.018307255264463182, + "flos": 20303495208960.0, + "grad_norm": 2.562111615777435, + "learning_rate": 3.659142996720576e-06, + "loss": 0.9125, + "num_input_tokens_seen": 6428180, + "step": 203 + }, + { + "epoch": 0.018397438787933446, + "flos": 20565433473120.0, + "grad_norm": 16.191238030640022, + "learning_rate": 3.6625272214092135e-06, + "loss": 0.8223, + "num_input_tokens_seen": 6455440, + "step": 204 + }, + { + "epoch": 0.018487622311403707, + "flos": 69057702972480.0, + "grad_norm": 0.6901464461324848, + "learning_rate": 3.6658948972557535e-06, + "loss": 0.5874, + "num_input_tokens_seen": 6559170, + "step": 205 + }, + { + "epoch": 0.018577805834873968, + "flos": 22642063081920.0, + "grad_norm": 1.8043749650110203, + "learning_rate": 3.6692461853204432e-06, + "loss": 0.7736, + "num_input_tokens_seen": 6587570, + "step": 206 + }, + { + "epoch": 0.018667989358344232, + "flos": 20929791626880.0, + "grad_norm": 1.989508888755896, + "learning_rate": 3.672581244323656e-06, + "loss": 0.7932, + "num_input_tokens_seen": 6616235, + "step": 207 + }, + { + "epoch": 0.018758172881814493, + "flos": 33426261566400.0, + "grad_norm": 1.9046878980553252, + "learning_rate": 3.6759002306909926e-06, + "loss": 0.7274, + "num_input_tokens_seen": 6650150, + "step": 208 + }, + { + "epoch": 0.018848356405284754, + "flos": 19836791505120.0, + "grad_norm": 1.7841109339303154, + "learning_rate": 3.67920329859731e-06, + "loss": 0.8628, + "num_input_tokens_seen": 6677980, + "step": 209 + }, + { + "epoch": 0.018938539928755018, + "flos": 20784509270400.0, + "grad_norm": 3.079956594136537, + "learning_rate": 3.6824906000096923e-06, + "loss": 0.8522, + "num_input_tokens_seen": 6706700, + "step": 210 + }, + { + "epoch": 0.01902872345222528, + "flos": 28247365299360.0, + "grad_norm": 5.074474348835768, + "learning_rate": 3.6857622847294067e-06, + "loss": 0.8585, + "num_input_tokens_seen": 6734750, + "step": 211 + }, + { + "epoch": 0.01911890697569554, + "flos": 17315356914720.0, + "grad_norm": 3.2053753647018777, + "learning_rate": 3.6890185004328626e-06, + "loss": 0.8478, + "num_input_tokens_seen": 6760115, + "step": 212 + }, + { + "epoch": 0.019209090499165804, + "flos": 38162062660800.0, + "grad_norm": 2.0227029417092295, + "learning_rate": 3.6922593927116113e-06, + "loss": 0.7912, + "num_input_tokens_seen": 6791665, + "step": 213 + }, + { + "epoch": 0.019299274022636065, + "flos": 21184370278560.0, + "grad_norm": 2.617883481532759, + "learning_rate": 3.695485105111406e-06, + "loss": 0.8372, + "num_input_tokens_seen": 6816890, + "step": 214 + }, + { + "epoch": 0.019389457546106326, + "flos": 22132496911200.0, + "grad_norm": 1.7343066777699303, + "learning_rate": 3.698695779170352e-06, + "loss": 0.8194, + "num_input_tokens_seen": 6844875, + "step": 215 + }, + { + "epoch": 0.01947964106957659, + "flos": 20820941368800.0, + "grad_norm": 1.999391224715717, + "learning_rate": 3.7018915544561744e-06, + "loss": 0.8391, + "num_input_tokens_seen": 6873510, + "step": 216 + }, + { + "epoch": 0.01956982459304685, + "flos": 20857187618400.0, + "grad_norm": 10.134203836628892, + "learning_rate": 3.7050725686026164e-06, + "loss": 0.7976, + "num_input_tokens_seen": 6900735, + "step": 217 + }, + { + "epoch": 0.01966000811651711, + "flos": 23256911213760.0, + "grad_norm": 2.2682201386778273, + "learning_rate": 3.708238957345014e-06, + "loss": 0.91, + "num_input_tokens_seen": 6925130, + "step": 218 + }, + { + "epoch": 0.019750191639987376, + "flos": 23516544952800.0, + "grad_norm": 1.8013248574706684, + "learning_rate": 3.7113908545550482e-06, + "loss": 0.8882, + "num_input_tokens_seen": 6953215, + "step": 219 + }, + { + "epoch": 0.019840375163457637, + "flos": 22019111942400.0, + "grad_norm": 1.6637470501559022, + "learning_rate": 3.7145283922747028e-06, + "loss": 0.8423, + "num_input_tokens_seen": 6981205, + "step": 220 + }, + { + "epoch": 0.019930558686927898, + "flos": 21286455640320.0, + "grad_norm": 1.606210727434194, + "learning_rate": 3.7176517007494612e-06, + "loss": 0.9359, + "num_input_tokens_seen": 7010190, + "step": 221 + }, + { + "epoch": 0.020020742210398162, + "flos": 36121679301600.0, + "grad_norm": 1.987828231195153, + "learning_rate": 3.7207609084607496e-06, + "loss": 0.7871, + "num_input_tokens_seen": 7044110, + "step": 222 + }, + { + "epoch": 0.020110925733868423, + "flos": 22816417924320.0, + "grad_norm": 2.648233681547019, + "learning_rate": 3.723856142157645e-06, + "loss": 0.8958, + "num_input_tokens_seen": 7070420, + "step": 223 + }, + { + "epoch": 0.020201109257338683, + "flos": 27084883428960.0, + "grad_norm": 1.61571446656186, + "learning_rate": 3.726937526887885e-06, + "loss": 0.8547, + "num_input_tokens_seen": 7100240, + "step": 224 + }, + { + "epoch": 0.020291292780808948, + "flos": 28909127401920.0, + "grad_norm": 3.149932884833434, + "learning_rate": 3.7300051860281798e-06, + "loss": 0.8078, + "num_input_tokens_seen": 7130385, + "step": 225 + }, + { + "epoch": 0.02038147630427921, + "flos": 16411839727680.0, + "grad_norm": 2.5254692661346474, + "learning_rate": 3.733059241313852e-06, + "loss": 0.9273, + "num_input_tokens_seen": 7152920, + "step": 226 + }, + { + "epoch": 0.02047165982774947, + "flos": 26577212916000.0, + "grad_norm": 1.9321812257968711, + "learning_rate": 3.736099812867827e-06, + "loss": 0.8359, + "num_input_tokens_seen": 7182140, + "step": 227 + }, + { + "epoch": 0.020561843351219734, + "flos": 20857596485760.0, + "grad_norm": 2.233252716981778, + "learning_rate": 3.73912701922898e-06, + "loss": 0.9051, + "num_input_tokens_seen": 7211450, + "step": 228 + }, + { + "epoch": 0.020652026874689994, + "flos": 65837999841600.0, + "grad_norm": 0.878677236602209, + "learning_rate": 3.742140977379868e-06, + "loss": 0.6176, + "num_input_tokens_seen": 7299215, + "step": 229 + }, + { + "epoch": 0.020742210398160255, + "flos": 21112100797920.0, + "grad_norm": 2.632496800724125, + "learning_rate": 3.745141802773854e-06, + "loss": 0.8779, + "num_input_tokens_seen": 7326385, + "step": 230 + }, + { + "epoch": 0.02083239392163052, + "flos": 24353851330080.0, + "grad_norm": 2.3925921755008357, + "learning_rate": 3.748129609361645e-06, + "loss": 0.8379, + "num_input_tokens_seen": 7353150, + "step": 231 + }, + { + "epoch": 0.02092257744510078, + "flos": 29824836270240.0, + "grad_norm": 1.7210877905425153, + "learning_rate": 3.7511045096172555e-06, + "loss": 0.8216, + "num_input_tokens_seen": 7386195, + "step": 232 + }, + { + "epoch": 0.02101276096857104, + "flos": 19800693934560.0, + "grad_norm": 2.1314582297885294, + "learning_rate": 3.7540666145634137e-06, + "loss": 0.8385, + "num_input_tokens_seen": 7414470, + "step": 233 + }, + { + "epoch": 0.021102944492041305, + "flos": 21586201284000.0, + "grad_norm": 1.696132198057856, + "learning_rate": 3.7570160337964225e-06, + "loss": 0.8215, + "num_input_tokens_seen": 7443970, + "step": 234 + }, + { + "epoch": 0.021193128015511566, + "flos": 21253034292480.0, + "grad_norm": 2.982908198008954, + "learning_rate": 3.7599528755104913e-06, + "loss": 0.8769, + "num_input_tokens_seen": 7471075, + "step": 235 + }, + { + "epoch": 0.021283311538981827, + "flos": 23146351146720.0, + "grad_norm": 1.836263684454838, + "learning_rate": 3.7628772465215515e-06, + "loss": 0.8432, + "num_input_tokens_seen": 7499375, + "step": 236 + }, + { + "epoch": 0.02137349506245209, + "flos": 28939129131840.0, + "grad_norm": 3.8432023813569045, + "learning_rate": 3.7657892522905666e-06, + "loss": 0.8378, + "num_input_tokens_seen": 7528425, + "step": 237 + }, + { + "epoch": 0.021463678585922352, + "flos": 37579743802560.0, + "grad_norm": 5.455281790247938, + "learning_rate": 3.7686889969463542e-06, + "loss": 0.7592, + "num_input_tokens_seen": 7558670, + "step": 238 + }, + { + "epoch": 0.021553862109392613, + "flos": 22860618502560.0, + "grad_norm": 2.107107908037579, + "learning_rate": 3.771576583307928e-06, + "loss": 0.8177, + "num_input_tokens_seen": 7585190, + "step": 239 + }, + { + "epoch": 0.021644045632862877, + "flos": 23115308663520.0, + "grad_norm": 2.26964039630377, + "learning_rate": 3.7744521129063722e-06, + "loss": 0.8778, + "num_input_tokens_seen": 7612515, + "step": 240 + }, + { + "epoch": 0.021734229156333138, + "flos": 19726937663520.0, + "grad_norm": 1.8173092802116093, + "learning_rate": 3.7773156860062653e-06, + "loss": 0.8682, + "num_input_tokens_seen": 7641000, + "step": 241 + }, + { + "epoch": 0.0218244126798034, + "flos": 25519269611520.0, + "grad_norm": 2.2485852562109945, + "learning_rate": 3.7801674016266554e-06, + "loss": 0.9411, + "num_input_tokens_seen": 7667140, + "step": 242 + }, + { + "epoch": 0.021914596203273663, + "flos": 69281127631680.0, + "grad_norm": 0.6194952060754128, + "learning_rate": 3.7830073575616035e-06, + "loss": 0.563, + "num_input_tokens_seen": 7767010, + "step": 243 + }, + { + "epoch": 0.022004779726743924, + "flos": 20378701100640.0, + "grad_norm": 2.0583892083660165, + "learning_rate": 3.785835650400313e-06, + "loss": 0.8827, + "num_input_tokens_seen": 7794790, + "step": 244 + }, + { + "epoch": 0.022094963250214185, + "flos": 32114520175200.0, + "grad_norm": 2.2163708248003364, + "learning_rate": 3.7886523755468334e-06, + "loss": 0.751, + "num_input_tokens_seen": 7823310, + "step": 245 + }, + { + "epoch": 0.02218514677368445, + "flos": 26431112824800.0, + "grad_norm": 1.804629106113587, + "learning_rate": 3.7914576272393746e-06, + "loss": 0.8358, + "num_input_tokens_seen": 7853825, + "step": 246 + }, + { + "epoch": 0.02227533029715471, + "flos": 22568715678240.0, + "grad_norm": 3.768930211352993, + "learning_rate": 3.7942514985692284e-06, + "loss": 0.9506, + "num_input_tokens_seen": 7879165, + "step": 247 + }, + { + "epoch": 0.02236551382062497, + "flos": 21076560773760.0, + "grad_norm": 2.4852748177001733, + "learning_rate": 3.797034081499296e-06, + "loss": 0.8347, + "num_input_tokens_seen": 7907175, + "step": 248 + }, + { + "epoch": 0.022455697344095235, + "flos": 18080877018240.0, + "grad_norm": 2.0085113140506414, + "learning_rate": 3.7998054668822595e-06, + "loss": 0.8227, + "num_input_tokens_seen": 7935370, + "step": 249 + }, + { + "epoch": 0.022545880867565496, + "flos": 24935947169760.0, + "grad_norm": 2.2053081976962248, + "learning_rate": 3.8025657444783776e-06, + "loss": 0.8198, + "num_input_tokens_seen": 7964420, + "step": 250 + }, + { + "epoch": 0.022636064391035757, + "flos": 24275114311200.0, + "grad_norm": 2.394180346816224, + "learning_rate": 3.80531500297293e-06, + "loss": 0.7552, + "num_input_tokens_seen": 7991905, + "step": 251 + }, + { + "epoch": 0.02272624791450602, + "flos": 14549044980000.0, + "grad_norm": 3.1318161443091403, + "learning_rate": 3.8080533299933147e-06, + "loss": 0.8124, + "num_input_tokens_seen": 8017095, + "step": 252 + }, + { + "epoch": 0.022816431437976282, + "flos": 28762952971200.0, + "grad_norm": 2.612097775830152, + "learning_rate": 3.8107808121258067e-06, + "loss": 0.808, + "num_input_tokens_seen": 8047070, + "step": 253 + }, + { + "epoch": 0.022906614961446543, + "flos": 21585457888800.0, + "grad_norm": 2.2667908924318994, + "learning_rate": 3.813497534931985e-06, + "loss": 0.91, + "num_input_tokens_seen": 8076250, + "step": 254 + }, + { + "epoch": 0.022996798484916807, + "flos": 24245001072000.0, + "grad_norm": 2.367477651005186, + "learning_rate": 3.816203582964841e-06, + "loss": 0.8724, + "num_input_tokens_seen": 8103985, + "step": 255 + }, + { + "epoch": 0.023086982008387068, + "flos": 21039719808000.0, + "grad_norm": 1.8080245042887335, + "learning_rate": 3.818899039784565e-06, + "loss": 0.8986, + "num_input_tokens_seen": 8130890, + "step": 256 + }, + { + "epoch": 0.02317716553185733, + "flos": 21947808875520.0, + "grad_norm": 2.0095007066575947, + "learning_rate": 3.821583987974031e-06, + "loss": 0.9026, + "num_input_tokens_seen": 8158980, + "step": 257 + }, + { + "epoch": 0.023267349055327593, + "flos": 26176311154560.0, + "grad_norm": 2.890607844121387, + "learning_rate": 3.8242585091539755e-06, + "loss": 0.7827, + "num_input_tokens_seen": 8188740, + "step": 258 + }, + { + "epoch": 0.023357532578797854, + "flos": 22387224241920.0, + "grad_norm": 2.7243666236505386, + "learning_rate": 3.8269226839978895e-06, + "loss": 0.9287, + "num_input_tokens_seen": 8216920, + "step": 259 + }, + { + "epoch": 0.023447716102268115, + "flos": 15681190592640.0, + "grad_norm": 2.8791250407130695, + "learning_rate": 3.82957659224662e-06, + "loss": 0.8021, + "num_input_tokens_seen": 8242085, + "step": 260 + }, + { + "epoch": 0.02353789962573838, + "flos": 34040440642560.0, + "grad_norm": 2.9812175565029904, + "learning_rate": 3.8322203127226855e-06, + "loss": 0.7322, + "num_input_tokens_seen": 8272885, + "step": 261 + }, + { + "epoch": 0.02362808314920864, + "flos": 18598025820000.0, + "grad_norm": 1.7952523243831562, + "learning_rate": 3.834853923344326e-06, + "loss": 0.9535, + "num_input_tokens_seen": 8300230, + "step": 262 + }, + { + "epoch": 0.0237182666726789, + "flos": 30365890961280.0, + "grad_norm": 2.1692057062075754, + "learning_rate": 3.837477501139285e-06, + "loss": 0.8467, + "num_input_tokens_seen": 8330580, + "step": 263 + }, + { + "epoch": 0.023808450196149165, + "flos": 27525042190560.0, + "grad_norm": 2.0377223262131565, + "learning_rate": 3.840091122258324e-06, + "loss": 0.8585, + "num_input_tokens_seen": 8360325, + "step": 264 + }, + { + "epoch": 0.023898633719619426, + "flos": 35280321420480.0, + "grad_norm": 1.840233698702878, + "learning_rate": 3.84269486198849e-06, + "loss": 0.767, + "num_input_tokens_seen": 8393910, + "step": 265 + }, + { + "epoch": 0.023988817243089686, + "flos": 22204506203520.0, + "grad_norm": 2.370090731240115, + "learning_rate": 3.845288794766121e-06, + "loss": 0.8639, + "num_input_tokens_seen": 8421665, + "step": 266 + }, + { + "epoch": 0.02407900076655995, + "flos": 17432979236160.0, + "grad_norm": 1.913177223228062, + "learning_rate": 3.847872994189619e-06, + "loss": 0.8415, + "num_input_tokens_seen": 8447780, + "step": 267 + }, + { + "epoch": 0.02416918429003021, + "flos": 24681740215680.0, + "grad_norm": 3.883740688345962, + "learning_rate": 3.8504475330319805e-06, + "loss": 0.9056, + "num_input_tokens_seen": 8476950, + "step": 268 + }, + { + "epoch": 0.024259367813500472, + "flos": 20201186828640.0, + "grad_norm": 1.9575666049426939, + "learning_rate": 3.853012483253093e-06, + "loss": 0.826, + "num_input_tokens_seen": 8504470, + "step": 269 + }, + { + "epoch": 0.024349551336970737, + "flos": 31676851787520.0, + "grad_norm": 2.0102205197274627, + "learning_rate": 3.855567916011802e-06, + "loss": 0.8541, + "num_input_tokens_seen": 8535815, + "step": 270 + }, + { + "epoch": 0.024439734860440997, + "flos": 19107889348800.0, + "grad_norm": 2.056890430838823, + "learning_rate": 3.858113901677755e-06, + "loss": 0.8528, + "num_input_tokens_seen": 8563690, + "step": 271 + }, + { + "epoch": 0.024529918383911258, + "flos": 24131318745120.0, + "grad_norm": 2.2269162616445533, + "learning_rate": 3.860650509843034e-06, + "loss": 0.8071, + "num_input_tokens_seen": 8594790, + "step": 272 + }, + { + "epoch": 0.024620101907381522, + "flos": 21185336692320.0, + "grad_norm": 2.080194461058651, + "learning_rate": 3.863177809333563e-06, + "loss": 0.8303, + "num_input_tokens_seen": 8623485, + "step": 273 + }, + { + "epoch": 0.024710285430851783, + "flos": 20784026063520.0, + "grad_norm": 1.9570416052386341, + "learning_rate": 3.86569586822032e-06, + "loss": 0.8208, + "num_input_tokens_seen": 8652075, + "step": 274 + }, + { + "epoch": 0.024800468954322044, + "flos": 29710224699360.0, + "grad_norm": 2.2312636687557292, + "learning_rate": 3.868204753830331e-06, + "loss": 0.8108, + "num_input_tokens_seen": 8681245, + "step": 275 + }, + { + "epoch": 0.02489065247779231, + "flos": 32110245652800.0, + "grad_norm": 2.229176803518284, + "learning_rate": 3.870704532757476e-06, + "loss": 0.7635, + "num_input_tokens_seen": 8712735, + "step": 276 + }, + { + "epoch": 0.02498083600126257, + "flos": 24719138727840.0, + "grad_norm": 1.8908812924530847, + "learning_rate": 3.8731952708730974e-06, + "loss": 0.8737, + "num_input_tokens_seen": 8741885, + "step": 277 + }, + { + "epoch": 0.02507101952473283, + "flos": 23437659254880.0, + "grad_norm": 2.1883671812803676, + "learning_rate": 3.8756770333364085e-06, + "loss": 0.8693, + "num_input_tokens_seen": 8767765, + "step": 278 + }, + { + "epoch": 0.025161203048203094, + "flos": 25955711397120.0, + "grad_norm": 3.75070552546874, + "learning_rate": 3.878149884604725e-06, + "loss": 0.7674, + "num_input_tokens_seen": 8797395, + "step": 279 + }, + { + "epoch": 0.025251386571673355, + "flos": 21221099735040.0, + "grad_norm": 2.363844118791325, + "learning_rate": 3.8806138884435125e-06, + "loss": 0.8194, + "num_input_tokens_seen": 8824825, + "step": 280 + }, + { + "epoch": 0.025341570095143616, + "flos": 24714492507840.0, + "grad_norm": 1.9126386679449001, + "learning_rate": 3.883069107936248e-06, + "loss": 0.7746, + "num_input_tokens_seen": 8856240, + "step": 281 + }, + { + "epoch": 0.02543175361861388, + "flos": 23698705444800.0, + "grad_norm": 2.3004738775609543, + "learning_rate": 3.885515605494114e-06, + "loss": 0.9022, + "num_input_tokens_seen": 8881215, + "step": 282 + }, + { + "epoch": 0.02552193714208414, + "flos": 23439666421920.0, + "grad_norm": 2.192839400360342, + "learning_rate": 3.8879534428655145e-06, + "loss": 0.9004, + "num_input_tokens_seen": 8908075, + "step": 283 + }, + { + "epoch": 0.025612120665554402, + "flos": 21511478599200.0, + "grad_norm": 1.8136968935888833, + "learning_rate": 3.890382681145432e-06, + "loss": 0.8822, + "num_input_tokens_seen": 8938480, + "step": 284 + }, + { + "epoch": 0.025702304189024666, + "flos": 29852830833120.0, + "grad_norm": 2.05636099194057, + "learning_rate": 3.892803380784608e-06, + "loss": 0.8763, + "num_input_tokens_seen": 8969985, + "step": 285 + }, + { + "epoch": 0.025792487712494927, + "flos": 53529823404480.0, + "grad_norm": 0.9740359161538354, + "learning_rate": 3.8952156015985725e-06, + "loss": 0.6414, + "num_input_tokens_seen": 9057285, + "step": 286 + }, + { + "epoch": 0.025882671235965188, + "flos": 21803678781600.0, + "grad_norm": 2.3734173771982707, + "learning_rate": 3.897619402776516e-06, + "loss": 0.844, + "num_input_tokens_seen": 9086520, + "step": 287 + }, + { + "epoch": 0.025972854759435452, + "flos": 19356966876000.0, + "grad_norm": 2.126785986352267, + "learning_rate": 3.900014842889995e-06, + "loss": 0.9181, + "num_input_tokens_seen": 9112620, + "step": 288 + }, + { + "epoch": 0.026063038282905713, + "flos": 30329793390720.0, + "grad_norm": 2.0534386846631536, + "learning_rate": 3.902401979901503e-06, + "loss": 0.7632, + "num_input_tokens_seen": 9144740, + "step": 289 + }, + { + "epoch": 0.026153221806375974, + "flos": 19363099886400.0, + "grad_norm": 5.898054946721081, + "learning_rate": 3.904780871172884e-06, + "loss": 0.8866, + "num_input_tokens_seen": 9172585, + "step": 290 + }, + { + "epoch": 0.026243405329846238, + "flos": 23071033745760.0, + "grad_norm": 3.0675356200496577, + "learning_rate": 3.907151573473601e-06, + "loss": 0.7185, + "num_input_tokens_seen": 9199655, + "step": 291 + }, + { + "epoch": 0.0263335888533165, + "flos": 24246264843840.0, + "grad_norm": 1.6077713615411307, + "learning_rate": 3.909514142988868e-06, + "loss": 0.8503, + "num_input_tokens_seen": 9232010, + "step": 292 + }, + { + "epoch": 0.02642377237678676, + "flos": 30617384522880.0, + "grad_norm": 2.209071365923621, + "learning_rate": 3.911868635327639e-06, + "loss": 0.8084, + "num_input_tokens_seen": 9262825, + "step": 293 + }, + { + "epoch": 0.026513955900257024, + "flos": 29562489138720.0, + "grad_norm": 1.4789743253634353, + "learning_rate": 3.914215105530455e-06, + "loss": 0.8182, + "num_input_tokens_seen": 9294745, + "step": 294 + }, + { + "epoch": 0.026604139423727285, + "flos": 25337592326400.0, + "grad_norm": 1.7327303950376143, + "learning_rate": 3.916553608077179e-06, + "loss": 0.8326, + "num_input_tokens_seen": 9325680, + "step": 295 + }, + { + "epoch": 0.026694322947197546, + "flos": 29964320144160.0, + "grad_norm": 1.9830907435543692, + "learning_rate": 3.91888419689457e-06, + "loss": 0.8284, + "num_input_tokens_seen": 9357525, + "step": 296 + }, + { + "epoch": 0.02678450647066781, + "flos": 22239340002240.0, + "grad_norm": 3.073695835103232, + "learning_rate": 3.921206925363754e-06, + "loss": 0.7786, + "num_input_tokens_seen": 9387045, + "step": 297 + }, + { + "epoch": 0.02687468999413807, + "flos": 15866213156160.0, + "grad_norm": 2.612375236342927, + "learning_rate": 3.923521846327559e-06, + "loss": 0.7545, + "num_input_tokens_seen": 9414210, + "step": 298 + }, + { + "epoch": 0.02696487351760833, + "flos": 13825569608640.0, + "grad_norm": 2.569050116537577, + "learning_rate": 3.925829012097725e-06, + "loss": 0.7959, + "num_input_tokens_seen": 9440680, + "step": 299 + }, + { + "epoch": 0.027055057041078596, + "flos": 25954559134560.0, + "grad_norm": 2.4982755435556583, + "learning_rate": 3.928128474462e-06, + "loss": 0.7585, + "num_input_tokens_seen": 9467725, + "step": 300 + }, + { + "epoch": 0.027145240564548857, + "flos": 24354037178880.0, + "grad_norm": 2.770711929606323, + "learning_rate": 3.930420284691115e-06, + "loss": 0.8462, + "num_input_tokens_seen": 9493175, + "step": 301 + }, + { + "epoch": 0.027235424088019117, + "flos": 18485681604480.0, + "grad_norm": 2.272465239071187, + "learning_rate": 3.932704493545644e-06, + "loss": 0.8715, + "num_input_tokens_seen": 9520760, + "step": 302 + }, + { + "epoch": 0.02732560761148938, + "flos": 24860555429280.0, + "grad_norm": 2.2129621637266954, + "learning_rate": 3.934981151282745e-06, + "loss": 0.9255, + "num_input_tokens_seen": 9547230, + "step": 303 + }, + { + "epoch": 0.027415791134959643, + "flos": 25848793966560.0, + "grad_norm": 1.8985724381578752, + "learning_rate": 3.9372503076628006e-06, + "loss": 0.8341, + "num_input_tokens_seen": 9576465, + "step": 304 + }, + { + "epoch": 0.027505974658429903, + "flos": 21148904593920.0, + "grad_norm": 1.7949266568557087, + "learning_rate": 3.939512011955941e-06, + "loss": 0.8578, + "num_input_tokens_seen": 9605080, + "step": 305 + }, + { + "epoch": 0.027596158181900168, + "flos": 14882026122720.0, + "grad_norm": 5.410046605951233, + "learning_rate": 3.941766312948463e-06, + "loss": 0.8341, + "num_input_tokens_seen": 9632400, + "step": 306 + }, + { + "epoch": 0.02768634170537043, + "flos": 17686256946240.0, + "grad_norm": 1.8840323647515385, + "learning_rate": 3.944013258949147e-06, + "loss": 0.8083, + "num_input_tokens_seen": 9660480, + "step": 307 + }, + { + "epoch": 0.02777652522884069, + "flos": 22824743950560.0, + "grad_norm": 2.0784272856024733, + "learning_rate": 3.946252897795465e-06, + "loss": 0.6582, + "num_input_tokens_seen": 9687235, + "step": 308 + }, + { + "epoch": 0.027866708752310954, + "flos": 15464939697120.0, + "grad_norm": 2.9189654056371825, + "learning_rate": 3.9484852768596935e-06, + "loss": 0.8472, + "num_input_tokens_seen": 9713295, + "step": 309 + }, + { + "epoch": 0.027956892275781214, + "flos": 23005566331200.0, + "grad_norm": 2.117278375118142, + "learning_rate": 3.950710443054923e-06, + "loss": 0.7386, + "num_input_tokens_seen": 9743365, + "step": 310 + }, + { + "epoch": 0.028047075799251475, + "flos": 33208338031680.0, + "grad_norm": 1.9730788580810192, + "learning_rate": 3.952928442840981e-06, + "loss": 0.7461, + "num_input_tokens_seen": 9771565, + "step": 311 + }, + { + "epoch": 0.02813725932272174, + "flos": 20703170368320.0, + "grad_norm": 1.8452763462643316, + "learning_rate": 3.955139322230243e-06, + "loss": 0.8566, + "num_input_tokens_seen": 9799445, + "step": 312 + }, + { + "epoch": 0.028227442846192, + "flos": 32588806510080.0, + "grad_norm": 2.5057544284885327, + "learning_rate": 3.957343126793365e-06, + "loss": 0.765, + "num_input_tokens_seen": 9832070, + "step": 313 + }, + { + "epoch": 0.02831762636966226, + "flos": 34300223060640.0, + "grad_norm": 1.8404356051089465, + "learning_rate": 3.959539901664921e-06, + "loss": 0.8152, + "num_input_tokens_seen": 9862895, + "step": 314 + }, + { + "epoch": 0.028407809893132525, + "flos": 19472619200160.0, + "grad_norm": 3.35537524198507, + "learning_rate": 3.9617296915489425e-06, + "loss": 0.9288, + "num_input_tokens_seen": 9886995, + "step": 315 + }, + { + "epoch": 0.028497993416602786, + "flos": 33455854428960.0, + "grad_norm": 4.523756052241371, + "learning_rate": 3.963912540724387e-06, + "loss": 0.775, + "num_input_tokens_seen": 9914340, + "step": 316 + }, + { + "epoch": 0.028588176940073047, + "flos": 22314248535840.0, + "grad_norm": 2.340356124285931, + "learning_rate": 3.966088493050501e-06, + "loss": 0.8338, + "num_input_tokens_seen": 9942000, + "step": 317 + }, + { + "epoch": 0.02867836046354331, + "flos": 24062989259040.0, + "grad_norm": 2.0107087497978595, + "learning_rate": 3.968257591972113e-06, + "loss": 0.8161, + "num_input_tokens_seen": 9970080, + "step": 318 + }, + { + "epoch": 0.028768543987013572, + "flos": 26249286860640.0, + "grad_norm": 2.087699231902369, + "learning_rate": 3.970419880524835e-06, + "loss": 0.912, + "num_input_tokens_seen": 10001425, + "step": 319 + }, + { + "epoch": 0.028858727510483833, + "flos": 25520570553120.0, + "grad_norm": 2.226687243798634, + "learning_rate": 3.972575401340192e-06, + "loss": 0.7954, + "num_input_tokens_seen": 10030445, + "step": 320 + }, + { + "epoch": 0.028948911033954097, + "flos": 23954845226400.0, + "grad_norm": 2.1820507295902005, + "learning_rate": 3.974724196650656e-06, + "loss": 0.8704, + "num_input_tokens_seen": 10060870, + "step": 321 + }, + { + "epoch": 0.029039094557424358, + "flos": 26649296547840.0, + "grad_norm": 3.152502573543167, + "learning_rate": 3.976866308294617e-06, + "loss": 0.8196, + "num_input_tokens_seen": 10089795, + "step": 322 + }, + { + "epoch": 0.02912927808089462, + "flos": 33206145015840.0, + "grad_norm": 4.300794998133112, + "learning_rate": 3.979001777721269e-06, + "loss": 0.7875, + "num_input_tokens_seen": 10124185, + "step": 323 + }, + { + "epoch": 0.029219461604364883, + "flos": 21695720597760.0, + "grad_norm": 2.709936655296617, + "learning_rate": 3.981130645995424e-06, + "loss": 0.7954, + "num_input_tokens_seen": 10153590, + "step": 324 + }, + { + "epoch": 0.029309645127835144, + "flos": 22569979450080.0, + "grad_norm": 2.013480176412692, + "learning_rate": 3.983252953802248e-06, + "loss": 0.8999, + "num_input_tokens_seen": 10183050, + "step": 325 + }, + { + "epoch": 0.029399828651305405, + "flos": 68919222682080.0, + "grad_norm": 0.7203496050431835, + "learning_rate": 3.9853687414519285e-06, + "loss": 0.6339, + "num_input_tokens_seen": 10285075, + "step": 326 + }, + { + "epoch": 0.02949001217477567, + "flos": 21144481392480.0, + "grad_norm": 2.0434896414101984, + "learning_rate": 3.987478048884265e-06, + "loss": 0.8717, + "num_input_tokens_seen": 10310665, + "step": 327 + }, + { + "epoch": 0.02958019569824593, + "flos": 26212631743680.0, + "grad_norm": 1.9421343598956295, + "learning_rate": 3.989580915673196e-06, + "loss": 0.802, + "num_input_tokens_seen": 10340905, + "step": 328 + }, + { + "epoch": 0.02967037922171619, + "flos": 24135890625600.0, + "grad_norm": 1.937562572330232, + "learning_rate": 3.991677381031255e-06, + "loss": 0.8029, + "num_input_tokens_seen": 10368060, + "step": 329 + }, + { + "epoch": 0.029760562745186455, + "flos": 21076151906400.0, + "grad_norm": 1.8536871457906818, + "learning_rate": 3.993767483813953e-06, + "loss": 0.8974, + "num_input_tokens_seen": 10397615, + "step": 330 + }, + { + "epoch": 0.029850746268656716, + "flos": 27376934932320.0, + "grad_norm": 2.039819920202332, + "learning_rate": 3.995851262524104e-06, + "loss": 0.741, + "num_input_tokens_seen": 10427625, + "step": 331 + }, + { + "epoch": 0.02994092979212698, + "flos": 26977482791520.0, + "grad_norm": 1.9822803766273553, + "learning_rate": 3.997928755316079e-06, + "loss": 0.8206, + "num_input_tokens_seen": 10457155, + "step": 332 + }, + { + "epoch": 0.03003111331559724, + "flos": 64155984571200.0, + "grad_norm": 0.8252025355638019, + "learning_rate": 4e-06, + "loss": 0.6197, + "num_input_tokens_seen": 10544555, + "step": 333 + }, + { + "epoch": 0.030121296839067502, + "flos": 26213598157440.0, + "grad_norm": 3.470577961173401, + "learning_rate": 3.999999914674486e-06, + "loss": 0.7899, + "num_input_tokens_seen": 10573780, + "step": 334 + }, + { + "epoch": 0.030211480362537766, + "flos": 32368058073600.0, + "grad_norm": 1.9985208405989234, + "learning_rate": 3.999999658697952e-06, + "loss": 0.8054, + "num_input_tokens_seen": 10605560, + "step": 335 + }, + { + "epoch": 0.030301663886008027, + "flos": 65416500299520.0, + "grad_norm": 0.6758771618822085, + "learning_rate": 3.9999992320704185e-06, + "loss": 0.5577, + "num_input_tokens_seen": 10693430, + "step": 336 + }, + { + "epoch": 0.030391847409478288, + "flos": 24718655520960.0, + "grad_norm": 2.6720779074201326, + "learning_rate": 3.999998634791922e-06, + "loss": 0.8409, + "num_input_tokens_seen": 10718025, + "step": 337 + }, + { + "epoch": 0.030482030932948552, + "flos": 33754670828640.0, + "grad_norm": 2.7957995174388715, + "learning_rate": 3.999997866862515e-06, + "loss": 0.8064, + "num_input_tokens_seen": 10748735, + "step": 338 + }, + { + "epoch": 0.030572214456418813, + "flos": 19616414766240.0, + "grad_norm": 1.998132131478497, + "learning_rate": 3.999996928282262e-06, + "loss": 0.9095, + "num_input_tokens_seen": 10776155, + "step": 339 + }, + { + "epoch": 0.030662397979889074, + "flos": 16084322539680.0, + "grad_norm": 2.3870830351534758, + "learning_rate": 3.999995819051244e-06, + "loss": 0.8191, + "num_input_tokens_seen": 10801910, + "step": 340 + }, + { + "epoch": 0.030752581503359338, + "flos": 29557768579200.0, + "grad_norm": 1.7349367348957068, + "learning_rate": 3.9999945391695536e-06, + "loss": 0.7827, + "num_input_tokens_seen": 10830420, + "step": 341 + }, + { + "epoch": 0.0308427650268296, + "flos": 23074713552000.0, + "grad_norm": 3.7318267427645533, + "learning_rate": 3.999993088637302e-06, + "loss": 0.9072, + "num_input_tokens_seen": 10860090, + "step": 342 + }, + { + "epoch": 0.03093294855029986, + "flos": 25807715648160.0, + "grad_norm": 2.018336047552139, + "learning_rate": 3.999991467454612e-06, + "loss": 0.8867, + "num_input_tokens_seen": 10888715, + "step": 343 + }, + { + "epoch": 0.031023132073770124, + "flos": 30182726885760.0, + "grad_norm": 1.538285964795855, + "learning_rate": 3.999989675621622e-06, + "loss": 0.7365, + "num_input_tokens_seen": 10922060, + "step": 344 + }, + { + "epoch": 0.031113315597240385, + "flos": 27342621510240.0, + "grad_norm": 1.7207967189314175, + "learning_rate": 3.999987713138485e-06, + "loss": 0.8265, + "num_input_tokens_seen": 10952525, + "step": 345 + }, + { + "epoch": 0.031203499120710645, + "flos": 71623152292320.0, + "grad_norm": 0.788415208970884, + "learning_rate": 3.999985580005369e-06, + "loss": 0.666, + "num_input_tokens_seen": 11051690, + "step": 346 + }, + { + "epoch": 0.031293682644180906, + "flos": 21257234475360.0, + "grad_norm": 2.1021194570844512, + "learning_rate": 3.999983276222455e-06, + "loss": 0.8934, + "num_input_tokens_seen": 11077485, + "step": 347 + }, + { + "epoch": 0.03138386616765117, + "flos": 21804608025600.0, + "grad_norm": 1.9100899714105242, + "learning_rate": 3.999980801789941e-06, + "loss": 0.8321, + "num_input_tokens_seen": 11107585, + "step": 348 + }, + { + "epoch": 0.031474049691121435, + "flos": 30948135480000.0, + "grad_norm": 1.8550812969232857, + "learning_rate": 3.999978156708036e-06, + "loss": 0.8263, + "num_input_tokens_seen": 11138590, + "step": 349 + }, + { + "epoch": 0.031564233214591696, + "flos": 18925505838240.0, + "grad_norm": 2.123710128652395, + "learning_rate": 3.9999753409769675e-06, + "loss": 0.9303, + "num_input_tokens_seen": 11162785, + "step": 350 + }, + { + "epoch": 0.031654416738061956, + "flos": 66610842387840.0, + "grad_norm": 0.7179795987676258, + "learning_rate": 3.999972354596975e-06, + "loss": 0.6088, + "num_input_tokens_seen": 11252365, + "step": 351 + }, + { + "epoch": 0.03174460026153222, + "flos": 25083831409440.0, + "grad_norm": 2.1526203289358983, + "learning_rate": 3.999969197568314e-06, + "loss": 0.773, + "num_input_tokens_seen": 11280120, + "step": 352 + }, + { + "epoch": 0.03183478378500248, + "flos": 17209926274560.0, + "grad_norm": 2.434739824216788, + "learning_rate": 3.999965869891253e-06, + "loss": 0.8726, + "num_input_tokens_seen": 11305605, + "step": 353 + }, + { + "epoch": 0.03192496730847274, + "flos": 22712065207200.0, + "grad_norm": 2.0777755099242827, + "learning_rate": 3.999962371566075e-06, + "loss": 0.8496, + "num_input_tokens_seen": 11334870, + "step": 354 + }, + { + "epoch": 0.03201515083194301, + "flos": 20966112216000.0, + "grad_norm": 2.0328001709548777, + "learning_rate": 3.999958702593082e-06, + "loss": 0.8908, + "num_input_tokens_seen": 11362935, + "step": 355 + }, + { + "epoch": 0.03210533435541327, + "flos": 27525636906720.0, + "grad_norm": 1.820950637751659, + "learning_rate": 3.999954862972583e-06, + "loss": 0.8473, + "num_input_tokens_seen": 11394775, + "step": 356 + }, + { + "epoch": 0.03219551787888353, + "flos": 24100685129280.0, + "grad_norm": 1.8005149131311957, + "learning_rate": 3.999950852704908e-06, + "loss": 0.7695, + "num_input_tokens_seen": 11424210, + "step": 357 + }, + { + "epoch": 0.03228570140235379, + "flos": 19982817256800.0, + "grad_norm": 1.5785104262038168, + "learning_rate": 3.9999466717903995e-06, + "loss": 0.8811, + "num_input_tokens_seen": 11450810, + "step": 358 + }, + { + "epoch": 0.03237588492582405, + "flos": 57576536889120.0, + "grad_norm": 0.7129580531173719, + "learning_rate": 3.999942320229413e-06, + "loss": 0.594, + "num_input_tokens_seen": 11542735, + "step": 359 + }, + { + "epoch": 0.03246606844929431, + "flos": 17140890563040.0, + "grad_norm": 2.3964779083488037, + "learning_rate": 3.99993779802232e-06, + "loss": 0.9526, + "num_input_tokens_seen": 11569130, + "step": 360 + }, + { + "epoch": 0.03255625197276458, + "flos": 24641888499360.0, + "grad_norm": 1.9390903836688447, + "learning_rate": 3.999933105169506e-06, + "loss": 0.8128, + "num_input_tokens_seen": 11595935, + "step": 361 + }, + { + "epoch": 0.03264643549623484, + "flos": 20020438787520.0, + "grad_norm": 1.6018745676576625, + "learning_rate": 3.999928241671373e-06, + "loss": 0.8216, + "num_input_tokens_seen": 11626830, + "step": 362 + }, + { + "epoch": 0.0327366190197051, + "flos": 50767568707200.0, + "grad_norm": 2.4771690227637047, + "learning_rate": 3.999923207528334e-06, + "loss": 0.7173, + "num_input_tokens_seen": 11658320, + "step": 363 + }, + { + "epoch": 0.03282680254317536, + "flos": 21694940032800.0, + "grad_norm": 1.5110708814919378, + "learning_rate": 3.9999180027408196e-06, + "loss": 0.8187, + "num_input_tokens_seen": 11688520, + "step": 364 + }, + { + "epoch": 0.03291698606664562, + "flos": 28719309939360.0, + "grad_norm": 2.0930819228424817, + "learning_rate": 3.9999126273092735e-06, + "loss": 0.8002, + "num_input_tokens_seen": 11713595, + "step": 365 + }, + { + "epoch": 0.03300716959011588, + "flos": 25990805384160.0, + "grad_norm": 5.2071228248230295, + "learning_rate": 3.999907081234156e-06, + "loss": 0.9126, + "num_input_tokens_seen": 11742855, + "step": 366 + }, + { + "epoch": 0.03309735311358615, + "flos": 22568975866560.0, + "grad_norm": 2.4107196464680807, + "learning_rate": 3.999901364515938e-06, + "loss": 0.7908, + "num_input_tokens_seen": 11770265, + "step": 367 + }, + { + "epoch": 0.03318753663705641, + "flos": 21617243767200.0, + "grad_norm": 2.1171958331968153, + "learning_rate": 3.999895477155108e-06, + "loss": 0.8709, + "num_input_tokens_seen": 11797585, + "step": 368 + }, + { + "epoch": 0.03327772016052667, + "flos": 57111840352320.0, + "grad_norm": 0.7218602100525112, + "learning_rate": 3.999889419152169e-06, + "loss": 0.5753, + "num_input_tokens_seen": 11890860, + "step": 369 + }, + { + "epoch": 0.03336790368399693, + "flos": 16813001677440.0, + "grad_norm": 1.8524861024034103, + "learning_rate": 3.999883190507638e-06, + "loss": 0.8192, + "num_input_tokens_seen": 11919400, + "step": 370 + }, + { + "epoch": 0.033458087207467194, + "flos": 21658879632000.0, + "grad_norm": 2.169901877790957, + "learning_rate": 3.999876791222044e-06, + "loss": 0.8047, + "num_input_tokens_seen": 11948130, + "step": 371 + }, + { + "epoch": 0.033548270730937454, + "flos": 21731855338080.0, + "grad_norm": 2.144203702999903, + "learning_rate": 3.999870221295936e-06, + "loss": 0.7738, + "num_input_tokens_seen": 11976850, + "step": 372 + }, + { + "epoch": 0.03363845425440772, + "flos": 34263456434400.0, + "grad_norm": 1.8254511262551822, + "learning_rate": 3.999863480729875e-06, + "loss": 0.787, + "num_input_tokens_seen": 12009675, + "step": 373 + }, + { + "epoch": 0.03372863777787798, + "flos": 27630063963360.0, + "grad_norm": 3.2919265708845487, + "learning_rate": 3.999856569524433e-06, + "loss": 0.7589, + "num_input_tokens_seen": 12038820, + "step": 374 + }, + { + "epoch": 0.033818821301348244, + "flos": 21073438513920.0, + "grad_norm": 1.7463905713702752, + "learning_rate": 3.999849487680202e-06, + "loss": 0.7604, + "num_input_tokens_seen": 12067815, + "step": 375 + }, + { + "epoch": 0.033909004824818505, + "flos": 39620498859360.0, + "grad_norm": 1.6053789768040532, + "learning_rate": 3.999842235197786e-06, + "loss": 0.7538, + "num_input_tokens_seen": 12100470, + "step": 376 + }, + { + "epoch": 0.033999188348288766, + "flos": 28399040854560.0, + "grad_norm": 2.4347875125610536, + "learning_rate": 3.999834812077803e-06, + "loss": 0.7529, + "num_input_tokens_seen": 12130575, + "step": 377 + }, + { + "epoch": 0.034089371871759026, + "flos": 25557337179360.0, + "grad_norm": 2.113565647922343, + "learning_rate": 3.999827218320886e-06, + "loss": 0.8727, + "num_input_tokens_seen": 12160305, + "step": 378 + }, + { + "epoch": 0.034179555395229294, + "flos": 19360795361280.0, + "grad_norm": 2.32607860362178, + "learning_rate": 3.999819453927685e-06, + "loss": 0.7336, + "num_input_tokens_seen": 12188015, + "step": 379 + }, + { + "epoch": 0.034269738918699555, + "flos": 18998890411680.0, + "grad_norm": 1.9398712905975162, + "learning_rate": 3.999811518898861e-06, + "loss": 0.8195, + "num_input_tokens_seen": 12214610, + "step": 380 + }, + { + "epoch": 0.034359922442169816, + "flos": 16229716405440.0, + "grad_norm": 2.45263074038379, + "learning_rate": 3.999803413235092e-06, + "loss": 0.834, + "num_input_tokens_seen": 12242205, + "step": 381 + }, + { + "epoch": 0.03445010596564008, + "flos": 21185150843520.0, + "grad_norm": 1.99027762024441, + "learning_rate": 3.999795136937068e-06, + "loss": 0.9289, + "num_input_tokens_seen": 12270525, + "step": 382 + }, + { + "epoch": 0.03454028948911034, + "flos": 21694679844480.0, + "grad_norm": 1.999522450177246, + "learning_rate": 3.999786690005496e-06, + "loss": 0.8077, + "num_input_tokens_seen": 12298770, + "step": 383 + }, + { + "epoch": 0.0346304730125806, + "flos": 24317567910720.0, + "grad_norm": 2.2362134079474143, + "learning_rate": 3.999778072441098e-06, + "loss": 0.8026, + "num_input_tokens_seen": 12329845, + "step": 384 + }, + { + "epoch": 0.034720656536050866, + "flos": 20054900888640.0, + "grad_norm": 2.693393012729577, + "learning_rate": 3.999769284244608e-06, + "loss": 0.8488, + "num_input_tokens_seen": 12356150, + "step": 385 + }, + { + "epoch": 0.03481084005952113, + "flos": 28976787832320.0, + "grad_norm": 1.5027439927263628, + "learning_rate": 3.999760325416775e-06, + "loss": 0.836, + "num_input_tokens_seen": 12387385, + "step": 386 + }, + { + "epoch": 0.03490102358299139, + "flos": 22460162778240.0, + "grad_norm": 1.770667109448213, + "learning_rate": 3.999751195958366e-06, + "loss": 0.9061, + "num_input_tokens_seen": 12416085, + "step": 387 + }, + { + "epoch": 0.03499120710646165, + "flos": 59128912538400.0, + "grad_norm": 0.792734459151157, + "learning_rate": 3.999741895870157e-06, + "loss": 0.5985, + "num_input_tokens_seen": 12507880, + "step": 388 + }, + { + "epoch": 0.03508139062993191, + "flos": 21216788042880.0, + "grad_norm": 4.057710142023415, + "learning_rate": 3.999732425152944e-06, + "loss": 0.8639, + "num_input_tokens_seen": 12533705, + "step": 389 + }, + { + "epoch": 0.03517157415340217, + "flos": 23406914129760.0, + "grad_norm": 1.6440745827354675, + "learning_rate": 3.999722783807533e-06, + "loss": 0.8791, + "num_input_tokens_seen": 12564120, + "step": 390 + }, + { + "epoch": 0.03526175767687244, + "flos": 33025657163040.0, + "grad_norm": 2.655119442165785, + "learning_rate": 3.999712971834748e-06, + "loss": 0.7522, + "num_input_tokens_seen": 12593590, + "step": 391 + }, + { + "epoch": 0.0353519412003427, + "flos": 17578373101920.0, + "grad_norm": 1.8292318400573409, + "learning_rate": 3.999702989235427e-06, + "loss": 0.8361, + "num_input_tokens_seen": 12621770, + "step": 392 + }, + { + "epoch": 0.03544212472381296, + "flos": 22093946136480.0, + "grad_norm": 2.9838943493832435, + "learning_rate": 3.999692836010419e-06, + "loss": 0.9127, + "num_input_tokens_seen": 12648465, + "step": 393 + }, + { + "epoch": 0.03553230824728322, + "flos": 17577406688160.0, + "grad_norm": 2.351039181900796, + "learning_rate": 3.999682512160593e-06, + "loss": 0.9621, + "num_input_tokens_seen": 12672705, + "step": 394 + }, + { + "epoch": 0.03562249177075348, + "flos": 59439591825120.0, + "grad_norm": 0.7735432253930699, + "learning_rate": 3.99967201768683e-06, + "loss": 0.5766, + "num_input_tokens_seen": 12757070, + "step": 395 + }, + { + "epoch": 0.03571267529422374, + "flos": 21294298459680.0, + "grad_norm": 3.3547015662002786, + "learning_rate": 3.999661352590023e-06, + "loss": 0.8091, + "num_input_tokens_seen": 12785635, + "step": 396 + }, + { + "epoch": 0.03580285881769401, + "flos": 25407111244800.0, + "grad_norm": 3.514502587754212, + "learning_rate": 3.999650516871083e-06, + "loss": 0.8261, + "num_input_tokens_seen": 12815180, + "step": 397 + }, + { + "epoch": 0.03589304234116427, + "flos": 22532543768160.0, + "grad_norm": 1.9090574883966422, + "learning_rate": 3.9996395105309365e-06, + "loss": 0.8487, + "num_input_tokens_seen": 12843385, + "step": 398 + }, + { + "epoch": 0.03598322586463453, + "flos": 28468411093920.0, + "grad_norm": 1.665958023417146, + "learning_rate": 3.99962833357052e-06, + "loss": 0.8488, + "num_input_tokens_seen": 12873435, + "step": 399 + }, + { + "epoch": 0.03607340938810479, + "flos": 26905250480640.0, + "grad_norm": 2.2534806562075502, + "learning_rate": 3.999616985990789e-06, + "loss": 0.8048, + "num_input_tokens_seen": 12903180, + "step": 400 + }, + { + "epoch": 0.03616359291157505, + "flos": 24099421357440.0, + "grad_norm": 2.011855630014064, + "learning_rate": 3.9996054677927104e-06, + "loss": 0.7611, + "num_input_tokens_seen": 12933930, + "step": 401 + }, + { + "epoch": 0.03625377643504532, + "flos": 26210178539520.0, + "grad_norm": 2.786760593075569, + "learning_rate": 3.9995937789772675e-06, + "loss": 0.7645, + "num_input_tokens_seen": 12962935, + "step": 402 + }, + { + "epoch": 0.03634395995851558, + "flos": 72012196900320.0, + "grad_norm": 0.726054475435064, + "learning_rate": 3.999581919545458e-06, + "loss": 0.602, + "num_input_tokens_seen": 13055140, + "step": 403 + }, + { + "epoch": 0.03643414348198584, + "flos": 18197086888800.0, + "grad_norm": 3.023555309750527, + "learning_rate": 3.9995698894982935e-06, + "loss": 0.8476, + "num_input_tokens_seen": 13078670, + "step": 404 + }, + { + "epoch": 0.0365243270054561, + "flos": 23734059620160.0, + "grad_norm": 1.6815251433698617, + "learning_rate": 3.9995576888368e-06, + "loss": 0.8738, + "num_input_tokens_seen": 13109050, + "step": 405 + }, + { + "epoch": 0.036614510528926364, + "flos": 21694717014240.0, + "grad_norm": 2.0490315266348524, + "learning_rate": 3.9995453175620194e-06, + "loss": 0.8119, + "num_input_tokens_seen": 13136915, + "step": 406 + }, + { + "epoch": 0.036704694052396625, + "flos": 64390831813920.0, + "grad_norm": 3.310617883871567, + "learning_rate": 3.999532775675007e-06, + "loss": 0.6794, + "num_input_tokens_seen": 13164455, + "step": 407 + }, + { + "epoch": 0.03679487757586689, + "flos": 35903607087840.0, + "grad_norm": 1.9575332332825202, + "learning_rate": 3.9995200631768326e-06, + "loss": 0.7956, + "num_input_tokens_seen": 13195475, + "step": 408 + }, + { + "epoch": 0.03688506109933715, + "flos": 16922595330720.0, + "grad_norm": 2.4724462338610107, + "learning_rate": 3.9995071800685815e-06, + "loss": 0.9498, + "num_input_tokens_seen": 13220215, + "step": 409 + }, + { + "epoch": 0.036975244622807414, + "flos": 24937433960160.0, + "grad_norm": 2.632725745336702, + "learning_rate": 3.999494126351352e-06, + "loss": 0.7943, + "num_input_tokens_seen": 13246735, + "step": 410 + }, + { + "epoch": 0.037065428146277675, + "flos": 23807964570240.0, + "grad_norm": 3.3851792159315006, + "learning_rate": 3.99948090202626e-06, + "loss": 0.829, + "num_input_tokens_seen": 13278145, + "step": 411 + }, + { + "epoch": 0.037155611669747936, + "flos": 36776007452160.0, + "grad_norm": 1.7333985750440086, + "learning_rate": 3.999467507094431e-06, + "loss": 0.8263, + "num_input_tokens_seen": 13310430, + "step": 412 + }, + { + "epoch": 0.0372457951932182, + "flos": 32841601013280.0, + "grad_norm": 1.8716034961388144, + "learning_rate": 3.999453941557011e-06, + "loss": 0.7279, + "num_input_tokens_seen": 13342040, + "step": 413 + }, + { + "epoch": 0.037335978716688464, + "flos": 60179496230400.0, + "grad_norm": 0.7299851229209908, + "learning_rate": 3.999440205415154e-06, + "loss": 0.5552, + "num_input_tokens_seen": 13432250, + "step": 414 + }, + { + "epoch": 0.037426162240158725, + "flos": 18051507174240.0, + "grad_norm": 2.4360804214293976, + "learning_rate": 3.999426298670035e-06, + "loss": 0.8706, + "num_input_tokens_seen": 13458855, + "step": 415 + }, + { + "epoch": 0.037516345763628986, + "flos": 22271497578240.0, + "grad_norm": 1.9300256965586835, + "learning_rate": 3.9994122213228385e-06, + "loss": 0.8661, + "num_input_tokens_seen": 13486285, + "step": 416 + }, + { + "epoch": 0.03760652928709925, + "flos": 24604006780320.0, + "grad_norm": 1.9917303420971657, + "learning_rate": 3.9993979733747675e-06, + "loss": 0.8527, + "num_input_tokens_seen": 13515230, + "step": 417 + }, + { + "epoch": 0.03769671281056951, + "flos": 26212483064640.0, + "grad_norm": 2.2215309445400098, + "learning_rate": 3.999383554827037e-06, + "loss": 0.8266, + "num_input_tokens_seen": 13543590, + "step": 418 + }, + { + "epoch": 0.03778689633403977, + "flos": 23917075016640.0, + "grad_norm": 1.9638398717254328, + "learning_rate": 3.999368965680876e-06, + "loss": 0.7834, + "num_input_tokens_seen": 13572800, + "step": 419 + }, + { + "epoch": 0.037877079857510036, + "flos": 22678755368640.0, + "grad_norm": 2.1140663254076624, + "learning_rate": 3.999354205937531e-06, + "loss": 0.8675, + "num_input_tokens_seen": 13602120, + "step": 420 + }, + { + "epoch": 0.0379672633809803, + "flos": 21986805687360.0, + "grad_norm": 1.984358900110008, + "learning_rate": 3.999339275598261e-06, + "loss": 0.8683, + "num_input_tokens_seen": 13628700, + "step": 421 + }, + { + "epoch": 0.03805744690445056, + "flos": 32913982003200.0, + "grad_norm": 2.8668661743054824, + "learning_rate": 3.99932417466434e-06, + "loss": 0.8048, + "num_input_tokens_seen": 13661060, + "step": 422 + }, + { + "epoch": 0.03814763042792082, + "flos": 21288722995680.0, + "grad_norm": 3.2177304625354823, + "learning_rate": 3.999308903137056e-06, + "loss": 0.7354, + "num_input_tokens_seen": 13688725, + "step": 423 + }, + { + "epoch": 0.03823781395139108, + "flos": 23079768639360.0, + "grad_norm": 2.290654901202669, + "learning_rate": 3.999293461017711e-06, + "loss": 0.8736, + "num_input_tokens_seen": 13718415, + "step": 424 + }, + { + "epoch": 0.03832799747486134, + "flos": 19249789257120.0, + "grad_norm": 1.9150503917866692, + "learning_rate": 3.9992778483076255e-06, + "loss": 0.8123, + "num_input_tokens_seen": 13744975, + "step": 425 + }, + { + "epoch": 0.03841818099833161, + "flos": 19867202102400.0, + "grad_norm": 1.8733764613041428, + "learning_rate": 3.99926206500813e-06, + "loss": 0.8798, + "num_input_tokens_seen": 13772530, + "step": 426 + }, + { + "epoch": 0.03850836452180187, + "flos": 16557493781760.0, + "grad_norm": 2.164931250310965, + "learning_rate": 3.999246111120571e-06, + "loss": 0.7509, + "num_input_tokens_seen": 13798285, + "step": 427 + }, + { + "epoch": 0.03859854804527213, + "flos": 25770094117440.0, + "grad_norm": 1.5918898853673924, + "learning_rate": 3.999229986646311e-06, + "loss": 0.8956, + "num_input_tokens_seen": 13826625, + "step": 428 + }, + { + "epoch": 0.03868873156874239, + "flos": 29309583126240.0, + "grad_norm": 4.19330888636932, + "learning_rate": 3.999213691586723e-06, + "loss": 0.7926, + "num_input_tokens_seen": 13858335, + "step": 429 + }, + { + "epoch": 0.03877891509221265, + "flos": 31494616956000.0, + "grad_norm": 1.7634225797993768, + "learning_rate": 3.9991972259432e-06, + "loss": 0.8658, + "num_input_tokens_seen": 13888455, + "step": 430 + }, + { + "epoch": 0.03886909861568291, + "flos": 24458798763360.0, + "grad_norm": 1.5614415207513985, + "learning_rate": 3.999180589717147e-06, + "loss": 0.9067, + "num_input_tokens_seen": 13918300, + "step": 431 + }, + { + "epoch": 0.03895928213915318, + "flos": 19400052361440.0, + "grad_norm": 2.8620447921286765, + "learning_rate": 3.999163782909983e-06, + "loss": 0.7365, + "num_input_tokens_seen": 13945050, + "step": 432 + }, + { + "epoch": 0.03904946566262344, + "flos": 27596716955040.0, + "grad_norm": 1.7731045313753953, + "learning_rate": 3.99914680552314e-06, + "loss": 0.8573, + "num_input_tokens_seen": 13974120, + "step": 433 + }, + { + "epoch": 0.0391396491860937, + "flos": 27816052940640.0, + "grad_norm": 2.172912130476062, + "learning_rate": 3.999129657558069e-06, + "loss": 0.7524, + "num_input_tokens_seen": 14005540, + "step": 434 + }, + { + "epoch": 0.03922983270956396, + "flos": 64380864584640.0, + "grad_norm": 0.7215456289612991, + "learning_rate": 3.999112339016234e-06, + "loss": 0.5683, + "num_input_tokens_seen": 14088300, + "step": 435 + }, + { + "epoch": 0.03932001623303422, + "flos": 20200926640320.0, + "grad_norm": 2.031814848784808, + "learning_rate": 3.999094849899109e-06, + "loss": 0.8357, + "num_input_tokens_seen": 14116265, + "step": 436 + }, + { + "epoch": 0.039410199756504484, + "flos": 20784211912320.0, + "grad_norm": 1.9750025326676366, + "learning_rate": 3.99907719020819e-06, + "loss": 0.7423, + "num_input_tokens_seen": 14144870, + "step": 437 + }, + { + "epoch": 0.03950038327997475, + "flos": 19290718896480.0, + "grad_norm": 2.4890127020174972, + "learning_rate": 3.999059359944982e-06, + "loss": 0.8511, + "num_input_tokens_seen": 14173715, + "step": 438 + }, + { + "epoch": 0.03959056680344501, + "flos": 35464972286400.0, + "grad_norm": 1.6192051192485415, + "learning_rate": 3.999041359111007e-06, + "loss": 0.8775, + "num_input_tokens_seen": 14205700, + "step": 439 + }, + { + "epoch": 0.03968075032691527, + "flos": 23771532471840.0, + "grad_norm": 2.6832826036985065, + "learning_rate": 3.999023187707801e-06, + "loss": 0.8954, + "num_input_tokens_seen": 14232815, + "step": 440 + }, + { + "epoch": 0.039770933850385534, + "flos": 21293815252800.0, + "grad_norm": 1.711819147241302, + "learning_rate": 3.999004845736913e-06, + "loss": 0.8864, + "num_input_tokens_seen": 14264895, + "step": 441 + }, + { + "epoch": 0.039861117373855795, + "flos": 25885226064960.0, + "grad_norm": 2.296494017683128, + "learning_rate": 3.9989863331999096e-06, + "loss": 0.8979, + "num_input_tokens_seen": 14295700, + "step": 442 + }, + { + "epoch": 0.039951300897326056, + "flos": 21476050084320.0, + "grad_norm": 2.4577111787761496, + "learning_rate": 3.99896765009837e-06, + "loss": 0.7556, + "num_input_tokens_seen": 14324090, + "step": 443 + }, + { + "epoch": 0.040041484420796324, + "flos": 23951053910880.0, + "grad_norm": 1.7063015353625772, + "learning_rate": 3.998948796433888e-06, + "loss": 0.7913, + "num_input_tokens_seen": 14354845, + "step": 444 + }, + { + "epoch": 0.040131667944266584, + "flos": 23843616103680.0, + "grad_norm": 4.255044523873574, + "learning_rate": 3.998929772208073e-06, + "loss": 0.8646, + "num_input_tokens_seen": 14384780, + "step": 445 + }, + { + "epoch": 0.040221851467736845, + "flos": 70347173943840.0, + "grad_norm": 0.6805027643367614, + "learning_rate": 3.998910577422547e-06, + "loss": 0.5952, + "num_input_tokens_seen": 14483575, + "step": 446 + }, + { + "epoch": 0.040312034991207106, + "flos": 25516853577120.0, + "grad_norm": 1.653001153990521, + "learning_rate": 3.99889121207895e-06, + "loss": 0.8638, + "num_input_tokens_seen": 14513215, + "step": 447 + }, + { + "epoch": 0.04040221851467737, + "flos": 17286172919520.0, + "grad_norm": 1.7836166854900934, + "learning_rate": 3.9988716761789324e-06, + "loss": 0.8485, + "num_input_tokens_seen": 14541015, + "step": 448 + }, + { + "epoch": 0.04049240203814763, + "flos": 21112026458400.0, + "grad_norm": 2.2272432056374325, + "learning_rate": 3.998851969724161e-06, + "loss": 0.7624, + "num_input_tokens_seen": 14568470, + "step": 449 + }, + { + "epoch": 0.040582585561617895, + "flos": 20492346257760.0, + "grad_norm": 2.245456983447726, + "learning_rate": 3.998832092716319e-06, + "loss": 0.8731, + "num_input_tokens_seen": 14593630, + "step": 450 + }, + { + "epoch": 0.040672769085088156, + "flos": 26539925913120.0, + "grad_norm": 1.6287822721948026, + "learning_rate": 3.998812045157102e-06, + "loss": 0.802, + "num_input_tokens_seen": 14627175, + "step": 451 + }, + { + "epoch": 0.04076295260855842, + "flos": 19435964083200.0, + "grad_norm": 2.15720324246817, + "learning_rate": 3.998791827048219e-06, + "loss": 0.873, + "num_input_tokens_seen": 14652825, + "step": 452 + }, + { + "epoch": 0.04085313613202868, + "flos": 28321344588960.0, + "grad_norm": 2.279692130969797, + "learning_rate": 3.998771438391396e-06, + "loss": 0.7655, + "num_input_tokens_seen": 14682550, + "step": 453 + }, + { + "epoch": 0.04094331965549894, + "flos": 65365311806400.0, + "grad_norm": 0.6604202916219326, + "learning_rate": 3.9987508791883725e-06, + "loss": 0.5738, + "num_input_tokens_seen": 14775700, + "step": 454 + }, + { + "epoch": 0.0410335031789692, + "flos": 16557828309600.0, + "grad_norm": 3.198801090239434, + "learning_rate": 3.998730149440904e-06, + "loss": 0.9436, + "num_input_tokens_seen": 14800335, + "step": 455 + }, + { + "epoch": 0.04112368670243947, + "flos": 24386938150080.0, + "grad_norm": 2.31800126071267, + "learning_rate": 3.998709249150758e-06, + "loss": 0.8739, + "num_input_tokens_seen": 14827205, + "step": 456 + }, + { + "epoch": 0.04121387022590973, + "flos": 23218137420480.0, + "grad_norm": 3.01170608379308, + "learning_rate": 3.998688178319717e-06, + "loss": 0.8486, + "num_input_tokens_seen": 14855235, + "step": 457 + }, + { + "epoch": 0.04130405374937999, + "flos": 20857968183360.0, + "grad_norm": 1.783888351966769, + "learning_rate": 3.9986669369495805e-06, + "loss": 0.8091, + "num_input_tokens_seen": 14881760, + "step": 458 + }, + { + "epoch": 0.04139423727285025, + "flos": 22095507266400.0, + "grad_norm": 2.3715296035196087, + "learning_rate": 3.998645525042161e-06, + "loss": 0.8788, + "num_input_tokens_seen": 14909805, + "step": 459 + }, + { + "epoch": 0.04148442079632051, + "flos": 13128416160960.0, + "grad_norm": 6.614443159245135, + "learning_rate": 3.998623942599284e-06, + "loss": 0.8937, + "num_input_tokens_seen": 14935840, + "step": 460 + }, + { + "epoch": 0.04157460431979077, + "flos": 25188853182240.0, + "grad_norm": 1.989781621136046, + "learning_rate": 3.998602189622793e-06, + "loss": 0.8184, + "num_input_tokens_seen": 14964945, + "step": 461 + }, + { + "epoch": 0.04166478784326104, + "flos": 18525533320800.0, + "grad_norm": 1.8938307892842914, + "learning_rate": 3.998580266114542e-06, + "loss": 0.8609, + "num_input_tokens_seen": 14992295, + "step": 462 + }, + { + "epoch": 0.0417549713667313, + "flos": 25994225002080.0, + "grad_norm": 1.703717959712149, + "learning_rate": 3.998558172076404e-06, + "loss": 0.7283, + "num_input_tokens_seen": 15021380, + "step": 463 + }, + { + "epoch": 0.04184515489020156, + "flos": 24639658313760.0, + "grad_norm": 1.8644099305858055, + "learning_rate": 3.998535907510262e-06, + "loss": 0.7549, + "num_input_tokens_seen": 15047520, + "step": 464 + }, + { + "epoch": 0.04193533841367182, + "flos": 25812138849600.0, + "grad_norm": 2.2854228068025555, + "learning_rate": 3.998513472418016e-06, + "loss": 0.8359, + "num_input_tokens_seen": 15077640, + "step": 465 + }, + { + "epoch": 0.04202552193714208, + "flos": 23079471281280.0, + "grad_norm": 2.1808235735365717, + "learning_rate": 3.998490866801582e-06, + "loss": 0.8586, + "num_input_tokens_seen": 15103940, + "step": 466 + }, + { + "epoch": 0.04211570546061234, + "flos": 18227869183680.0, + "grad_norm": 1.7720572860577026, + "learning_rate": 3.998468090662886e-06, + "loss": 0.8139, + "num_input_tokens_seen": 15131270, + "step": 467 + }, + { + "epoch": 0.04220588898408261, + "flos": 26139879056160.0, + "grad_norm": 1.7930704058915279, + "learning_rate": 3.998445144003874e-06, + "loss": 0.7204, + "num_input_tokens_seen": 15160850, + "step": 468 + }, + { + "epoch": 0.04229607250755287, + "flos": 21804645195360.0, + "grad_norm": 2.265201818472664, + "learning_rate": 3.998422026826504e-06, + "loss": 0.849, + "num_input_tokens_seen": 15189305, + "step": 469 + }, + { + "epoch": 0.04238625603102313, + "flos": 19253654912160.0, + "grad_norm": 2.6816241770196143, + "learning_rate": 3.998398739132746e-06, + "loss": 0.8538, + "num_input_tokens_seen": 15216675, + "step": 470 + }, + { + "epoch": 0.04247643955449339, + "flos": 25299859286400.0, + "grad_norm": 1.9391307735796102, + "learning_rate": 3.99837528092459e-06, + "loss": 0.7811, + "num_input_tokens_seen": 15244490, + "step": 471 + }, + { + "epoch": 0.042566623077963654, + "flos": 23257394420640.0, + "grad_norm": 2.3744726097405207, + "learning_rate": 3.998351652204034e-06, + "loss": 0.8726, + "num_input_tokens_seen": 15272295, + "step": 472 + }, + { + "epoch": 0.042656806601433915, + "flos": 29746582458240.0, + "grad_norm": 1.9926227168986694, + "learning_rate": 3.998327852973098e-06, + "loss": 0.867, + "num_input_tokens_seen": 15304985, + "step": 473 + }, + { + "epoch": 0.04274699012490418, + "flos": 14299261227360.0, + "grad_norm": 2.5783564608160723, + "learning_rate": 3.99830388323381e-06, + "loss": 0.7734, + "num_input_tokens_seen": 15330850, + "step": 474 + }, + { + "epoch": 0.042837173648374444, + "flos": 22202090169120.0, + "grad_norm": 1.866587969562931, + "learning_rate": 3.998279742988216e-06, + "loss": 0.8554, + "num_input_tokens_seen": 15359080, + "step": 475 + }, + { + "epoch": 0.042927357171844704, + "flos": 19655783275680.0, + "grad_norm": 1.638442849851347, + "learning_rate": 3.998255432238377e-06, + "loss": 0.7923, + "num_input_tokens_seen": 15389510, + "step": 476 + }, + { + "epoch": 0.043017540695314965, + "flos": 22751285037600.0, + "grad_norm": 4.17776587689736, + "learning_rate": 3.9982309509863656e-06, + "loss": 0.8984, + "num_input_tokens_seen": 15417930, + "step": 477 + }, + { + "epoch": 0.043107724218785226, + "flos": 26394940914720.0, + "grad_norm": 3.164716789507094, + "learning_rate": 3.998206299234272e-06, + "loss": 0.7719, + "num_input_tokens_seen": 15447615, + "step": 478 + }, + { + "epoch": 0.04319790774225549, + "flos": 25265099827200.0, + "grad_norm": 2.322170245984706, + "learning_rate": 3.998181476984198e-06, + "loss": 0.841, + "num_input_tokens_seen": 15477915, + "step": 479 + }, + { + "epoch": 0.043288091265725755, + "flos": 21294112610880.0, + "grad_norm": 2.011371115860162, + "learning_rate": 3.998156484238263e-06, + "loss": 0.7503, + "num_input_tokens_seen": 15506375, + "step": 480 + }, + { + "epoch": 0.043378274789196015, + "flos": 25045169125440.0, + "grad_norm": 2.385690920415955, + "learning_rate": 3.998131320998599e-06, + "loss": 0.7611, + "num_input_tokens_seen": 15534660, + "step": 481 + }, + { + "epoch": 0.043468458312666276, + "flos": 27925386405600.0, + "grad_norm": 2.4892592491843146, + "learning_rate": 3.998105987267353e-06, + "loss": 0.8712, + "num_input_tokens_seen": 15563860, + "step": 482 + }, + { + "epoch": 0.04355864183613654, + "flos": 25009926459360.0, + "grad_norm": 1.7553372057533845, + "learning_rate": 3.998080483046687e-06, + "loss": 0.7504, + "num_input_tokens_seen": 15593625, + "step": 483 + }, + { + "epoch": 0.0436488253596068, + "flos": 38925798615840.0, + "grad_norm": 1.7755549892975806, + "learning_rate": 3.998054808338776e-06, + "loss": 0.7739, + "num_input_tokens_seen": 15626815, + "step": 484 + }, + { + "epoch": 0.04373900888307706, + "flos": 22459493722560.0, + "grad_norm": 2.102157338417814, + "learning_rate": 3.998028963145812e-06, + "loss": 0.7849, + "num_input_tokens_seen": 15654870, + "step": 485 + }, + { + "epoch": 0.043829192406547327, + "flos": 20529410242080.0, + "grad_norm": 1.9944381865483218, + "learning_rate": 3.99800294747e-06, + "loss": 0.837, + "num_input_tokens_seen": 15682775, + "step": 486 + }, + { + "epoch": 0.04391937593001759, + "flos": 28289744559360.0, + "grad_norm": 1.7139315948991103, + "learning_rate": 3.99797676131356e-06, + "loss": 0.8592, + "num_input_tokens_seen": 15713090, + "step": 487 + }, + { + "epoch": 0.04400955945348785, + "flos": 25046172708960.0, + "grad_norm": 3.047981100559463, + "learning_rate": 3.997950404678726e-06, + "loss": 0.6813, + "num_input_tokens_seen": 15738190, + "step": 488 + }, + { + "epoch": 0.04409974297695811, + "flos": 18561259193760.0, + "grad_norm": 2.4225707008587345, + "learning_rate": 3.997923877567746e-06, + "loss": 0.9383, + "num_input_tokens_seen": 15765720, + "step": 489 + }, + { + "epoch": 0.04418992650042837, + "flos": 35172809273760.0, + "grad_norm": 1.6000627147781255, + "learning_rate": 3.9978971799828855e-06, + "loss": 0.653, + "num_input_tokens_seen": 15798950, + "step": 490 + }, + { + "epoch": 0.04428011002389863, + "flos": 23225422693440.0, + "grad_norm": 5.9463576251641355, + "learning_rate": 3.997870311926421e-06, + "loss": 0.8143, + "num_input_tokens_seen": 15827285, + "step": 491 + }, + { + "epoch": 0.0443702935473689, + "flos": 25448152393440.0, + "grad_norm": 3.8271207776813694, + "learning_rate": 3.997843273400645e-06, + "loss": 0.7809, + "num_input_tokens_seen": 15855320, + "step": 492 + }, + { + "epoch": 0.04446047707083916, + "flos": 28835631319200.0, + "grad_norm": 2.0840514153491627, + "learning_rate": 3.997816064407865e-06, + "loss": 0.7999, + "num_input_tokens_seen": 15882870, + "step": 493 + }, + { + "epoch": 0.04455066059430942, + "flos": 42169258956960.0, + "grad_norm": 1.6271111639412925, + "learning_rate": 3.997788684950402e-06, + "loss": 0.8434, + "num_input_tokens_seen": 15915905, + "step": 494 + }, + { + "epoch": 0.04464084411777968, + "flos": 19909767211200.0, + "grad_norm": 1.9417271799745779, + "learning_rate": 3.997761135030593e-06, + "loss": 0.8669, + "num_input_tokens_seen": 15943965, + "step": 495 + }, + { + "epoch": 0.04473102764124994, + "flos": 23331299370720.0, + "grad_norm": 1.8343139612287653, + "learning_rate": 3.997733414650789e-06, + "loss": 0.7238, + "num_input_tokens_seen": 15972710, + "step": 496 + }, + { + "epoch": 0.0448212111647202, + "flos": 28070594422560.0, + "grad_norm": 2.907186541860079, + "learning_rate": 3.9977055238133554e-06, + "loss": 0.7827, + "num_input_tokens_seen": 16000410, + "step": 497 + }, + { + "epoch": 0.04491139468819047, + "flos": 26431261503840.0, + "grad_norm": 4.136940061360457, + "learning_rate": 3.99767746252067e-06, + "loss": 0.7028, + "num_input_tokens_seen": 16031125, + "step": 498 + }, + { + "epoch": 0.04500157821166073, + "flos": 26139581698080.0, + "grad_norm": 1.9556654620472746, + "learning_rate": 3.997649230775129e-06, + "loss": 0.8262, + "num_input_tokens_seen": 16059995, + "step": 499 + }, + { + "epoch": 0.04509176173513099, + "flos": 24719919292800.0, + "grad_norm": 2.2343294365778847, + "learning_rate": 3.9976208285791395e-06, + "loss": 0.7614, + "num_input_tokens_seen": 16089250, + "step": 500 + }, + { + "epoch": 0.04518194525860125, + "flos": 70225277100000.0, + "grad_norm": 0.7747091762869349, + "learning_rate": 3.997592255935127e-06, + "loss": 0.6706, + "num_input_tokens_seen": 16178085, + "step": 501 + }, + { + "epoch": 0.045272128782071513, + "flos": 25552542280320.0, + "grad_norm": 1.982557290248797, + "learning_rate": 3.997563512845529e-06, + "loss": 0.8487, + "num_input_tokens_seen": 16209510, + "step": 502 + }, + { + "epoch": 0.045362312305541774, + "flos": 22315177779840.0, + "grad_norm": 1.8946492895733915, + "learning_rate": 3.9975345993127975e-06, + "loss": 0.7804, + "num_input_tokens_seen": 16237985, + "step": 503 + }, + { + "epoch": 0.04545249582901204, + "flos": 20529410242080.0, + "grad_norm": 2.24987888010181, + "learning_rate": 3.9975055153393985e-06, + "loss": 0.835, + "num_input_tokens_seen": 16266455, + "step": 504 + }, + { + "epoch": 0.0455426793524823, + "flos": 23731717925280.0, + "grad_norm": 1.84351075770046, + "learning_rate": 3.997476260927816e-06, + "loss": 0.842, + "num_input_tokens_seen": 16293850, + "step": 505 + }, + { + "epoch": 0.045632862875952564, + "flos": 39616447355520.0, + "grad_norm": 1.8385902881046712, + "learning_rate": 3.997446836080545e-06, + "loss": 0.7973, + "num_input_tokens_seen": 16325260, + "step": 506 + }, + { + "epoch": 0.045723046399422825, + "flos": 24317716589760.0, + "grad_norm": 2.381265907534834, + "learning_rate": 3.997417240800095e-06, + "loss": 0.8745, + "num_input_tokens_seen": 16352195, + "step": 507 + }, + { + "epoch": 0.045813229922893085, + "flos": 23001923694720.0, + "grad_norm": 3.579384527951417, + "learning_rate": 3.997387475088994e-06, + "loss": 0.9224, + "num_input_tokens_seen": 16379090, + "step": 508 + }, + { + "epoch": 0.045903413446363346, + "flos": 18124408540800.0, + "grad_norm": 2.5983316791065065, + "learning_rate": 3.99735753894978e-06, + "loss": 0.9206, + "num_input_tokens_seen": 16405720, + "step": 509 + }, + { + "epoch": 0.045993596969833614, + "flos": 26030657100480.0, + "grad_norm": 1.9886189531827463, + "learning_rate": 3.997327432385006e-06, + "loss": 0.8231, + "num_input_tokens_seen": 16436105, + "step": 510 + }, + { + "epoch": 0.046083780493303875, + "flos": 19727383700640.0, + "grad_norm": 1.9542723680660343, + "learning_rate": 3.997297155397244e-06, + "loss": 0.7535, + "num_input_tokens_seen": 16463890, + "step": 511 + }, + { + "epoch": 0.046173964016774136, + "flos": 25083013674720.0, + "grad_norm": 1.8373392800348842, + "learning_rate": 3.997266707989074e-06, + "loss": 0.743, + "num_input_tokens_seen": 16492680, + "step": 512 + }, + { + "epoch": 0.046264147540244396, + "flos": 18998853241920.0, + "grad_norm": 2.4312260408373194, + "learning_rate": 3.997236090163097e-06, + "loss": 0.7451, + "num_input_tokens_seen": 16520870, + "step": 513 + }, + { + "epoch": 0.04635433106371466, + "flos": 20602423117920.0, + "grad_norm": 2.633357322760654, + "learning_rate": 3.9972053019219235e-06, + "loss": 0.9553, + "num_input_tokens_seen": 16549885, + "step": 514 + }, + { + "epoch": 0.04644451458718492, + "flos": 18342852452160.0, + "grad_norm": 2.7009105983543282, + "learning_rate": 3.997174343268181e-06, + "loss": 0.8386, + "num_input_tokens_seen": 16575935, + "step": 515 + }, + { + "epoch": 0.046534698110655186, + "flos": 33571172225280.0, + "grad_norm": 3.1599191923311882, + "learning_rate": 3.9971432142045115e-06, + "loss": 0.8306, + "num_input_tokens_seen": 16607545, + "step": 516 + }, + { + "epoch": 0.04662488163412545, + "flos": 68849480745120.0, + "grad_norm": 0.7989540280686571, + "learning_rate": 3.99711191473357e-06, + "loss": 0.6619, + "num_input_tokens_seen": 16691105, + "step": 517 + }, + { + "epoch": 0.04671506515759571, + "flos": 26573384430720.0, + "grad_norm": 2.0380792695371874, + "learning_rate": 3.99708044485803e-06, + "loss": 0.8386, + "num_input_tokens_seen": 16721380, + "step": 518 + }, + { + "epoch": 0.04680524868106597, + "flos": 22237221325920.0, + "grad_norm": 2.1331283838450354, + "learning_rate": 3.997048804580574e-06, + "loss": 0.8292, + "num_input_tokens_seen": 16749795, + "step": 519 + }, + { + "epoch": 0.04689543220453623, + "flos": 38957993361600.0, + "grad_norm": 2.187312417843124, + "learning_rate": 3.997016993903901e-06, + "loss": 0.7719, + "num_input_tokens_seen": 16780335, + "step": 520 + }, + { + "epoch": 0.04698561572800649, + "flos": 22929542704800.0, + "grad_norm": 2.073992068445679, + "learning_rate": 3.996985012830728e-06, + "loss": 0.9012, + "num_input_tokens_seen": 16808390, + "step": 521 + }, + { + "epoch": 0.04707579925147676, + "flos": 37688185193280.0, + "grad_norm": 1.605576857293893, + "learning_rate": 3.996952861363782e-06, + "loss": 0.7671, + "num_input_tokens_seen": 16843245, + "step": 522 + }, + { + "epoch": 0.04716598277494702, + "flos": 23953953152160.0, + "grad_norm": 1.9795300066534003, + "learning_rate": 3.9969205395058064e-06, + "loss": 0.827, + "num_input_tokens_seen": 16873525, + "step": 523 + }, + { + "epoch": 0.04725616629841728, + "flos": 22423879358880.0, + "grad_norm": 2.627313865652736, + "learning_rate": 3.99688804725956e-06, + "loss": 0.8192, + "num_input_tokens_seen": 16902335, + "step": 524 + }, + { + "epoch": 0.04734634982188754, + "flos": 57547990513440.0, + "grad_norm": 1.3947278052542031, + "learning_rate": 3.996855384627815e-06, + "loss": 0.7766, + "num_input_tokens_seen": 16941655, + "step": 525 + }, + { + "epoch": 0.0474365333453578, + "flos": 16623444403200.0, + "grad_norm": 2.2233056592958564, + "learning_rate": 3.996822551613357e-06, + "loss": 0.8581, + "num_input_tokens_seen": 16966820, + "step": 526 + }, + { + "epoch": 0.04752671686882806, + "flos": 34985630864160.0, + "grad_norm": 2.1838588383087356, + "learning_rate": 3.996789548218989e-06, + "loss": 0.7541, + "num_input_tokens_seen": 16997650, + "step": 527 + }, + { + "epoch": 0.04761690039229833, + "flos": 26686509211200.0, + "grad_norm": 1.7718391105340698, + "learning_rate": 3.996756374447526e-06, + "loss": 0.7095, + "num_input_tokens_seen": 17029415, + "step": 528 + }, + { + "epoch": 0.04770708391576859, + "flos": 33461206874400.0, + "grad_norm": 2.516809693960593, + "learning_rate": 3.9967230303018005e-06, + "loss": 0.8214, + "num_input_tokens_seen": 17056565, + "step": 529 + }, + { + "epoch": 0.04779726743923885, + "flos": 23328771827040.0, + "grad_norm": 2.0506348766587834, + "learning_rate": 3.996689515784655e-06, + "loss": 0.805, + "num_input_tokens_seen": 17083450, + "step": 530 + }, + { + "epoch": 0.04788745096270911, + "flos": 23400446591520.0, + "grad_norm": 2.7577838842187914, + "learning_rate": 3.996655830898951e-06, + "loss": 0.7949, + "num_input_tokens_seen": 17111230, + "step": 531 + }, + { + "epoch": 0.04797763448617937, + "flos": 16481358646080.0, + "grad_norm": 2.9191078505787598, + "learning_rate": 3.996621975647562e-06, + "loss": 0.8218, + "num_input_tokens_seen": 17137020, + "step": 532 + }, + { + "epoch": 0.04806781800964964, + "flos": 39400382308800.0, + "grad_norm": 2.100467893155766, + "learning_rate": 3.996587950033377e-06, + "loss": 0.7552, + "num_input_tokens_seen": 17171870, + "step": 533 + }, + { + "epoch": 0.0481580015331199, + "flos": 19035136661280.0, + "grad_norm": 2.7081358226257897, + "learning_rate": 3.996553754059299e-06, + "loss": 0.7464, + "num_input_tokens_seen": 17199960, + "step": 534 + }, + { + "epoch": 0.04824818505659016, + "flos": 21328240184160.0, + "grad_norm": 2.9546196601458883, + "learning_rate": 3.996519387728245e-06, + "loss": 0.8214, + "num_input_tokens_seen": 17229140, + "step": 535 + }, + { + "epoch": 0.04833836858006042, + "flos": 26758630012800.0, + "grad_norm": 1.3859565466912758, + "learning_rate": 3.9964848510431495e-06, + "loss": 0.8042, + "num_input_tokens_seen": 17259620, + "step": 536 + }, + { + "epoch": 0.048428552103530684, + "flos": 31308070432320.0, + "grad_norm": 1.982175494137675, + "learning_rate": 3.996450144006957e-06, + "loss": 0.8786, + "num_input_tokens_seen": 17289170, + "step": 537 + }, + { + "epoch": 0.048518735627000945, + "flos": 23987225820960.0, + "grad_norm": 2.128178631391025, + "learning_rate": 3.99641526662263e-06, + "loss": 0.7991, + "num_input_tokens_seen": 17319385, + "step": 538 + }, + { + "epoch": 0.04860891915047121, + "flos": 20163267939840.0, + "grad_norm": 1.9388118732547666, + "learning_rate": 3.996380218893145e-06, + "loss": 0.9033, + "num_input_tokens_seen": 17348555, + "step": 539 + }, + { + "epoch": 0.04869910267394147, + "flos": 28107100860480.0, + "grad_norm": 2.2093422462377217, + "learning_rate": 3.996345000821491e-06, + "loss": 0.8454, + "num_input_tokens_seen": 17378850, + "step": 540 + }, + { + "epoch": 0.048789286197411734, + "flos": 26431596031680.0, + "grad_norm": 2.055515021133563, + "learning_rate": 3.996309612410674e-06, + "loss": 0.9291, + "num_input_tokens_seen": 17408205, + "step": 541 + }, + { + "epoch": 0.048879469720881995, + "flos": 25156732776000.0, + "grad_norm": 1.9349810931048441, + "learning_rate": 3.996274053663713e-06, + "loss": 0.7561, + "num_input_tokens_seen": 17436690, + "step": 542 + }, + { + "epoch": 0.048969653244352256, + "flos": 69534591190560.0, + "grad_norm": 0.7895070325140813, + "learning_rate": 3.996238324583643e-06, + "loss": 0.618, + "num_input_tokens_seen": 17527420, + "step": 543 + }, + { + "epoch": 0.049059836767822516, + "flos": 21038753394240.0, + "grad_norm": 2.4513809905196604, + "learning_rate": 3.996202425173512e-06, + "loss": 0.8856, + "num_input_tokens_seen": 17553550, + "step": 544 + }, + { + "epoch": 0.049150020291292784, + "flos": 22095804624480.0, + "grad_norm": 2.0674652674201144, + "learning_rate": 3.996166355436383e-06, + "loss": 0.7768, + "num_input_tokens_seen": 17582665, + "step": 545 + }, + { + "epoch": 0.049240203814763045, + "flos": 21512519352480.0, + "grad_norm": 1.8750185011904599, + "learning_rate": 3.996130115375333e-06, + "loss": 0.8301, + "num_input_tokens_seen": 17611975, + "step": 546 + }, + { + "epoch": 0.049330387338233306, + "flos": 70656663798240.0, + "grad_norm": 0.8792182877076518, + "learning_rate": 3.996093704993456e-06, + "loss": 0.6335, + "num_input_tokens_seen": 17698215, + "step": 547 + }, + { + "epoch": 0.04942057086170357, + "flos": 19035545528640.0, + "grad_norm": 2.960934440188475, + "learning_rate": 3.996057124293857e-06, + "loss": 0.9031, + "num_input_tokens_seen": 17723965, + "step": 548 + }, + { + "epoch": 0.04951075438517383, + "flos": 26905361989920.0, + "grad_norm": 1.3661720646637594, + "learning_rate": 3.996020373279659e-06, + "loss": 0.8713, + "num_input_tokens_seen": 17755430, + "step": 549 + }, + { + "epoch": 0.04960093790864409, + "flos": 16848950568960.0, + "grad_norm": 3.685465364510014, + "learning_rate": 3.995983451953996e-06, + "loss": 0.8303, + "num_input_tokens_seen": 17782940, + "step": 550 + }, + { + "epoch": 0.049691121432114356, + "flos": 24972193419360.0, + "grad_norm": 1.7847763091260356, + "learning_rate": 3.99594636032002e-06, + "loss": 0.8424, + "num_input_tokens_seen": 17808990, + "step": 551 + }, + { + "epoch": 0.04978130495558462, + "flos": 29124374713920.0, + "grad_norm": 3.834275516876539, + "learning_rate": 3.995909098380894e-06, + "loss": 0.7069, + "num_input_tokens_seen": 17839470, + "step": 552 + }, + { + "epoch": 0.04987148847905488, + "flos": 29018944073760.0, + "grad_norm": 1.536908031107113, + "learning_rate": 3.995871666139799e-06, + "loss": 0.7056, + "num_input_tokens_seen": 17872720, + "step": 553 + }, + { + "epoch": 0.04996167200252514, + "flos": 23298732927360.0, + "grad_norm": 2.008812918588595, + "learning_rate": 3.995834063599928e-06, + "loss": 0.7988, + "num_input_tokens_seen": 17901855, + "step": 554 + }, + { + "epoch": 0.0500518555259954, + "flos": 25406553698400.0, + "grad_norm": 2.2844965577561944, + "learning_rate": 3.99579629076449e-06, + "loss": 0.8498, + "num_input_tokens_seen": 17931865, + "step": 555 + }, + { + "epoch": 0.05014203904946566, + "flos": 25776041279040.0, + "grad_norm": 2.0755695894099526, + "learning_rate": 3.9957583476367084e-06, + "loss": 0.7943, + "num_input_tokens_seen": 17960645, + "step": 556 + }, + { + "epoch": 0.05023222257293593, + "flos": 21658136236800.0, + "grad_norm": 2.5433604324050365, + "learning_rate": 3.995720234219819e-06, + "loss": 0.8184, + "num_input_tokens_seen": 17989625, + "step": 557 + }, + { + "epoch": 0.05032240609640619, + "flos": 18270397122720.0, + "grad_norm": 2.1070780363111226, + "learning_rate": 3.995681950517075e-06, + "loss": 0.7268, + "num_input_tokens_seen": 18018595, + "step": 558 + }, + { + "epoch": 0.05041258961987645, + "flos": 20602200099360.0, + "grad_norm": 2.241684169299162, + "learning_rate": 3.995643496531743e-06, + "loss": 0.8367, + "num_input_tokens_seen": 18045025, + "step": 559 + }, + { + "epoch": 0.05050277314334671, + "flos": 18926100554400.0, + "grad_norm": 2.3464603265984567, + "learning_rate": 3.9956048722671044e-06, + "loss": 0.8306, + "num_input_tokens_seen": 18070625, + "step": 560 + }, + { + "epoch": 0.05059295666681697, + "flos": 26358806174400.0, + "grad_norm": 1.7672412192951767, + "learning_rate": 3.995566077726454e-06, + "loss": 0.7993, + "num_input_tokens_seen": 18101340, + "step": 561 + }, + { + "epoch": 0.05068314019028723, + "flos": 22715187467040.0, + "grad_norm": 2.3031928954345795, + "learning_rate": 3.995527112913103e-06, + "loss": 0.7816, + "num_input_tokens_seen": 18130560, + "step": 562 + }, + { + "epoch": 0.0507733237137575, + "flos": 23917223695680.0, + "grad_norm": 2.5878310171496, + "learning_rate": 3.995487977830375e-06, + "loss": 0.7713, + "num_input_tokens_seen": 18160205, + "step": 563 + }, + { + "epoch": 0.05086350723722776, + "flos": 20232043463040.0, + "grad_norm": 2.476648082517961, + "learning_rate": 3.9954486724816105e-06, + "loss": 0.9139, + "num_input_tokens_seen": 18185020, + "step": 564 + }, + { + "epoch": 0.05095369076069802, + "flos": 21111506081760.0, + "grad_norm": 2.157704070375769, + "learning_rate": 3.995409196870161e-06, + "loss": 0.8596, + "num_input_tokens_seen": 18212095, + "step": 565 + }, + { + "epoch": 0.05104387428416828, + "flos": 64237886753280.0, + "grad_norm": 0.7778164753672437, + "learning_rate": 3.995369550999398e-06, + "loss": 0.5782, + "num_input_tokens_seen": 18296240, + "step": 566 + }, + { + "epoch": 0.05113405780763854, + "flos": 16776978446400.0, + "grad_norm": 3.626067826416779, + "learning_rate": 3.995329734872702e-06, + "loss": 0.8439, + "num_input_tokens_seen": 18322715, + "step": 567 + }, + { + "epoch": 0.051224241331108804, + "flos": 20523611759520.0, + "grad_norm": 1.8584949750978583, + "learning_rate": 3.9952897484934706e-06, + "loss": 0.9338, + "num_input_tokens_seen": 18350315, + "step": 568 + }, + { + "epoch": 0.05131442485457907, + "flos": 18962606992320.0, + "grad_norm": 2.118743887473322, + "learning_rate": 3.995249591865115e-06, + "loss": 0.8233, + "num_input_tokens_seen": 18378870, + "step": 569 + }, + { + "epoch": 0.05140460837804933, + "flos": 26905436329440.0, + "grad_norm": 1.785632047121898, + "learning_rate": 3.995209264991063e-06, + "loss": 0.7795, + "num_input_tokens_seen": 18408970, + "step": 570 + }, + { + "epoch": 0.05149479190151959, + "flos": 21583785249600.0, + "grad_norm": 2.5506844144377534, + "learning_rate": 3.995168767874756e-06, + "loss": 0.8674, + "num_input_tokens_seen": 18437680, + "step": 571 + }, + { + "epoch": 0.051584975424989854, + "flos": 43518547539360.0, + "grad_norm": 1.76689622065428, + "learning_rate": 3.995128100519648e-06, + "loss": 0.8025, + "num_input_tokens_seen": 18472300, + "step": 572 + }, + { + "epoch": 0.051675158948460115, + "flos": 25258372100640.0, + "grad_norm": 2.2002291213281246, + "learning_rate": 3.995087262929209e-06, + "loss": 0.8339, + "num_input_tokens_seen": 18500300, + "step": 573 + }, + { + "epoch": 0.051765342471930376, + "flos": 22825747534080.0, + "grad_norm": 2.426529183929723, + "learning_rate": 3.995046255106925e-06, + "loss": 0.8285, + "num_input_tokens_seen": 18530095, + "step": 574 + }, + { + "epoch": 0.05185552599540064, + "flos": 20344870885440.0, + "grad_norm": 2.1782045412796114, + "learning_rate": 3.995005077056293e-06, + "loss": 0.8844, + "num_input_tokens_seen": 18558290, + "step": 575 + }, + { + "epoch": 0.051945709518870904, + "flos": 18416497213920.0, + "grad_norm": 2.141133825710301, + "learning_rate": 3.9949637287808284e-06, + "loss": 0.7096, + "num_input_tokens_seen": 18586870, + "step": 576 + }, + { + "epoch": 0.052035893042341165, + "flos": 25265545864320.0, + "grad_norm": 1.9189504686436771, + "learning_rate": 3.994922210284057e-06, + "loss": 0.7695, + "num_input_tokens_seen": 18616450, + "step": 577 + }, + { + "epoch": 0.052126076565811426, + "flos": 17609675773440.0, + "grad_norm": 2.251472521115055, + "learning_rate": 3.994880521569524e-06, + "loss": 0.8527, + "num_input_tokens_seen": 18642510, + "step": 578 + }, + { + "epoch": 0.05221626008928169, + "flos": 25229411124000.0, + "grad_norm": 21.78760933774738, + "learning_rate": 3.994838662640785e-06, + "loss": 0.8298, + "num_input_tokens_seen": 18672235, + "step": 579 + }, + { + "epoch": 0.05230644361275195, + "flos": 17323534261920.0, + "grad_norm": 2.7353508652043024, + "learning_rate": 3.9947966335014116e-06, + "loss": 0.8564, + "num_input_tokens_seen": 18698720, + "step": 580 + }, + { + "epoch": 0.052396627136222215, + "flos": 13970740455840.0, + "grad_norm": 2.6888628320858894, + "learning_rate": 3.99475443415499e-06, + "loss": 0.7958, + "num_input_tokens_seen": 18724255, + "step": 581 + }, + { + "epoch": 0.052486810659692476, + "flos": 21002507144640.0, + "grad_norm": 2.1700406458339745, + "learning_rate": 3.994712064605121e-06, + "loss": 0.814, + "num_input_tokens_seen": 18752580, + "step": 582 + }, + { + "epoch": 0.05257699418316274, + "flos": 19472433351360.0, + "grad_norm": 2.134501029855422, + "learning_rate": 3.99466952485542e-06, + "loss": 0.9693, + "num_input_tokens_seen": 18778215, + "step": 583 + }, + { + "epoch": 0.052667177706633, + "flos": 23589223300800.0, + "grad_norm": 2.2352164224727207, + "learning_rate": 3.994626814909518e-06, + "loss": 0.8328, + "num_input_tokens_seen": 18806600, + "step": 584 + }, + { + "epoch": 0.05275736123010326, + "flos": 20201521356480.0, + "grad_norm": 2.8589188675738297, + "learning_rate": 3.994583934771056e-06, + "loss": 0.7992, + "num_input_tokens_seen": 18832980, + "step": 585 + }, + { + "epoch": 0.05284754475357352, + "flos": 23729562079200.0, + "grad_norm": 2.0170168543703864, + "learning_rate": 3.9945408844436955e-06, + "loss": 0.8793, + "num_input_tokens_seen": 18862840, + "step": 586 + }, + { + "epoch": 0.05293772827704379, + "flos": 24464076869280.0, + "grad_norm": 2.256141862893215, + "learning_rate": 3.994497663931109e-06, + "loss": 0.8453, + "num_input_tokens_seen": 18892360, + "step": 587 + }, + { + "epoch": 0.05302791180051405, + "flos": 22452691656480.0, + "grad_norm": 1.9657459045972956, + "learning_rate": 3.994454273236984e-06, + "loss": 0.8394, + "num_input_tokens_seen": 18919000, + "step": 588 + }, + { + "epoch": 0.05311809532398431, + "flos": 23443792265280.0, + "grad_norm": 1.674629438034046, + "learning_rate": 3.994410712365023e-06, + "loss": 0.8244, + "num_input_tokens_seen": 18948680, + "step": 589 + }, + { + "epoch": 0.05320827884745457, + "flos": 19508530921920.0, + "grad_norm": 2.392727394897192, + "learning_rate": 3.994366981318943e-06, + "loss": 0.9249, + "num_input_tokens_seen": 18974560, + "step": 590 + }, + { + "epoch": 0.05329846237092483, + "flos": 21796467848160.0, + "grad_norm": 2.3341078721576705, + "learning_rate": 3.9943230801024765e-06, + "loss": 0.8299, + "num_input_tokens_seen": 19002050, + "step": 591 + }, + { + "epoch": 0.05338864589439509, + "flos": 19756270337760.0, + "grad_norm": 2.1716792615328635, + "learning_rate": 3.9942790087193666e-06, + "loss": 0.8652, + "num_input_tokens_seen": 19025650, + "step": 592 + }, + { + "epoch": 0.05347882941786536, + "flos": 18124371371040.0, + "grad_norm": 2.0398845550184292, + "learning_rate": 3.994234767173376e-06, + "loss": 0.8125, + "num_input_tokens_seen": 19053555, + "step": 593 + }, + { + "epoch": 0.05356901294133562, + "flos": 19545074529600.0, + "grad_norm": 2.021438376967609, + "learning_rate": 3.994190355468279e-06, + "loss": 0.7105, + "num_input_tokens_seen": 19082490, + "step": 594 + }, + { + "epoch": 0.05365919646480588, + "flos": 33425295152640.0, + "grad_norm": 1.8277217061499915, + "learning_rate": 3.994145773607865e-06, + "loss": 0.821, + "num_input_tokens_seen": 19114300, + "step": 595 + }, + { + "epoch": 0.05374937998827614, + "flos": 21583041854400.0, + "grad_norm": 1.7956079410920394, + "learning_rate": 3.994101021595938e-06, + "loss": 0.7914, + "num_input_tokens_seen": 19143660, + "step": 596 + }, + { + "epoch": 0.0538395635117464, + "flos": 20529930618720.0, + "grad_norm": 2.368043851652585, + "learning_rate": 3.9940560994363165e-06, + "loss": 0.8384, + "num_input_tokens_seen": 19171945, + "step": 597 + }, + { + "epoch": 0.05392974703521666, + "flos": 23407322997120.0, + "grad_norm": 2.2885053041116206, + "learning_rate": 3.994011007132833e-06, + "loss": 0.6978, + "num_input_tokens_seen": 19200735, + "step": 598 + }, + { + "epoch": 0.05401993055868693, + "flos": 62850270414720.0, + "grad_norm": 0.7054222725294166, + "learning_rate": 3.993965744689337e-06, + "loss": 0.6482, + "num_input_tokens_seen": 19277815, + "step": 599 + }, + { + "epoch": 0.05411011408215719, + "flos": 39292424124960.0, + "grad_norm": 2.0145294997535275, + "learning_rate": 3.993920312109687e-06, + "loss": 0.8042, + "num_input_tokens_seen": 19307315, + "step": 600 + }, + { + "epoch": 0.05420029760562745, + "flos": 25917569489760.0, + "grad_norm": 1.9076729763290377, + "learning_rate": 3.993874709397764e-06, + "loss": 0.8165, + "num_input_tokens_seen": 19334050, + "step": 601 + }, + { + "epoch": 0.05429048112909771, + "flos": 23079396941760.0, + "grad_norm": 2.1596937460485437, + "learning_rate": 3.993828936557454e-06, + "loss": 0.8063, + "num_input_tokens_seen": 19362170, + "step": 602 + }, + { + "epoch": 0.054380664652567974, + "flos": 36886753368000.0, + "grad_norm": 2.135059556331399, + "learning_rate": 3.993782993592667e-06, + "loss": 0.7827, + "num_input_tokens_seen": 19393805, + "step": 603 + }, + { + "epoch": 0.054470848176038235, + "flos": 25119780300960.0, + "grad_norm": 1.989318996531614, + "learning_rate": 3.993736880507321e-06, + "loss": 0.8616, + "num_input_tokens_seen": 19423325, + "step": 604 + }, + { + "epoch": 0.0545610316995085, + "flos": 33712997794080.0, + "grad_norm": 2.4352898970722117, + "learning_rate": 3.99369059730535e-06, + "loss": 0.766, + "num_input_tokens_seen": 19454750, + "step": 605 + }, + { + "epoch": 0.05465121522297876, + "flos": 24680810971680.0, + "grad_norm": 1.509603119033559, + "learning_rate": 3.993644143990706e-06, + "loss": 0.8243, + "num_input_tokens_seen": 19486300, + "step": 606 + }, + { + "epoch": 0.054741398746449024, + "flos": 17429262260160.0, + "grad_norm": 1.9683374245084557, + "learning_rate": 3.99359752056735e-06, + "loss": 0.742, + "num_input_tokens_seen": 19514880, + "step": 607 + }, + { + "epoch": 0.054831582269919285, + "flos": 25775929769760.0, + "grad_norm": 2.0567793840076187, + "learning_rate": 3.993550727039261e-06, + "loss": 0.6757, + "num_input_tokens_seen": 19542830, + "step": 608 + }, + { + "epoch": 0.054921765793389546, + "flos": 22496371858080.0, + "grad_norm": 2.2501044778938573, + "learning_rate": 3.993503763410431e-06, + "loss": 0.8405, + "num_input_tokens_seen": 19572060, + "step": 609 + }, + { + "epoch": 0.05501194931685981, + "flos": 29236756099200.0, + "grad_norm": 1.9398283965125007, + "learning_rate": 3.9934566296848686e-06, + "loss": 0.7378, + "num_input_tokens_seen": 19600805, + "step": 610 + }, + { + "epoch": 0.055102132840330074, + "flos": 21216602194080.0, + "grad_norm": 5.100078559719215, + "learning_rate": 3.993409325866595e-06, + "loss": 0.8989, + "num_input_tokens_seen": 19625255, + "step": 611 + }, + { + "epoch": 0.055192316363800335, + "flos": 60503110593600.0, + "grad_norm": 0.6288683388223979, + "learning_rate": 3.993361851959645e-06, + "loss": 0.5966, + "num_input_tokens_seen": 19727255, + "step": 612 + }, + { + "epoch": 0.055282499887270596, + "flos": 24427310243040.0, + "grad_norm": 2.6400764392132503, + "learning_rate": 3.993314207968071e-06, + "loss": 0.8654, + "num_input_tokens_seen": 19759040, + "step": 613 + }, + { + "epoch": 0.05537268341074086, + "flos": 31714027281120.0, + "grad_norm": 1.8575803285996322, + "learning_rate": 3.993266393895938e-06, + "loss": 0.8079, + "num_input_tokens_seen": 19789525, + "step": 614 + }, + { + "epoch": 0.05546286693421112, + "flos": 17503464568320.0, + "grad_norm": 1.8914618413445652, + "learning_rate": 3.993218409747326e-06, + "loss": 0.826, + "num_input_tokens_seen": 19816050, + "step": 615 + }, + { + "epoch": 0.05555305045768138, + "flos": 26831382700320.0, + "grad_norm": 1.9805532584133205, + "learning_rate": 3.993170255526328e-06, + "loss": 0.8125, + "num_input_tokens_seen": 19845155, + "step": 616 + }, + { + "epoch": 0.055643233981151646, + "flos": 22566076625280.0, + "grad_norm": 3.316244779684054, + "learning_rate": 3.993121931237054e-06, + "loss": 0.8473, + "num_input_tokens_seen": 19873765, + "step": 617 + }, + { + "epoch": 0.05573341750462191, + "flos": 21841300312320.0, + "grad_norm": 1.8144068717193997, + "learning_rate": 3.993073436883627e-06, + "loss": 0.8383, + "num_input_tokens_seen": 19902620, + "step": 618 + }, + { + "epoch": 0.05582360102809217, + "flos": 23549334414720.0, + "grad_norm": 2.1564867193784694, + "learning_rate": 3.993024772470184e-06, + "loss": 0.7995, + "num_input_tokens_seen": 19931765, + "step": 619 + }, + { + "epoch": 0.05591378455156243, + "flos": 19181013733920.0, + "grad_norm": 2.0946059114438143, + "learning_rate": 3.992975938000878e-06, + "loss": 0.804, + "num_input_tokens_seen": 19959440, + "step": 620 + }, + { + "epoch": 0.05600396807503269, + "flos": 66215367411360.0, + "grad_norm": 0.6545705194593023, + "learning_rate": 3.992926933479876e-06, + "loss": 0.586, + "num_input_tokens_seen": 20054830, + "step": 621 + }, + { + "epoch": 0.05609415159850295, + "flos": 25666670644320.0, + "grad_norm": 1.9176649797033092, + "learning_rate": 3.9928777589113595e-06, + "loss": 0.8456, + "num_input_tokens_seen": 20082440, + "step": 622 + }, + { + "epoch": 0.05618433512197322, + "flos": 35830073835360.0, + "grad_norm": 2.4262289963942676, + "learning_rate": 3.992828414299524e-06, + "loss": 0.7815, + "num_input_tokens_seen": 20116225, + "step": 623 + }, + { + "epoch": 0.05627451864544348, + "flos": 23043150692160.0, + "grad_norm": 1.8564684293011016, + "learning_rate": 3.992778899648579e-06, + "loss": 0.8188, + "num_input_tokens_seen": 20147370, + "step": 624 + }, + { + "epoch": 0.05636470216891374, + "flos": 20456397366240.0, + "grad_norm": 2.0449939266944743, + "learning_rate": 3.992729214962751e-06, + "loss": 0.8388, + "num_input_tokens_seen": 20174535, + "step": 625 + }, + { + "epoch": 0.056454885692384, + "flos": 25812399037920.0, + "grad_norm": 1.9622505928512028, + "learning_rate": 3.992679360246279e-06, + "loss": 0.857, + "num_input_tokens_seen": 20201480, + "step": 626 + }, + { + "epoch": 0.05654506921585426, + "flos": 23553311579040.0, + "grad_norm": 4.515712644789039, + "learning_rate": 3.992629335503416e-06, + "loss": 0.8506, + "num_input_tokens_seen": 20232140, + "step": 627 + }, + { + "epoch": 0.05663525273932452, + "flos": 36666190780320.0, + "grad_norm": 1.9408208014199784, + "learning_rate": 3.9925791407384304e-06, + "loss": 0.8021, + "num_input_tokens_seen": 20261360, + "step": 628 + }, + { + "epoch": 0.05672543626279479, + "flos": 20711496394560.0, + "grad_norm": 4.641648304946897, + "learning_rate": 3.992528775955606e-06, + "loss": 0.8601, + "num_input_tokens_seen": 20287825, + "step": 629 + }, + { + "epoch": 0.05681561978626505, + "flos": 19727866907520.0, + "grad_norm": 2.618656850537447, + "learning_rate": 3.992478241159239e-06, + "loss": 0.7891, + "num_input_tokens_seen": 20317400, + "step": 630 + }, + { + "epoch": 0.05690580330973531, + "flos": 20893768395840.0, + "grad_norm": 2.214761937697555, + "learning_rate": 3.992427536353643e-06, + "loss": 0.8415, + "num_input_tokens_seen": 20346965, + "step": 631 + }, + { + "epoch": 0.05699598683320557, + "flos": 25593286070880.0, + "grad_norm": 2.796421035494757, + "learning_rate": 3.992376661543143e-06, + "loss": 0.8058, + "num_input_tokens_seen": 20373590, + "step": 632 + }, + { + "epoch": 0.05708617035667583, + "flos": 17686926001920.0, + "grad_norm": 2.2952751764284307, + "learning_rate": 3.992325616732081e-06, + "loss": 0.793, + "num_input_tokens_seen": 20398630, + "step": 633 + }, + { + "epoch": 0.057176353880146094, + "flos": 25516444709760.0, + "grad_norm": 3.4166825146811535, + "learning_rate": 3.992274401924811e-06, + "loss": 0.9222, + "num_input_tokens_seen": 20426720, + "step": 634 + }, + { + "epoch": 0.05726653740361636, + "flos": 21403185887520.0, + "grad_norm": 1.772917616679934, + "learning_rate": 3.992223017125704e-06, + "loss": 0.7406, + "num_input_tokens_seen": 20454095, + "step": 635 + }, + { + "epoch": 0.05735672092708662, + "flos": 19108112367360.0, + "grad_norm": 1.9534549892132662, + "learning_rate": 3.992171462339145e-06, + "loss": 0.848, + "num_input_tokens_seen": 20480590, + "step": 636 + }, + { + "epoch": 0.057446904450556883, + "flos": 21216007477920.0, + "grad_norm": 2.1201956799006396, + "learning_rate": 3.992119737569532e-06, + "loss": 0.8866, + "num_input_tokens_seen": 20508885, + "step": 637 + }, + { + "epoch": 0.057537087974027144, + "flos": 21258312398400.0, + "grad_norm": 2.1916737656665704, + "learning_rate": 3.992067842821277e-06, + "loss": 0.8164, + "num_input_tokens_seen": 20536730, + "step": 638 + }, + { + "epoch": 0.057627271497497405, + "flos": 19909544192640.0, + "grad_norm": 3.3438178795315108, + "learning_rate": 3.99201577809881e-06, + "loss": 0.9493, + "num_input_tokens_seen": 20562815, + "step": 639 + }, + { + "epoch": 0.057717455020967666, + "flos": 18890002983840.0, + "grad_norm": 2.450964200770434, + "learning_rate": 3.991963543406574e-06, + "loss": 0.8386, + "num_input_tokens_seen": 20589290, + "step": 640 + }, + { + "epoch": 0.057807638544437934, + "flos": 21983534748480.0, + "grad_norm": 2.1097471685931857, + "learning_rate": 3.991911138749024e-06, + "loss": 0.8156, + "num_input_tokens_seen": 20618200, + "step": 641 + }, + { + "epoch": 0.057897822067908195, + "flos": 19179378264480.0, + "grad_norm": 2.265219185099528, + "learning_rate": 3.991858564130633e-06, + "loss": 0.8431, + "num_input_tokens_seen": 20646845, + "step": 642 + }, + { + "epoch": 0.057988005591378455, + "flos": 17170706444160.0, + "grad_norm": 2.4122073257882852, + "learning_rate": 3.991805819555885e-06, + "loss": 0.7976, + "num_input_tokens_seen": 20670870, + "step": 643 + }, + { + "epoch": 0.058078189114848716, + "flos": 21364635112800.0, + "grad_norm": 1.9918386714939207, + "learning_rate": 3.991752905029283e-06, + "loss": 0.8216, + "num_input_tokens_seen": 20700685, + "step": 644 + }, + { + "epoch": 0.05816837263831898, + "flos": 24099235508640.0, + "grad_norm": 2.0900196607477195, + "learning_rate": 3.991699820555341e-06, + "loss": 0.6953, + "num_input_tokens_seen": 20732560, + "step": 645 + }, + { + "epoch": 0.05825855616178924, + "flos": 16340127793440.0, + "grad_norm": 2.1916063186788355, + "learning_rate": 3.991646566138588e-06, + "loss": 0.8214, + "num_input_tokens_seen": 20760085, + "step": 646 + }, + { + "epoch": 0.058348739685259506, + "flos": 23224493449440.0, + "grad_norm": 1.7934400151061127, + "learning_rate": 3.991593141783567e-06, + "loss": 0.7415, + "num_input_tokens_seen": 20790995, + "step": 647 + }, + { + "epoch": 0.058438923208729766, + "flos": 19254398307360.0, + "grad_norm": 1.8206369659313464, + "learning_rate": 3.991539547494839e-06, + "loss": 0.8442, + "num_input_tokens_seen": 20818825, + "step": 648 + }, + { + "epoch": 0.05852910673220003, + "flos": 23116052058720.0, + "grad_norm": 2.130872465909891, + "learning_rate": 3.991485783276974e-06, + "loss": 0.8147, + "num_input_tokens_seen": 20847585, + "step": 649 + }, + { + "epoch": 0.05861929025567029, + "flos": 31130184462720.0, + "grad_norm": 2.9598331858050666, + "learning_rate": 3.991431849134563e-06, + "loss": 0.8731, + "num_input_tokens_seen": 20879715, + "step": 650 + }, + { + "epoch": 0.05870947377914055, + "flos": 28289298522240.0, + "grad_norm": 2.72403072652418, + "learning_rate": 3.991377745072205e-06, + "loss": 0.713, + "num_input_tokens_seen": 20909100, + "step": 651 + }, + { + "epoch": 0.05879965730261081, + "flos": 19582064174400.0, + "grad_norm": 2.418453432970395, + "learning_rate": 3.991323471094517e-06, + "loss": 0.8471, + "num_input_tokens_seen": 20937330, + "step": 652 + }, + { + "epoch": 0.05888984082608108, + "flos": 29888705385120.0, + "grad_norm": 1.9041441870878206, + "learning_rate": 3.991269027206131e-06, + "loss": 0.7918, + "num_input_tokens_seen": 20965385, + "step": 653 + }, + { + "epoch": 0.05898002434955134, + "flos": 63859844394240.0, + "grad_norm": 0.7935056700837415, + "learning_rate": 3.9912144134116916e-06, + "loss": 0.6377, + "num_input_tokens_seen": 21062515, + "step": 654 + }, + { + "epoch": 0.0590702078730216, + "flos": 23079842978880.0, + "grad_norm": 2.2320897079679622, + "learning_rate": 3.99115962971586e-06, + "loss": 0.7426, + "num_input_tokens_seen": 21090305, + "step": 655 + }, + { + "epoch": 0.05916039139649186, + "flos": 25805076595200.0, + "grad_norm": 2.2703809608381644, + "learning_rate": 3.991104676123308e-06, + "loss": 0.8526, + "num_input_tokens_seen": 21115670, + "step": 656 + }, + { + "epoch": 0.05925057491996212, + "flos": 70386951320640.0, + "grad_norm": 0.6586139200726209, + "learning_rate": 3.991049552638727e-06, + "loss": 0.5725, + "num_input_tokens_seen": 21212170, + "step": 657 + }, + { + "epoch": 0.05934075844343238, + "flos": 17432124331680.0, + "grad_norm": 2.1939318241943266, + "learning_rate": 3.99099425926682e-06, + "loss": 0.7815, + "num_input_tokens_seen": 21237675, + "step": 658 + }, + { + "epoch": 0.05943094196690265, + "flos": 26722941309600.0, + "grad_norm": 2.2795238496269494, + "learning_rate": 3.990938796012304e-06, + "loss": 0.787, + "num_input_tokens_seen": 21266395, + "step": 659 + }, + { + "epoch": 0.05952112549037291, + "flos": 14403279416640.0, + "grad_norm": 2.0054509267992087, + "learning_rate": 3.990883162879912e-06, + "loss": 0.9717, + "num_input_tokens_seen": 21290875, + "step": 660 + }, + { + "epoch": 0.05961130901384317, + "flos": 25120263507840.0, + "grad_norm": 2.6634931655145877, + "learning_rate": 3.990827359874391e-06, + "loss": 0.8836, + "num_input_tokens_seen": 21318575, + "step": 661 + }, + { + "epoch": 0.05970149253731343, + "flos": 23803838726880.0, + "grad_norm": 1.51025674784529, + "learning_rate": 3.990771387000503e-06, + "loss": 0.8201, + "num_input_tokens_seen": 21349215, + "step": 662 + }, + { + "epoch": 0.05979167606078369, + "flos": 28289038333920.0, + "grad_norm": 2.9662018129571748, + "learning_rate": 3.990715244263023e-06, + "loss": 0.849, + "num_input_tokens_seen": 21380170, + "step": 663 + }, + { + "epoch": 0.05988185958425396, + "flos": 71870668689600.0, + "grad_norm": 0.7100865647776274, + "learning_rate": 3.990658931666741e-06, + "loss": 0.6051, + "num_input_tokens_seen": 21473305, + "step": 664 + }, + { + "epoch": 0.05997204310772422, + "flos": 23880977446080.0, + "grad_norm": 1.8618913820998777, + "learning_rate": 3.990602449216463e-06, + "loss": 0.7371, + "num_input_tokens_seen": 21501045, + "step": 665 + }, + { + "epoch": 0.06006222663119448, + "flos": 29813350814400.0, + "grad_norm": 2.2833912872236435, + "learning_rate": 3.990545796917008e-06, + "loss": 0.7226, + "num_input_tokens_seen": 21528980, + "step": 666 + }, + { + "epoch": 0.06015241015466474, + "flos": 19138299946080.0, + "grad_norm": 1.8006149513099294, + "learning_rate": 3.99048897477321e-06, + "loss": 0.8635, + "num_input_tokens_seen": 21556390, + "step": 667 + }, + { + "epoch": 0.060242593678135004, + "flos": 20925145406880.0, + "grad_norm": 2.081459296072949, + "learning_rate": 3.990431982789917e-06, + "loss": 0.8378, + "num_input_tokens_seen": 21583160, + "step": 668 + }, + { + "epoch": 0.060332777201605264, + "flos": 24427235903520.0, + "grad_norm": 1.8677862635826998, + "learning_rate": 3.990374820971992e-06, + "loss": 0.753, + "num_input_tokens_seen": 21610510, + "step": 669 + }, + { + "epoch": 0.06042296072507553, + "flos": 49529620756800.0, + "grad_norm": 2.280014957914309, + "learning_rate": 3.990317489324312e-06, + "loss": 0.704, + "num_input_tokens_seen": 21643140, + "step": 670 + }, + { + "epoch": 0.06051314424854579, + "flos": 45266693546400.0, + "grad_norm": 2.530246017466489, + "learning_rate": 3.99025998785177e-06, + "loss": 0.8007, + "num_input_tokens_seen": 21674615, + "step": 671 + }, + { + "epoch": 0.060603327772016054, + "flos": 19910436266880.0, + "grad_norm": 1.7677985122448259, + "learning_rate": 3.990202316559271e-06, + "loss": 0.7587, + "num_input_tokens_seen": 21701840, + "step": 672 + }, + { + "epoch": 0.060693511295486315, + "flos": 25520496213600.0, + "grad_norm": 2.2189996068540676, + "learning_rate": 3.990144475451738e-06, + "loss": 0.8667, + "num_input_tokens_seen": 21730880, + "step": 673 + }, + { + "epoch": 0.060783694818956575, + "flos": 21472816315200.0, + "grad_norm": 1.663289925054071, + "learning_rate": 3.9900864645341036e-06, + "loss": 0.8169, + "num_input_tokens_seen": 21759050, + "step": 674 + }, + { + "epoch": 0.060873878342426836, + "flos": 22205472617280.0, + "grad_norm": 1.7287341653204849, + "learning_rate": 3.990028283811319e-06, + "loss": 0.8233, + "num_input_tokens_seen": 21786480, + "step": 675 + }, + { + "epoch": 0.060964061865897104, + "flos": 26975996001120.0, + "grad_norm": 1.7104433643830015, + "learning_rate": 3.989969933288348e-06, + "loss": 0.8129, + "num_input_tokens_seen": 21817470, + "step": 676 + }, + { + "epoch": 0.061054245389367365, + "flos": 61331756416800.0, + "grad_norm": 0.6583589318485473, + "learning_rate": 3.98991141297017e-06, + "loss": 0.5502, + "num_input_tokens_seen": 21906585, + "step": 677 + }, + { + "epoch": 0.061144428912837626, + "flos": 23917186525920.0, + "grad_norm": 2.396016343355561, + "learning_rate": 3.989852722861778e-06, + "loss": 0.7651, + "num_input_tokens_seen": 21936300, + "step": 678 + }, + { + "epoch": 0.061234612436307886, + "flos": 49768136338560.0, + "grad_norm": 0.7164350214747846, + "learning_rate": 3.98979386296818e-06, + "loss": 0.578, + "num_input_tokens_seen": 22018330, + "step": 679 + }, + { + "epoch": 0.06132479595977815, + "flos": 40561600407360.0, + "grad_norm": 1.8661101996252487, + "learning_rate": 3.989734833294398e-06, + "loss": 0.7964, + "num_input_tokens_seen": 22050905, + "step": 680 + }, + { + "epoch": 0.06141497948324841, + "flos": 20018840487840.0, + "grad_norm": 1.8126150292036223, + "learning_rate": 3.989675633845469e-06, + "loss": 0.8832, + "num_input_tokens_seen": 22077845, + "step": 681 + }, + { + "epoch": 0.061505163006718676, + "flos": 16704188589120.0, + "grad_norm": 4.136264332999801, + "learning_rate": 3.989616264626443e-06, + "loss": 0.7959, + "num_input_tokens_seen": 22105475, + "step": 682 + }, + { + "epoch": 0.06159534653018894, + "flos": 24935686981440.0, + "grad_norm": 1.9920464559254907, + "learning_rate": 3.989556725642388e-06, + "loss": 0.8522, + "num_input_tokens_seen": 22137655, + "step": 683 + }, + { + "epoch": 0.0616855300536592, + "flos": 25185024696960.0, + "grad_norm": 2.672108868631039, + "learning_rate": 3.989497016898382e-06, + "loss": 0.8697, + "num_input_tokens_seen": 22167180, + "step": 684 + }, + { + "epoch": 0.06177571357712946, + "flos": 27374964935040.0, + "grad_norm": 1.7349025164654595, + "learning_rate": 3.98943713839952e-06, + "loss": 0.8118, + "num_input_tokens_seen": 22198775, + "step": 685 + }, + { + "epoch": 0.06186589710059972, + "flos": 71348613479520.0, + "grad_norm": 0.6731368347374361, + "learning_rate": 3.9893770901509125e-06, + "loss": 0.6063, + "num_input_tokens_seen": 22300785, + "step": 686 + }, + { + "epoch": 0.06195608062406998, + "flos": 26654165786400.0, + "grad_norm": 1.8004992878707753, + "learning_rate": 3.989316872157682e-06, + "loss": 0.8085, + "num_input_tokens_seen": 22333970, + "step": 687 + }, + { + "epoch": 0.06204626414754025, + "flos": 15865841458560.0, + "grad_norm": 1.8845608140419638, + "learning_rate": 3.989256484424968e-06, + "loss": 0.8934, + "num_input_tokens_seen": 22359480, + "step": 688 + }, + { + "epoch": 0.06213644767101051, + "flos": 23115345833280.0, + "grad_norm": 2.7315559023856104, + "learning_rate": 3.98919592695792e-06, + "loss": 0.7821, + "num_input_tokens_seen": 22387800, + "step": 689 + }, + { + "epoch": 0.06222663119448077, + "flos": 24464076869280.0, + "grad_norm": 1.9309336233401089, + "learning_rate": 3.9891351997617096e-06, + "loss": 0.8452, + "num_input_tokens_seen": 22415570, + "step": 690 + }, + { + "epoch": 0.06231681471795103, + "flos": 30908395272960.0, + "grad_norm": 2.4301080370461534, + "learning_rate": 3.989074302841514e-06, + "loss": 0.757, + "num_input_tokens_seen": 22447750, + "step": 691 + }, + { + "epoch": 0.06240699824142129, + "flos": 17319408418560.0, + "grad_norm": 2.1903826657068404, + "learning_rate": 3.989013236202533e-06, + "loss": 0.9272, + "num_input_tokens_seen": 22473460, + "step": 692 + }, + { + "epoch": 0.06249718176489155, + "flos": 24317456401440.0, + "grad_norm": 1.5309549942776843, + "learning_rate": 3.988951999849974e-06, + "loss": 0.8541, + "num_input_tokens_seen": 22502300, + "step": 693 + }, + { + "epoch": 0.06258736528836181, + "flos": 26905064631840.0, + "grad_norm": 2.2408583685782872, + "learning_rate": 3.988890593789064e-06, + "loss": 0.7328, + "num_input_tokens_seen": 22529620, + "step": 694 + }, + { + "epoch": 0.06267754881183207, + "flos": 24390692295840.0, + "grad_norm": 2.1510548687422744, + "learning_rate": 3.9888290180250415e-06, + "loss": 0.8059, + "num_input_tokens_seen": 22561640, + "step": 695 + }, + { + "epoch": 0.06276773233530233, + "flos": 30293732989920.0, + "grad_norm": 3.8411819439068866, + "learning_rate": 3.988767272563161e-06, + "loss": 0.7833, + "num_input_tokens_seen": 22591225, + "step": 696 + }, + { + "epoch": 0.06285791585877261, + "flos": 17650010696640.0, + "grad_norm": 2.0516952948144302, + "learning_rate": 3.988705357408691e-06, + "loss": 0.7592, + "num_input_tokens_seen": 22614965, + "step": 697 + }, + { + "epoch": 0.06294809938224287, + "flos": 21876542978400.0, + "grad_norm": 2.5976850694477456, + "learning_rate": 3.9886432725669146e-06, + "loss": 0.7787, + "num_input_tokens_seen": 22644270, + "step": 698 + }, + { + "epoch": 0.06303828290571313, + "flos": 21368203409760.0, + "grad_norm": 2.0659437617488488, + "learning_rate": 3.988581018043128e-06, + "loss": 0.7753, + "num_input_tokens_seen": 22670560, + "step": 699 + }, + { + "epoch": 0.06312846642918339, + "flos": 27196075381920.0, + "grad_norm": 2.4720287521134607, + "learning_rate": 3.988518593842645e-06, + "loss": 0.8243, + "num_input_tokens_seen": 22699805, + "step": 700 + }, + { + "epoch": 0.06321864995265365, + "flos": 19909692871680.0, + "grad_norm": 1.8545457331041817, + "learning_rate": 3.9884559999707906e-06, + "loss": 0.8129, + "num_input_tokens_seen": 22725890, + "step": 701 + }, + { + "epoch": 0.06330883347612391, + "flos": 22889430800160.0, + "grad_norm": 4.299205688922555, + "learning_rate": 3.988393236432906e-06, + "loss": 0.8291, + "num_input_tokens_seen": 22755285, + "step": 702 + }, + { + "epoch": 0.06339901699959417, + "flos": 23406876960000.0, + "grad_norm": 2.0030999002059477, + "learning_rate": 3.988330303234347e-06, + "loss": 0.8911, + "num_input_tokens_seen": 22782965, + "step": 703 + }, + { + "epoch": 0.06348920052306443, + "flos": 39728865910560.0, + "grad_norm": 1.7094089874291964, + "learning_rate": 3.988267200380483e-06, + "loss": 0.7687, + "num_input_tokens_seen": 22818335, + "step": 704 + }, + { + "epoch": 0.0635793840465347, + "flos": 24354817743840.0, + "grad_norm": 1.8579525034185094, + "learning_rate": 3.988203927876698e-06, + "loss": 0.8647, + "num_input_tokens_seen": 22848610, + "step": 705 + }, + { + "epoch": 0.06366956757000496, + "flos": 28653768185280.0, + "grad_norm": 1.9472489121924157, + "learning_rate": 3.988140485728391e-06, + "loss": 0.7063, + "num_input_tokens_seen": 22877915, + "step": 706 + }, + { + "epoch": 0.06375975109347522, + "flos": 18598397517600.0, + "grad_norm": 1.8632216214089616, + "learning_rate": 3.988076873940975e-06, + "loss": 0.7662, + "num_input_tokens_seen": 22904915, + "step": 707 + }, + { + "epoch": 0.06384993461694548, + "flos": 31604247779040.0, + "grad_norm": 2.155208059476525, + "learning_rate": 3.9880130925198786e-06, + "loss": 0.8121, + "num_input_tokens_seen": 22934890, + "step": 708 + }, + { + "epoch": 0.06394011814041575, + "flos": 20128471310880.0, + "grad_norm": 2.9232383444072587, + "learning_rate": 3.987949141470543e-06, + "loss": 0.8267, + "num_input_tokens_seen": 22961320, + "step": 709 + }, + { + "epoch": 0.06403030166388601, + "flos": 25627190625600.0, + "grad_norm": 1.8715805025838397, + "learning_rate": 3.987885020798425e-06, + "loss": 0.8166, + "num_input_tokens_seen": 22990670, + "step": 710 + }, + { + "epoch": 0.06412048518735627, + "flos": 27225296546880.0, + "grad_norm": 2.6582338890520956, + "learning_rate": 3.987820730508996e-06, + "loss": 0.7847, + "num_input_tokens_seen": 23017890, + "step": 711 + }, + { + "epoch": 0.06421066871082654, + "flos": 21694865693280.0, + "grad_norm": 1.803478612224605, + "learning_rate": 3.987756270607742e-06, + "loss": 0.9102, + "num_input_tokens_seen": 23042690, + "step": 712 + }, + { + "epoch": 0.0643008522342968, + "flos": 71756205797760.0, + "grad_norm": 0.6637011274028455, + "learning_rate": 3.987691641100162e-06, + "loss": 0.5649, + "num_input_tokens_seen": 23137420, + "step": 713 + }, + { + "epoch": 0.06439103575776706, + "flos": 20893694056320.0, + "grad_norm": 1.8944575659197105, + "learning_rate": 3.987626841991771e-06, + "loss": 0.9015, + "num_input_tokens_seen": 23165910, + "step": 714 + }, + { + "epoch": 0.06448121928123732, + "flos": 20966558253120.0, + "grad_norm": 3.125926681449495, + "learning_rate": 3.987561873288099e-06, + "loss": 0.8498, + "num_input_tokens_seen": 23192440, + "step": 715 + }, + { + "epoch": 0.06457140280470758, + "flos": 31130742009120.0, + "grad_norm": 2.066183557751762, + "learning_rate": 3.987496734994688e-06, + "loss": 0.7638, + "num_input_tokens_seen": 23222060, + "step": 716 + }, + { + "epoch": 0.06466158632817784, + "flos": 24172545742560.0, + "grad_norm": 5.06942031134634, + "learning_rate": 3.987431427117097e-06, + "loss": 0.847, + "num_input_tokens_seen": 23246460, + "step": 717 + }, + { + "epoch": 0.0647517698516481, + "flos": 22277965116480.0, + "grad_norm": 2.1358196591601732, + "learning_rate": 3.9873659496608985e-06, + "loss": 0.8274, + "num_input_tokens_seen": 23273705, + "step": 718 + }, + { + "epoch": 0.06484195337511836, + "flos": 36850953155520.0, + "grad_norm": 3.2239932714160155, + "learning_rate": 3.987300302631678e-06, + "loss": 0.7707, + "num_input_tokens_seen": 23305395, + "step": 719 + }, + { + "epoch": 0.06493213689858862, + "flos": 20492346257760.0, + "grad_norm": 2.0384459751016712, + "learning_rate": 3.987234486035039e-06, + "loss": 0.8533, + "num_input_tokens_seen": 23334705, + "step": 720 + }, + { + "epoch": 0.0650223204220589, + "flos": 20814845528160.0, + "grad_norm": 2.4411649294796005, + "learning_rate": 3.987168499876595e-06, + "loss": 0.8664, + "num_input_tokens_seen": 23360195, + "step": 721 + }, + { + "epoch": 0.06511250394552916, + "flos": 15683680966560.0, + "grad_norm": 1.9994929982334875, + "learning_rate": 3.987102344161978e-06, + "loss": 0.8279, + "num_input_tokens_seen": 23387990, + "step": 722 + }, + { + "epoch": 0.06520268746899942, + "flos": 16047667422720.0, + "grad_norm": 2.186984301133536, + "learning_rate": 3.987036018896832e-06, + "loss": 0.8531, + "num_input_tokens_seen": 23412510, + "step": 723 + }, + { + "epoch": 0.06529287099246968, + "flos": 22314545893920.0, + "grad_norm": 1.8381529731269257, + "learning_rate": 3.986969524086817e-06, + "loss": 0.8776, + "num_input_tokens_seen": 23440810, + "step": 724 + }, + { + "epoch": 0.06538305451593994, + "flos": 22124208054720.0, + "grad_norm": 2.4339404402618285, + "learning_rate": 3.986902859737605e-06, + "loss": 0.7954, + "num_input_tokens_seen": 23467540, + "step": 725 + }, + { + "epoch": 0.0654732380394102, + "flos": 37652719508640.0, + "grad_norm": 2.3633975670303484, + "learning_rate": 3.986836025854886e-06, + "loss": 0.7502, + "num_input_tokens_seen": 23500970, + "step": 726 + }, + { + "epoch": 0.06556342156288046, + "flos": 26249361200160.0, + "grad_norm": 2.1753697442785467, + "learning_rate": 3.986769022444362e-06, + "loss": 0.7687, + "num_input_tokens_seen": 23530465, + "step": 727 + }, + { + "epoch": 0.06565360508635072, + "flos": 21068123238240.0, + "grad_norm": 7.700531869329328, + "learning_rate": 3.986701849511751e-06, + "loss": 0.8106, + "num_input_tokens_seen": 23558330, + "step": 728 + }, + { + "epoch": 0.06574378860982098, + "flos": 19109041611360.0, + "grad_norm": 2.3688642650411538, + "learning_rate": 3.986634507062782e-06, + "loss": 0.7952, + "num_input_tokens_seen": 23585035, + "step": 729 + }, + { + "epoch": 0.06583397213329124, + "flos": 19800136388160.0, + "grad_norm": 3.3981091570910227, + "learning_rate": 3.986566995103204e-06, + "loss": 0.8327, + "num_input_tokens_seen": 23612175, + "step": 730 + }, + { + "epoch": 0.0659241556567615, + "flos": 27338532836640.0, + "grad_norm": 2.5607654935037933, + "learning_rate": 3.986499313638776e-06, + "loss": 0.7473, + "num_input_tokens_seen": 23641580, + "step": 731 + }, + { + "epoch": 0.06601433918023177, + "flos": 24425860622400.0, + "grad_norm": 2.1522162481021674, + "learning_rate": 3.986431462675272e-06, + "loss": 0.7415, + "num_input_tokens_seen": 23671700, + "step": 732 + }, + { + "epoch": 0.06610452270370204, + "flos": 22857384733440.0, + "grad_norm": 1.526645432467539, + "learning_rate": 3.9863634422184835e-06, + "loss": 0.8332, + "num_input_tokens_seen": 23699905, + "step": 733 + }, + { + "epoch": 0.0661947062271723, + "flos": 19181385431520.0, + "grad_norm": 2.4845399621861493, + "learning_rate": 3.986295252274213e-06, + "loss": 0.8486, + "num_input_tokens_seen": 23727200, + "step": 734 + }, + { + "epoch": 0.06628488975064256, + "flos": 27814975017600.0, + "grad_norm": 1.7732619144045352, + "learning_rate": 3.9862268928482796e-06, + "loss": 0.8634, + "num_input_tokens_seen": 23759215, + "step": 735 + }, + { + "epoch": 0.06637507327411282, + "flos": 28725703138080.0, + "grad_norm": 3.7816884440813134, + "learning_rate": 3.986158363946515e-06, + "loss": 0.7258, + "num_input_tokens_seen": 23789460, + "step": 736 + }, + { + "epoch": 0.06646525679758308, + "flos": 22824520932000.0, + "grad_norm": 1.8279148209215208, + "learning_rate": 3.9860896655747685e-06, + "loss": 0.8157, + "num_input_tokens_seen": 23818195, + "step": 737 + }, + { + "epoch": 0.06655544032105334, + "flos": 27050384158080.0, + "grad_norm": 3.1672704188027194, + "learning_rate": 3.9860207977388994e-06, + "loss": 0.8907, + "num_input_tokens_seen": 23847240, + "step": 738 + }, + { + "epoch": 0.0666456238445236, + "flos": 22751879753760.0, + "grad_norm": 2.8202349792139976, + "learning_rate": 3.9859517604447854e-06, + "loss": 0.7433, + "num_input_tokens_seen": 23876450, + "step": 739 + }, + { + "epoch": 0.06673580736799387, + "flos": 16699988406240.0, + "grad_norm": 1.5282982382840764, + "learning_rate": 3.985882553698317e-06, + "loss": 0.7837, + "num_input_tokens_seen": 23903725, + "step": 740 + }, + { + "epoch": 0.06682599089146413, + "flos": 19144767484320.0, + "grad_norm": 2.3404922223596527, + "learning_rate": 3.985813177505399e-06, + "loss": 0.8755, + "num_input_tokens_seen": 23930865, + "step": 741 + }, + { + "epoch": 0.06691617441493439, + "flos": 29055115983840.0, + "grad_norm": 1.7099702095203002, + "learning_rate": 3.985743631871951e-06, + "loss": 0.7651, + "num_input_tokens_seen": 23959335, + "step": 742 + }, + { + "epoch": 0.06700635793840465, + "flos": 25884817197600.0, + "grad_norm": 2.264987513863545, + "learning_rate": 3.985673916803907e-06, + "loss": 0.812, + "num_input_tokens_seen": 23986685, + "step": 743 + }, + { + "epoch": 0.06709654146187491, + "flos": 22889839667520.0, + "grad_norm": 2.379785534948957, + "learning_rate": 3.985604032307215e-06, + "loss": 0.838, + "num_input_tokens_seen": 24008255, + "step": 744 + }, + { + "epoch": 0.06718672498534518, + "flos": 25265174166720.0, + "grad_norm": 2.6610166975814207, + "learning_rate": 3.985533978387839e-06, + "loss": 0.8218, + "num_input_tokens_seen": 24037790, + "step": 745 + }, + { + "epoch": 0.06727690850881544, + "flos": 15392818895520.0, + "grad_norm": 1.9198803979656327, + "learning_rate": 3.985463755051756e-06, + "loss": 0.855, + "num_input_tokens_seen": 24063690, + "step": 746 + }, + { + "epoch": 0.0673670920322857, + "flos": 26577547443840.0, + "grad_norm": 2.2860143370892145, + "learning_rate": 3.9853933623049575e-06, + "loss": 0.8122, + "num_input_tokens_seen": 24093415, + "step": 747 + }, + { + "epoch": 0.06745727555575597, + "flos": 26540371950240.0, + "grad_norm": 1.7968845533034088, + "learning_rate": 3.98532280015345e-06, + "loss": 0.8094, + "num_input_tokens_seen": 24122685, + "step": 748 + }, + { + "epoch": 0.06754745907922623, + "flos": 43007643257280.0, + "grad_norm": 2.5292532309518787, + "learning_rate": 3.985252068603254e-06, + "loss": 0.7349, + "num_input_tokens_seen": 24154835, + "step": 749 + }, + { + "epoch": 0.06763764260269649, + "flos": 19177482606720.0, + "grad_norm": 11.204046363132678, + "learning_rate": 3.985181167660406e-06, + "loss": 0.8452, + "num_input_tokens_seen": 24182745, + "step": 750 + }, + { + "epoch": 0.06772782612616675, + "flos": 20856407053440.0, + "grad_norm": 3.755769995361356, + "learning_rate": 3.985110097330953e-06, + "loss": 0.8388, + "num_input_tokens_seen": 24211680, + "step": 751 + }, + { + "epoch": 0.06781800964963701, + "flos": 25264430771520.0, + "grad_norm": 1.6147265298983051, + "learning_rate": 3.985038857620962e-06, + "loss": 0.806, + "num_input_tokens_seen": 24244075, + "step": 752 + }, + { + "epoch": 0.06790819317310727, + "flos": 70900834917120.0, + "grad_norm": 0.6786357151857163, + "learning_rate": 3.9849674485365094e-06, + "loss": 0.5889, + "num_input_tokens_seen": 24334200, + "step": 753 + }, + { + "epoch": 0.06799837669657753, + "flos": 14335618986240.0, + "grad_norm": 2.076265119466557, + "learning_rate": 3.98489587008369e-06, + "loss": 0.7395, + "num_input_tokens_seen": 24359125, + "step": 754 + }, + { + "epoch": 0.06808856022004779, + "flos": 23807927400480.0, + "grad_norm": 2.477358620898829, + "learning_rate": 3.98482412226861e-06, + "loss": 0.6573, + "num_input_tokens_seen": 24386210, + "step": 755 + }, + { + "epoch": 0.06817874374351805, + "flos": 16373437632000.0, + "grad_norm": 1.9671823830269095, + "learning_rate": 3.984752205097391e-06, + "loss": 0.8517, + "num_input_tokens_seen": 24410160, + "step": 756 + }, + { + "epoch": 0.06826892726698833, + "flos": 18051953211360.0, + "grad_norm": 1.9798197689294468, + "learning_rate": 3.984680118576171e-06, + "loss": 0.8223, + "num_input_tokens_seen": 24437620, + "step": 757 + }, + { + "epoch": 0.06835911079045859, + "flos": 23916591809760.0, + "grad_norm": 1.8006023061554295, + "learning_rate": 3.984607862711099e-06, + "loss": 0.7695, + "num_input_tokens_seen": 24467580, + "step": 758 + }, + { + "epoch": 0.06844929431392885, + "flos": 21075185492640.0, + "grad_norm": 3.673515050954582, + "learning_rate": 3.984535437508341e-06, + "loss": 0.8507, + "num_input_tokens_seen": 24494785, + "step": 759 + }, + { + "epoch": 0.06853947783739911, + "flos": 69445186450560.0, + "grad_norm": 0.6598908381786133, + "learning_rate": 3.984462842974078e-06, + "loss": 0.5937, + "num_input_tokens_seen": 24581350, + "step": 760 + }, + { + "epoch": 0.06862966136086937, + "flos": 21875242036800.0, + "grad_norm": 3.3740959369472043, + "learning_rate": 3.984390079114502e-06, + "loss": 0.8252, + "num_input_tokens_seen": 24609370, + "step": 761 + }, + { + "epoch": 0.06871984488433963, + "flos": 14954444282400.0, + "grad_norm": 1.9724768226528568, + "learning_rate": 3.984317145935824e-06, + "loss": 0.931, + "num_input_tokens_seen": 24636840, + "step": 762 + }, + { + "epoch": 0.06881002840780989, + "flos": 26139098491200.0, + "grad_norm": 1.9687181500477393, + "learning_rate": 3.984244043444264e-06, + "loss": 0.7867, + "num_input_tokens_seen": 24666720, + "step": 763 + }, + { + "epoch": 0.06890021193128015, + "flos": 24282027886560.0, + "grad_norm": 1.7677185340037043, + "learning_rate": 3.984170771646062e-06, + "loss": 0.7648, + "num_input_tokens_seen": 24695255, + "step": 764 + }, + { + "epoch": 0.06899039545475041, + "flos": 18926360742720.0, + "grad_norm": 2.1018557432564693, + "learning_rate": 3.9840973305474695e-06, + "loss": 0.8621, + "num_input_tokens_seen": 24722870, + "step": 765 + }, + { + "epoch": 0.06908057897822067, + "flos": 20201335507680.0, + "grad_norm": 1.7500059945140076, + "learning_rate": 3.984023720154752e-06, + "loss": 0.8656, + "num_input_tokens_seen": 24749620, + "step": 766 + }, + { + "epoch": 0.06917076250169094, + "flos": 21622447533600.0, + "grad_norm": 1.7114801112542535, + "learning_rate": 3.9839499404741915e-06, + "loss": 0.8257, + "num_input_tokens_seen": 24779265, + "step": 767 + }, + { + "epoch": 0.0692609460251612, + "flos": 65853908498880.0, + "grad_norm": 0.740461933517555, + "learning_rate": 3.983875991512082e-06, + "loss": 0.6089, + "num_input_tokens_seen": 24870130, + "step": 768 + }, + { + "epoch": 0.06935112954863147, + "flos": 23001143129760.0, + "grad_norm": 2.514115822690666, + "learning_rate": 3.9838018732747345e-06, + "loss": 0.8278, + "num_input_tokens_seen": 24897810, + "step": 769 + }, + { + "epoch": 0.06944131307210173, + "flos": 24136299492960.0, + "grad_norm": 1.846010535742223, + "learning_rate": 3.9837275857684716e-06, + "loss": 0.8045, + "num_input_tokens_seen": 24926915, + "step": 770 + }, + { + "epoch": 0.06953149659557199, + "flos": 18670741337760.0, + "grad_norm": 4.195377740050813, + "learning_rate": 3.983653128999634e-06, + "loss": 0.9482, + "num_input_tokens_seen": 24950860, + "step": 771 + }, + { + "epoch": 0.06962168011904225, + "flos": 26539963082880.0, + "grad_norm": 2.0597877007872567, + "learning_rate": 3.983578502974574e-06, + "loss": 0.845, + "num_input_tokens_seen": 24979985, + "step": 772 + }, + { + "epoch": 0.06971186364251251, + "flos": 19217520171840.0, + "grad_norm": 1.9241646254528526, + "learning_rate": 3.983503707699658e-06, + "loss": 0.8022, + "num_input_tokens_seen": 25005175, + "step": 773 + }, + { + "epoch": 0.06980204716598278, + "flos": 21039422449920.0, + "grad_norm": 1.6260132931521336, + "learning_rate": 3.983428743181268e-06, + "loss": 0.844, + "num_input_tokens_seen": 25032040, + "step": 774 + }, + { + "epoch": 0.06989223068945304, + "flos": 20958603924480.0, + "grad_norm": 1.8389551564070292, + "learning_rate": 3.983353609425802e-06, + "loss": 0.8796, + "num_input_tokens_seen": 25056770, + "step": 775 + }, + { + "epoch": 0.0699824142129233, + "flos": 25846972648320.0, + "grad_norm": 1.5709124884996133, + "learning_rate": 3.983278306439671e-06, + "loss": 0.809, + "num_input_tokens_seen": 25088485, + "step": 776 + }, + { + "epoch": 0.07007259773639356, + "flos": 28361679512160.0, + "grad_norm": 1.525914999573122, + "learning_rate": 3.983202834229297e-06, + "loss": 0.8472, + "num_input_tokens_seen": 25117175, + "step": 777 + }, + { + "epoch": 0.07016278125986382, + "flos": 49673044625280.0, + "grad_norm": 1.9790534806930709, + "learning_rate": 3.983127192801123e-06, + "loss": 0.8922, + "num_input_tokens_seen": 25151220, + "step": 778 + }, + { + "epoch": 0.07025296478333408, + "flos": 26314382577600.0, + "grad_norm": 2.3595989196826603, + "learning_rate": 3.983051382161602e-06, + "loss": 0.8484, + "num_input_tokens_seen": 25180425, + "step": 779 + }, + { + "epoch": 0.07034314830680434, + "flos": 21257643342720.0, + "grad_norm": 1.7043194563193964, + "learning_rate": 3.982975402317203e-06, + "loss": 0.837, + "num_input_tokens_seen": 25211490, + "step": 780 + }, + { + "epoch": 0.07043333183027461, + "flos": 20164420202400.0, + "grad_norm": 2.181424341093293, + "learning_rate": 3.982899253274409e-06, + "loss": 0.8346, + "num_input_tokens_seen": 25238970, + "step": 781 + }, + { + "epoch": 0.07052351535374488, + "flos": 19906979479200.0, + "grad_norm": 2.379698272299978, + "learning_rate": 3.982822935039717e-06, + "loss": 0.7996, + "num_input_tokens_seen": 25265625, + "step": 782 + }, + { + "epoch": 0.07061369887721514, + "flos": 19654631013120.0, + "grad_norm": 2.0974429002521418, + "learning_rate": 3.982746447619638e-06, + "loss": 0.7483, + "num_input_tokens_seen": 25293240, + "step": 783 + }, + { + "epoch": 0.0707038824006854, + "flos": 72866012384640.0, + "grad_norm": 0.7085152327031379, + "learning_rate": 3.9826697910207e-06, + "loss": 0.5827, + "num_input_tokens_seen": 25381345, + "step": 784 + }, + { + "epoch": 0.07079406592415566, + "flos": 28179965057280.0, + "grad_norm": 1.791827667750143, + "learning_rate": 3.982592965249442e-06, + "loss": 0.795, + "num_input_tokens_seen": 25409230, + "step": 785 + }, + { + "epoch": 0.07088424944762592, + "flos": 27664749083040.0, + "grad_norm": 2.3480636573233644, + "learning_rate": 3.982515970312422e-06, + "loss": 0.8573, + "num_input_tokens_seen": 25435575, + "step": 786 + }, + { + "epoch": 0.07097443297109618, + "flos": 20273642158080.0, + "grad_norm": 2.5101220260077435, + "learning_rate": 3.982438806216207e-06, + "loss": 0.7852, + "num_input_tokens_seen": 25460700, + "step": 787 + }, + { + "epoch": 0.07106461649456644, + "flos": 39838645412640.0, + "grad_norm": 2.041937310581958, + "learning_rate": 3.982361472967382e-06, + "loss": 0.7992, + "num_input_tokens_seen": 25492560, + "step": 788 + }, + { + "epoch": 0.0711548000180367, + "flos": 23225013826080.0, + "grad_norm": 1.9177134046706805, + "learning_rate": 3.982283970572546e-06, + "loss": 0.8858, + "num_input_tokens_seen": 25518260, + "step": 789 + }, + { + "epoch": 0.07124498354150696, + "flos": 12951385095840.0, + "grad_norm": 2.527436008567731, + "learning_rate": 3.982206299038311e-06, + "loss": 0.7895, + "num_input_tokens_seen": 25544100, + "step": 790 + }, + { + "epoch": 0.07133516706497722, + "flos": 17064904106400.0, + "grad_norm": 1.835321525968535, + "learning_rate": 3.9821284583713054e-06, + "loss": 0.888, + "num_input_tokens_seen": 25570860, + "step": 791 + }, + { + "epoch": 0.07142535058844748, + "flos": 17980166937600.0, + "grad_norm": 2.7546667902054933, + "learning_rate": 3.98205044857817e-06, + "loss": 0.7955, + "num_input_tokens_seen": 25597555, + "step": 792 + }, + { + "epoch": 0.07151553411191776, + "flos": 69745601149920.0, + "grad_norm": 0.7417931964897562, + "learning_rate": 3.981972269665561e-06, + "loss": 0.6117, + "num_input_tokens_seen": 25686020, + "step": 793 + }, + { + "epoch": 0.07160571763538802, + "flos": 26723164328160.0, + "grad_norm": 1.8184253914196056, + "learning_rate": 3.98189392164015e-06, + "loss": 0.8065, + "num_input_tokens_seen": 25718720, + "step": 794 + }, + { + "epoch": 0.07169590115885828, + "flos": 27265259772480.0, + "grad_norm": 1.85028946648132, + "learning_rate": 3.981815404508621e-06, + "loss": 0.854, + "num_input_tokens_seen": 25744515, + "step": 795 + }, + { + "epoch": 0.07178608468232854, + "flos": 21476310272640.0, + "grad_norm": 2.2086293603909213, + "learning_rate": 3.981736718277674e-06, + "loss": 0.8194, + "num_input_tokens_seen": 25773010, + "step": 796 + }, + { + "epoch": 0.0718762682057988, + "flos": 15683457948000.0, + "grad_norm": 2.160046096945659, + "learning_rate": 3.9816578629540235e-06, + "loss": 0.8478, + "num_input_tokens_seen": 25798695, + "step": 797 + }, + { + "epoch": 0.07196645172926906, + "flos": 28939352150400.0, + "grad_norm": 2.605844542069865, + "learning_rate": 3.981578838544398e-06, + "loss": 0.8231, + "num_input_tokens_seen": 25826695, + "step": 798 + }, + { + "epoch": 0.07205663525273932, + "flos": 66627308591520.0, + "grad_norm": 0.6683741895058267, + "learning_rate": 3.981499645055539e-06, + "loss": 0.566, + "num_input_tokens_seen": 25927685, + "step": 799 + }, + { + "epoch": 0.07214681877620958, + "flos": 60721368656160.0, + "grad_norm": 0.6514686379645989, + "learning_rate": 3.981420282494204e-06, + "loss": 0.6001, + "num_input_tokens_seen": 26024045, + "step": 800 + }, + { + "epoch": 0.07223700229967984, + "flos": 70613095105920.0, + "grad_norm": 0.7441136312720067, + "learning_rate": 3.981340750867166e-06, + "loss": 0.5993, + "num_input_tokens_seen": 26109055, + "step": 801 + }, + { + "epoch": 0.0723271858231501, + "flos": 23878635751200.0, + "grad_norm": 1.6217453715536356, + "learning_rate": 3.981261050181209e-06, + "loss": 0.7995, + "num_input_tokens_seen": 26138125, + "step": 802 + }, + { + "epoch": 0.07241736934662037, + "flos": 25301977962720.0, + "grad_norm": 2.0498223669735434, + "learning_rate": 3.9811811804431355e-06, + "loss": 0.7667, + "num_input_tokens_seen": 26166785, + "step": 803 + }, + { + "epoch": 0.07250755287009064, + "flos": 24281396000640.0, + "grad_norm": 2.0129495059491624, + "learning_rate": 3.981101141659759e-06, + "loss": 0.8529, + "num_input_tokens_seen": 26195365, + "step": 804 + }, + { + "epoch": 0.0725977363935609, + "flos": 19069896120480.0, + "grad_norm": 1.8044621050230885, + "learning_rate": 3.98102093383791e-06, + "loss": 0.8408, + "num_input_tokens_seen": 26223310, + "step": 805 + }, + { + "epoch": 0.07268791991703116, + "flos": 27956280209760.0, + "grad_norm": 5.54675481961576, + "learning_rate": 3.9809405569844315e-06, + "loss": 0.7335, + "num_input_tokens_seen": 26251745, + "step": 806 + }, + { + "epoch": 0.07277810344050142, + "flos": 20489335507200.0, + "grad_norm": 1.682567097543211, + "learning_rate": 3.980860011106182e-06, + "loss": 0.7746, + "num_input_tokens_seen": 26279365, + "step": 807 + }, + { + "epoch": 0.07286828696397168, + "flos": 23297617834560.0, + "grad_norm": 1.756388787049668, + "learning_rate": 3.980779296210033e-06, + "loss": 0.7503, + "num_input_tokens_seen": 26310470, + "step": 808 + }, + { + "epoch": 0.07295847048744195, + "flos": 19764224666400.0, + "grad_norm": 2.8354308461309574, + "learning_rate": 3.980698412302874e-06, + "loss": 0.8912, + "num_input_tokens_seen": 26337455, + "step": 809 + }, + { + "epoch": 0.0730486540109122, + "flos": 21585383549280.0, + "grad_norm": 1.9911978264227805, + "learning_rate": 3.980617359391604e-06, + "loss": 0.8262, + "num_input_tokens_seen": 26364510, + "step": 810 + }, + { + "epoch": 0.07313883753438247, + "flos": 30873747323040.0, + "grad_norm": 1.9883403656870613, + "learning_rate": 3.98053613748314e-06, + "loss": 0.7691, + "num_input_tokens_seen": 26395895, + "step": 811 + }, + { + "epoch": 0.07322902105785273, + "flos": 43882310976960.0, + "grad_norm": 1.5394512789215968, + "learning_rate": 3.980454746584413e-06, + "loss": 0.6743, + "num_input_tokens_seen": 26426045, + "step": 812 + }, + { + "epoch": 0.07331920458132299, + "flos": 20274534232320.0, + "grad_norm": 2.971641818584707, + "learning_rate": 3.9803731867023665e-06, + "loss": 0.8981, + "num_input_tokens_seen": 26452705, + "step": 813 + }, + { + "epoch": 0.07340938810479325, + "flos": 22601914007520.0, + "grad_norm": 2.103363966721277, + "learning_rate": 3.9802914578439596e-06, + "loss": 0.7979, + "num_input_tokens_seen": 26481080, + "step": 814 + }, + { + "epoch": 0.07349957162826351, + "flos": 24537052575360.0, + "grad_norm": 1.8534308555842822, + "learning_rate": 3.980209560016167e-06, + "loss": 0.9172, + "num_input_tokens_seen": 26509240, + "step": 815 + }, + { + "epoch": 0.07358975515173379, + "flos": 27158565360480.0, + "grad_norm": 1.9296369661826405, + "learning_rate": 3.980127493225975e-06, + "loss": 0.904, + "num_input_tokens_seen": 26539635, + "step": 816 + }, + { + "epoch": 0.07367993867520405, + "flos": 23735806598880.0, + "grad_norm": 2.8457384807665527, + "learning_rate": 3.980045257480387e-06, + "loss": 0.8436, + "num_input_tokens_seen": 26568420, + "step": 817 + }, + { + "epoch": 0.0737701221986743, + "flos": 24026817348960.0, + "grad_norm": 2.542446569220783, + "learning_rate": 3.9799628527864205e-06, + "loss": 0.7707, + "num_input_tokens_seen": 26595545, + "step": 818 + }, + { + "epoch": 0.07386030572214457, + "flos": 23917929921120.0, + "grad_norm": 1.990036584910802, + "learning_rate": 3.979880279151106e-06, + "loss": 0.7968, + "num_input_tokens_seen": 26624950, + "step": 819 + }, + { + "epoch": 0.07395048924561483, + "flos": 26030136723840.0, + "grad_norm": 1.92123414806808, + "learning_rate": 3.979797536581489e-06, + "loss": 0.7548, + "num_input_tokens_seen": 26655695, + "step": 820 + }, + { + "epoch": 0.07404067276908509, + "flos": 26905250480640.0, + "grad_norm": 2.721056431205836, + "learning_rate": 3.97971462508463e-06, + "loss": 0.6517, + "num_input_tokens_seen": 26682665, + "step": 821 + }, + { + "epoch": 0.07413085629255535, + "flos": 27924828859200.0, + "grad_norm": 2.4042053453142773, + "learning_rate": 3.979631544667603e-06, + "loss": 0.7959, + "num_input_tokens_seen": 26713390, + "step": 822 + }, + { + "epoch": 0.07422103981602561, + "flos": 38230801014240.0, + "grad_norm": 2.6770994628737, + "learning_rate": 3.979548295337496e-06, + "loss": 0.8239, + "num_input_tokens_seen": 26740670, + "step": 823 + }, + { + "epoch": 0.07431122333949587, + "flos": 26208654579360.0, + "grad_norm": 1.806425834465787, + "learning_rate": 3.9794648771014146e-06, + "loss": 0.7777, + "num_input_tokens_seen": 26772125, + "step": 824 + }, + { + "epoch": 0.07440140686296613, + "flos": 22350903652800.0, + "grad_norm": 1.9473580019971943, + "learning_rate": 3.9793812899664745e-06, + "loss": 0.8144, + "num_input_tokens_seen": 26799605, + "step": 825 + }, + { + "epoch": 0.0744915903864364, + "flos": 29161736056320.0, + "grad_norm": 2.946695261193683, + "learning_rate": 3.979297533939809e-06, + "loss": 0.7694, + "num_input_tokens_seen": 26827190, + "step": 826 + }, + { + "epoch": 0.07458177390990665, + "flos": 23480521721760.0, + "grad_norm": 2.181609588101744, + "learning_rate": 3.979213609028564e-06, + "loss": 0.788, + "num_input_tokens_seen": 26855695, + "step": 827 + }, + { + "epoch": 0.07467195743337693, + "flos": 20857336297440.0, + "grad_norm": 1.8694883718646744, + "learning_rate": 3.979129515239901e-06, + "loss": 0.8258, + "num_input_tokens_seen": 26884355, + "step": 828 + }, + { + "epoch": 0.07476214095684719, + "flos": 26867517440640.0, + "grad_norm": 1.8729428132141233, + "learning_rate": 3.979045252580994e-06, + "loss": 0.8165, + "num_input_tokens_seen": 26915595, + "step": 829 + }, + { + "epoch": 0.07485232448031745, + "flos": 37615767033600.0, + "grad_norm": 2.2098663215425804, + "learning_rate": 3.978960821059034e-06, + "loss": 0.7551, + "num_input_tokens_seen": 26946505, + "step": 830 + }, + { + "epoch": 0.07494250800378771, + "flos": 23695731864000.0, + "grad_norm": 2.0371921189676097, + "learning_rate": 3.978876220681225e-06, + "loss": 0.8504, + "num_input_tokens_seen": 26973305, + "step": 831 + }, + { + "epoch": 0.07503269152725797, + "flos": 21549100129920.0, + "grad_norm": 1.928913470873135, + "learning_rate": 3.978791451454786e-06, + "loss": 0.7399, + "num_input_tokens_seen": 27001795, + "step": 832 + }, + { + "epoch": 0.07512287505072823, + "flos": 24934163021280.0, + "grad_norm": 1.8941111305286393, + "learning_rate": 3.978706513386949e-06, + "loss": 0.771, + "num_input_tokens_seen": 27031210, + "step": 833 + }, + { + "epoch": 0.0752130585741985, + "flos": 54843020149920.0, + "grad_norm": 2.7602820213636465, + "learning_rate": 3.978621406484962e-06, + "loss": 0.8213, + "num_input_tokens_seen": 27065670, + "step": 834 + }, + { + "epoch": 0.07530324209766875, + "flos": 28143681637920.0, + "grad_norm": 3.3147366727841145, + "learning_rate": 3.978536130756086e-06, + "loss": 0.8009, + "num_input_tokens_seen": 27096090, + "step": 835 + }, + { + "epoch": 0.07539342562113902, + "flos": 22931698550880.0, + "grad_norm": 2.0545988462635187, + "learning_rate": 3.978450686207599e-06, + "loss": 0.8743, + "num_input_tokens_seen": 27125370, + "step": 836 + }, + { + "epoch": 0.07548360914460928, + "flos": 17104569973920.0, + "grad_norm": 2.1742913505777612, + "learning_rate": 3.978365072846789e-06, + "loss": 0.791, + "num_input_tokens_seen": 27151560, + "step": 837 + }, + { + "epoch": 0.07557379266807954, + "flos": 25702247838240.0, + "grad_norm": 9.479424531105144, + "learning_rate": 3.9782792906809625e-06, + "loss": 0.7594, + "num_input_tokens_seen": 27180425, + "step": 838 + }, + { + "epoch": 0.0756639761915498, + "flos": 20675250144960.0, + "grad_norm": 2.3906603043885184, + "learning_rate": 3.97819333971744e-06, + "loss": 0.7943, + "num_input_tokens_seen": 27207455, + "step": 839 + }, + { + "epoch": 0.07575415971502007, + "flos": 23147540579040.0, + "grad_norm": 3.5926348739079006, + "learning_rate": 3.978107219963553e-06, + "loss": 0.8211, + "num_input_tokens_seen": 27235885, + "step": 840 + }, + { + "epoch": 0.07584434323849033, + "flos": 19210755275520.0, + "grad_norm": 2.551057809321985, + "learning_rate": 3.978020931426651e-06, + "loss": 0.718, + "num_input_tokens_seen": 27263610, + "step": 841 + }, + { + "epoch": 0.0759345267619606, + "flos": 28145168428320.0, + "grad_norm": 2.1969911937997257, + "learning_rate": 3.977934474114096e-06, + "loss": 0.7589, + "num_input_tokens_seen": 27294335, + "step": 842 + }, + { + "epoch": 0.07602471028543085, + "flos": 70408552684800.0, + "grad_norm": 0.6404249762908303, + "learning_rate": 3.977847848033267e-06, + "loss": 0.5779, + "num_input_tokens_seen": 27390535, + "step": 843 + }, + { + "epoch": 0.07611489380890112, + "flos": 27669692661120.0, + "grad_norm": 1.7542349118740048, + "learning_rate": 3.977761053191553e-06, + "loss": 0.8523, + "num_input_tokens_seen": 27417140, + "step": 844 + }, + { + "epoch": 0.07620507733237138, + "flos": 25518154518720.0, + "grad_norm": 2.8495926415290045, + "learning_rate": 3.977674089596361e-06, + "loss": 0.7894, + "num_input_tokens_seen": 27448470, + "step": 845 + }, + { + "epoch": 0.07629526085584164, + "flos": 52979773631520.0, + "grad_norm": 0.6889488602263723, + "learning_rate": 3.97758695725511e-06, + "loss": 0.5748, + "num_input_tokens_seen": 27533360, + "step": 846 + }, + { + "epoch": 0.0763854443793119, + "flos": 31822877539200.0, + "grad_norm": 2.631057018171106, + "learning_rate": 3.977499656175236e-06, + "loss": 0.8038, + "num_input_tokens_seen": 27566515, + "step": 847 + }, + { + "epoch": 0.07647562790278216, + "flos": 25478042614080.0, + "grad_norm": 2.7337026586911852, + "learning_rate": 3.977412186364187e-06, + "loss": 0.717, + "num_input_tokens_seen": 27596320, + "step": 848 + }, + { + "epoch": 0.07656581142625242, + "flos": 19764113157120.0, + "grad_norm": 2.0678200743595454, + "learning_rate": 3.977324547829428e-06, + "loss": 0.8328, + "num_input_tokens_seen": 27623210, + "step": 849 + }, + { + "epoch": 0.07665599494972268, + "flos": 16988954819520.0, + "grad_norm": 2.28700950658613, + "learning_rate": 3.977236740578435e-06, + "loss": 0.8326, + "num_input_tokens_seen": 27648160, + "step": 850 + }, + { + "epoch": 0.07674617847319294, + "flos": 23553460258080.0, + "grad_norm": 2.959456181444925, + "learning_rate": 3.9771487646187015e-06, + "loss": 0.7957, + "num_input_tokens_seen": 27677070, + "step": 851 + }, + { + "epoch": 0.07683636199666322, + "flos": 17246135354400.0, + "grad_norm": 4.0982408424749375, + "learning_rate": 3.9770606199577325e-06, + "loss": 0.8456, + "num_input_tokens_seen": 27702665, + "step": 852 + }, + { + "epoch": 0.07692654552013348, + "flos": 19545706415520.0, + "grad_norm": 2.8382022946435344, + "learning_rate": 3.9769723066030505e-06, + "loss": 0.7394, + "num_input_tokens_seen": 27730085, + "step": 853 + }, + { + "epoch": 0.07701672904360374, + "flos": 20122115281920.0, + "grad_norm": 3.2459840170701253, + "learning_rate": 3.976883824562191e-06, + "loss": 0.8679, + "num_input_tokens_seen": 27755805, + "step": 854 + }, + { + "epoch": 0.077106912567074, + "flos": 19581729646560.0, + "grad_norm": 2.025307163442801, + "learning_rate": 3.976795173842703e-06, + "loss": 0.807, + "num_input_tokens_seen": 27782290, + "step": 855 + }, + { + "epoch": 0.07719709609054426, + "flos": 32401107723840.0, + "grad_norm": 4.0048055392523825, + "learning_rate": 3.97670635445215e-06, + "loss": 0.7852, + "num_input_tokens_seen": 27810420, + "step": 856 + }, + { + "epoch": 0.07728727961401452, + "flos": 23297766513600.0, + "grad_norm": 2.079061770293525, + "learning_rate": 3.976617366398112e-06, + "loss": 0.7781, + "num_input_tokens_seen": 27837010, + "step": 857 + }, + { + "epoch": 0.07737746313748478, + "flos": 24463444983360.0, + "grad_norm": 1.418401814894529, + "learning_rate": 3.976528209688181e-06, + "loss": 0.8371, + "num_input_tokens_seen": 27866500, + "step": 858 + }, + { + "epoch": 0.07746764666095504, + "flos": 24024512823840.0, + "grad_norm": 1.737964661303684, + "learning_rate": 3.976438884329965e-06, + "loss": 0.8125, + "num_input_tokens_seen": 27898645, + "step": 859 + }, + { + "epoch": 0.0775578301844253, + "flos": 25443394664160.0, + "grad_norm": 1.82928990475363, + "learning_rate": 3.976349390331085e-06, + "loss": 0.9374, + "num_input_tokens_seen": 27928350, + "step": 860 + }, + { + "epoch": 0.07764801370789556, + "flos": 23913432380160.0, + "grad_norm": 2.528453082285461, + "learning_rate": 3.976259727699178e-06, + "loss": 0.7743, + "num_input_tokens_seen": 27955385, + "step": 861 + }, + { + "epoch": 0.07773819723136582, + "flos": 18338169062400.0, + "grad_norm": 2.4403667708976586, + "learning_rate": 3.976169896441895e-06, + "loss": 0.7625, + "num_input_tokens_seen": 27980565, + "step": 862 + }, + { + "epoch": 0.07782838075483609, + "flos": 23881497822720.0, + "grad_norm": 2.3610849815368913, + "learning_rate": 3.976079896566898e-06, + "loss": 0.8401, + "num_input_tokens_seen": 28009140, + "step": 863 + }, + { + "epoch": 0.07791856427830636, + "flos": 29706805081440.0, + "grad_norm": 2.0940087002024224, + "learning_rate": 3.97598972808187e-06, + "loss": 0.8733, + "num_input_tokens_seen": 28038040, + "step": 864 + }, + { + "epoch": 0.07800874780177662, + "flos": 22788497700960.0, + "grad_norm": 2.127885318204249, + "learning_rate": 3.975899390994501e-06, + "loss": 0.808, + "num_input_tokens_seen": 28067820, + "step": 865 + }, + { + "epoch": 0.07809893132524688, + "flos": 22349231013600.0, + "grad_norm": 1.6350397862251127, + "learning_rate": 3.975808885312502e-06, + "loss": 0.8713, + "num_input_tokens_seen": 28097390, + "step": 866 + }, + { + "epoch": 0.07818911484871714, + "flos": 22132534080960.0, + "grad_norm": 1.755623908797971, + "learning_rate": 3.975718211043594e-06, + "loss": 0.8553, + "num_input_tokens_seen": 28126580, + "step": 867 + }, + { + "epoch": 0.0782792983721874, + "flos": 22095321417600.0, + "grad_norm": 2.5973598684010315, + "learning_rate": 3.975627368195515e-06, + "loss": 0.7237, + "num_input_tokens_seen": 28153765, + "step": 868 + }, + { + "epoch": 0.07836948189565766, + "flos": 27635565087840.0, + "grad_norm": 1.6951607491216132, + "learning_rate": 3.975536356776015e-06, + "loss": 0.7772, + "num_input_tokens_seen": 28186050, + "step": 869 + }, + { + "epoch": 0.07845966541912792, + "flos": 66662997294720.0, + "grad_norm": 0.7400085185179259, + "learning_rate": 3.975445176792861e-06, + "loss": 0.6516, + "num_input_tokens_seen": 28282075, + "step": 870 + }, + { + "epoch": 0.07854984894259819, + "flos": 20963733351360.0, + "grad_norm": 2.0388874548832803, + "learning_rate": 3.975353828253831e-06, + "loss": 0.815, + "num_input_tokens_seen": 28312830, + "step": 871 + }, + { + "epoch": 0.07864003246606845, + "flos": 24791148020160.0, + "grad_norm": 4.035827167345328, + "learning_rate": 3.97526231116672e-06, + "loss": 0.8093, + "num_input_tokens_seen": 28342420, + "step": 872 + }, + { + "epoch": 0.07873021598953871, + "flos": 26321630680800.0, + "grad_norm": 3.554651281151334, + "learning_rate": 3.975170625539338e-06, + "loss": 0.7496, + "num_input_tokens_seen": 28371795, + "step": 873 + }, + { + "epoch": 0.07882039951300897, + "flos": 24536086161600.0, + "grad_norm": 1.6224494137394339, + "learning_rate": 3.975078771379507e-06, + "loss": 0.8062, + "num_input_tokens_seen": 28404155, + "step": 874 + }, + { + "epoch": 0.07891058303647923, + "flos": 21294335629440.0, + "grad_norm": 3.093818124348594, + "learning_rate": 3.974986748695064e-06, + "loss": 0.7186, + "num_input_tokens_seen": 28429285, + "step": 875 + }, + { + "epoch": 0.0790007665599495, + "flos": 61866492248640.0, + "grad_norm": 0.7207659355436961, + "learning_rate": 3.974894557493862e-06, + "loss": 0.5894, + "num_input_tokens_seen": 28510930, + "step": 876 + }, + { + "epoch": 0.07909095008341976, + "flos": 18598546196640.0, + "grad_norm": 2.012769465443201, + "learning_rate": 3.974802197783768e-06, + "loss": 0.8524, + "num_input_tokens_seen": 28538705, + "step": 877 + }, + { + "epoch": 0.07918113360689003, + "flos": 13607051357760.0, + "grad_norm": 2.15857626334586, + "learning_rate": 3.974709669572661e-06, + "loss": 0.8665, + "num_input_tokens_seen": 28564580, + "step": 878 + }, + { + "epoch": 0.07927131713036029, + "flos": 17687409208800.0, + "grad_norm": 1.7951466084487906, + "learning_rate": 3.974616972868436e-06, + "loss": 0.799, + "num_input_tokens_seen": 28592085, + "step": 879 + }, + { + "epoch": 0.07936150065383055, + "flos": 23405538848640.0, + "grad_norm": 1.7979503565078077, + "learning_rate": 3.974524107679003e-06, + "loss": 0.8137, + "num_input_tokens_seen": 28618260, + "step": 880 + }, + { + "epoch": 0.07945168417730081, + "flos": 31093938213120.0, + "grad_norm": 1.8351025569851185, + "learning_rate": 3.974431074012286e-06, + "loss": 0.7984, + "num_input_tokens_seen": 28650475, + "step": 881 + }, + { + "epoch": 0.07954186770077107, + "flos": 24172359893760.0, + "grad_norm": 2.7073299799095762, + "learning_rate": 3.974337871876223e-06, + "loss": 0.8322, + "num_input_tokens_seen": 28676020, + "step": 882 + }, + { + "epoch": 0.07963205122424133, + "flos": 29778256827360.0, + "grad_norm": 1.5512948903144426, + "learning_rate": 3.974244501278766e-06, + "loss": 0.7748, + "num_input_tokens_seen": 28705870, + "step": 883 + }, + { + "epoch": 0.07972223474771159, + "flos": 31058398188960.0, + "grad_norm": 1.719952110910203, + "learning_rate": 3.974150962227883e-06, + "loss": 0.7575, + "num_input_tokens_seen": 28740180, + "step": 884 + }, + { + "epoch": 0.07981241827118185, + "flos": 19837014523680.0, + "grad_norm": 2.0535579230770717, + "learning_rate": 3.9740572547315535e-06, + "loss": 0.7342, + "num_input_tokens_seen": 28768565, + "step": 885 + }, + { + "epoch": 0.07990260179465211, + "flos": 22860321144480.0, + "grad_norm": 2.458758839997156, + "learning_rate": 3.973963378797775e-06, + "loss": 0.8462, + "num_input_tokens_seen": 28797515, + "step": 886 + }, + { + "epoch": 0.07999278531812237, + "flos": 24243551451360.0, + "grad_norm": 1.775859573590467, + "learning_rate": 3.973869334434556e-06, + "loss": 0.8221, + "num_input_tokens_seen": 28828660, + "step": 887 + }, + { + "epoch": 0.08008296884159265, + "flos": 47853112344480.0, + "grad_norm": 1.8202007734204222, + "learning_rate": 3.973775121649922e-06, + "loss": 0.8226, + "num_input_tokens_seen": 28862345, + "step": 888 + }, + { + "epoch": 0.08017315236506291, + "flos": 23336726155680.0, + "grad_norm": 1.7856026723585332, + "learning_rate": 3.973680740451911e-06, + "loss": 0.7212, + "num_input_tokens_seen": 28894665, + "step": 889 + }, + { + "epoch": 0.08026333588853317, + "flos": 19655077050240.0, + "grad_norm": 2.08504624203419, + "learning_rate": 3.9735861908485776e-06, + "loss": 0.7691, + "num_input_tokens_seen": 28922125, + "step": 890 + }, + { + "epoch": 0.08035351941200343, + "flos": 38413184524800.0, + "grad_norm": 3.135304209233298, + "learning_rate": 3.973491472847987e-06, + "loss": 0.8286, + "num_input_tokens_seen": 28952670, + "step": 891 + }, + { + "epoch": 0.08044370293547369, + "flos": 34152859197600.0, + "grad_norm": 1.6937555995721914, + "learning_rate": 3.973396586458222e-06, + "loss": 0.738, + "num_input_tokens_seen": 28985555, + "step": 892 + }, + { + "epoch": 0.08053388645894395, + "flos": 24680253425280.0, + "grad_norm": 4.32823649238731, + "learning_rate": 3.97330153168738e-06, + "loss": 0.7929, + "num_input_tokens_seen": 29016095, + "step": 893 + }, + { + "epoch": 0.08062406998241421, + "flos": 25808198855040.0, + "grad_norm": 1.7307702484290888, + "learning_rate": 3.973206308543571e-06, + "loss": 0.7519, + "num_input_tokens_seen": 29048180, + "step": 894 + }, + { + "epoch": 0.08071425350588447, + "flos": 22460497306080.0, + "grad_norm": 2.145541927464039, + "learning_rate": 3.973110917034918e-06, + "loss": 0.7792, + "num_input_tokens_seen": 29076770, + "step": 895 + }, + { + "epoch": 0.08080443702935473, + "flos": 29455125671040.0, + "grad_norm": 1.7170659224397251, + "learning_rate": 3.973015357169563e-06, + "loss": 0.7156, + "num_input_tokens_seen": 29105410, + "step": 896 + }, + { + "epoch": 0.080894620552825, + "flos": 40602604386240.0, + "grad_norm": 2.5607225720077125, + "learning_rate": 3.972919628955659e-06, + "loss": 0.6998, + "num_input_tokens_seen": 29135255, + "step": 897 + }, + { + "epoch": 0.08098480407629526, + "flos": 19172092991520.0, + "grad_norm": 2.4355246535987587, + "learning_rate": 3.972823732401373e-06, + "loss": 0.7948, + "num_input_tokens_seen": 29158075, + "step": 898 + }, + { + "epoch": 0.08107498759976552, + "flos": 26431112824800.0, + "grad_norm": 1.7873505643255667, + "learning_rate": 3.972727667514888e-06, + "loss": 0.7994, + "num_input_tokens_seen": 29185440, + "step": 899 + }, + { + "epoch": 0.08116517112323579, + "flos": 22314917591520.0, + "grad_norm": 1.6284353915859549, + "learning_rate": 3.972631434304402e-06, + "loss": 0.7667, + "num_input_tokens_seen": 29216030, + "step": 900 + }, + { + "epoch": 0.08125535464670605, + "flos": 18597914310720.0, + "grad_norm": 2.1150796116844934, + "learning_rate": 3.972535032778124e-06, + "loss": 0.9202, + "num_input_tokens_seen": 29241495, + "step": 901 + }, + { + "epoch": 0.08134553817017631, + "flos": 27632888865120.0, + "grad_norm": 2.251274807383989, + "learning_rate": 3.97243846294428e-06, + "loss": 0.8914, + "num_input_tokens_seen": 29271195, + "step": 902 + }, + { + "epoch": 0.08143572169364657, + "flos": 20346840882720.0, + "grad_norm": 1.6176341448997686, + "learning_rate": 3.972341724811111e-06, + "loss": 0.9421, + "num_input_tokens_seen": 29299090, + "step": 903 + }, + { + "epoch": 0.08152590521711683, + "flos": 22932925152960.0, + "grad_norm": 2.0009858562650202, + "learning_rate": 3.972244818386872e-06, + "loss": 0.7491, + "num_input_tokens_seen": 29327530, + "step": 904 + }, + { + "epoch": 0.0816160887405871, + "flos": 22605407964960.0, + "grad_norm": 1.782753360178617, + "learning_rate": 3.972147743679828e-06, + "loss": 0.7455, + "num_input_tokens_seen": 29354465, + "step": 905 + }, + { + "epoch": 0.08170627226405736, + "flos": 21221545772160.0, + "grad_norm": 2.1277871646568913, + "learning_rate": 3.972050500698265e-06, + "loss": 0.7689, + "num_input_tokens_seen": 29384960, + "step": 906 + }, + { + "epoch": 0.08179645578752762, + "flos": 25447669186560.0, + "grad_norm": 1.8350451309516598, + "learning_rate": 3.971953089450481e-06, + "loss": 0.7965, + "num_input_tokens_seen": 29412875, + "step": 907 + }, + { + "epoch": 0.08188663931099788, + "flos": 21221545772160.0, + "grad_norm": 1.8200103116005346, + "learning_rate": 3.971855509944784e-06, + "loss": 0.876, + "num_input_tokens_seen": 29440935, + "step": 908 + }, + { + "epoch": 0.08197682283446814, + "flos": 27592070735040.0, + "grad_norm": 1.662810235010306, + "learning_rate": 3.971757762189504e-06, + "loss": 0.791, + "num_input_tokens_seen": 29472000, + "step": 909 + }, + { + "epoch": 0.0820670063579384, + "flos": 27232693329120.0, + "grad_norm": 2.495047814287545, + "learning_rate": 3.9716598461929785e-06, + "loss": 0.7925, + "num_input_tokens_seen": 29501575, + "step": 910 + }, + { + "epoch": 0.08215718988140866, + "flos": 23844210819840.0, + "grad_norm": 1.7272191934868149, + "learning_rate": 3.971561761963563e-06, + "loss": 0.7786, + "num_input_tokens_seen": 29529775, + "step": 911 + }, + { + "epoch": 0.08224737340487893, + "flos": 15610147714080.0, + "grad_norm": 2.057406638751551, + "learning_rate": 3.971463509509628e-06, + "loss": 0.8818, + "num_input_tokens_seen": 29557115, + "step": 912 + }, + { + "epoch": 0.0823375569283492, + "flos": 16557939818880.0, + "grad_norm": 2.8614366113712872, + "learning_rate": 3.9713650888395555e-06, + "loss": 0.7934, + "num_input_tokens_seen": 29581215, + "step": 913 + }, + { + "epoch": 0.08242774045181946, + "flos": 19508828280000.0, + "grad_norm": 1.9445990054448061, + "learning_rate": 3.9712664999617425e-06, + "loss": 0.7925, + "num_input_tokens_seen": 29605060, + "step": 914 + }, + { + "epoch": 0.08251792397528972, + "flos": 24464411397120.0, + "grad_norm": 2.0632417674333143, + "learning_rate": 3.971167742884603e-06, + "loss": 0.7766, + "num_input_tokens_seen": 29634025, + "step": 915 + }, + { + "epoch": 0.08260810749875998, + "flos": 19212948291360.0, + "grad_norm": 1.9772674750813373, + "learning_rate": 3.971068817616564e-06, + "loss": 0.8532, + "num_input_tokens_seen": 29661645, + "step": 916 + }, + { + "epoch": 0.08269829102223024, + "flos": 28289558710560.0, + "grad_norm": 2.0481083772116864, + "learning_rate": 3.970969724166064e-06, + "loss": 0.8271, + "num_input_tokens_seen": 29693300, + "step": 917 + }, + { + "epoch": 0.0827884745457005, + "flos": 69651841814400.0, + "grad_norm": 0.8297232553712384, + "learning_rate": 3.970870462541559e-06, + "loss": 0.5777, + "num_input_tokens_seen": 29785140, + "step": 918 + }, + { + "epoch": 0.08287865806917076, + "flos": 24718246653600.0, + "grad_norm": 1.5500944908627432, + "learning_rate": 3.97077103275152e-06, + "loss": 0.821, + "num_input_tokens_seen": 29816060, + "step": 919 + }, + { + "epoch": 0.08296884159264102, + "flos": 31891392874080.0, + "grad_norm": 2.222156971397337, + "learning_rate": 3.970671434804428e-06, + "loss": 0.788, + "num_input_tokens_seen": 29848965, + "step": 920 + }, + { + "epoch": 0.08305902511611128, + "flos": 24898251299520.0, + "grad_norm": 1.8294752954033264, + "learning_rate": 3.970571668708784e-06, + "loss": 0.7985, + "num_input_tokens_seen": 29879670, + "step": 921 + }, + { + "epoch": 0.08314920863958154, + "flos": 67144606072320.0, + "grad_norm": 0.6616010036444748, + "learning_rate": 3.9704717344731e-06, + "loss": 0.5871, + "num_input_tokens_seen": 29980395, + "step": 922 + }, + { + "epoch": 0.0832393921630518, + "flos": 66261463647360.0, + "grad_norm": 0.718787358183867, + "learning_rate": 3.9703716321059026e-06, + "loss": 0.603, + "num_input_tokens_seen": 30081175, + "step": 923 + }, + { + "epoch": 0.08332957568652208, + "flos": 26066457312960.0, + "grad_norm": 2.3480193861724814, + "learning_rate": 3.9702713616157325e-06, + "loss": 0.891, + "num_input_tokens_seen": 30110635, + "step": 924 + }, + { + "epoch": 0.08341975920999234, + "flos": 24063026428800.0, + "grad_norm": 2.269593229777068, + "learning_rate": 3.9701709230111455e-06, + "loss": 0.8155, + "num_input_tokens_seen": 30137155, + "step": 925 + }, + { + "epoch": 0.0835099427334626, + "flos": 17687818076160.0, + "grad_norm": 2.5700549892375997, + "learning_rate": 3.970070316300713e-06, + "loss": 0.7696, + "num_input_tokens_seen": 30164485, + "step": 926 + }, + { + "epoch": 0.08360012625693286, + "flos": 35576164239360.0, + "grad_norm": 2.1045901813394408, + "learning_rate": 3.969969541493017e-06, + "loss": 0.8216, + "num_input_tokens_seen": 30194195, + "step": 927 + }, + { + "epoch": 0.08369030978040312, + "flos": 20050737875520.0, + "grad_norm": 2.0605723302633834, + "learning_rate": 3.969868598596658e-06, + "loss": 0.8541, + "num_input_tokens_seen": 30222150, + "step": 928 + }, + { + "epoch": 0.08378049330387338, + "flos": 34081556130720.0, + "grad_norm": 1.500275824936222, + "learning_rate": 3.969767487620249e-06, + "loss": 0.7429, + "num_input_tokens_seen": 30254750, + "step": 929 + }, + { + "epoch": 0.08387067682734364, + "flos": 23257505929920.0, + "grad_norm": 1.9687469119093155, + "learning_rate": 3.969666208572416e-06, + "loss": 0.8504, + "num_input_tokens_seen": 30284795, + "step": 930 + }, + { + "epoch": 0.0839608603508139, + "flos": 23224976656320.0, + "grad_norm": 1.957541576250899, + "learning_rate": 3.969564761461802e-06, + "loss": 0.8678, + "num_input_tokens_seen": 30310190, + "step": 931 + }, + { + "epoch": 0.08405104387428416, + "flos": 26030582760960.0, + "grad_norm": 1.6644169246422738, + "learning_rate": 3.969463146297062e-06, + "loss": 0.7815, + "num_input_tokens_seen": 30342865, + "step": 932 + }, + { + "epoch": 0.08414122739775443, + "flos": 23915216528640.0, + "grad_norm": 4.297376568737624, + "learning_rate": 3.969361363086867e-06, + "loss": 0.8263, + "num_input_tokens_seen": 30366530, + "step": 933 + }, + { + "epoch": 0.08423141092122469, + "flos": 31855890019680.0, + "grad_norm": 3.053481075954685, + "learning_rate": 3.9692594118399014e-06, + "loss": 0.7701, + "num_input_tokens_seen": 30397145, + "step": 934 + }, + { + "epoch": 0.08432159444469496, + "flos": 22715038788000.0, + "grad_norm": 2.238093343306863, + "learning_rate": 3.969157292564865e-06, + "loss": 0.8378, + "num_input_tokens_seen": 30424185, + "step": 935 + }, + { + "epoch": 0.08441177796816522, + "flos": 33059896245600.0, + "grad_norm": 1.762896564861614, + "learning_rate": 3.96905500527047e-06, + "loss": 0.7911, + "num_input_tokens_seen": 30453595, + "step": 936 + }, + { + "epoch": 0.08450196149163548, + "flos": 25229150935680.0, + "grad_norm": 1.9748865191535807, + "learning_rate": 3.968952549965445e-06, + "loss": 0.7897, + "num_input_tokens_seen": 30483890, + "step": 937 + }, + { + "epoch": 0.08459214501510574, + "flos": 20566102528800.0, + "grad_norm": 1.7107471377582792, + "learning_rate": 3.968849926658532e-06, + "loss": 0.8352, + "num_input_tokens_seen": 30511970, + "step": 938 + }, + { + "epoch": 0.084682328538576, + "flos": 23185942674720.0, + "grad_norm": 2.4441740170600346, + "learning_rate": 3.9687471353584866e-06, + "loss": 0.7179, + "num_input_tokens_seen": 30541910, + "step": 939 + }, + { + "epoch": 0.08477251206204627, + "flos": 21549806355360.0, + "grad_norm": 2.420708652334288, + "learning_rate": 3.9686441760740795e-06, + "loss": 0.8203, + "num_input_tokens_seen": 30571305, + "step": 940 + }, + { + "epoch": 0.08486269558551653, + "flos": 31676219901600.0, + "grad_norm": 2.1435737070269534, + "learning_rate": 3.968541048814098e-06, + "loss": 0.7459, + "num_input_tokens_seen": 30604665, + "step": 941 + }, + { + "epoch": 0.08495287910898679, + "flos": 22933148171520.0, + "grad_norm": 2.7751805805183354, + "learning_rate": 3.968437753587339e-06, + "loss": 0.7328, + "num_input_tokens_seen": 30630315, + "step": 942 + }, + { + "epoch": 0.08504306263245705, + "flos": 19653961957440.0, + "grad_norm": 1.590137876254722, + "learning_rate": 3.968334290402616e-06, + "loss": 0.85, + "num_input_tokens_seen": 30658965, + "step": 943 + }, + { + "epoch": 0.08513324615592731, + "flos": 33788203685760.0, + "grad_norm": 1.7279051656396227, + "learning_rate": 3.968230659268759e-06, + "loss": 0.7023, + "num_input_tokens_seen": 30692780, + "step": 944 + }, + { + "epoch": 0.08522342967939757, + "flos": 20237990624640.0, + "grad_norm": 2.0413350854580017, + "learning_rate": 3.968126860194609e-06, + "loss": 0.8769, + "num_input_tokens_seen": 30720260, + "step": 945 + }, + { + "epoch": 0.08531361320286783, + "flos": 27046890200640.0, + "grad_norm": 1.794794819084258, + "learning_rate": 3.968022893189025e-06, + "loss": 0.8591, + "num_input_tokens_seen": 30749690, + "step": 946 + }, + { + "epoch": 0.0854037967263381, + "flos": 20857447806720.0, + "grad_norm": 1.680511201670801, + "learning_rate": 3.967918758260874e-06, + "loss": 0.8396, + "num_input_tokens_seen": 30778415, + "step": 947 + }, + { + "epoch": 0.08549398024980837, + "flos": 17285987070720.0, + "grad_norm": 2.2047760100140548, + "learning_rate": 3.967814455419044e-06, + "loss": 0.8514, + "num_input_tokens_seen": 30804665, + "step": 948 + }, + { + "epoch": 0.08558416377327863, + "flos": 24055741155840.0, + "grad_norm": 2.0376613670058465, + "learning_rate": 3.967709984672434e-06, + "loss": 0.8491, + "num_input_tokens_seen": 30832875, + "step": 949 + }, + { + "epoch": 0.08567434729674889, + "flos": 26722792630560.0, + "grad_norm": 3.0658941339379506, + "learning_rate": 3.967605346029959e-06, + "loss": 0.8368, + "num_input_tokens_seen": 30862650, + "step": 950 + }, + { + "epoch": 0.08576453082021915, + "flos": 21877174864320.0, + "grad_norm": 1.6468727676405426, + "learning_rate": 3.9675005395005466e-06, + "loss": 0.8855, + "num_input_tokens_seen": 30891145, + "step": 951 + }, + { + "epoch": 0.08585471434368941, + "flos": 23115159984480.0, + "grad_norm": 1.9170709750230315, + "learning_rate": 3.967395565093139e-06, + "loss": 0.8285, + "num_input_tokens_seen": 30921200, + "step": 952 + }, + { + "epoch": 0.08594489786715967, + "flos": 22934709301440.0, + "grad_norm": 1.6337768043473195, + "learning_rate": 3.967290422816693e-06, + "loss": 0.8144, + "num_input_tokens_seen": 30949205, + "step": 953 + }, + { + "epoch": 0.08603508139062993, + "flos": 21840371068320.0, + "grad_norm": 2.315642618583341, + "learning_rate": 3.967185112680183e-06, + "loss": 0.7716, + "num_input_tokens_seen": 30976660, + "step": 954 + }, + { + "epoch": 0.08612526491410019, + "flos": 21949518684480.0, + "grad_norm": 2.1240506014786, + "learning_rate": 3.96707963469259e-06, + "loss": 0.8843, + "num_input_tokens_seen": 31002525, + "step": 955 + }, + { + "epoch": 0.08621544843757045, + "flos": 23699114312160.0, + "grad_norm": 1.8161040680260807, + "learning_rate": 3.966973988862917e-06, + "loss": 0.8001, + "num_input_tokens_seen": 31031415, + "step": 956 + }, + { + "epoch": 0.08630563196104071, + "flos": 25079928584640.0, + "grad_norm": 1.5636800506095991, + "learning_rate": 3.966868175200178e-06, + "loss": 0.862, + "num_input_tokens_seen": 31062195, + "step": 957 + }, + { + "epoch": 0.08639581548451097, + "flos": 25809611305920.0, + "grad_norm": 2.143371157077604, + "learning_rate": 3.9667621937134e-06, + "loss": 0.8684, + "num_input_tokens_seen": 31091295, + "step": 958 + }, + { + "epoch": 0.08648599900798125, + "flos": 33935121511680.0, + "grad_norm": 3.6948907784926788, + "learning_rate": 3.966656044411627e-06, + "loss": 0.7202, + "num_input_tokens_seen": 31122600, + "step": 959 + }, + { + "epoch": 0.08657618253145151, + "flos": 19248897182880.0, + "grad_norm": 1.5848735378609748, + "learning_rate": 3.966549727303918e-06, + "loss": 0.7805, + "num_input_tokens_seen": 31149590, + "step": 960 + }, + { + "epoch": 0.08666636605492177, + "flos": 23297989532160.0, + "grad_norm": 2.728778364983836, + "learning_rate": 3.966443242399341e-06, + "loss": 0.8414, + "num_input_tokens_seen": 31178055, + "step": 961 + }, + { + "epoch": 0.08675654957839203, + "flos": 61322575486080.0, + "grad_norm": 0.717374937542389, + "learning_rate": 3.966336589706985e-06, + "loss": 0.5985, + "num_input_tokens_seen": 31271395, + "step": 962 + }, + { + "epoch": 0.08684673310186229, + "flos": 26939043526080.0, + "grad_norm": 2.5027386923261985, + "learning_rate": 3.966229769235948e-06, + "loss": 0.8246, + "num_input_tokens_seen": 31299855, + "step": 963 + }, + { + "epoch": 0.08693691662533255, + "flos": 17177248321920.0, + "grad_norm": 2.0614161099785275, + "learning_rate": 3.966122780995345e-06, + "loss": 0.8887, + "num_input_tokens_seen": 31325450, + "step": 964 + }, + { + "epoch": 0.08702710014880281, + "flos": 19144098428640.0, + "grad_norm": 2.2229535753736744, + "learning_rate": 3.966015624994306e-06, + "loss": 0.8412, + "num_input_tokens_seen": 31352000, + "step": 965 + }, + { + "epoch": 0.08711728367227307, + "flos": 21148124028960.0, + "grad_norm": 2.0767623908491926, + "learning_rate": 3.9659083012419735e-06, + "loss": 0.7837, + "num_input_tokens_seen": 31379710, + "step": 966 + }, + { + "epoch": 0.08720746719574334, + "flos": 42679717201920.0, + "grad_norm": 2.5716774513511296, + "learning_rate": 3.965800809747505e-06, + "loss": 0.7422, + "num_input_tokens_seen": 31410870, + "step": 967 + }, + { + "epoch": 0.0872976507192136, + "flos": 15282890714400.0, + "grad_norm": 1.712343935572707, + "learning_rate": 3.965693150520071e-06, + "loss": 0.779, + "num_input_tokens_seen": 31437990, + "step": 968 + }, + { + "epoch": 0.08738783424268386, + "flos": 14044273708320.0, + "grad_norm": 3.536080895995925, + "learning_rate": 3.96558532356886e-06, + "loss": 0.7475, + "num_input_tokens_seen": 31463975, + "step": 969 + }, + { + "epoch": 0.08747801776615412, + "flos": 17760422084640.0, + "grad_norm": 5.890506166428189, + "learning_rate": 3.9654773289030704e-06, + "loss": 0.8568, + "num_input_tokens_seen": 31488785, + "step": 970 + }, + { + "epoch": 0.08756820128962439, + "flos": 26613124637760.0, + "grad_norm": 2.0117580965508965, + "learning_rate": 3.9653691665319176e-06, + "loss": 0.8244, + "num_input_tokens_seen": 31518530, + "step": 971 + }, + { + "epoch": 0.08765838481309465, + "flos": 39473841221760.0, + "grad_norm": 1.9908009809140945, + "learning_rate": 3.96526083646463e-06, + "loss": 0.739, + "num_input_tokens_seen": 31551320, + "step": 972 + }, + { + "epoch": 0.08774856833656491, + "flos": 34444055796480.0, + "grad_norm": 2.4269113657091097, + "learning_rate": 3.9651523387104526e-06, + "loss": 0.7625, + "num_input_tokens_seen": 31584265, + "step": 973 + }, + { + "epoch": 0.08783875186003517, + "flos": 19910287587840.0, + "grad_norm": 2.2250946024810347, + "learning_rate": 3.965043673278641e-06, + "loss": 0.8237, + "num_input_tokens_seen": 31613170, + "step": 974 + }, + { + "epoch": 0.08792893538350544, + "flos": 26904432745920.0, + "grad_norm": 3.416777554705223, + "learning_rate": 3.964934840178469e-06, + "loss": 0.8166, + "num_input_tokens_seen": 31643465, + "step": 975 + }, + { + "epoch": 0.0880191189069757, + "flos": 21689736266400.0, + "grad_norm": 1.784693618360481, + "learning_rate": 3.964825839419221e-06, + "loss": 0.8606, + "num_input_tokens_seen": 31670050, + "step": 976 + }, + { + "epoch": 0.08810930243044596, + "flos": 21148495726560.0, + "grad_norm": 2.2179746750674867, + "learning_rate": 3.964716671010199e-06, + "loss": 0.7944, + "num_input_tokens_seen": 31696740, + "step": 977 + }, + { + "epoch": 0.08819948595391622, + "flos": 27779583672480.0, + "grad_norm": 1.5910451038232394, + "learning_rate": 3.9646073349607165e-06, + "loss": 0.7759, + "num_input_tokens_seen": 31729405, + "step": 978 + }, + { + "epoch": 0.08828966947738648, + "flos": 16302246074400.0, + "grad_norm": 2.1875080653450647, + "learning_rate": 3.964497831280105e-06, + "loss": 0.9417, + "num_input_tokens_seen": 31756330, + "step": 979 + }, + { + "epoch": 0.08837985300085674, + "flos": 18525421811520.0, + "grad_norm": 3.2764875455258986, + "learning_rate": 3.964388159977705e-06, + "loss": 0.8134, + "num_input_tokens_seen": 31780990, + "step": 980 + }, + { + "epoch": 0.088470036524327, + "flos": 29451631713600.0, + "grad_norm": 1.4916533467783177, + "learning_rate": 3.964278321062876e-06, + "loss": 0.7154, + "num_input_tokens_seen": 31812755, + "step": 981 + }, + { + "epoch": 0.08856022004779726, + "flos": 24604898854560.0, + "grad_norm": 1.7068179627997786, + "learning_rate": 3.96416831454499e-06, + "loss": 0.8303, + "num_input_tokens_seen": 31839880, + "step": 982 + }, + { + "epoch": 0.08865040357126754, + "flos": 68337164012160.0, + "grad_norm": 0.6991876350171508, + "learning_rate": 3.964058140433434e-06, + "loss": 0.6297, + "num_input_tokens_seen": 31932035, + "step": 983 + }, + { + "epoch": 0.0887405870947378, + "flos": 20237433078240.0, + "grad_norm": 1.7839410097094148, + "learning_rate": 3.963947798737606e-06, + "loss": 0.8727, + "num_input_tokens_seen": 31958775, + "step": 984 + }, + { + "epoch": 0.08883077061820806, + "flos": 13969365174720.0, + "grad_norm": 1.8431336617355192, + "learning_rate": 3.963837289466923e-06, + "loss": 0.83, + "num_input_tokens_seen": 31980445, + "step": 985 + }, + { + "epoch": 0.08892095414167832, + "flos": 29345308999200.0, + "grad_norm": 2.0477719270996664, + "learning_rate": 3.9637266126308145e-06, + "loss": 0.784, + "num_input_tokens_seen": 32012105, + "step": 986 + }, + { + "epoch": 0.08901113766514858, + "flos": 61511203516320.0, + "grad_norm": 0.7208950128424861, + "learning_rate": 3.963615768238724e-06, + "loss": 0.6184, + "num_input_tokens_seen": 32104065, + "step": 987 + }, + { + "epoch": 0.08910132118861884, + "flos": 67234728504960.0, + "grad_norm": 1.6911580738423142, + "learning_rate": 3.963504756300107e-06, + "loss": 0.6979, + "num_input_tokens_seen": 32138965, + "step": 988 + }, + { + "epoch": 0.0891915047120891, + "flos": 26645802590400.0, + "grad_norm": 1.6698277140165332, + "learning_rate": 3.96339357682444e-06, + "loss": 0.8497, + "num_input_tokens_seen": 32168365, + "step": 989 + }, + { + "epoch": 0.08928168823555936, + "flos": 33021903017280.0, + "grad_norm": 1.40589226305058, + "learning_rate": 3.963282229821206e-06, + "loss": 0.8272, + "num_input_tokens_seen": 32201585, + "step": 990 + }, + { + "epoch": 0.08937187175902962, + "flos": 18890226002400.0, + "grad_norm": 1.689455735750331, + "learning_rate": 3.963170715299906e-06, + "loss": 0.8234, + "num_input_tokens_seen": 32229700, + "step": 991 + }, + { + "epoch": 0.08946205528249988, + "flos": 26467507753440.0, + "grad_norm": 1.6099747216271951, + "learning_rate": 3.963059033270056e-06, + "loss": 0.7347, + "num_input_tokens_seen": 32259460, + "step": 992 + }, + { + "epoch": 0.08955223880597014, + "flos": 24312884520960.0, + "grad_norm": 1.722437353780498, + "learning_rate": 3.9629471837411855e-06, + "loss": 0.7591, + "num_input_tokens_seen": 32288865, + "step": 993 + }, + { + "epoch": 0.0896424223294404, + "flos": 24791965754880.0, + "grad_norm": 1.6984364780679135, + "learning_rate": 3.962835166722838e-06, + "loss": 0.7558, + "num_input_tokens_seen": 32318975, + "step": 994 + }, + { + "epoch": 0.08973260585291068, + "flos": 22714927278720.0, + "grad_norm": 1.8169685395854778, + "learning_rate": 3.96272298222457e-06, + "loss": 0.8193, + "num_input_tokens_seen": 32346540, + "step": 995 + }, + { + "epoch": 0.08982278937638094, + "flos": 22237109816640.0, + "grad_norm": 1.8239783183741654, + "learning_rate": 3.962610630255956e-06, + "loss": 0.9378, + "num_input_tokens_seen": 32374140, + "step": 996 + }, + { + "epoch": 0.0899129728998512, + "flos": 54662495127360.0, + "grad_norm": 2.070910673277125, + "learning_rate": 3.96249811082658e-06, + "loss": 0.7789, + "num_input_tokens_seen": 32408295, + "step": 997 + }, + { + "epoch": 0.09000315642332146, + "flos": 50555623770240.0, + "grad_norm": 0.8063006850878286, + "learning_rate": 3.962385423946046e-06, + "loss": 0.6734, + "num_input_tokens_seen": 32492565, + "step": 998 + }, + { + "epoch": 0.09009333994679172, + "flos": 29927739366720.0, + "grad_norm": 1.4734049950227228, + "learning_rate": 3.962272569623966e-06, + "loss": 0.7846, + "num_input_tokens_seen": 32522910, + "step": 999 + }, + { + "epoch": 0.09018352347026198, + "flos": 20196800796960.0, + "grad_norm": 1.806524753463259, + "learning_rate": 3.9621595478699704e-06, + "loss": 0.7023, + "num_input_tokens_seen": 32548530, + "step": 1000 + }, + { + "epoch": 0.09027370699373224, + "flos": 15902831103360.0, + "grad_norm": 1.7090777039248328, + "learning_rate": 3.962046358693703e-06, + "loss": 0.8051, + "num_input_tokens_seen": 32576780, + "step": 1001 + }, + { + "epoch": 0.0903638905172025, + "flos": 15464976866880.0, + "grad_norm": 1.8998943751238375, + "learning_rate": 3.961933002104822e-06, + "loss": 0.8911, + "num_input_tokens_seen": 32602345, + "step": 1002 + }, + { + "epoch": 0.09045407404067277, + "flos": 24791333868960.0, + "grad_norm": 2.396981956353546, + "learning_rate": 3.961819478112999e-06, + "loss": 0.7361, + "num_input_tokens_seen": 32630660, + "step": 1003 + }, + { + "epoch": 0.09054425756414303, + "flos": 22203651299040.0, + "grad_norm": 1.9466961455508194, + "learning_rate": 3.961705786727921e-06, + "loss": 0.8332, + "num_input_tokens_seen": 32659095, + "step": 1004 + }, + { + "epoch": 0.09063444108761329, + "flos": 29528324395680.0, + "grad_norm": 2.090451275231894, + "learning_rate": 3.961591927959288e-06, + "loss": 0.734, + "num_input_tokens_seen": 32689065, + "step": 1005 + }, + { + "epoch": 0.09072462461108355, + "flos": 68435904095520.0, + "grad_norm": 0.7148235297339802, + "learning_rate": 3.961477901816816e-06, + "loss": 0.5937, + "num_input_tokens_seen": 32782835, + "step": 1006 + }, + { + "epoch": 0.09081480813455382, + "flos": 16740137480640.0, + "grad_norm": 1.9895372302187853, + "learning_rate": 3.961363708310233e-06, + "loss": 0.874, + "num_input_tokens_seen": 32808945, + "step": 1007 + }, + { + "epoch": 0.09090499165802408, + "flos": 20929940305920.0, + "grad_norm": 2.0379331537875918, + "learning_rate": 3.961249347449286e-06, + "loss": 0.8369, + "num_input_tokens_seen": 32837745, + "step": 1008 + }, + { + "epoch": 0.09099517518149434, + "flos": 20998418471040.0, + "grad_norm": 1.6141587796442796, + "learning_rate": 3.961134819243728e-06, + "loss": 0.7937, + "num_input_tokens_seen": 32867425, + "step": 1009 + }, + { + "epoch": 0.0910853587049646, + "flos": 21768101587680.0, + "grad_norm": 2.221937656726313, + "learning_rate": 3.961020123703335e-06, + "loss": 0.8174, + "num_input_tokens_seen": 32897060, + "step": 1010 + }, + { + "epoch": 0.09117554222843487, + "flos": 21403966452480.0, + "grad_norm": 2.0163321031005523, + "learning_rate": 3.960905260837892e-06, + "loss": 0.8209, + "num_input_tokens_seen": 32925585, + "step": 1011 + }, + { + "epoch": 0.09126572575190513, + "flos": 25483655247840.0, + "grad_norm": 1.8251770900349378, + "learning_rate": 3.960790230657199e-06, + "loss": 0.78, + "num_input_tokens_seen": 32954945, + "step": 1012 + }, + { + "epoch": 0.09135590927537539, + "flos": 40672011795360.0, + "grad_norm": 1.4227139020840514, + "learning_rate": 3.960675033171072e-06, + "loss": 0.7566, + "num_input_tokens_seen": 32987665, + "step": 1013 + }, + { + "epoch": 0.09144609279884565, + "flos": 31381417836000.0, + "grad_norm": 1.5469972966718573, + "learning_rate": 3.960559668389341e-06, + "loss": 0.7535, + "num_input_tokens_seen": 33018775, + "step": 1014 + }, + { + "epoch": 0.09153627632231591, + "flos": 20200926640320.0, + "grad_norm": 1.7757106110754333, + "learning_rate": 3.960444136321847e-06, + "loss": 0.8117, + "num_input_tokens_seen": 33044725, + "step": 1015 + }, + { + "epoch": 0.09162645984578617, + "flos": 22714815769440.0, + "grad_norm": 2.0211031494393916, + "learning_rate": 3.960328436978451e-06, + "loss": 0.8955, + "num_input_tokens_seen": 33072980, + "step": 1016 + }, + { + "epoch": 0.09171664336925643, + "flos": 62461709013600.0, + "grad_norm": 0.655175623863157, + "learning_rate": 3.960212570369024e-06, + "loss": 0.5943, + "num_input_tokens_seen": 33160125, + "step": 1017 + }, + { + "epoch": 0.09180682689272669, + "flos": 21731446470720.0, + "grad_norm": 1.7464560318508893, + "learning_rate": 3.9600965365034515e-06, + "loss": 0.7479, + "num_input_tokens_seen": 33189685, + "step": 1018 + }, + { + "epoch": 0.09189701041619697, + "flos": 20673614675520.0, + "grad_norm": 1.9332047504657934, + "learning_rate": 3.959980335391634e-06, + "loss": 0.8605, + "num_input_tokens_seen": 33218220, + "step": 1019 + }, + { + "epoch": 0.09198719393966723, + "flos": 20561753666880.0, + "grad_norm": 1.7399904912929953, + "learning_rate": 3.959863967043487e-06, + "loss": 0.7481, + "num_input_tokens_seen": 33246165, + "step": 1020 + }, + { + "epoch": 0.09207737746313749, + "flos": 22163093357280.0, + "grad_norm": 1.8166287849862988, + "learning_rate": 3.9597474314689405e-06, + "loss": 0.8371, + "num_input_tokens_seen": 33274855, + "step": 1021 + }, + { + "epoch": 0.09216756098660775, + "flos": 20925554274240.0, + "grad_norm": 1.9180326431726544, + "learning_rate": 3.959630728677937e-06, + "loss": 0.8284, + "num_input_tokens_seen": 33302430, + "step": 1022 + }, + { + "epoch": 0.09225774451007801, + "flos": 19581246439680.0, + "grad_norm": 1.6507557843072451, + "learning_rate": 3.959513858680434e-06, + "loss": 0.7857, + "num_input_tokens_seen": 33329455, + "step": 1023 + }, + { + "epoch": 0.09234792803354827, + "flos": 18846731649600.0, + "grad_norm": 1.9879376727260114, + "learning_rate": 3.959396821486405e-06, + "loss": 0.8988, + "num_input_tokens_seen": 33353565, + "step": 1024 + }, + { + "epoch": 0.09243811155701853, + "flos": 66527830846560.0, + "grad_norm": 0.8170644124038668, + "learning_rate": 3.959279617105835e-06, + "loss": 0.6243, + "num_input_tokens_seen": 33429600, + "step": 1025 + }, + { + "epoch": 0.09252829508048879, + "flos": 32369321845440.0, + "grad_norm": 1.9051549298821469, + "learning_rate": 3.9591622455487235e-06, + "loss": 0.8174, + "num_input_tokens_seen": 33458965, + "step": 1026 + }, + { + "epoch": 0.09261847860395905, + "flos": 18120877413600.0, + "grad_norm": 1.7494604777679588, + "learning_rate": 3.959044706825087e-06, + "loss": 0.8826, + "num_input_tokens_seen": 33485085, + "step": 1027 + }, + { + "epoch": 0.09270866212742931, + "flos": 29414604899040.0, + "grad_norm": 3.2836719112455683, + "learning_rate": 3.958927000944954e-06, + "loss": 0.7158, + "num_input_tokens_seen": 33517580, + "step": 1028 + }, + { + "epoch": 0.09279884565089958, + "flos": 27014509606080.0, + "grad_norm": 2.034276497561301, + "learning_rate": 3.958809127918368e-06, + "loss": 0.8949, + "num_input_tokens_seen": 33547820, + "step": 1029 + }, + { + "epoch": 0.09288902917436984, + "flos": 21149313461280.0, + "grad_norm": 1.5828315876810577, + "learning_rate": 3.958691087755387e-06, + "loss": 0.7456, + "num_input_tokens_seen": 33577265, + "step": 1030 + }, + { + "epoch": 0.09297921269784011, + "flos": 27415225518720.0, + "grad_norm": 1.9416649260397163, + "learning_rate": 3.958572880466081e-06, + "loss": 0.6838, + "num_input_tokens_seen": 33606580, + "step": 1031 + }, + { + "epoch": 0.09306939622131037, + "flos": 23334272951520.0, + "grad_norm": 1.6738357454747403, + "learning_rate": 3.9584545060605385e-06, + "loss": 0.7942, + "num_input_tokens_seen": 33636595, + "step": 1032 + }, + { + "epoch": 0.09315957974478063, + "flos": 22204915070880.0, + "grad_norm": 1.6373621147284203, + "learning_rate": 3.958335964548859e-06, + "loss": 0.774, + "num_input_tokens_seen": 33663610, + "step": 1033 + }, + { + "epoch": 0.0932497632682509, + "flos": 20234087799840.0, + "grad_norm": 1.9228314136633204, + "learning_rate": 3.958217255941156e-06, + "loss": 0.858, + "num_input_tokens_seen": 33691620, + "step": 1034 + }, + { + "epoch": 0.09333994679172115, + "flos": 37179325248000.0, + "grad_norm": 1.7053371376912798, + "learning_rate": 3.95809838024756e-06, + "loss": 0.7595, + "num_input_tokens_seen": 33725895, + "step": 1035 + }, + { + "epoch": 0.09343013031519141, + "flos": 21731855338080.0, + "grad_norm": 1.7527225977733851, + "learning_rate": 3.957979337478212e-06, + "loss": 0.7755, + "num_input_tokens_seen": 33753810, + "step": 1036 + }, + { + "epoch": 0.09352031383866168, + "flos": 16849508115360.0, + "grad_norm": 3.0437325729717193, + "learning_rate": 3.957860127643272e-06, + "loss": 0.827, + "num_input_tokens_seen": 33781825, + "step": 1037 + }, + { + "epoch": 0.09361049736213194, + "flos": 21293815252800.0, + "grad_norm": 2.0054345205645516, + "learning_rate": 3.95774075075291e-06, + "loss": 0.8127, + "num_input_tokens_seen": 33810785, + "step": 1038 + }, + { + "epoch": 0.0937006808856022, + "flos": 20820681180480.0, + "grad_norm": 1.9765890971569264, + "learning_rate": 3.957621206817312e-06, + "loss": 0.8661, + "num_input_tokens_seen": 33837630, + "step": 1039 + }, + { + "epoch": 0.09379086440907246, + "flos": 23773205111040.0, + "grad_norm": 2.724527381018707, + "learning_rate": 3.957501495846679e-06, + "loss": 0.7922, + "num_input_tokens_seen": 33867465, + "step": 1040 + }, + { + "epoch": 0.09388104793254272, + "flos": 68154631822560.0, + "grad_norm": 0.6870542077954159, + "learning_rate": 3.957381617851225e-06, + "loss": 0.5991, + "num_input_tokens_seen": 33951650, + "step": 1041 + }, + { + "epoch": 0.09397123145601298, + "flos": 42169742163840.0, + "grad_norm": 1.5771722467001212, + "learning_rate": 3.9572615728411776e-06, + "loss": 0.7318, + "num_input_tokens_seen": 33984345, + "step": 1042 + }, + { + "epoch": 0.09406141497948325, + "flos": 21840817105440.0, + "grad_norm": 7.576439738916402, + "learning_rate": 3.957141360826781e-06, + "loss": 0.6945, + "num_input_tokens_seen": 34009450, + "step": 1043 + }, + { + "epoch": 0.09415159850295352, + "flos": 21767655550560.0, + "grad_norm": 3.163328636683149, + "learning_rate": 3.957020981818292e-06, + "loss": 0.7916, + "num_input_tokens_seen": 34036540, + "step": 1044 + }, + { + "epoch": 0.09424178202642378, + "flos": 29782233991680.0, + "grad_norm": 2.6409308803747575, + "learning_rate": 3.956900435825982e-06, + "loss": 0.7993, + "num_input_tokens_seen": 34069285, + "step": 1045 + }, + { + "epoch": 0.09433196554989404, + "flos": 23114862626400.0, + "grad_norm": 1.5165866278743896, + "learning_rate": 3.9567797228601364e-06, + "loss": 0.7599, + "num_input_tokens_seen": 34099545, + "step": 1046 + }, + { + "epoch": 0.0944221490733643, + "flos": 23371336935840.0, + "grad_norm": 1.7179088268090346, + "learning_rate": 3.956658842931055e-06, + "loss": 0.8816, + "num_input_tokens_seen": 34128905, + "step": 1047 + }, + { + "epoch": 0.09451233259683456, + "flos": 23443197549120.0, + "grad_norm": 1.5434252099990053, + "learning_rate": 3.956537796049052e-06, + "loss": 0.8235, + "num_input_tokens_seen": 34158920, + "step": 1048 + }, + { + "epoch": 0.09460251612030482, + "flos": 68588434555200.0, + "grad_norm": 0.7183814140287046, + "learning_rate": 3.956416582224457e-06, + "loss": 0.6188, + "num_input_tokens_seen": 34259480, + "step": 1049 + }, + { + "epoch": 0.09469269964377508, + "flos": 20529447411840.0, + "grad_norm": 2.2048335383202198, + "learning_rate": 3.956295201467611e-06, + "loss": 0.8454, + "num_input_tokens_seen": 34285505, + "step": 1050 + }, + { + "epoch": 0.09478288316724534, + "flos": 21766205929920.0, + "grad_norm": 1.5770358113327136, + "learning_rate": 3.956173653788872e-06, + "loss": 0.8386, + "num_input_tokens_seen": 34315150, + "step": 1051 + }, + { + "epoch": 0.0948730666907156, + "flos": 29381443739520.0, + "grad_norm": 2.2559159176828887, + "learning_rate": 3.95605193919861e-06, + "loss": 0.8453, + "num_input_tokens_seen": 34341465, + "step": 1052 + }, + { + "epoch": 0.09496325021418586, + "flos": 26832274774560.0, + "grad_norm": 1.7050968619209628, + "learning_rate": 3.955930057707211e-06, + "loss": 0.7962, + "num_input_tokens_seen": 34369440, + "step": 1053 + }, + { + "epoch": 0.09505343373765612, + "flos": 22673625941760.0, + "grad_norm": 2.0405819664896803, + "learning_rate": 3.955808009325075e-06, + "loss": 0.8209, + "num_input_tokens_seen": 34398550, + "step": 1054 + }, + { + "epoch": 0.0951436172611264, + "flos": 39000595640160.0, + "grad_norm": 2.499743448445147, + "learning_rate": 3.955685794062615e-06, + "loss": 0.7806, + "num_input_tokens_seen": 34429920, + "step": 1055 + }, + { + "epoch": 0.09523380078459666, + "flos": 21585346379520.0, + "grad_norm": 2.210864062488129, + "learning_rate": 3.95556341193026e-06, + "loss": 0.9079, + "num_input_tokens_seen": 34458980, + "step": 1056 + }, + { + "epoch": 0.09532398430806692, + "flos": 26502007024320.0, + "grad_norm": 1.8692834175758364, + "learning_rate": 3.955440862938452e-06, + "loss": 0.9041, + "num_input_tokens_seen": 34487540, + "step": 1057 + }, + { + "epoch": 0.09541416783153718, + "flos": 22605556644000.0, + "grad_norm": 2.0522663660929203, + "learning_rate": 3.955318147097647e-06, + "loss": 0.8077, + "num_input_tokens_seen": 34512720, + "step": 1058 + }, + { + "epoch": 0.09550435135500744, + "flos": 20193195330240.0, + "grad_norm": 1.888577448033752, + "learning_rate": 3.955195264418316e-06, + "loss": 0.7729, + "num_input_tokens_seen": 34538510, + "step": 1059 + }, + { + "epoch": 0.0955945348784777, + "flos": 24353962839360.0, + "grad_norm": 1.5654016636215888, + "learning_rate": 3.955072214910944e-06, + "loss": 0.6684, + "num_input_tokens_seen": 34567395, + "step": 1060 + }, + { + "epoch": 0.09568471840194796, + "flos": 25229150935680.0, + "grad_norm": 2.451756478312839, + "learning_rate": 3.954948998586032e-06, + "loss": 0.7321, + "num_input_tokens_seen": 34595090, + "step": 1061 + }, + { + "epoch": 0.09577490192541822, + "flos": 32040838243680.0, + "grad_norm": 1.895626671674314, + "learning_rate": 3.954825615454089e-06, + "loss": 0.8159, + "num_input_tokens_seen": 34626550, + "step": 1062 + }, + { + "epoch": 0.09586508544888848, + "flos": 21432630071040.0, + "grad_norm": 1.7714925847814862, + "learning_rate": 3.954702065525649e-06, + "loss": 0.8038, + "num_input_tokens_seen": 34653420, + "step": 1063 + }, + { + "epoch": 0.09595526897235875, + "flos": 18489249901440.0, + "grad_norm": 2.061899849243839, + "learning_rate": 3.954578348811248e-06, + "loss": 0.835, + "num_input_tokens_seen": 34679070, + "step": 1064 + }, + { + "epoch": 0.096045452495829, + "flos": 32879519902080.0, + "grad_norm": 1.8257104300857587, + "learning_rate": 3.954454465321447e-06, + "loss": 0.7834, + "num_input_tokens_seen": 34707800, + "step": 1065 + }, + { + "epoch": 0.09613563601929928, + "flos": 15719667027840.0, + "grad_norm": 2.660832659391451, + "learning_rate": 3.954330415066813e-06, + "loss": 0.8906, + "num_input_tokens_seen": 34733180, + "step": 1066 + }, + { + "epoch": 0.09622581954276954, + "flos": 20196317590080.0, + "grad_norm": 2.322643509158473, + "learning_rate": 3.954206198057932e-06, + "loss": 0.7945, + "num_input_tokens_seen": 34762870, + "step": 1067 + }, + { + "epoch": 0.0963160030662398, + "flos": 28689791416320.0, + "grad_norm": 1.659487457200406, + "learning_rate": 3.954081814305403e-06, + "loss": 0.8664, + "num_input_tokens_seen": 34792750, + "step": 1068 + }, + { + "epoch": 0.09640618658971006, + "flos": 23768001344640.0, + "grad_norm": 1.7116777046885268, + "learning_rate": 3.953957263819839e-06, + "loss": 0.7465, + "num_input_tokens_seen": 34823260, + "step": 1069 + }, + { + "epoch": 0.09649637011318032, + "flos": 28908755704320.0, + "grad_norm": 1.6746594780344466, + "learning_rate": 3.953832546611867e-06, + "loss": 0.7436, + "num_input_tokens_seen": 34853175, + "step": 1070 + }, + { + "epoch": 0.09658655363665059, + "flos": 23808447777120.0, + "grad_norm": 2.3535542211889693, + "learning_rate": 3.953707662692129e-06, + "loss": 0.8358, + "num_input_tokens_seen": 34883135, + "step": 1071 + }, + { + "epoch": 0.09667673716012085, + "flos": 25630089866880.0, + "grad_norm": 2.385418176997566, + "learning_rate": 3.95358261207128e-06, + "loss": 0.739, + "num_input_tokens_seen": 34913695, + "step": 1072 + }, + { + "epoch": 0.0967669206835911, + "flos": 21727878173760.0, + "grad_norm": 1.647665412483044, + "learning_rate": 3.953457394759992e-06, + "loss": 0.8008, + "num_input_tokens_seen": 34943570, + "step": 1073 + }, + { + "epoch": 0.09685710420706137, + "flos": 24900778843200.0, + "grad_norm": 1.6714307961431656, + "learning_rate": 3.953332010768947e-06, + "loss": 0.8227, + "num_input_tokens_seen": 34970055, + "step": 1074 + }, + { + "epoch": 0.09694728773053163, + "flos": 25229002256640.0, + "grad_norm": 2.6235432897143838, + "learning_rate": 3.9532064601088436e-06, + "loss": 0.7732, + "num_input_tokens_seen": 34995740, + "step": 1075 + }, + { + "epoch": 0.09703747125400189, + "flos": 18196938209760.0, + "grad_norm": 3.7796347454752954, + "learning_rate": 3.953080742790396e-06, + "loss": 0.9563, + "num_input_tokens_seen": 35020250, + "step": 1076 + }, + { + "epoch": 0.09712765477747215, + "flos": 66301835740320.0, + "grad_norm": 0.7631252508139608, + "learning_rate": 3.95295485882433e-06, + "loss": 0.6606, + "num_input_tokens_seen": 35115530, + "step": 1077 + }, + { + "epoch": 0.09721783830094242, + "flos": 70827450343680.0, + "grad_norm": 0.6775273603284836, + "learning_rate": 3.952828808221387e-06, + "loss": 0.6279, + "num_input_tokens_seen": 35209785, + "step": 1078 + }, + { + "epoch": 0.09730802182441269, + "flos": 33389160412320.0, + "grad_norm": 4.907389125883224, + "learning_rate": 3.9527025909923225e-06, + "loss": 0.7874, + "num_input_tokens_seen": 35240030, + "step": 1079 + }, + { + "epoch": 0.09739820534788295, + "flos": 23916926337600.0, + "grad_norm": 1.6099551738398439, + "learning_rate": 3.952576207147906e-06, + "loss": 0.9007, + "num_input_tokens_seen": 35267820, + "step": 1080 + }, + { + "epoch": 0.09748838887135321, + "flos": 26103521297280.0, + "grad_norm": 1.8801847616674783, + "learning_rate": 3.95244965669892e-06, + "loss": 0.8467, + "num_input_tokens_seen": 35299050, + "step": 1081 + }, + { + "epoch": 0.09757857239482347, + "flos": 40822051881120.0, + "grad_norm": 2.763490877189231, + "learning_rate": 3.952322939656165e-06, + "loss": 0.7637, + "num_input_tokens_seen": 35331340, + "step": 1082 + }, + { + "epoch": 0.09766875591829373, + "flos": 20781461350080.0, + "grad_norm": 1.8719412331812153, + "learning_rate": 3.952196056030451e-06, + "loss": 0.8048, + "num_input_tokens_seen": 35361095, + "step": 1083 + }, + { + "epoch": 0.09775893944176399, + "flos": 22607117773920.0, + "grad_norm": 2.488034030963684, + "learning_rate": 3.952069005832605e-06, + "loss": 0.7749, + "num_input_tokens_seen": 35391975, + "step": 1084 + }, + { + "epoch": 0.09784912296523425, + "flos": 24901336389600.0, + "grad_norm": 1.9333283829899692, + "learning_rate": 3.951941789073468e-06, + "loss": 0.8319, + "num_input_tokens_seen": 35421600, + "step": 1085 + }, + { + "epoch": 0.09793930648870451, + "flos": 22168259953920.0, + "grad_norm": 1.7913136541073307, + "learning_rate": 3.9518144057638955e-06, + "loss": 0.8338, + "num_input_tokens_seen": 35449130, + "step": 1086 + }, + { + "epoch": 0.09802949001217477, + "flos": 22095618775680.0, + "grad_norm": 1.8324394503344021, + "learning_rate": 3.951686855914755e-06, + "loss": 0.8014, + "num_input_tokens_seen": 35476465, + "step": 1087 + }, + { + "epoch": 0.09811967353564503, + "flos": 30803224821120.0, + "grad_norm": 2.25980381780873, + "learning_rate": 3.9515591395369305e-06, + "loss": 0.8328, + "num_input_tokens_seen": 35505745, + "step": 1088 + }, + { + "epoch": 0.0982098570591153, + "flos": 24900592994400.0, + "grad_norm": 1.5219385842703284, + "learning_rate": 3.95143125664132e-06, + "loss": 0.7939, + "num_input_tokens_seen": 35535255, + "step": 1089 + }, + { + "epoch": 0.09830004058258557, + "flos": 26430964145760.0, + "grad_norm": 1.975275638515803, + "learning_rate": 3.951303207238833e-06, + "loss": 0.8326, + "num_input_tokens_seen": 35563650, + "step": 1090 + }, + { + "epoch": 0.09839022410605583, + "flos": 21838735598880.0, + "grad_norm": 1.939849843673568, + "learning_rate": 3.951174991340399e-06, + "loss": 0.7815, + "num_input_tokens_seen": 35592225, + "step": 1091 + }, + { + "epoch": 0.09848040762952609, + "flos": 24609545074560.0, + "grad_norm": 1.8081227424573205, + "learning_rate": 3.9510466089569546e-06, + "loss": 0.7683, + "num_input_tokens_seen": 35623955, + "step": 1092 + }, + { + "epoch": 0.09857059115299635, + "flos": 21471478203840.0, + "grad_norm": 1.618556967754744, + "learning_rate": 3.950918060099456e-06, + "loss": 0.7713, + "num_input_tokens_seen": 35652505, + "step": 1093 + }, + { + "epoch": 0.09866077467646661, + "flos": 18197755944480.0, + "grad_norm": 1.6865670486227704, + "learning_rate": 3.950789344778871e-06, + "loss": 0.8575, + "num_input_tokens_seen": 35680665, + "step": 1094 + }, + { + "epoch": 0.09875095819993687, + "flos": 24347904168480.0, + "grad_norm": 1.6289958974928274, + "learning_rate": 3.950660463006184e-06, + "loss": 0.7485, + "num_input_tokens_seen": 35710405, + "step": 1095 + }, + { + "epoch": 0.09884114172340713, + "flos": 29234600253120.0, + "grad_norm": 1.4939117056814128, + "learning_rate": 3.950531414792389e-06, + "loss": 0.8741, + "num_input_tokens_seen": 35739640, + "step": 1096 + }, + { + "epoch": 0.0989313252468774, + "flos": 22751619565440.0, + "grad_norm": 1.639008661336863, + "learning_rate": 3.950402200148498e-06, + "loss": 0.8122, + "num_input_tokens_seen": 35769110, + "step": 1097 + }, + { + "epoch": 0.09902150877034765, + "flos": 25624514402880.0, + "grad_norm": 4.600139715311984, + "learning_rate": 3.950272819085538e-06, + "loss": 0.7833, + "num_input_tokens_seen": 35798030, + "step": 1098 + }, + { + "epoch": 0.09911169229381792, + "flos": 15355420383360.0, + "grad_norm": 3.7405414823738905, + "learning_rate": 3.9501432716145474e-06, + "loss": 0.8923, + "num_input_tokens_seen": 35825300, + "step": 1099 + }, + { + "epoch": 0.09920187581728818, + "flos": 20091816193920.0, + "grad_norm": 1.8976657381915947, + "learning_rate": 3.950013557746579e-06, + "loss": 0.7023, + "num_input_tokens_seen": 35853555, + "step": 1100 + }, + { + "epoch": 0.09929205934075844, + "flos": 22824483762240.0, + "grad_norm": 1.6314927019241872, + "learning_rate": 3.949883677492703e-06, + "loss": 0.8425, + "num_input_tokens_seen": 35883485, + "step": 1101 + }, + { + "epoch": 0.09938224286422871, + "flos": 22933928736480.0, + "grad_norm": 2.4406216512698595, + "learning_rate": 3.9497536308639994e-06, + "loss": 0.825, + "num_input_tokens_seen": 35912730, + "step": 1102 + }, + { + "epoch": 0.09947242638769897, + "flos": 23294607084000.0, + "grad_norm": 2.353285351643218, + "learning_rate": 3.949623417871565e-06, + "loss": 0.8566, + "num_input_tokens_seen": 35941465, + "step": 1103 + }, + { + "epoch": 0.09956260991116923, + "flos": 21618210180960.0, + "grad_norm": 1.6100532806062635, + "learning_rate": 3.949493038526511e-06, + "loss": 0.7874, + "num_input_tokens_seen": 35969995, + "step": 1104 + }, + { + "epoch": 0.0996527934346395, + "flos": 21622521873120.0, + "grad_norm": 1.7310832635345148, + "learning_rate": 3.949362492839961e-06, + "loss": 0.7583, + "num_input_tokens_seen": 35997810, + "step": 1105 + }, + { + "epoch": 0.09974297695810976, + "flos": 24280578265920.0, + "grad_norm": 1.824915554757995, + "learning_rate": 3.949231780823054e-06, + "loss": 0.7663, + "num_input_tokens_seen": 36027465, + "step": 1106 + }, + { + "epoch": 0.09983316048158002, + "flos": 23698705444800.0, + "grad_norm": 1.8351556289951123, + "learning_rate": 3.949100902486945e-06, + "loss": 0.753, + "num_input_tokens_seen": 36055650, + "step": 1107 + }, + { + "epoch": 0.09992334400505028, + "flos": 23917595393280.0, + "grad_norm": 1.8488668141527176, + "learning_rate": 3.948969857842799e-06, + "loss": 0.7744, + "num_input_tokens_seen": 36084280, + "step": 1108 + }, + { + "epoch": 0.10001352752852054, + "flos": 25483989775680.0, + "grad_norm": 1.6190882767519656, + "learning_rate": 3.948838646901798e-06, + "loss": 0.7951, + "num_input_tokens_seen": 36113175, + "step": 1109 + }, + { + "epoch": 0.1001037110519908, + "flos": 21905764143360.0, + "grad_norm": 2.057720657882815, + "learning_rate": 3.948707269675138e-06, + "loss": 0.7935, + "num_input_tokens_seen": 36140155, + "step": 1110 + }, + { + "epoch": 0.10019389457546106, + "flos": 23444349811680.0, + "grad_norm": 1.8959918182977162, + "learning_rate": 3.948575726174028e-06, + "loss": 0.7933, + "num_input_tokens_seen": 36168130, + "step": 1111 + }, + { + "epoch": 0.10028407809893132, + "flos": 18815503317600.0, + "grad_norm": 2.0280699265442315, + "learning_rate": 3.9484440164096935e-06, + "loss": 0.8811, + "num_input_tokens_seen": 36194215, + "step": 1112 + }, + { + "epoch": 0.10037426162240158, + "flos": 20201186828640.0, + "grad_norm": 2.0862207113304856, + "learning_rate": 3.948312140393372e-06, + "loss": 0.827, + "num_input_tokens_seen": 36221370, + "step": 1113 + }, + { + "epoch": 0.10046444514587186, + "flos": 16630915524960.0, + "grad_norm": 2.154159527686247, + "learning_rate": 3.948180098136316e-06, + "loss": 0.7762, + "num_input_tokens_seen": 36245940, + "step": 1114 + }, + { + "epoch": 0.10055462866934212, + "flos": 28617744954240.0, + "grad_norm": 8.60109726758993, + "learning_rate": 3.948047889649791e-06, + "loss": 0.7735, + "num_input_tokens_seen": 36276015, + "step": 1115 + }, + { + "epoch": 0.10064481219281238, + "flos": 22532469428640.0, + "grad_norm": 1.9969348665674767, + "learning_rate": 3.947915514945079e-06, + "loss": 0.7875, + "num_input_tokens_seen": 36304965, + "step": 1116 + }, + { + "epoch": 0.10073499571628264, + "flos": 24206487467040.0, + "grad_norm": 1.677080268936043, + "learning_rate": 3.947782974033474e-06, + "loss": 0.8682, + "num_input_tokens_seen": 36335460, + "step": 1117 + }, + { + "epoch": 0.1008251792397529, + "flos": 23735174712960.0, + "grad_norm": 2.0151753478510672, + "learning_rate": 3.9476502669262866e-06, + "loss": 0.7666, + "num_input_tokens_seen": 36364535, + "step": 1118 + }, + { + "epoch": 0.10091536276322316, + "flos": 17833249111680.0, + "grad_norm": 1.706936342582102, + "learning_rate": 3.947517393634839e-06, + "loss": 0.7779, + "num_input_tokens_seen": 36391570, + "step": 1119 + }, + { + "epoch": 0.10100554628669342, + "flos": 25921249296000.0, + "grad_norm": 1.7928353719198216, + "learning_rate": 3.947384354170469e-06, + "loss": 0.8424, + "num_input_tokens_seen": 36420985, + "step": 1120 + }, + { + "epoch": 0.10109572981016368, + "flos": 20705214705120.0, + "grad_norm": 2.059699729590267, + "learning_rate": 3.947251148544528e-06, + "loss": 0.844, + "num_input_tokens_seen": 36450750, + "step": 1121 + }, + { + "epoch": 0.10118591333363394, + "flos": 15937627732320.0, + "grad_norm": 2.294352489206173, + "learning_rate": 3.947117776768382e-06, + "loss": 0.8327, + "num_input_tokens_seen": 36474995, + "step": 1122 + }, + { + "epoch": 0.1012760968571042, + "flos": 23737144710240.0, + "grad_norm": 1.4728865304298246, + "learning_rate": 3.9469842388534105e-06, + "loss": 0.7869, + "num_input_tokens_seen": 36508735, + "step": 1123 + }, + { + "epoch": 0.10136628038057446, + "flos": 25447334658720.0, + "grad_norm": 2.548484808355822, + "learning_rate": 3.946850534811009e-06, + "loss": 0.8177, + "num_input_tokens_seen": 36537690, + "step": 1124 + }, + { + "epoch": 0.10145646390404472, + "flos": 29856027432480.0, + "grad_norm": 1.779014456117165, + "learning_rate": 3.946716664652585e-06, + "loss": 0.7919, + "num_input_tokens_seen": 36568675, + "step": 1125 + }, + { + "epoch": 0.101546647427515, + "flos": 14407814127360.0, + "grad_norm": 2.905221244279074, + "learning_rate": 3.94658262838956e-06, + "loss": 0.7161, + "num_input_tokens_seen": 36593260, + "step": 1126 + }, + { + "epoch": 0.10163683095098526, + "flos": 25448226732960.0, + "grad_norm": 1.728715684595681, + "learning_rate": 3.946448426033373e-06, + "loss": 0.7251, + "num_input_tokens_seen": 36622405, + "step": 1127 + }, + { + "epoch": 0.10172701447445552, + "flos": 25040671584480.0, + "grad_norm": 1.7992148382332367, + "learning_rate": 3.946314057595473e-06, + "loss": 0.843, + "num_input_tokens_seen": 36649955, + "step": 1128 + }, + { + "epoch": 0.10181719799792578, + "flos": 30001607147040.0, + "grad_norm": 1.9435906200750208, + "learning_rate": 3.946179523087326e-06, + "loss": 0.8669, + "num_input_tokens_seen": 36679080, + "step": 1129 + }, + { + "epoch": 0.10190738152139604, + "flos": 28140596547840.0, + "grad_norm": 1.5875316556642913, + "learning_rate": 3.9460448225204104e-06, + "loss": 0.72, + "num_input_tokens_seen": 36709150, + "step": 1130 + }, + { + "epoch": 0.1019975650448663, + "flos": 22023497974080.0, + "grad_norm": 2.022088031659844, + "learning_rate": 3.945909955906221e-06, + "loss": 0.7962, + "num_input_tokens_seen": 36738045, + "step": 1131 + }, + { + "epoch": 0.10208774856833656, + "flos": 29782977386880.0, + "grad_norm": 2.040465358313883, + "learning_rate": 3.945774923256264e-06, + "loss": 0.8017, + "num_input_tokens_seen": 36767910, + "step": 1132 + }, + { + "epoch": 0.10217793209180683, + "flos": 15606021870720.0, + "grad_norm": 2.9397100697499834, + "learning_rate": 3.945639724582062e-06, + "loss": 0.8736, + "num_input_tokens_seen": 36792320, + "step": 1133 + }, + { + "epoch": 0.10226811561527709, + "flos": 26249286860640.0, + "grad_norm": 2.070798651278821, + "learning_rate": 3.94550435989515e-06, + "loss": 0.7708, + "num_input_tokens_seen": 36819935, + "step": 1134 + }, + { + "epoch": 0.10235829913874735, + "flos": 25739572010880.0, + "grad_norm": 1.9691006181735806, + "learning_rate": 3.945368829207079e-06, + "loss": 0.8236, + "num_input_tokens_seen": 36851120, + "step": 1135 + }, + { + "epoch": 0.10244848266221761, + "flos": 24751221964320.0, + "grad_norm": 2.6591007866783425, + "learning_rate": 3.945233132529414e-06, + "loss": 0.7611, + "num_input_tokens_seen": 36879395, + "step": 1136 + }, + { + "epoch": 0.10253866618568787, + "flos": 25260081909600.0, + "grad_norm": 1.704314839917385, + "learning_rate": 3.9450972698737304e-06, + "loss": 0.8017, + "num_input_tokens_seen": 36907670, + "step": 1137 + }, + { + "epoch": 0.10262884970915814, + "flos": 26249435539680.0, + "grad_norm": 1.975344195733967, + "learning_rate": 3.944961241251623e-06, + "loss": 0.7876, + "num_input_tokens_seen": 36936540, + "step": 1138 + }, + { + "epoch": 0.1027190332326284, + "flos": 30001123940160.0, + "grad_norm": 1.4546043495154104, + "learning_rate": 3.9448250466746985e-06, + "loss": 0.771, + "num_input_tokens_seen": 36968260, + "step": 1139 + }, + { + "epoch": 0.10280921675609866, + "flos": 27852410699520.0, + "grad_norm": 2.335733666345827, + "learning_rate": 3.944688686154578e-06, + "loss": 0.8327, + "num_input_tokens_seen": 36998375, + "step": 1140 + }, + { + "epoch": 0.10289940027956893, + "flos": 27043693601280.0, + "grad_norm": 1.7488490134315184, + "learning_rate": 3.944552159702894e-06, + "loss": 0.763, + "num_input_tokens_seen": 37025685, + "step": 1141 + }, + { + "epoch": 0.10298958380303919, + "flos": 23115717530880.0, + "grad_norm": 1.668552769924751, + "learning_rate": 3.944415467331299e-06, + "loss": 0.793, + "num_input_tokens_seen": 37053080, + "step": 1142 + }, + { + "epoch": 0.10307976732650945, + "flos": 23630004261120.0, + "grad_norm": 1.3309338207601553, + "learning_rate": 3.944278609051455e-06, + "loss": 0.7575, + "num_input_tokens_seen": 37085560, + "step": 1143 + }, + { + "epoch": 0.10316995084997971, + "flos": 25227069429120.0, + "grad_norm": 1.6562959916548872, + "learning_rate": 3.944141584875039e-06, + "loss": 0.7426, + "num_input_tokens_seen": 37115570, + "step": 1144 + }, + { + "epoch": 0.10326013437344997, + "flos": 21112769853600.0, + "grad_norm": 1.5911836343906574, + "learning_rate": 3.944004394813743e-06, + "loss": 0.8059, + "num_input_tokens_seen": 37143530, + "step": 1145 + }, + { + "epoch": 0.10335031789692023, + "flos": 17687483548320.0, + "grad_norm": 1.7037777830840675, + "learning_rate": 3.943867038879273e-06, + "loss": 0.8281, + "num_input_tokens_seen": 37171695, + "step": 1146 + }, + { + "epoch": 0.10344050142039049, + "flos": 41658763542240.0, + "grad_norm": 1.749582308921626, + "learning_rate": 3.943729517083349e-06, + "loss": 0.8386, + "num_input_tokens_seen": 37201090, + "step": 1147 + }, + { + "epoch": 0.10353068494386075, + "flos": 24390878144640.0, + "grad_norm": 1.7796969419367894, + "learning_rate": 3.943591829437705e-06, + "loss": 0.8333, + "num_input_tokens_seen": 37230455, + "step": 1148 + }, + { + "epoch": 0.10362086846733101, + "flos": 11967904287840.0, + "grad_norm": 1.795184990958847, + "learning_rate": 3.9434539759540895e-06, + "loss": 0.7485, + "num_input_tokens_seen": 37256395, + "step": 1149 + }, + { + "epoch": 0.10371105199080129, + "flos": 29089912612800.0, + "grad_norm": 1.605437192583517, + "learning_rate": 3.943315956644264e-06, + "loss": 0.7414, + "num_input_tokens_seen": 37285590, + "step": 1150 + }, + { + "epoch": 0.10380123551427155, + "flos": 17468816618400.0, + "grad_norm": 2.2068109723107203, + "learning_rate": 3.943177771520006e-06, + "loss": 0.8717, + "num_input_tokens_seen": 37313935, + "step": 1151 + }, + { + "epoch": 0.10389141903774181, + "flos": 24391063993440.0, + "grad_norm": 1.9090242678706875, + "learning_rate": 3.9430394205931065e-06, + "loss": 0.8467, + "num_input_tokens_seen": 37339075, + "step": 1152 + }, + { + "epoch": 0.10398160256121207, + "flos": 45189703506240.0, + "grad_norm": 1.692869051614899, + "learning_rate": 3.942900903875369e-06, + "loss": 0.6768, + "num_input_tokens_seen": 37373410, + "step": 1153 + }, + { + "epoch": 0.10407178608468233, + "flos": 25046878934400.0, + "grad_norm": 1.6429157068938431, + "learning_rate": 3.942762221378614e-06, + "loss": 0.8065, + "num_input_tokens_seen": 37403820, + "step": 1154 + }, + { + "epoch": 0.10416196960815259, + "flos": 25669161018240.0, + "grad_norm": 1.5117222985514311, + "learning_rate": 3.942623373114673e-06, + "loss": 0.8369, + "num_input_tokens_seen": 37437965, + "step": 1155 + }, + { + "epoch": 0.10425215313162285, + "flos": 20784843798240.0, + "grad_norm": 2.662684646436222, + "learning_rate": 3.942484359095396e-06, + "loss": 0.7905, + "num_input_tokens_seen": 37465450, + "step": 1156 + }, + { + "epoch": 0.10434233665509311, + "flos": 20929419929280.0, + "grad_norm": 1.718466163086814, + "learning_rate": 3.942345179332642e-06, + "loss": 0.7382, + "num_input_tokens_seen": 37494625, + "step": 1157 + }, + { + "epoch": 0.10443252017856337, + "flos": 26795024941440.0, + "grad_norm": 1.4553178218574687, + "learning_rate": 3.942205833838287e-06, + "loss": 0.7531, + "num_input_tokens_seen": 37526875, + "step": 1158 + }, + { + "epoch": 0.10452270370203363, + "flos": 20748337360320.0, + "grad_norm": 1.9142802834687944, + "learning_rate": 3.9420663226242204e-06, + "loss": 0.8954, + "num_input_tokens_seen": 37554690, + "step": 1159 + }, + { + "epoch": 0.1046128872255039, + "flos": 23006532744960.0, + "grad_norm": 2.3398094111041923, + "learning_rate": 3.941926645702348e-06, + "loss": 0.8514, + "num_input_tokens_seen": 37583090, + "step": 1160 + }, + { + "epoch": 0.10470307074897416, + "flos": 19217594511360.0, + "grad_norm": 5.466827403395913, + "learning_rate": 3.941786803084586e-06, + "loss": 0.7907, + "num_input_tokens_seen": 37610635, + "step": 1161 + }, + { + "epoch": 0.10479325427244443, + "flos": 24682260592320.0, + "grad_norm": 1.853779105783027, + "learning_rate": 3.941646794782867e-06, + "loss": 0.7518, + "num_input_tokens_seen": 37638865, + "step": 1162 + }, + { + "epoch": 0.10488343779591469, + "flos": 18999336448800.0, + "grad_norm": 2.543975865039662, + "learning_rate": 3.941506620809137e-06, + "loss": 0.8173, + "num_input_tokens_seen": 37664580, + "step": 1163 + }, + { + "epoch": 0.10497362131938495, + "flos": 27596977143360.0, + "grad_norm": 2.2755553353743316, + "learning_rate": 3.941366281175357e-06, + "loss": 0.7941, + "num_input_tokens_seen": 37692470, + "step": 1164 + }, + { + "epoch": 0.10506380484285521, + "flos": 23407843373760.0, + "grad_norm": 1.374019880364754, + "learning_rate": 3.941225775893502e-06, + "loss": 0.7399, + "num_input_tokens_seen": 37723460, + "step": 1165 + }, + { + "epoch": 0.10515398836632547, + "flos": 31967751028320.0, + "grad_norm": 1.6667583281424974, + "learning_rate": 3.941085104975559e-06, + "loss": 0.7792, + "num_input_tokens_seen": 37754475, + "step": 1166 + }, + { + "epoch": 0.10524417188979573, + "flos": 59940385932480.0, + "grad_norm": 0.7532567997121308, + "learning_rate": 3.9409442684335325e-06, + "loss": 0.6227, + "num_input_tokens_seen": 37842490, + "step": 1167 + }, + { + "epoch": 0.105334355413266, + "flos": 53210712315840.0, + "grad_norm": 2.8564040795635113, + "learning_rate": 3.940803266279438e-06, + "loss": 0.794, + "num_input_tokens_seen": 37877185, + "step": 1168 + }, + { + "epoch": 0.10542453893673626, + "flos": 23444015283840.0, + "grad_norm": 1.764875539870767, + "learning_rate": 3.9406620985253076e-06, + "loss": 0.7424, + "num_input_tokens_seen": 37903955, + "step": 1169 + }, + { + "epoch": 0.10551472246020652, + "flos": 21876282790080.0, + "grad_norm": 1.9174174275828701, + "learning_rate": 3.940520765183187e-06, + "loss": 0.7754, + "num_input_tokens_seen": 37933355, + "step": 1170 + }, + { + "epoch": 0.10560490598367678, + "flos": 22678606689600.0, + "grad_norm": 2.0014913482288432, + "learning_rate": 3.940379266265134e-06, + "loss": 0.8391, + "num_input_tokens_seen": 37960870, + "step": 1171 + }, + { + "epoch": 0.10569508950714704, + "flos": 31126764844800.0, + "grad_norm": 3.255368713497049, + "learning_rate": 3.940237601783223e-06, + "loss": 0.769, + "num_input_tokens_seen": 37989010, + "step": 1172 + }, + { + "epoch": 0.1057852730306173, + "flos": 27778877447040.0, + "grad_norm": 1.6687153701903734, + "learning_rate": 3.940095771749542e-06, + "loss": 0.8038, + "num_input_tokens_seen": 38018105, + "step": 1173 + }, + { + "epoch": 0.10587545655408757, + "flos": 22642843646880.0, + "grad_norm": 2.1431727040772874, + "learning_rate": 3.939953776176192e-06, + "loss": 0.8612, + "num_input_tokens_seen": 38046555, + "step": 1174 + }, + { + "epoch": 0.10596564007755783, + "flos": 40057944228480.0, + "grad_norm": 1.3666075430075837, + "learning_rate": 3.939811615075288e-06, + "loss": 0.8291, + "num_input_tokens_seen": 38081210, + "step": 1175 + }, + { + "epoch": 0.1060558236010281, + "flos": 21178869154080.0, + "grad_norm": 2.4536661899612655, + "learning_rate": 3.9396692884589616e-06, + "loss": 0.8689, + "num_input_tokens_seen": 38108575, + "step": 1176 + }, + { + "epoch": 0.10614600712449836, + "flos": 25519827157920.0, + "grad_norm": 2.3028743529868345, + "learning_rate": 3.9395267963393565e-06, + "loss": 0.7871, + "num_input_tokens_seen": 38137205, + "step": 1177 + }, + { + "epoch": 0.10623619064796862, + "flos": 19612103074080.0, + "grad_norm": 1.889808541545929, + "learning_rate": 3.939384138728631e-06, + "loss": 0.8412, + "num_input_tokens_seen": 38163340, + "step": 1178 + }, + { + "epoch": 0.10632637417143888, + "flos": 31783992236640.0, + "grad_norm": 1.7987245411824124, + "learning_rate": 3.939241315638956e-06, + "loss": 0.7901, + "num_input_tokens_seen": 38192810, + "step": 1179 + }, + { + "epoch": 0.10641655769490914, + "flos": 31456623727680.0, + "grad_norm": 2.1544184991945348, + "learning_rate": 3.93909832708252e-06, + "loss": 0.7877, + "num_input_tokens_seen": 38221975, + "step": 1180 + }, + { + "epoch": 0.1065067412183794, + "flos": 24825089744640.0, + "grad_norm": 1.5513028821761161, + "learning_rate": 3.938955173071523e-06, + "loss": 0.8096, + "num_input_tokens_seen": 38249905, + "step": 1181 + }, + { + "epoch": 0.10659692474184966, + "flos": 68405642177280.0, + "grad_norm": 0.6261193741880557, + "learning_rate": 3.938811853618179e-06, + "loss": 0.6089, + "num_input_tokens_seen": 38349945, + "step": 1182 + }, + { + "epoch": 0.10668710826531992, + "flos": 23951834475840.0, + "grad_norm": 4.11843488353192, + "learning_rate": 3.938668368734717e-06, + "loss": 0.8395, + "num_input_tokens_seen": 38376455, + "step": 1183 + }, + { + "epoch": 0.10677729178879018, + "flos": 23258732532000.0, + "grad_norm": 2.8964616564980585, + "learning_rate": 3.93852471843338e-06, + "loss": 0.6797, + "num_input_tokens_seen": 38406510, + "step": 1184 + }, + { + "epoch": 0.10686747531226044, + "flos": 20565544982400.0, + "grad_norm": 1.794744273169021, + "learning_rate": 3.9383809027264254e-06, + "loss": 0.7853, + "num_input_tokens_seen": 38433050, + "step": 1185 + }, + { + "epoch": 0.10695765883573072, + "flos": 23006978782080.0, + "grad_norm": 2.1754705731193096, + "learning_rate": 3.938236921626124e-06, + "loss": 0.8435, + "num_input_tokens_seen": 38462230, + "step": 1186 + }, + { + "epoch": 0.10704784235920098, + "flos": 20233455913920.0, + "grad_norm": 2.4903793232937748, + "learning_rate": 3.938092775144761e-06, + "loss": 0.8007, + "num_input_tokens_seen": 38488745, + "step": 1187 + }, + { + "epoch": 0.10713802588267124, + "flos": 24967621538880.0, + "grad_norm": 2.14187073483695, + "learning_rate": 3.9379484632946355e-06, + "loss": 0.7255, + "num_input_tokens_seen": 38517485, + "step": 1188 + }, + { + "epoch": 0.1072282094061415, + "flos": 24026594330400.0, + "grad_norm": 2.0839777394897876, + "learning_rate": 3.937803986088062e-06, + "loss": 0.8489, + "num_input_tokens_seen": 38545845, + "step": 1189 + }, + { + "epoch": 0.10731839292961176, + "flos": 20274051025440.0, + "grad_norm": 2.784554479360936, + "learning_rate": 3.937659343537367e-06, + "loss": 0.7589, + "num_input_tokens_seen": 38574335, + "step": 1190 + }, + { + "epoch": 0.10740857645308202, + "flos": 21549062960160.0, + "grad_norm": 1.795758041196013, + "learning_rate": 3.937514535654893e-06, + "loss": 0.8745, + "num_input_tokens_seen": 38602780, + "step": 1191 + }, + { + "epoch": 0.10749875997655228, + "flos": 22969282911840.0, + "grad_norm": 2.564876106051146, + "learning_rate": 3.937369562452996e-06, + "loss": 0.8074, + "num_input_tokens_seen": 38629650, + "step": 1192 + }, + { + "epoch": 0.10758894350002254, + "flos": 21841560500640.0, + "grad_norm": 1.8170231253622493, + "learning_rate": 3.937224423944044e-06, + "loss": 0.8576, + "num_input_tokens_seen": 38656985, + "step": 1193 + }, + { + "epoch": 0.1076791270234928, + "flos": 27919550753280.0, + "grad_norm": 1.7784468026086309, + "learning_rate": 3.937079120140423e-06, + "loss": 0.7341, + "num_input_tokens_seen": 38686445, + "step": 1194 + }, + { + "epoch": 0.10776931054696307, + "flos": 21585755246880.0, + "grad_norm": 1.633163122468688, + "learning_rate": 3.936933651054531e-06, + "loss": 0.8415, + "num_input_tokens_seen": 38715725, + "step": 1195 + }, + { + "epoch": 0.10785949407043333, + "flos": 23843393085120.0, + "grad_norm": 2.214675708565502, + "learning_rate": 3.936788016698779e-06, + "loss": 0.9186, + "num_input_tokens_seen": 38745120, + "step": 1196 + }, + { + "epoch": 0.1079496775939036, + "flos": 21584937512160.0, + "grad_norm": 1.609257443773359, + "learning_rate": 3.936642217085594e-06, + "loss": 0.8325, + "num_input_tokens_seen": 38775235, + "step": 1197 + }, + { + "epoch": 0.10803986111737386, + "flos": 21285452056800.0, + "grad_norm": 1.9583597451992372, + "learning_rate": 3.936496252227417e-06, + "loss": 0.7125, + "num_input_tokens_seen": 38801845, + "step": 1198 + }, + { + "epoch": 0.10813004464084412, + "flos": 20013450872640.0, + "grad_norm": 4.497784174984523, + "learning_rate": 3.936350122136703e-06, + "loss": 0.7831, + "num_input_tokens_seen": 38829085, + "step": 1199 + }, + { + "epoch": 0.10822022816431438, + "flos": 66403958271840.0, + "grad_norm": 0.6856375305574693, + "learning_rate": 3.936203826825919e-06, + "loss": 0.6296, + "num_input_tokens_seen": 38924855, + "step": 1200 + }, + { + "epoch": 0.10831041168778464, + "flos": 21073512853440.0, + "grad_norm": 1.745855191381485, + "learning_rate": 3.9360573663075475e-06, + "loss": 0.7623, + "num_input_tokens_seen": 38954710, + "step": 1201 + }, + { + "epoch": 0.1084005952112549, + "flos": 21075966057600.0, + "grad_norm": 2.1557257194580575, + "learning_rate": 3.935910740594087e-06, + "loss": 0.8255, + "num_input_tokens_seen": 38981855, + "step": 1202 + }, + { + "epoch": 0.10849077873472517, + "flos": 27775420659360.0, + "grad_norm": 2.5220327335081754, + "learning_rate": 3.935763949698047e-06, + "loss": 0.8111, + "num_input_tokens_seen": 39011690, + "step": 1203 + }, + { + "epoch": 0.10858096225819543, + "flos": 20524541003520.0, + "grad_norm": 1.774302872242882, + "learning_rate": 3.935616993631954e-06, + "loss": 0.8627, + "num_input_tokens_seen": 39038165, + "step": 1204 + }, + { + "epoch": 0.10867114578166569, + "flos": 28287700222560.0, + "grad_norm": 2.288877928912312, + "learning_rate": 3.935469872408345e-06, + "loss": 0.7954, + "num_input_tokens_seen": 39067405, + "step": 1205 + }, + { + "epoch": 0.10876132930513595, + "flos": 21403520415360.0, + "grad_norm": 1.898546820110704, + "learning_rate": 3.935322586039776e-06, + "loss": 0.8254, + "num_input_tokens_seen": 39095850, + "step": 1206 + }, + { + "epoch": 0.10885151282860621, + "flos": 35357199951360.0, + "grad_norm": 1.6783054077239514, + "learning_rate": 3.935175134538811e-06, + "loss": 0.7375, + "num_input_tokens_seen": 39130685, + "step": 1207 + }, + { + "epoch": 0.10894169635207647, + "flos": 14189630404320.0, + "grad_norm": 3.5924416192075483, + "learning_rate": 3.935027517918034e-06, + "loss": 0.7446, + "num_input_tokens_seen": 39153435, + "step": 1208 + }, + { + "epoch": 0.10903187987554674, + "flos": 17900723693280.0, + "grad_norm": 2.453175538143829, + "learning_rate": 3.93487973619004e-06, + "loss": 0.8341, + "num_input_tokens_seen": 39179770, + "step": 1209 + }, + { + "epoch": 0.109122063399017, + "flos": 13861444160640.0, + "grad_norm": 2.2605846243014063, + "learning_rate": 3.934731789367438e-06, + "loss": 0.8106, + "num_input_tokens_seen": 39206280, + "step": 1210 + }, + { + "epoch": 0.10921224692248727, + "flos": 63469393068960.0, + "grad_norm": 0.7607109641252708, + "learning_rate": 3.9345836774628505e-06, + "loss": 0.6084, + "num_input_tokens_seen": 39282975, + "step": 1211 + }, + { + "epoch": 0.10930243044595753, + "flos": 21585978265440.0, + "grad_norm": 1.8282359814987317, + "learning_rate": 3.934435400488917e-06, + "loss": 0.7939, + "num_input_tokens_seen": 39311545, + "step": 1212 + }, + { + "epoch": 0.10939261396942779, + "flos": 15355346043840.0, + "grad_norm": 2.036736927017205, + "learning_rate": 3.934286958458289e-06, + "loss": 0.8387, + "num_input_tokens_seen": 39337400, + "step": 1213 + }, + { + "epoch": 0.10948279749289805, + "flos": 21039979996320.0, + "grad_norm": 1.636098011088895, + "learning_rate": 3.934138351383632e-06, + "loss": 0.8643, + "num_input_tokens_seen": 39366110, + "step": 1214 + }, + { + "epoch": 0.10957298101636831, + "flos": 39364990963680.0, + "grad_norm": 1.6067145124938043, + "learning_rate": 3.933989579277626e-06, + "loss": 0.7352, + "num_input_tokens_seen": 39399135, + "step": 1215 + }, + { + "epoch": 0.10966316453983857, + "flos": 22313728159200.0, + "grad_norm": 6.26646512837007, + "learning_rate": 3.933840642152966e-06, + "loss": 0.7943, + "num_input_tokens_seen": 39430820, + "step": 1216 + }, + { + "epoch": 0.10975334806330883, + "flos": 20857001769600.0, + "grad_norm": 2.254542639421568, + "learning_rate": 3.933691540022359e-06, + "loss": 0.8282, + "num_input_tokens_seen": 39456750, + "step": 1217 + }, + { + "epoch": 0.10984353158677909, + "flos": 22751359377120.0, + "grad_norm": 1.8479738506886345, + "learning_rate": 3.933542272898527e-06, + "loss": 0.7707, + "num_input_tokens_seen": 39484670, + "step": 1218 + }, + { + "epoch": 0.10993371511024935, + "flos": 65117572390560.0, + "grad_norm": 0.7584193444642928, + "learning_rate": 3.933392840794207e-06, + "loss": 0.5656, + "num_input_tokens_seen": 39572555, + "step": 1219 + }, + { + "epoch": 0.11002389863371961, + "flos": 13314851175360.0, + "grad_norm": 2.305652170825168, + "learning_rate": 3.93324324372215e-06, + "loss": 0.9066, + "num_input_tokens_seen": 39597525, + "step": 1220 + }, + { + "epoch": 0.11011408215718989, + "flos": 37761272408640.0, + "grad_norm": 1.888808962433688, + "learning_rate": 3.9330934816951185e-06, + "loss": 0.8265, + "num_input_tokens_seen": 39628215, + "step": 1221 + }, + { + "epoch": 0.11020426568066015, + "flos": 20158956247680.0, + "grad_norm": 2.0556044472501727, + "learning_rate": 3.932943554725893e-06, + "loss": 0.8301, + "num_input_tokens_seen": 39655915, + "step": 1222 + }, + { + "epoch": 0.11029444920413041, + "flos": 57507798535680.0, + "grad_norm": 0.7611089759952951, + "learning_rate": 3.932793462827265e-06, + "loss": 0.6396, + "num_input_tokens_seen": 39746620, + "step": 1223 + }, + { + "epoch": 0.11038463272760067, + "flos": 25154205232320.0, + "grad_norm": 1.5816507517687584, + "learning_rate": 3.932643206012041e-06, + "loss": 0.8611, + "num_input_tokens_seen": 39776965, + "step": 1224 + }, + { + "epoch": 0.11047481625107093, + "flos": 20820755520000.0, + "grad_norm": 2.12638764794223, + "learning_rate": 3.932492784293043e-06, + "loss": 0.7321, + "num_input_tokens_seen": 39804025, + "step": 1225 + }, + { + "epoch": 0.11056499977454119, + "flos": 26467619262720.0, + "grad_norm": 1.9344425072545555, + "learning_rate": 3.932342197683104e-06, + "loss": 0.792, + "num_input_tokens_seen": 39834255, + "step": 1226 + }, + { + "epoch": 0.11065518329801145, + "flos": 27487903866720.0, + "grad_norm": 1.8977596501353924, + "learning_rate": 3.932191446195075e-06, + "loss": 0.6797, + "num_input_tokens_seen": 39865110, + "step": 1227 + }, + { + "epoch": 0.11074536682148171, + "flos": 20383495999680.0, + "grad_norm": 1.9888840756396557, + "learning_rate": 3.9320405298418175e-06, + "loss": 0.8094, + "num_input_tokens_seen": 39892430, + "step": 1228 + }, + { + "epoch": 0.11083555034495197, + "flos": 24063137938080.0, + "grad_norm": 2.06067183418494, + "learning_rate": 3.9318894486362076e-06, + "loss": 0.7224, + "num_input_tokens_seen": 39922905, + "step": 1229 + }, + { + "epoch": 0.11092573386842224, + "flos": 23152781515200.0, + "grad_norm": 2.1576378820039284, + "learning_rate": 3.9317382025911395e-06, + "loss": 0.8775, + "num_input_tokens_seen": 39951070, + "step": 1230 + }, + { + "epoch": 0.1110159173918925, + "flos": 20929196910720.0, + "grad_norm": 1.850004933324886, + "learning_rate": 3.9315867917195145e-06, + "loss": 0.8994, + "num_input_tokens_seen": 39977830, + "step": 1231 + }, + { + "epoch": 0.11110610091536276, + "flos": 22823554518240.0, + "grad_norm": 1.5626685242370615, + "learning_rate": 3.931435216034256e-06, + "loss": 0.806, + "num_input_tokens_seen": 40007795, + "step": 1232 + }, + { + "epoch": 0.11119628443883303, + "flos": 22860618502560.0, + "grad_norm": 2.23996659185837, + "learning_rate": 3.931283475548293e-06, + "loss": 0.8059, + "num_input_tokens_seen": 40036940, + "step": 1233 + }, + { + "epoch": 0.11128646796230329, + "flos": 22539011306400.0, + "grad_norm": 1.4247768260801872, + "learning_rate": 3.931131570274576e-06, + "loss": 0.8043, + "num_input_tokens_seen": 40068410, + "step": 1234 + }, + { + "epoch": 0.11137665148577355, + "flos": 26540260440960.0, + "grad_norm": 1.5220922546097424, + "learning_rate": 3.930979500226065e-06, + "loss": 0.8386, + "num_input_tokens_seen": 40100440, + "step": 1235 + }, + { + "epoch": 0.11146683500924381, + "flos": 20346915222240.0, + "grad_norm": 2.1975238563535924, + "learning_rate": 3.930827265415736e-06, + "loss": 0.6409, + "num_input_tokens_seen": 40128205, + "step": 1236 + }, + { + "epoch": 0.11155701853271408, + "flos": 21549471827520.0, + "grad_norm": 8.582187711932871, + "learning_rate": 3.930674865856578e-06, + "loss": 0.8151, + "num_input_tokens_seen": 40154990, + "step": 1237 + }, + { + "epoch": 0.11164720205618434, + "flos": 23407099978560.0, + "grad_norm": 1.7354612643162095, + "learning_rate": 3.930522301561595e-06, + "loss": 0.8338, + "num_input_tokens_seen": 40184595, + "step": 1238 + }, + { + "epoch": 0.1117373855796546, + "flos": 21986842857120.0, + "grad_norm": 2.1654699272792794, + "learning_rate": 3.930369572543804e-06, + "loss": 0.8486, + "num_input_tokens_seen": 40212450, + "step": 1239 + }, + { + "epoch": 0.11182756910312486, + "flos": 21618433199520.0, + "grad_norm": 1.631010384335371, + "learning_rate": 3.930216678816237e-06, + "loss": 0.7478, + "num_input_tokens_seen": 40243600, + "step": 1240 + }, + { + "epoch": 0.11191775262659512, + "flos": 71294928878880.0, + "grad_norm": 0.6668669553734479, + "learning_rate": 3.930063620391941e-06, + "loss": 0.6136, + "num_input_tokens_seen": 40334185, + "step": 1241 + }, + { + "epoch": 0.11200793615006538, + "flos": 22314248535840.0, + "grad_norm": 2.084909936071177, + "learning_rate": 3.9299103972839735e-06, + "loss": 0.8101, + "num_input_tokens_seen": 40363385, + "step": 1242 + }, + { + "epoch": 0.11209811967353564, + "flos": 65990678980320.0, + "grad_norm": 0.6588383221637237, + "learning_rate": 3.92975700950541e-06, + "loss": 0.5776, + "num_input_tokens_seen": 40459670, + "step": 1243 + }, + { + "epoch": 0.1121883031970059, + "flos": 25043682335040.0, + "grad_norm": 1.9915210282242504, + "learning_rate": 3.929603457069338e-06, + "loss": 0.836, + "num_input_tokens_seen": 40488960, + "step": 1244 + }, + { + "epoch": 0.11227848672047618, + "flos": 22275734930880.0, + "grad_norm": 2.103959089055486, + "learning_rate": 3.929449739988859e-06, + "loss": 0.77, + "num_input_tokens_seen": 40517290, + "step": 1245 + }, + { + "epoch": 0.11236867024394644, + "flos": 25373615557440.0, + "grad_norm": 2.2575175478771783, + "learning_rate": 3.929295858277089e-06, + "loss": 0.7997, + "num_input_tokens_seen": 40547460, + "step": 1246 + }, + { + "epoch": 0.1124588537674167, + "flos": 20602497457440.0, + "grad_norm": 3.0239251110754344, + "learning_rate": 3.9291418119471585e-06, + "loss": 0.8006, + "num_input_tokens_seen": 40574400, + "step": 1247 + }, + { + "epoch": 0.11254903729088696, + "flos": 17870350265760.0, + "grad_norm": 2.3882887650790536, + "learning_rate": 3.928987601012212e-06, + "loss": 0.8249, + "num_input_tokens_seen": 40600665, + "step": 1248 + }, + { + "epoch": 0.11263922081435722, + "flos": 16703631042720.0, + "grad_norm": 1.8784832071706663, + "learning_rate": 3.928833225485407e-06, + "loss": 0.7947, + "num_input_tokens_seen": 40626195, + "step": 1249 + }, + { + "epoch": 0.11272940433782748, + "flos": 23548814038080.0, + "grad_norm": 2.013080854290116, + "learning_rate": 3.928678685379915e-06, + "loss": 0.7916, + "num_input_tokens_seen": 40654825, + "step": 1250 + }, + { + "epoch": 0.11281958786129774, + "flos": 25004425334880.0, + "grad_norm": 1.6349983004173287, + "learning_rate": 3.928523980708924e-06, + "loss": 0.8533, + "num_input_tokens_seen": 40682835, + "step": 1251 + }, + { + "epoch": 0.112909771384768, + "flos": 35903161050720.0, + "grad_norm": 2.3850500388664235, + "learning_rate": 3.928369111485632e-06, + "loss": 0.7604, + "num_input_tokens_seen": 40712655, + "step": 1252 + }, + { + "epoch": 0.11299995490823826, + "flos": 21984240973920.0, + "grad_norm": 2.1088683473564522, + "learning_rate": 3.928214077723255e-06, + "loss": 0.7931, + "num_input_tokens_seen": 40742450, + "step": 1253 + }, + { + "epoch": 0.11309013843170852, + "flos": 22606225699680.0, + "grad_norm": 3.387003850321444, + "learning_rate": 3.928058879435021e-06, + "loss": 0.7917, + "num_input_tokens_seen": 40769535, + "step": 1254 + }, + { + "epoch": 0.11318032195517878, + "flos": 16667868000000.0, + "grad_norm": 1.930901520037785, + "learning_rate": 3.9279035166341725e-06, + "loss": 0.7877, + "num_input_tokens_seen": 40795365, + "step": 1255 + }, + { + "epoch": 0.11327050547864904, + "flos": 30509166150720.0, + "grad_norm": 1.8304343350179546, + "learning_rate": 3.927747989333965e-06, + "loss": 0.7844, + "num_input_tokens_seen": 40826205, + "step": 1256 + }, + { + "epoch": 0.11336068900211932, + "flos": 24095815890720.0, + "grad_norm": 2.1972403427364924, + "learning_rate": 3.927592297547669e-06, + "loss": 0.7632, + "num_input_tokens_seen": 40854495, + "step": 1257 + }, + { + "epoch": 0.11345087252558958, + "flos": 19035322510080.0, + "grad_norm": 2.338670745436455, + "learning_rate": 3.927436441288571e-06, + "loss": 0.8297, + "num_input_tokens_seen": 40879485, + "step": 1258 + }, + { + "epoch": 0.11354105604905984, + "flos": 29488175321280.0, + "grad_norm": 1.7826298505613971, + "learning_rate": 3.927280420569968e-06, + "loss": 0.8206, + "num_input_tokens_seen": 40909320, + "step": 1259 + }, + { + "epoch": 0.1136312395725301, + "flos": 29856213281280.0, + "grad_norm": 1.6902402598677742, + "learning_rate": 3.927124235405171e-06, + "loss": 0.8295, + "num_input_tokens_seen": 40937920, + "step": 1260 + }, + { + "epoch": 0.11372142309600036, + "flos": 24170575745280.0, + "grad_norm": 2.2487819341541164, + "learning_rate": 3.92696788580751e-06, + "loss": 0.6613, + "num_input_tokens_seen": 40968650, + "step": 1261 + }, + { + "epoch": 0.11381160661947062, + "flos": 34847670950400.0, + "grad_norm": 1.7290222460019997, + "learning_rate": 3.9268113717903225e-06, + "loss": 0.6397, + "num_input_tokens_seen": 40998960, + "step": 1262 + }, + { + "epoch": 0.11390179014294088, + "flos": 21440324211360.0, + "grad_norm": 1.8480319715764524, + "learning_rate": 3.926654693366965e-06, + "loss": 0.8705, + "num_input_tokens_seen": 41027500, + "step": 1263 + }, + { + "epoch": 0.11399197366641114, + "flos": 21288760165440.0, + "grad_norm": 2.663836696150142, + "learning_rate": 3.926497850550805e-06, + "loss": 0.8167, + "num_input_tokens_seen": 41054710, + "step": 1264 + }, + { + "epoch": 0.1140821571898814, + "flos": 20820420992160.0, + "grad_norm": 2.550370701790768, + "learning_rate": 3.926340843355226e-06, + "loss": 0.7958, + "num_input_tokens_seen": 41083125, + "step": 1265 + }, + { + "epoch": 0.11417234071335167, + "flos": 19727160682080.0, + "grad_norm": 2.188493448555727, + "learning_rate": 3.926183671793625e-06, + "loss": 0.8145, + "num_input_tokens_seen": 41111455, + "step": 1266 + }, + { + "epoch": 0.11426252423682193, + "flos": 22205509787040.0, + "grad_norm": 1.8769959316870082, + "learning_rate": 3.926026335879412e-06, + "loss": 0.768, + "num_input_tokens_seen": 41138295, + "step": 1267 + }, + { + "epoch": 0.11435270776029219, + "flos": 70576291309920.0, + "grad_norm": 0.6283909125679903, + "learning_rate": 3.925868835626012e-06, + "loss": 0.5503, + "num_input_tokens_seen": 41231875, + "step": 1268 + }, + { + "epoch": 0.11444289128376246, + "flos": 20784397761120.0, + "grad_norm": 2.193025640029023, + "learning_rate": 3.925711171046864e-06, + "loss": 0.8114, + "num_input_tokens_seen": 41261915, + "step": 1269 + }, + { + "epoch": 0.11453307480723272, + "flos": 23078988074400.0, + "grad_norm": 1.655476227437838, + "learning_rate": 3.925553342155421e-06, + "loss": 0.7638, + "num_input_tokens_seen": 41293070, + "step": 1270 + }, + { + "epoch": 0.11462325833070298, + "flos": 60978883718880.0, + "grad_norm": 0.6776138550292737, + "learning_rate": 3.9253953489651485e-06, + "loss": 0.6248, + "num_input_tokens_seen": 41381965, + "step": 1271 + }, + { + "epoch": 0.11471344185417325, + "flos": 63014918216160.0, + "grad_norm": 0.7069609169726109, + "learning_rate": 3.925237191489529e-06, + "loss": 0.6001, + "num_input_tokens_seen": 41477040, + "step": 1272 + }, + { + "epoch": 0.1148036253776435, + "flos": 23371336935840.0, + "grad_norm": 1.7902510247485293, + "learning_rate": 3.925078869742056e-06, + "loss": 0.7695, + "num_input_tokens_seen": 41506920, + "step": 1273 + }, + { + "epoch": 0.11489380890111377, + "flos": 21731372131200.0, + "grad_norm": 2.375013150630481, + "learning_rate": 3.92492038373624e-06, + "loss": 0.7356, + "num_input_tokens_seen": 41537240, + "step": 1274 + }, + { + "epoch": 0.11498399242458403, + "flos": 22861101709440.0, + "grad_norm": 1.8532427796990312, + "learning_rate": 3.924761733485602e-06, + "loss": 0.8551, + "num_input_tokens_seen": 41566690, + "step": 1275 + }, + { + "epoch": 0.11507417594805429, + "flos": 25226734901280.0, + "grad_norm": 1.7184053942533535, + "learning_rate": 3.92460291900368e-06, + "loss": 0.7986, + "num_input_tokens_seen": 41597760, + "step": 1276 + }, + { + "epoch": 0.11516435947152455, + "flos": 21804645195360.0, + "grad_norm": 1.7451464240171268, + "learning_rate": 3.924443940304025e-06, + "loss": 0.8202, + "num_input_tokens_seen": 41625905, + "step": 1277 + }, + { + "epoch": 0.11525454299499481, + "flos": 19253915100480.0, + "grad_norm": 1.933044787227994, + "learning_rate": 3.924284797400202e-06, + "loss": 0.8491, + "num_input_tokens_seen": 41652105, + "step": 1278 + }, + { + "epoch": 0.11534472651846507, + "flos": 20455988498880.0, + "grad_norm": 2.6132661964427695, + "learning_rate": 3.924125490305789e-06, + "loss": 0.8206, + "num_input_tokens_seen": 41674525, + "step": 1279 + }, + { + "epoch": 0.11543491004193533, + "flos": 31200929983200.0, + "grad_norm": 2.6610748452688817, + "learning_rate": 3.923966019034381e-06, + "loss": 0.7825, + "num_input_tokens_seen": 41706520, + "step": 1280 + }, + { + "epoch": 0.1155250935654056, + "flos": 19326816467040.0, + "grad_norm": 1.8715013287086264, + "learning_rate": 3.923806383599583e-06, + "loss": 0.8092, + "num_input_tokens_seen": 41734305, + "step": 1281 + }, + { + "epoch": 0.11561527708887587, + "flos": 28216285646400.0, + "grad_norm": 2.1147254056761966, + "learning_rate": 3.923646584015017e-06, + "loss": 0.7684, + "num_input_tokens_seen": 41764370, + "step": 1282 + }, + { + "epoch": 0.11570546061234613, + "flos": 23216055913920.0, + "grad_norm": 3.0432260426567233, + "learning_rate": 3.923486620294316e-06, + "loss": 0.8034, + "num_input_tokens_seen": 41788965, + "step": 1283 + }, + { + "epoch": 0.11579564413581639, + "flos": 29707436967360.0, + "grad_norm": 1.9601844466284486, + "learning_rate": 3.923326492451132e-06, + "loss": 0.809, + "num_input_tokens_seen": 41815520, + "step": 1284 + }, + { + "epoch": 0.11588582765928665, + "flos": 23845288742880.0, + "grad_norm": 2.884894552171437, + "learning_rate": 3.923166200499125e-06, + "loss": 0.7815, + "num_input_tokens_seen": 41847555, + "step": 1285 + }, + { + "epoch": 0.11597601118275691, + "flos": 38634118810080.0, + "grad_norm": 1.9514609610513156, + "learning_rate": 3.923005744451975e-06, + "loss": 0.771, + "num_input_tokens_seen": 41880440, + "step": 1286 + }, + { + "epoch": 0.11606619470622717, + "flos": 20201112489120.0, + "grad_norm": 2.265204660293444, + "learning_rate": 3.9228451243233715e-06, + "loss": 0.7535, + "num_input_tokens_seen": 41906705, + "step": 1287 + }, + { + "epoch": 0.11615637822969743, + "flos": 25775260714080.0, + "grad_norm": 1.5153493728918848, + "learning_rate": 3.9226843401270195e-06, + "loss": 0.8302, + "num_input_tokens_seen": 41938110, + "step": 1288 + }, + { + "epoch": 0.1162465617531677, + "flos": 20674655428800.0, + "grad_norm": 1.9736127592839479, + "learning_rate": 3.9225233918766376e-06, + "loss": 0.8021, + "num_input_tokens_seen": 41967515, + "step": 1289 + }, + { + "epoch": 0.11633674527663795, + "flos": 27994161928800.0, + "grad_norm": 1.6689734548625657, + "learning_rate": 3.92236227958596e-06, + "loss": 0.8185, + "num_input_tokens_seen": 41996135, + "step": 1290 + }, + { + "epoch": 0.11642692880010821, + "flos": 23876331226080.0, + "grad_norm": 2.45925690926057, + "learning_rate": 3.922201003268731e-06, + "loss": 0.8205, + "num_input_tokens_seen": 42024025, + "step": 1291 + }, + { + "epoch": 0.11651711232357848, + "flos": 71612707589760.0, + "grad_norm": 0.7456358857811105, + "learning_rate": 3.922039562938715e-06, + "loss": 0.6257, + "num_input_tokens_seen": 42120780, + "step": 1292 + }, + { + "epoch": 0.11660729584704875, + "flos": 25702656705600.0, + "grad_norm": 2.1175048139893486, + "learning_rate": 3.921877958609685e-06, + "loss": 0.8519, + "num_input_tokens_seen": 42151405, + "step": 1293 + }, + { + "epoch": 0.11669747937051901, + "flos": 57940820703360.0, + "grad_norm": 0.7966285773340169, + "learning_rate": 3.921716190295431e-06, + "loss": 0.6906, + "num_input_tokens_seen": 42230740, + "step": 1294 + }, + { + "epoch": 0.11678766289398927, + "flos": 19721287860000.0, + "grad_norm": 1.8555997460954101, + "learning_rate": 3.921554258009755e-06, + "loss": 0.752, + "num_input_tokens_seen": 42257435, + "step": 1295 + }, + { + "epoch": 0.11687784641745953, + "flos": 20784732288960.0, + "grad_norm": 1.6612646374656777, + "learning_rate": 3.921392161766474e-06, + "loss": 0.7863, + "num_input_tokens_seen": 42286025, + "step": 1296 + }, + { + "epoch": 0.1169680299409298, + "flos": 34482643740960.0, + "grad_norm": 3.8644087238743587, + "learning_rate": 3.92122990157942e-06, + "loss": 0.7746, + "num_input_tokens_seen": 42317025, + "step": 1297 + }, + { + "epoch": 0.11705821346440005, + "flos": 21328388863200.0, + "grad_norm": 1.691020472275732, + "learning_rate": 3.921067477462437e-06, + "loss": 0.8431, + "num_input_tokens_seen": 42345975, + "step": 1298 + }, + { + "epoch": 0.11714839698787032, + "flos": 14954778810240.0, + "grad_norm": 2.2396709927695344, + "learning_rate": 3.920904889429385e-06, + "loss": 0.8518, + "num_input_tokens_seen": 42370800, + "step": 1299 + }, + { + "epoch": 0.11723858051134058, + "flos": 25883999462880.0, + "grad_norm": 1.7550473313863915, + "learning_rate": 3.920742137494135e-06, + "loss": 0.8363, + "num_input_tokens_seen": 42401245, + "step": 1300 + }, + { + "epoch": 0.11732876403481084, + "flos": 23443792265280.0, + "grad_norm": 1.5735906289677355, + "learning_rate": 3.920579221670575e-06, + "loss": 0.8586, + "num_input_tokens_seen": 42430225, + "step": 1301 + }, + { + "epoch": 0.1174189475582811, + "flos": 35714421511200.0, + "grad_norm": 1.9194215450524916, + "learning_rate": 3.920416141972606e-06, + "loss": 0.8036, + "num_input_tokens_seen": 42460335, + "step": 1302 + }, + { + "epoch": 0.11750913108175136, + "flos": 25629123453120.0, + "grad_norm": 1.592419012456684, + "learning_rate": 3.920252898414143e-06, + "loss": 0.7955, + "num_input_tokens_seen": 42491765, + "step": 1303 + }, + { + "epoch": 0.11759931460522162, + "flos": 21112844193120.0, + "grad_norm": 1.9487117497848134, + "learning_rate": 3.920089491009114e-06, + "loss": 0.8102, + "num_input_tokens_seen": 42519560, + "step": 1304 + }, + { + "epoch": 0.1176894981286919, + "flos": 22602434384160.0, + "grad_norm": 1.853035578250808, + "learning_rate": 3.919925919771463e-06, + "loss": 0.885, + "num_input_tokens_seen": 42546945, + "step": 1305 + }, + { + "epoch": 0.11777968165216215, + "flos": 19982891596320.0, + "grad_norm": 2.245250457352465, + "learning_rate": 3.919762184715146e-06, + "loss": 0.7147, + "num_input_tokens_seen": 42575245, + "step": 1306 + }, + { + "epoch": 0.11786986517563242, + "flos": 13970814795360.0, + "grad_norm": 3.871217050138081, + "learning_rate": 3.919598285854134e-06, + "loss": 0.8005, + "num_input_tokens_seen": 42601625, + "step": 1307 + }, + { + "epoch": 0.11796004869910268, + "flos": 22386406507200.0, + "grad_norm": 2.5026783900744234, + "learning_rate": 3.919434223202411e-06, + "loss": 0.8577, + "num_input_tokens_seen": 42628875, + "step": 1308 + }, + { + "epoch": 0.11805023222257294, + "flos": 23370667880160.0, + "grad_norm": 1.6279551704851627, + "learning_rate": 3.919269996773977e-06, + "loss": 0.8564, + "num_input_tokens_seen": 42657735, + "step": 1309 + }, + { + "epoch": 0.1181404157460432, + "flos": 22602174195840.0, + "grad_norm": 1.4130942892105909, + "learning_rate": 3.919105606582844e-06, + "loss": 0.8142, + "num_input_tokens_seen": 42687845, + "step": 1310 + }, + { + "epoch": 0.11823059926951346, + "flos": 23990348080800.0, + "grad_norm": 2.6585569349458886, + "learning_rate": 3.918941052643039e-06, + "loss": 0.7656, + "num_input_tokens_seen": 42715195, + "step": 1311 + }, + { + "epoch": 0.11832078279298372, + "flos": 20820012124800.0, + "grad_norm": 1.562500686645357, + "learning_rate": 3.918776334968602e-06, + "loss": 0.8111, + "num_input_tokens_seen": 42743885, + "step": 1312 + }, + { + "epoch": 0.11841096631645398, + "flos": 21068866633440.0, + "grad_norm": 2.0006120460517507, + "learning_rate": 3.918611453573589e-06, + "loss": 0.9091, + "num_input_tokens_seen": 42770800, + "step": 1313 + }, + { + "epoch": 0.11850114983992424, + "flos": 20054752209600.0, + "grad_norm": 1.774441607045773, + "learning_rate": 3.918446408472066e-06, + "loss": 0.736, + "num_input_tokens_seen": 42800090, + "step": 1314 + }, + { + "epoch": 0.1185913333633945, + "flos": 27305111488800.0, + "grad_norm": 2.36310651148939, + "learning_rate": 3.918281199678119e-06, + "loss": 0.8379, + "num_input_tokens_seen": 42828115, + "step": 1315 + }, + { + "epoch": 0.11868151688686476, + "flos": 68475346944480.0, + "grad_norm": 0.7530979705645425, + "learning_rate": 3.9181158272058414e-06, + "loss": 0.6091, + "num_input_tokens_seen": 42918440, + "step": 1316 + }, + { + "epoch": 0.11877170041033504, + "flos": 21439952513760.0, + "grad_norm": 1.5150862847360584, + "learning_rate": 3.9179502910693455e-06, + "loss": 0.8288, + "num_input_tokens_seen": 42946550, + "step": 1317 + }, + { + "epoch": 0.1188618839338053, + "flos": 38344185983040.0, + "grad_norm": 1.659467558784762, + "learning_rate": 3.917784591282756e-06, + "loss": 0.7607, + "num_input_tokens_seen": 42978910, + "step": 1318 + }, + { + "epoch": 0.11895206745727556, + "flos": 24789549720480.0, + "grad_norm": 1.8259723085098256, + "learning_rate": 3.9176187278602105e-06, + "loss": 0.8243, + "num_input_tokens_seen": 43009550, + "step": 1319 + }, + { + "epoch": 0.11904225098074582, + "flos": 18780446500320.0, + "grad_norm": 1.692348015411527, + "learning_rate": 3.9174527008158606e-06, + "loss": 0.7431, + "num_input_tokens_seen": 43039360, + "step": 1320 + }, + { + "epoch": 0.11913243450421608, + "flos": 23626175775840.0, + "grad_norm": 1.985432380648775, + "learning_rate": 3.917286510163874e-06, + "loss": 0.7192, + "num_input_tokens_seen": 43069065, + "step": 1321 + }, + { + "epoch": 0.11922261802768634, + "flos": 26066345803680.0, + "grad_norm": 1.1760208544763435, + "learning_rate": 3.917120155918431e-06, + "loss": 0.7733, + "num_input_tokens_seen": 43100595, + "step": 1322 + }, + { + "epoch": 0.1193128015511566, + "flos": 66698314300320.0, + "grad_norm": 0.7031930678586851, + "learning_rate": 3.916953638093725e-06, + "loss": 0.6392, + "num_input_tokens_seen": 43195775, + "step": 1323 + }, + { + "epoch": 0.11940298507462686, + "flos": 23185385128320.0, + "grad_norm": 2.615510226594873, + "learning_rate": 3.916786956703964e-06, + "loss": 0.7922, + "num_input_tokens_seen": 43224385, + "step": 1324 + }, + { + "epoch": 0.11949316859809712, + "flos": 21257383154400.0, + "grad_norm": 2.448851343368254, + "learning_rate": 3.916620111763372e-06, + "loss": 0.7889, + "num_input_tokens_seen": 43252655, + "step": 1325 + }, + { + "epoch": 0.11958335212156739, + "flos": 16120866147360.0, + "grad_norm": 1.6870051470792768, + "learning_rate": 3.916453103286183e-06, + "loss": 0.861, + "num_input_tokens_seen": 43280295, + "step": 1326 + }, + { + "epoch": 0.11967353564503765, + "flos": 20740197182880.0, + "grad_norm": 2.1169181173757825, + "learning_rate": 3.916285931286648e-06, + "loss": 0.8466, + "num_input_tokens_seen": 43306225, + "step": 1327 + }, + { + "epoch": 0.11976371916850792, + "flos": 31419076536480.0, + "grad_norm": 1.898615385284983, + "learning_rate": 3.916118595779031e-06, + "loss": 0.7437, + "num_input_tokens_seen": 43337040, + "step": 1328 + }, + { + "epoch": 0.11985390269197818, + "flos": 26139693207360.0, + "grad_norm": 1.5358710335740597, + "learning_rate": 3.915951096777611e-06, + "loss": 0.769, + "num_input_tokens_seen": 43365985, + "step": 1329 + }, + { + "epoch": 0.11994408621544844, + "flos": 18707768152320.0, + "grad_norm": 1.8927670788995197, + "learning_rate": 3.915783434296678e-06, + "loss": 0.7469, + "num_input_tokens_seen": 43395535, + "step": 1330 + }, + { + "epoch": 0.1200342697389187, + "flos": 16883747197920.0, + "grad_norm": 2.447540730496922, + "learning_rate": 3.91561560835054e-06, + "loss": 0.7794, + "num_input_tokens_seen": 43419620, + "step": 1331 + }, + { + "epoch": 0.12012445326238896, + "flos": 21477239516640.0, + "grad_norm": 1.6573540228864, + "learning_rate": 3.915447618953515e-06, + "loss": 0.8167, + "num_input_tokens_seen": 43449285, + "step": 1332 + }, + { + "epoch": 0.12021463678585922, + "flos": 24645902833440.0, + "grad_norm": 6.627634927994264, + "learning_rate": 3.915279466119937e-06, + "loss": 0.7538, + "num_input_tokens_seen": 43476975, + "step": 1333 + }, + { + "epoch": 0.12030482030932949, + "flos": 68820594108000.0, + "grad_norm": 0.7229693585045543, + "learning_rate": 3.9151111498641546e-06, + "loss": 0.6238, + "num_input_tokens_seen": 43569080, + "step": 1334 + }, + { + "epoch": 0.12039500383279975, + "flos": 38307530866080.0, + "grad_norm": 2.0739040100044175, + "learning_rate": 3.914942670200529e-06, + "loss": 0.6724, + "num_input_tokens_seen": 43597180, + "step": 1335 + }, + { + "epoch": 0.12048518735627001, + "flos": 25812064510080.0, + "grad_norm": 2.0967268776346986, + "learning_rate": 3.914774027143436e-06, + "loss": 0.8576, + "num_input_tokens_seen": 43626830, + "step": 1336 + }, + { + "epoch": 0.12057537087974027, + "flos": 16958990259360.0, + "grad_norm": 1.6901994941837961, + "learning_rate": 3.914605220707265e-06, + "loss": 0.7879, + "num_input_tokens_seen": 43650345, + "step": 1337 + }, + { + "epoch": 0.12066555440321053, + "flos": 16157038057440.0, + "grad_norm": 2.0291793833287786, + "learning_rate": 3.9144362509064194e-06, + "loss": 0.9456, + "num_input_tokens_seen": 43675870, + "step": 1338 + }, + { + "epoch": 0.12075573792668079, + "flos": 13934865903840.0, + "grad_norm": 2.3930864813896022, + "learning_rate": 3.914267117755317e-06, + "loss": 0.8657, + "num_input_tokens_seen": 43701045, + "step": 1339 + }, + { + "epoch": 0.12084592145015106, + "flos": 33899432808480.0, + "grad_norm": 1.6478007022197314, + "learning_rate": 3.914097821268389e-06, + "loss": 0.7653, + "num_input_tokens_seen": 43731235, + "step": 1340 + }, + { + "epoch": 0.12093610497362133, + "flos": 25301123058240.0, + "grad_norm": 1.9720119519442167, + "learning_rate": 3.913928361460081e-06, + "loss": 0.8127, + "num_input_tokens_seen": 43761750, + "step": 1341 + }, + { + "epoch": 0.12102628849709159, + "flos": 20887746894720.0, + "grad_norm": 1.7147992591039447, + "learning_rate": 3.913758738344851e-06, + "loss": 0.7855, + "num_input_tokens_seen": 43788180, + "step": 1342 + }, + { + "epoch": 0.12111647202056185, + "flos": 65378655750240.0, + "grad_norm": 0.730508446252286, + "learning_rate": 3.913588951937174e-06, + "loss": 0.6, + "num_input_tokens_seen": 43873345, + "step": 1343 + }, + { + "epoch": 0.12120665554403211, + "flos": 67920093405120.0, + "grad_norm": 0.7152368267356899, + "learning_rate": 3.9134190022515355e-06, + "loss": 0.6013, + "num_input_tokens_seen": 43961960, + "step": 1344 + }, + { + "epoch": 0.12129683906750237, + "flos": 20310668972640.0, + "grad_norm": 2.526629245232666, + "learning_rate": 3.913248889302438e-06, + "loss": 0.7563, + "num_input_tokens_seen": 43991560, + "step": 1345 + }, + { + "epoch": 0.12138702259097263, + "flos": 30618759804000.0, + "grad_norm": 1.7050326102038402, + "learning_rate": 3.913078613104395e-06, + "loss": 0.7321, + "num_input_tokens_seen": 44025320, + "step": 1346 + }, + { + "epoch": 0.12147720611444289, + "flos": 18780669518880.0, + "grad_norm": 2.60835812793521, + "learning_rate": 3.912908173671936e-06, + "loss": 0.6798, + "num_input_tokens_seen": 44052350, + "step": 1347 + }, + { + "epoch": 0.12156738963791315, + "flos": 23334681818880.0, + "grad_norm": 1.8871854279973088, + "learning_rate": 3.9127375710196044e-06, + "loss": 0.7258, + "num_input_tokens_seen": 44080130, + "step": 1348 + }, + { + "epoch": 0.12165757316138341, + "flos": 26358620325600.0, + "grad_norm": 1.798713075684306, + "learning_rate": 3.912566805161957e-06, + "loss": 0.7433, + "num_input_tokens_seen": 44109660, + "step": 1349 + }, + { + "epoch": 0.12174775668485367, + "flos": 19137296362560.0, + "grad_norm": 1.871892197264068, + "learning_rate": 3.912395876113564e-06, + "loss": 0.8155, + "num_input_tokens_seen": 44135555, + "step": 1350 + }, + { + "epoch": 0.12183794020832393, + "flos": 19618756461120.0, + "grad_norm": 1.6735960043844882, + "learning_rate": 3.912224783889009e-06, + "loss": 0.7997, + "num_input_tokens_seen": 44163430, + "step": 1351 + }, + { + "epoch": 0.12192812373179421, + "flos": 19720990501920.0, + "grad_norm": 1.77105889473277, + "learning_rate": 3.912053528502892e-06, + "loss": 0.8108, + "num_input_tokens_seen": 44188270, + "step": 1352 + }, + { + "epoch": 0.12201830725526447, + "flos": 43449623337120.0, + "grad_norm": 1.8881050529550156, + "learning_rate": 3.911882109969825e-06, + "loss": 0.7288, + "num_input_tokens_seen": 44225540, + "step": 1353 + }, + { + "epoch": 0.12210849077873473, + "flos": 19728647472480.0, + "grad_norm": 2.141623939157209, + "learning_rate": 3.911710528304435e-06, + "loss": 0.8068, + "num_input_tokens_seen": 44255165, + "step": 1354 + }, + { + "epoch": 0.12219867430220499, + "flos": 17906150478240.0, + "grad_norm": 2.433561811650875, + "learning_rate": 3.911538783521361e-06, + "loss": 0.7502, + "num_input_tokens_seen": 44282210, + "step": 1355 + }, + { + "epoch": 0.12228885782567525, + "flos": 25993481606880.0, + "grad_norm": 1.6724769720707469, + "learning_rate": 3.9113668756352575e-06, + "loss": 0.8098, + "num_input_tokens_seen": 44311225, + "step": 1356 + }, + { + "epoch": 0.12237904134914551, + "flos": 27995462870400.0, + "grad_norm": 1.5109894641986057, + "learning_rate": 3.911194804660793e-06, + "loss": 0.7244, + "num_input_tokens_seen": 44342735, + "step": 1357 + }, + { + "epoch": 0.12246922487261577, + "flos": 21694754184000.0, + "grad_norm": 1.7300033460154727, + "learning_rate": 3.91102257061265e-06, + "loss": 0.8802, + "num_input_tokens_seen": 44370620, + "step": 1358 + }, + { + "epoch": 0.12255940839608603, + "flos": 20891538210240.0, + "grad_norm": 1.5099082968194053, + "learning_rate": 3.910850173505524e-06, + "loss": 0.8374, + "num_input_tokens_seen": 44399555, + "step": 1359 + }, + { + "epoch": 0.1226495919195563, + "flos": 18521370307680.0, + "grad_norm": 2.931222416924876, + "learning_rate": 3.9106776133541255e-06, + "loss": 0.8723, + "num_input_tokens_seen": 44425645, + "step": 1360 + }, + { + "epoch": 0.12273977544302656, + "flos": 18483377079360.0, + "grad_norm": 1.5646705809680903, + "learning_rate": 3.9105048901731766e-06, + "loss": 0.7593, + "num_input_tokens_seen": 44451875, + "step": 1361 + }, + { + "epoch": 0.12282995896649682, + "flos": 25080002924160.0, + "grad_norm": 2.4937786894178147, + "learning_rate": 3.9103320039774165e-06, + "loss": 0.7896, + "num_input_tokens_seen": 44481580, + "step": 1362 + }, + { + "epoch": 0.12292014248996708, + "flos": 20996745831840.0, + "grad_norm": 1.6346923374169353, + "learning_rate": 3.9101589547815965e-06, + "loss": 0.8559, + "num_input_tokens_seen": 44509505, + "step": 1363 + }, + { + "epoch": 0.12301032601343735, + "flos": 22969357251360.0, + "grad_norm": 1.4144605213425956, + "learning_rate": 3.909985742600482e-06, + "loss": 0.7886, + "num_input_tokens_seen": 44538405, + "step": 1364 + }, + { + "epoch": 0.12310050953690761, + "flos": 32186938334880.0, + "grad_norm": 1.6130656913487198, + "learning_rate": 3.909812367448852e-06, + "loss": 0.7617, + "num_input_tokens_seen": 44568000, + "step": 1365 + }, + { + "epoch": 0.12319069306037787, + "flos": 25694628037440.0, + "grad_norm": 2.91404204578738, + "learning_rate": 3.909638829341501e-06, + "loss": 0.773, + "num_input_tokens_seen": 44591100, + "step": 1366 + }, + { + "epoch": 0.12328087658384813, + "flos": 21979148716800.0, + "grad_norm": 1.6916572789775777, + "learning_rate": 3.909465128293234e-06, + "loss": 0.8532, + "num_input_tokens_seen": 44618000, + "step": 1367 + }, + { + "epoch": 0.1233710601073184, + "flos": 22419865024800.0, + "grad_norm": 1.646546712217952, + "learning_rate": 3.9092912643188745e-06, + "loss": 0.8555, + "num_input_tokens_seen": 44648510, + "step": 1368 + }, + { + "epoch": 0.12346124363078866, + "flos": 23334495970080.0, + "grad_norm": 1.8916201533958334, + "learning_rate": 3.909117237433256e-06, + "loss": 0.7894, + "num_input_tokens_seen": 44675835, + "step": 1369 + }, + { + "epoch": 0.12355142715425892, + "flos": 21658693783200.0, + "grad_norm": 2.0833041252631896, + "learning_rate": 3.908943047651229e-06, + "loss": 0.8396, + "num_input_tokens_seen": 44703765, + "step": 1370 + }, + { + "epoch": 0.12364161067772918, + "flos": 21075297001920.0, + "grad_norm": 1.65090880074974, + "learning_rate": 3.908768694987655e-06, + "loss": 0.7847, + "num_input_tokens_seen": 44734660, + "step": 1371 + }, + { + "epoch": 0.12373179420119944, + "flos": 21654121902720.0, + "grad_norm": 2.6576855538902517, + "learning_rate": 3.908594179457411e-06, + "loss": 0.7207, + "num_input_tokens_seen": 44759515, + "step": 1372 + }, + { + "epoch": 0.1238219777246697, + "flos": 23619968425920.0, + "grad_norm": 2.194702824003117, + "learning_rate": 3.908419501075388e-06, + "loss": 0.8026, + "num_input_tokens_seen": 44788030, + "step": 1373 + }, + { + "epoch": 0.12391216124813996, + "flos": 28290153426720.0, + "grad_norm": 2.0147931652072066, + "learning_rate": 3.90824465985649e-06, + "loss": 0.87, + "num_input_tokens_seen": 44820615, + "step": 1374 + }, + { + "epoch": 0.12400234477161022, + "flos": 14991396757440.0, + "grad_norm": 1.8990207608149268, + "learning_rate": 3.908069655815636e-06, + "loss": 0.8398, + "num_input_tokens_seen": 44846450, + "step": 1375 + }, + { + "epoch": 0.1240925282950805, + "flos": 16193544495360.0, + "grad_norm": 2.110091250286037, + "learning_rate": 3.907894488967758e-06, + "loss": 0.7345, + "num_input_tokens_seen": 44872645, + "step": 1376 + }, + { + "epoch": 0.12418271181855076, + "flos": 22751396546880.0, + "grad_norm": 1.8314950642312267, + "learning_rate": 3.9077191593278005e-06, + "loss": 0.857, + "num_input_tokens_seen": 44900875, + "step": 1377 + }, + { + "epoch": 0.12427289534202102, + "flos": 67298926414080.0, + "grad_norm": 0.7134372653367621, + "learning_rate": 3.9075436669107265e-06, + "loss": 0.5773, + "num_input_tokens_seen": 44990410, + "step": 1378 + }, + { + "epoch": 0.12436307886549128, + "flos": 15173594419200.0, + "grad_norm": 1.7155098631532761, + "learning_rate": 3.90736801173151e-06, + "loss": 0.8064, + "num_input_tokens_seen": 45016875, + "step": 1379 + }, + { + "epoch": 0.12445326238896154, + "flos": 26504088530880.0, + "grad_norm": 1.523350092018704, + "learning_rate": 3.907192193805136e-06, + "loss": 0.8077, + "num_input_tokens_seen": 45047750, + "step": 1380 + }, + { + "epoch": 0.1245434459124318, + "flos": 25666447625760.0, + "grad_norm": 2.1797966605030243, + "learning_rate": 3.907016213146608e-06, + "loss": 0.849, + "num_input_tokens_seen": 45074830, + "step": 1381 + }, + { + "epoch": 0.12463362943590206, + "flos": 22750987679520.0, + "grad_norm": 1.8879789010137846, + "learning_rate": 3.906840069770942e-06, + "loss": 0.9099, + "num_input_tokens_seen": 45103020, + "step": 1382 + }, + { + "epoch": 0.12472381295937232, + "flos": 24317344892160.0, + "grad_norm": 1.5917323936086616, + "learning_rate": 3.906663763693167e-06, + "loss": 0.8292, + "num_input_tokens_seen": 45133210, + "step": 1383 + }, + { + "epoch": 0.12481399648284258, + "flos": 26099395453920.0, + "grad_norm": 1.7233004587285634, + "learning_rate": 3.906487294928327e-06, + "loss": 0.7785, + "num_input_tokens_seen": 45163740, + "step": 1384 + }, + { + "epoch": 0.12490418000631284, + "flos": 21729327794400.0, + "grad_norm": 1.851751454687133, + "learning_rate": 3.906310663491478e-06, + "loss": 0.8004, + "num_input_tokens_seen": 45191625, + "step": 1385 + }, + { + "epoch": 0.1249943635297831, + "flos": 21804347837280.0, + "grad_norm": 1.861986138035099, + "learning_rate": 3.906133869397692e-06, + "loss": 0.852, + "num_input_tokens_seen": 45218240, + "step": 1386 + }, + { + "epoch": 0.12508454705325336, + "flos": 35028827858880.0, + "grad_norm": 1.729927684507326, + "learning_rate": 3.905956912662054e-06, + "loss": 0.7667, + "num_input_tokens_seen": 45247410, + "step": 1387 + }, + { + "epoch": 0.12517473057672363, + "flos": 20126241125280.0, + "grad_norm": 2.1755989039389387, + "learning_rate": 3.905779793299662e-06, + "loss": 0.8625, + "num_input_tokens_seen": 45268790, + "step": 1388 + }, + { + "epoch": 0.12526491410019389, + "flos": 19472656369920.0, + "grad_norm": 1.8626494226105763, + "learning_rate": 3.905602511325631e-06, + "loss": 0.8319, + "num_input_tokens_seen": 45295905, + "step": 1389 + }, + { + "epoch": 0.12535509762366415, + "flos": 20492978143680.0, + "grad_norm": 1.6528731415677638, + "learning_rate": 3.905425066755086e-06, + "loss": 0.7602, + "num_input_tokens_seen": 45324400, + "step": 1390 + }, + { + "epoch": 0.1254452811471344, + "flos": 28505623757280.0, + "grad_norm": 8.005876766799236, + "learning_rate": 3.905247459603168e-06, + "loss": 0.7538, + "num_input_tokens_seen": 45356220, + "step": 1391 + }, + { + "epoch": 0.12553546467060467, + "flos": 24316415648160.0, + "grad_norm": 1.907638544373384, + "learning_rate": 3.905069689885031e-06, + "loss": 0.6804, + "num_input_tokens_seen": 45386375, + "step": 1392 + }, + { + "epoch": 0.12562564819407493, + "flos": 20456434536000.0, + "grad_norm": 1.6611648906321788, + "learning_rate": 3.904891757615843e-06, + "loss": 0.683, + "num_input_tokens_seen": 45415875, + "step": 1393 + }, + { + "epoch": 0.12571583171754522, + "flos": 69141718097280.0, + "grad_norm": 0.6304957756291206, + "learning_rate": 3.9047136628107874e-06, + "loss": 0.5629, + "num_input_tokens_seen": 45508200, + "step": 1394 + }, + { + "epoch": 0.12580601524101548, + "flos": 27518500312800.0, + "grad_norm": 1.759424309861077, + "learning_rate": 3.904535405485059e-06, + "loss": 0.9267, + "num_input_tokens_seen": 45537500, + "step": 1395 + }, + { + "epoch": 0.12589619876448574, + "flos": 24646683398400.0, + "grad_norm": 1.722392643765139, + "learning_rate": 3.90435698565387e-06, + "loss": 0.7472, + "num_input_tokens_seen": 45564920, + "step": 1396 + }, + { + "epoch": 0.125986382287956, + "flos": 39178890477120.0, + "grad_norm": 1.712085147366832, + "learning_rate": 3.904178403332441e-06, + "loss": 0.7612, + "num_input_tokens_seen": 45598040, + "step": 1397 + }, + { + "epoch": 0.12607656581142626, + "flos": 18926026214880.0, + "grad_norm": 2.041795674614328, + "learning_rate": 3.903999658536012e-06, + "loss": 0.8852, + "num_input_tokens_seen": 45624805, + "step": 1398 + }, + { + "epoch": 0.12616674933489652, + "flos": 24643561138560.0, + "grad_norm": 1.7540279762490203, + "learning_rate": 3.903820751279833e-06, + "loss": 0.7959, + "num_input_tokens_seen": 45654495, + "step": 1399 + }, + { + "epoch": 0.12625693285836678, + "flos": 17723655458400.0, + "grad_norm": 2.5337336094330354, + "learning_rate": 3.90364168157917e-06, + "loss": 0.833, + "num_input_tokens_seen": 45679910, + "step": 1400 + }, + { + "epoch": 0.12634711638183704, + "flos": 23006346896160.0, + "grad_norm": 2.0865933533770247, + "learning_rate": 3.903462449449302e-06, + "loss": 0.8304, + "num_input_tokens_seen": 45707615, + "step": 1401 + }, + { + "epoch": 0.1264372999053073, + "flos": 22605965511360.0, + "grad_norm": 1.5020947135391904, + "learning_rate": 3.903283054905522e-06, + "loss": 0.7688, + "num_input_tokens_seen": 45737435, + "step": 1402 + }, + { + "epoch": 0.12652748342877757, + "flos": 25812027340320.0, + "grad_norm": 1.8981363819071215, + "learning_rate": 3.9031034979631385e-06, + "loss": 0.708, + "num_input_tokens_seen": 45766840, + "step": 1403 + }, + { + "epoch": 0.12661766695224783, + "flos": 21654493600320.0, + "grad_norm": 1.5919781727899949, + "learning_rate": 3.902923778637469e-06, + "loss": 0.831, + "num_input_tokens_seen": 45794910, + "step": 1404 + }, + { + "epoch": 0.1267078504757181, + "flos": 29126196032160.0, + "grad_norm": 2.036608977408666, + "learning_rate": 3.902743896943852e-06, + "loss": 0.7297, + "num_input_tokens_seen": 45824230, + "step": 1405 + }, + { + "epoch": 0.12679803399918835, + "flos": 67781092738080.0, + "grad_norm": 0.7173128687502073, + "learning_rate": 3.902563852897633e-06, + "loss": 0.6319, + "num_input_tokens_seen": 45909875, + "step": 1406 + }, + { + "epoch": 0.1268882175226586, + "flos": 65404382957760.0, + "grad_norm": 0.6956210148466695, + "learning_rate": 3.9023836465141755e-06, + "loss": 0.6065, + "num_input_tokens_seen": 45998270, + "step": 1407 + }, + { + "epoch": 0.12697840104612887, + "flos": 23327916922560.0, + "grad_norm": 2.0285177782129193, + "learning_rate": 3.902203277808856e-06, + "loss": 0.8447, + "num_input_tokens_seen": 46023885, + "step": 1408 + }, + { + "epoch": 0.12706858456959913, + "flos": 21112695514080.0, + "grad_norm": 1.773772825560643, + "learning_rate": 3.902022746797064e-06, + "loss": 0.8536, + "num_input_tokens_seen": 46051155, + "step": 1409 + }, + { + "epoch": 0.1271587680930694, + "flos": 14371939575360.0, + "grad_norm": 3.0445767073043433, + "learning_rate": 3.9018420534942035e-06, + "loss": 0.9058, + "num_input_tokens_seen": 46076405, + "step": 1410 + }, + { + "epoch": 0.12724895161653965, + "flos": 23589148961280.0, + "grad_norm": 2.264531891304925, + "learning_rate": 3.9016611979156935e-06, + "loss": 0.7903, + "num_input_tokens_seen": 46105550, + "step": 1411 + }, + { + "epoch": 0.1273391351400099, + "flos": 24391026823680.0, + "grad_norm": 2.0568723484649305, + "learning_rate": 3.9014801800769635e-06, + "loss": 0.701, + "num_input_tokens_seen": 46133325, + "step": 1412 + }, + { + "epoch": 0.12742931866348017, + "flos": 15063926426400.0, + "grad_norm": 2.7662415760048193, + "learning_rate": 3.901298999993459e-06, + "loss": 0.8237, + "num_input_tokens_seen": 46157245, + "step": 1413 + }, + { + "epoch": 0.12751950218695043, + "flos": 13716050294880.0, + "grad_norm": 2.4759592952248037, + "learning_rate": 3.901117657680642e-06, + "loss": 0.728, + "num_input_tokens_seen": 46184445, + "step": 1414 + }, + { + "epoch": 0.1276096857104207, + "flos": 23735286222240.0, + "grad_norm": 2.725062091364972, + "learning_rate": 3.900936153153982e-06, + "loss": 0.7292, + "num_input_tokens_seen": 46211780, + "step": 1415 + }, + { + "epoch": 0.12769986923389096, + "flos": 68101659180960.0, + "grad_norm": 0.7147826705436181, + "learning_rate": 3.900754486428968e-06, + "loss": 0.5589, + "num_input_tokens_seen": 46294230, + "step": 1416 + }, + { + "epoch": 0.12779005275736122, + "flos": 23188953425280.0, + "grad_norm": 3.204455885382463, + "learning_rate": 3.900572657521102e-06, + "loss": 0.8848, + "num_input_tokens_seen": 46320940, + "step": 1417 + }, + { + "epoch": 0.1278802362808315, + "flos": 23004339729120.0, + "grad_norm": 1.4928719114380078, + "learning_rate": 3.900390666445896e-06, + "loss": 0.7809, + "num_input_tokens_seen": 46351370, + "step": 1418 + }, + { + "epoch": 0.12797041980430177, + "flos": 22820357918880.0, + "grad_norm": 1.8956488075861733, + "learning_rate": 3.9002085132188795e-06, + "loss": 0.764, + "num_input_tokens_seen": 46380080, + "step": 1419 + }, + { + "epoch": 0.12806060332777203, + "flos": 22496594876640.0, + "grad_norm": 1.7448954385292395, + "learning_rate": 3.9000261978555964e-06, + "loss": 0.8409, + "num_input_tokens_seen": 46409330, + "step": 1420 + }, + { + "epoch": 0.1281507868512423, + "flos": 28321418928480.0, + "grad_norm": 2.203860004728221, + "learning_rate": 3.8998437203716e-06, + "loss": 0.8634, + "num_input_tokens_seen": 46438655, + "step": 1421 + }, + { + "epoch": 0.12824097037471255, + "flos": 23444349811680.0, + "grad_norm": 1.893110737688627, + "learning_rate": 3.899661080782462e-06, + "loss": 0.8226, + "num_input_tokens_seen": 46467260, + "step": 1422 + }, + { + "epoch": 0.1283311538981828, + "flos": 20493015313440.0, + "grad_norm": 1.9438900936949488, + "learning_rate": 3.899478279103767e-06, + "loss": 0.8258, + "num_input_tokens_seen": 46495270, + "step": 1423 + }, + { + "epoch": 0.12842133742165307, + "flos": 23662087497600.0, + "grad_norm": 1.9891978251532965, + "learning_rate": 3.8992953153511105e-06, + "loss": 0.8256, + "num_input_tokens_seen": 46524965, + "step": 1424 + }, + { + "epoch": 0.12851152094512333, + "flos": 22310345711040.0, + "grad_norm": 1.6732027717330267, + "learning_rate": 3.899112189540106e-06, + "loss": 0.8619, + "num_input_tokens_seen": 46551900, + "step": 1425 + }, + { + "epoch": 0.1286017044685936, + "flos": 25265025487680.0, + "grad_norm": 2.4181649616850764, + "learning_rate": 3.898928901686377e-06, + "loss": 0.83, + "num_input_tokens_seen": 46578030, + "step": 1426 + }, + { + "epoch": 0.12869188799206385, + "flos": 60770481376320.0, + "grad_norm": 0.7160676162082797, + "learning_rate": 3.898745451805564e-06, + "loss": 0.5642, + "num_input_tokens_seen": 46667255, + "step": 1427 + }, + { + "epoch": 0.1287820715155341, + "flos": 25883479086240.0, + "grad_norm": 1.6725468934331953, + "learning_rate": 3.898561839913319e-06, + "loss": 0.7344, + "num_input_tokens_seen": 46700005, + "step": 1428 + }, + { + "epoch": 0.12887225503900437, + "flos": 22202907903840.0, + "grad_norm": 2.5856608908423473, + "learning_rate": 3.89837806602531e-06, + "loss": 0.6443, + "num_input_tokens_seen": 46728515, + "step": 1429 + }, + { + "epoch": 0.12896243856247463, + "flos": 19654705352640.0, + "grad_norm": 4.8085451596219855, + "learning_rate": 3.898194130157217e-06, + "loss": 0.807, + "num_input_tokens_seen": 46757275, + "step": 1430 + }, + { + "epoch": 0.1290526220859449, + "flos": 14481756247200.0, + "grad_norm": 2.2976123897720506, + "learning_rate": 3.8980100323247335e-06, + "loss": 0.7399, + "num_input_tokens_seen": 46783475, + "step": 1431 + }, + { + "epoch": 0.12914280560941516, + "flos": 22132199553120.0, + "grad_norm": 1.9639054289461073, + "learning_rate": 3.897825772543568e-06, + "loss": 0.73, + "num_input_tokens_seen": 46811195, + "step": 1432 + }, + { + "epoch": 0.12923298913288542, + "flos": 20091853363680.0, + "grad_norm": 2.991056462557508, + "learning_rate": 3.897641350829444e-06, + "loss": 0.8526, + "num_input_tokens_seen": 46838050, + "step": 1433 + }, + { + "epoch": 0.12932317265635568, + "flos": 22532357919360.0, + "grad_norm": 1.871607763609628, + "learning_rate": 3.897456767198096e-06, + "loss": 0.884, + "num_input_tokens_seen": 46866725, + "step": 1434 + }, + { + "epoch": 0.12941335617982594, + "flos": 34264162659840.0, + "grad_norm": 1.767224288004369, + "learning_rate": 3.897272021665275e-06, + "loss": 0.7367, + "num_input_tokens_seen": 46897635, + "step": 1435 + }, + { + "epoch": 0.1295035397032962, + "flos": 24791519717760.0, + "grad_norm": 1.7806684648987205, + "learning_rate": 3.897087114246743e-06, + "loss": 0.9098, + "num_input_tokens_seen": 46927650, + "step": 1436 + }, + { + "epoch": 0.12959372322676646, + "flos": 20383310150880.0, + "grad_norm": 5.193055636449673, + "learning_rate": 3.896902044958279e-06, + "loss": 0.9128, + "num_input_tokens_seen": 46955170, + "step": 1437 + }, + { + "epoch": 0.12968390675023672, + "flos": 23006235386880.0, + "grad_norm": 2.453661234372823, + "learning_rate": 3.896716813815672e-06, + "loss": 0.8518, + "num_input_tokens_seen": 46980820, + "step": 1438 + }, + { + "epoch": 0.12977409027370698, + "flos": 19067591595360.0, + "grad_norm": 2.3802848041250817, + "learning_rate": 3.896531420834728e-06, + "loss": 0.7468, + "num_input_tokens_seen": 47009715, + "step": 1439 + }, + { + "epoch": 0.12986427379717724, + "flos": 43441892027040.0, + "grad_norm": 1.781009523990618, + "learning_rate": 3.896345866031266e-06, + "loss": 0.711, + "num_input_tokens_seen": 47046045, + "step": 1440 + }, + { + "epoch": 0.1299544573206475, + "flos": 28244912095200.0, + "grad_norm": 1.7991587182853417, + "learning_rate": 3.896160149421119e-06, + "loss": 0.8622, + "num_input_tokens_seen": 47074685, + "step": 1441 + }, + { + "epoch": 0.1300446408441178, + "flos": 16995273678720.0, + "grad_norm": 1.8498428020189706, + "learning_rate": 3.8959742710201314e-06, + "loss": 0.8699, + "num_input_tokens_seen": 47098445, + "step": 1442 + }, + { + "epoch": 0.13013482436758805, + "flos": 25957049508480.0, + "grad_norm": 2.5174393833897466, + "learning_rate": 3.895788230844166e-06, + "loss": 0.6955, + "num_input_tokens_seen": 47124115, + "step": 1443 + }, + { + "epoch": 0.13022500789105831, + "flos": 29235269308800.0, + "grad_norm": 2.031680134933847, + "learning_rate": 3.895602028909095e-06, + "loss": 0.9556, + "num_input_tokens_seen": 47152995, + "step": 1444 + }, + { + "epoch": 0.13031519141452858, + "flos": 69645708804000.0, + "grad_norm": 0.6337786939622189, + "learning_rate": 3.895415665230807e-06, + "loss": 0.5969, + "num_input_tokens_seen": 47250825, + "step": 1445 + }, + { + "epoch": 0.13040537493799884, + "flos": 25080077263680.0, + "grad_norm": 1.703480464692289, + "learning_rate": 3.895229139825203e-06, + "loss": 0.8236, + "num_input_tokens_seen": 47278510, + "step": 1446 + }, + { + "epoch": 0.1304955584614691, + "flos": 29819558164320.0, + "grad_norm": 1.905130839522917, + "learning_rate": 3.895042452708198e-06, + "loss": 0.7104, + "num_input_tokens_seen": 47307640, + "step": 1447 + }, + { + "epoch": 0.13058574198493936, + "flos": 23002964448000.0, + "grad_norm": 1.9708719474230278, + "learning_rate": 3.894855603895723e-06, + "loss": 0.7083, + "num_input_tokens_seen": 47336490, + "step": 1448 + }, + { + "epoch": 0.13067592550840962, + "flos": 23116052058720.0, + "grad_norm": 1.8560939016758766, + "learning_rate": 3.894668593403718e-06, + "loss": 0.8785, + "num_input_tokens_seen": 47365030, + "step": 1449 + }, + { + "epoch": 0.13076610903187988, + "flos": 27229087862400.0, + "grad_norm": 2.0253885061313652, + "learning_rate": 3.8944814212481425e-06, + "loss": 0.8061, + "num_input_tokens_seen": 47393990, + "step": 1450 + }, + { + "epoch": 0.13085629255535014, + "flos": 18452520444960.0, + "grad_norm": 1.7846477293440246, + "learning_rate": 3.894294087444966e-06, + "loss": 0.7793, + "num_input_tokens_seen": 47421110, + "step": 1451 + }, + { + "epoch": 0.1309464760788204, + "flos": 34476064693440.0, + "grad_norm": 1.7209452394886728, + "learning_rate": 3.894106592010173e-06, + "loss": 0.689, + "num_input_tokens_seen": 47450085, + "step": 1452 + }, + { + "epoch": 0.13103665960229066, + "flos": 26467321904640.0, + "grad_norm": 1.9490879909407486, + "learning_rate": 3.893918934959762e-06, + "loss": 0.6735, + "num_input_tokens_seen": 47478910, + "step": 1453 + }, + { + "epoch": 0.13112684312576092, + "flos": 26830899493440.0, + "grad_norm": 2.127105231052509, + "learning_rate": 3.893731116309743e-06, + "loss": 0.7361, + "num_input_tokens_seen": 47504115, + "step": 1454 + }, + { + "epoch": 0.13121702664923118, + "flos": 32369805052320.0, + "grad_norm": 2.0697629955080514, + "learning_rate": 3.893543136076145e-06, + "loss": 0.8098, + "num_input_tokens_seen": 47534010, + "step": 1455 + }, + { + "epoch": 0.13130721017270144, + "flos": 16625302891200.0, + "grad_norm": 3.8133322635951465, + "learning_rate": 3.893354994275006e-06, + "loss": 0.8253, + "num_input_tokens_seen": 47558395, + "step": 1456 + }, + { + "epoch": 0.1313973936961717, + "flos": 24063546805440.0, + "grad_norm": 1.6512248261559272, + "learning_rate": 3.893166690922378e-06, + "loss": 0.767, + "num_input_tokens_seen": 47587150, + "step": 1457 + }, + { + "epoch": 0.13148757721964197, + "flos": 36231570312960.0, + "grad_norm": 2.207678559837386, + "learning_rate": 3.892978226034329e-06, + "loss": 0.8272, + "num_input_tokens_seen": 47617740, + "step": 1458 + }, + { + "epoch": 0.13157776074311223, + "flos": 21221694451200.0, + "grad_norm": 2.9751493849233626, + "learning_rate": 3.89278959962694e-06, + "loss": 0.8497, + "num_input_tokens_seen": 47644490, + "step": 1459 + }, + { + "epoch": 0.1316679442665825, + "flos": 48909494519040.0, + "grad_norm": 1.682234779868434, + "learning_rate": 3.8926008117163056e-06, + "loss": 0.7729, + "num_input_tokens_seen": 47677000, + "step": 1460 + }, + { + "epoch": 0.13175812779005275, + "flos": 26279102741760.0, + "grad_norm": 2.039672643351112, + "learning_rate": 3.892411862318535e-06, + "loss": 0.7862, + "num_input_tokens_seen": 47708135, + "step": 1461 + }, + { + "epoch": 0.131848311313523, + "flos": 21037935659520.0, + "grad_norm": 1.5167248068484516, + "learning_rate": 3.892222751449749e-06, + "loss": 0.8213, + "num_input_tokens_seen": 47735115, + "step": 1462 + }, + { + "epoch": 0.13193849483699327, + "flos": 19181088073440.0, + "grad_norm": 1.866335175519387, + "learning_rate": 3.892033479126084e-06, + "loss": 0.7446, + "num_input_tokens_seen": 47761565, + "step": 1463 + }, + { + "epoch": 0.13202867836046353, + "flos": 20164791900000.0, + "grad_norm": 2.535856975938705, + "learning_rate": 3.891844045363691e-06, + "loss": 0.7809, + "num_input_tokens_seen": 47789590, + "step": 1464 + }, + { + "epoch": 0.13211886188393382, + "flos": 21221025395520.0, + "grad_norm": 2.1549522807066523, + "learning_rate": 3.891654450178732e-06, + "loss": 0.7217, + "num_input_tokens_seen": 47817300, + "step": 1465 + }, + { + "epoch": 0.13220904540740408, + "flos": 20675435993760.0, + "grad_norm": 1.7323361346542867, + "learning_rate": 3.891464693587385e-06, + "loss": 0.8444, + "num_input_tokens_seen": 47846800, + "step": 1466 + }, + { + "epoch": 0.13229922893087434, + "flos": 27558872405760.0, + "grad_norm": 1.838334551880272, + "learning_rate": 3.89127477560584e-06, + "loss": 0.8369, + "num_input_tokens_seen": 47878320, + "step": 1467 + }, + { + "epoch": 0.1323894124543446, + "flos": 23407434506400.0, + "grad_norm": 1.7155777526908986, + "learning_rate": 3.891084696250304e-06, + "loss": 0.8359, + "num_input_tokens_seen": 47908895, + "step": 1468 + }, + { + "epoch": 0.13247959597781486, + "flos": 22788237512640.0, + "grad_norm": 1.8211148830711201, + "learning_rate": 3.890894455536993e-06, + "loss": 0.8374, + "num_input_tokens_seen": 47936830, + "step": 1469 + }, + { + "epoch": 0.13256977950128512, + "flos": 16442622022560.0, + "grad_norm": 2.669206145208286, + "learning_rate": 3.890704053482142e-06, + "loss": 0.8455, + "num_input_tokens_seen": 47959710, + "step": 1470 + }, + { + "epoch": 0.13265996302475538, + "flos": 22860767181600.0, + "grad_norm": 2.2110298056592708, + "learning_rate": 3.890513490101995e-06, + "loss": 0.7455, + "num_input_tokens_seen": 47987645, + "step": 1471 + }, + { + "epoch": 0.13275014654822564, + "flos": 65467763132160.0, + "grad_norm": 0.6217717482763064, + "learning_rate": 3.890322765412814e-06, + "loss": 0.6064, + "num_input_tokens_seen": 48086345, + "step": 1472 + }, + { + "epoch": 0.1328403300716959, + "flos": 25078739152320.0, + "grad_norm": 1.4154395885181772, + "learning_rate": 3.890131879430871e-06, + "loss": 0.8274, + "num_input_tokens_seen": 48115460, + "step": 1473 + }, + { + "epoch": 0.13293051359516617, + "flos": 22349751390240.0, + "grad_norm": 1.6394293651808725, + "learning_rate": 3.889940832172454e-06, + "loss": 0.8244, + "num_input_tokens_seen": 48143860, + "step": 1474 + }, + { + "epoch": 0.13302069711863643, + "flos": 18884055822240.0, + "grad_norm": 1.724443542180093, + "learning_rate": 3.889749623653864e-06, + "loss": 0.845, + "num_input_tokens_seen": 48169175, + "step": 1475 + }, + { + "epoch": 0.1331108806421067, + "flos": 21149053272960.0, + "grad_norm": 1.5413667885494555, + "learning_rate": 3.889558253891416e-06, + "loss": 0.7432, + "num_input_tokens_seen": 48198415, + "step": 1476 + }, + { + "epoch": 0.13320106416557695, + "flos": 67234128055200.0, + "grad_norm": 0.6227250658248039, + "learning_rate": 3.8893667229014385e-06, + "loss": 0.581, + "num_input_tokens_seen": 48296075, + "step": 1477 + }, + { + "epoch": 0.1332912476890472, + "flos": 23625395210880.0, + "grad_norm": 2.262599951290837, + "learning_rate": 3.8891750307002746e-06, + "loss": 0.7069, + "num_input_tokens_seen": 48325050, + "step": 1478 + }, + { + "epoch": 0.13338143121251747, + "flos": 28727784644640.0, + "grad_norm": 1.6317144470830238, + "learning_rate": 3.888983177304281e-06, + "loss": 0.73, + "num_input_tokens_seen": 48359000, + "step": 1479 + }, + { + "epoch": 0.13347161473598773, + "flos": 56377362732000.0, + "grad_norm": 0.7487607094790092, + "learning_rate": 3.888791162729826e-06, + "loss": 0.5802, + "num_input_tokens_seen": 48441340, + "step": 1480 + }, + { + "epoch": 0.133561798259458, + "flos": 25301085888480.0, + "grad_norm": 2.170132164160588, + "learning_rate": 3.888598986993295e-06, + "loss": 0.8204, + "num_input_tokens_seen": 48469400, + "step": 1481 + }, + { + "epoch": 0.13365198178292825, + "flos": 23152186799040.0, + "grad_norm": 1.848714126434887, + "learning_rate": 3.888406650111085e-06, + "loss": 0.8236, + "num_input_tokens_seen": 48501060, + "step": 1482 + }, + { + "epoch": 0.1337421653063985, + "flos": 26426801132640.0, + "grad_norm": 2.4229834204345444, + "learning_rate": 3.888214152099607e-06, + "loss": 0.8052, + "num_input_tokens_seen": 48529225, + "step": 1483 + }, + { + "epoch": 0.13383234882986877, + "flos": 10939367997120.0, + "grad_norm": 2.1582006946394685, + "learning_rate": 3.888021492975285e-06, + "loss": 0.8716, + "num_input_tokens_seen": 48552275, + "step": 1484 + }, + { + "epoch": 0.13392253235333904, + "flos": 20018951997120.0, + "grad_norm": 1.852183435035597, + "learning_rate": 3.88782867275456e-06, + "loss": 0.851, + "num_input_tokens_seen": 48579105, + "step": 1485 + }, + { + "epoch": 0.1340127158768093, + "flos": 33278674684800.0, + "grad_norm": 1.4571022650463565, + "learning_rate": 3.8876356914538824e-06, + "loss": 0.762, + "num_input_tokens_seen": 48611450, + "step": 1486 + }, + { + "epoch": 0.13410289940027956, + "flos": 55206820757280.0, + "grad_norm": 2.2452187069261544, + "learning_rate": 3.88744254908972e-06, + "loss": 0.753, + "num_input_tokens_seen": 48645345, + "step": 1487 + }, + { + "epoch": 0.13419308292374982, + "flos": 19946013460800.0, + "grad_norm": 2.462165550330665, + "learning_rate": 3.887249245678552e-06, + "loss": 0.7684, + "num_input_tokens_seen": 48673830, + "step": 1488 + }, + { + "epoch": 0.1342832664472201, + "flos": 23152855854720.0, + "grad_norm": 1.7857911297749072, + "learning_rate": 3.887055781236872e-06, + "loss": 0.8722, + "num_input_tokens_seen": 48703925, + "step": 1489 + }, + { + "epoch": 0.13437344997069037, + "flos": 28544025852960.0, + "grad_norm": 1.8196445926073452, + "learning_rate": 3.886862155781186e-06, + "loss": 0.6706, + "num_input_tokens_seen": 48734035, + "step": 1490 + }, + { + "epoch": 0.13446363349416063, + "flos": 68320660638720.0, + "grad_norm": 0.6692996680632344, + "learning_rate": 3.886668369328019e-06, + "loss": 0.6022, + "num_input_tokens_seen": 48823460, + "step": 1491 + }, + { + "epoch": 0.1345538170176309, + "flos": 24209312368800.0, + "grad_norm": 2.3335027860237, + "learning_rate": 3.886474421893904e-06, + "loss": 0.8256, + "num_input_tokens_seen": 48853660, + "step": 1492 + }, + { + "epoch": 0.13464400054110115, + "flos": 27051462081120.0, + "grad_norm": 1.9396770920373607, + "learning_rate": 3.886280313495388e-06, + "loss": 0.7214, + "num_input_tokens_seen": 48885950, + "step": 1493 + }, + { + "epoch": 0.1347341840645714, + "flos": 23589297640320.0, + "grad_norm": 2.0803613317694203, + "learning_rate": 3.886086044149035e-06, + "loss": 0.7667, + "num_input_tokens_seen": 48915010, + "step": 1494 + }, + { + "epoch": 0.13482436758804167, + "flos": 25446033717120.0, + "grad_norm": 1.4415399752977611, + "learning_rate": 3.885891613871421e-06, + "loss": 0.8049, + "num_input_tokens_seen": 48945080, + "step": 1495 + }, + { + "epoch": 0.13491455111151193, + "flos": 27342175473120.0, + "grad_norm": 2.3611088123185677, + "learning_rate": 3.885697022679136e-06, + "loss": 0.8412, + "num_input_tokens_seen": 48974155, + "step": 1496 + }, + { + "epoch": 0.1350047346349822, + "flos": 17432049992160.0, + "grad_norm": 1.8358699055666785, + "learning_rate": 3.885502270588784e-06, + "loss": 0.8255, + "num_input_tokens_seen": 49000390, + "step": 1497 + }, + { + "epoch": 0.13509491815845245, + "flos": 17064160711200.0, + "grad_norm": 2.0290739875114596, + "learning_rate": 3.885307357616981e-06, + "loss": 0.7878, + "num_input_tokens_seen": 49026070, + "step": 1498 + }, + { + "epoch": 0.13518510168192271, + "flos": 27815755582560.0, + "grad_norm": 1.7433088995763317, + "learning_rate": 3.885112283780359e-06, + "loss": 0.7941, + "num_input_tokens_seen": 49054860, + "step": 1499 + }, + { + "epoch": 0.13527528520539298, + "flos": 22022605899840.0, + "grad_norm": 1.6841630083912953, + "learning_rate": 3.8849170490955624e-06, + "loss": 0.6881, + "num_input_tokens_seen": 49083920, + "step": 1500 + }, + { + "epoch": 0.13536546872886324, + "flos": 30439052516160.0, + "grad_norm": 2.530880936455558, + "learning_rate": 3.88472165357925e-06, + "loss": 0.7905, + "num_input_tokens_seen": 49112265, + "step": 1501 + }, + { + "epoch": 0.1354556522523335, + "flos": 21148384217280.0, + "grad_norm": 1.8180580422145052, + "learning_rate": 3.884526097248093e-06, + "loss": 0.7509, + "num_input_tokens_seen": 49141760, + "step": 1502 + }, + { + "epoch": 0.13554583577580376, + "flos": 20018654639040.0, + "grad_norm": 1.9794701460677764, + "learning_rate": 3.884330380118779e-06, + "loss": 0.8166, + "num_input_tokens_seen": 49168645, + "step": 1503 + }, + { + "epoch": 0.13563601929927402, + "flos": 16558162837440.0, + "grad_norm": 2.3360342560699374, + "learning_rate": 3.884134502208007e-06, + "loss": 0.7488, + "num_input_tokens_seen": 49194775, + "step": 1504 + }, + { + "epoch": 0.13572620282274428, + "flos": 27705938910720.0, + "grad_norm": 1.9363761842022345, + "learning_rate": 3.88393846353249e-06, + "loss": 0.9034, + "num_input_tokens_seen": 49220760, + "step": 1505 + }, + { + "epoch": 0.13581638634621454, + "flos": 20125460560320.0, + "grad_norm": 1.6884980252756778, + "learning_rate": 3.883742264108955e-06, + "loss": 0.8487, + "num_input_tokens_seen": 49247950, + "step": 1506 + }, + { + "epoch": 0.1359065698696848, + "flos": 20565916680000.0, + "grad_norm": 2.0158541053857144, + "learning_rate": 3.883545903954145e-06, + "loss": 0.8035, + "num_input_tokens_seen": 49273695, + "step": 1507 + }, + { + "epoch": 0.13599675339315506, + "flos": 21914127339360.0, + "grad_norm": 1.6112539933231371, + "learning_rate": 3.883349383084811e-06, + "loss": 0.7892, + "num_input_tokens_seen": 49301025, + "step": 1508 + }, + { + "epoch": 0.13608693691662532, + "flos": 13274813610240.0, + "grad_norm": 2.0829917882060616, + "learning_rate": 3.883152701517723e-06, + "loss": 0.7169, + "num_input_tokens_seen": 49328090, + "step": 1509 + }, + { + "epoch": 0.13617712044009558, + "flos": 17140890563040.0, + "grad_norm": 1.7993976962853355, + "learning_rate": 3.882955859269664e-06, + "loss": 0.7859, + "num_input_tokens_seen": 49352945, + "step": 1510 + }, + { + "epoch": 0.13626730396356584, + "flos": 20966298064800.0, + "grad_norm": 1.7336516375372804, + "learning_rate": 3.882758856357428e-06, + "loss": 0.802, + "num_input_tokens_seen": 49382910, + "step": 1511 + }, + { + "epoch": 0.1363574874870361, + "flos": 15128873464320.0, + "grad_norm": 2.1286833760200783, + "learning_rate": 3.882561692797824e-06, + "loss": 0.7887, + "num_input_tokens_seen": 49405775, + "step": 1512 + }, + { + "epoch": 0.1364476710105064, + "flos": 24791928585120.0, + "grad_norm": 2.027129115979133, + "learning_rate": 3.882364368607677e-06, + "loss": 0.8021, + "num_input_tokens_seen": 49434540, + "step": 1513 + }, + { + "epoch": 0.13653785453397665, + "flos": 17464170398400.0, + "grad_norm": 1.9941168922857775, + "learning_rate": 3.8821668838038225e-06, + "loss": 0.8098, + "num_input_tokens_seen": 49460090, + "step": 1514 + }, + { + "epoch": 0.13662803805744692, + "flos": 18379358890080.0, + "grad_norm": 2.3055858865399923, + "learning_rate": 3.881969238403111e-06, + "loss": 0.7104, + "num_input_tokens_seen": 49486535, + "step": 1515 + }, + { + "epoch": 0.13671822158091718, + "flos": 13242767543520.0, + "grad_norm": 2.2326567118322074, + "learning_rate": 3.881771432422408e-06, + "loss": 0.9226, + "num_input_tokens_seen": 49512080, + "step": 1516 + }, + { + "epoch": 0.13680840510438744, + "flos": 20019100676160.0, + "grad_norm": 2.186965005030382, + "learning_rate": 3.88157346587859e-06, + "loss": 0.8515, + "num_input_tokens_seen": 49539060, + "step": 1517 + }, + { + "epoch": 0.1368985886278577, + "flos": 20157655306080.0, + "grad_norm": 1.9478839592729142, + "learning_rate": 3.881375338788549e-06, + "loss": 0.7745, + "num_input_tokens_seen": 49567405, + "step": 1518 + }, + { + "epoch": 0.13698877215132796, + "flos": 17359631832480.0, + "grad_norm": 3.4996392200354247, + "learning_rate": 3.88117705116919e-06, + "loss": 0.8071, + "num_input_tokens_seen": 49593625, + "step": 1519 + }, + { + "epoch": 0.13707895567479822, + "flos": 69079899052800.0, + "grad_norm": 0.82220154283769, + "learning_rate": 3.880978603037432e-06, + "loss": 0.6558, + "num_input_tokens_seen": 49675420, + "step": 1520 + }, + { + "epoch": 0.13716913919826848, + "flos": 26649519566400.0, + "grad_norm": 1.9604959788749414, + "learning_rate": 3.880779994410209e-06, + "loss": 0.7977, + "num_input_tokens_seen": 49705765, + "step": 1521 + }, + { + "epoch": 0.13725932272173874, + "flos": 24135816286080.0, + "grad_norm": 2.941194848395881, + "learning_rate": 3.880581225304466e-06, + "loss": 0.6727, + "num_input_tokens_seen": 49733900, + "step": 1522 + }, + { + "epoch": 0.137349506245209, + "flos": 21767321022720.0, + "grad_norm": 2.2226946514010195, + "learning_rate": 3.880382295737163e-06, + "loss": 0.8237, + "num_input_tokens_seen": 49762640, + "step": 1523 + }, + { + "epoch": 0.13743968976867926, + "flos": 16371021597600.0, + "grad_norm": 2.464360451206698, + "learning_rate": 3.880183205725274e-06, + "loss": 0.7679, + "num_input_tokens_seen": 49788450, + "step": 1524 + }, + { + "epoch": 0.13752987329214952, + "flos": 17323199734080.0, + "grad_norm": 2.195054334346009, + "learning_rate": 3.879983955285788e-06, + "loss": 0.7364, + "num_input_tokens_seen": 49815670, + "step": 1525 + }, + { + "epoch": 0.13762005681561978, + "flos": 25629532320480.0, + "grad_norm": 2.017568080467192, + "learning_rate": 3.879784544435703e-06, + "loss": 0.7684, + "num_input_tokens_seen": 49841475, + "step": 1526 + }, + { + "epoch": 0.13771024033909005, + "flos": 20528889865440.0, + "grad_norm": 2.027217677317971, + "learning_rate": 3.879584973192037e-06, + "loss": 0.7351, + "num_input_tokens_seen": 49869790, + "step": 1527 + }, + { + "epoch": 0.1378004238625603, + "flos": 13971669699840.0, + "grad_norm": 3.359755241698218, + "learning_rate": 3.8793852415718165e-06, + "loss": 0.7245, + "num_input_tokens_seen": 49897895, + "step": 1528 + }, + { + "epoch": 0.13789060738603057, + "flos": 19181050903680.0, + "grad_norm": 3.1278145894274108, + "learning_rate": 3.879185349592085e-06, + "loss": 0.8029, + "num_input_tokens_seen": 49924735, + "step": 1529 + }, + { + "epoch": 0.13798079090950083, + "flos": 19180344678240.0, + "grad_norm": 4.4923286747517785, + "learning_rate": 3.878985297269897e-06, + "loss": 0.8184, + "num_input_tokens_seen": 49954380, + "step": 1530 + }, + { + "epoch": 0.1380709744329711, + "flos": 24536532198720.0, + "grad_norm": 1.8425261104602184, + "learning_rate": 3.878785084622323e-06, + "loss": 0.8301, + "num_input_tokens_seen": 49982190, + "step": 1531 + }, + { + "epoch": 0.13816115795644135, + "flos": 14773027185600.0, + "grad_norm": 3.5001141665775917, + "learning_rate": 3.878584711666447e-06, + "loss": 0.9802, + "num_input_tokens_seen": 50007690, + "step": 1532 + }, + { + "epoch": 0.1382513414799116, + "flos": 20820941368800.0, + "grad_norm": 2.0391035404679134, + "learning_rate": 3.8783841784193635e-06, + "loss": 0.8136, + "num_input_tokens_seen": 50034230, + "step": 1533 + }, + { + "epoch": 0.13834152500338187, + "flos": 25630089866880.0, + "grad_norm": 2.4005713895270477, + "learning_rate": 3.8781834848981855e-06, + "loss": 0.7769, + "num_input_tokens_seen": 50065190, + "step": 1534 + }, + { + "epoch": 0.13843170852685213, + "flos": 28611760622880.0, + "grad_norm": 1.9487494322241934, + "learning_rate": 3.877982631120037e-06, + "loss": 0.7285, + "num_input_tokens_seen": 50097590, + "step": 1535 + }, + { + "epoch": 0.1385218920503224, + "flos": 14255543856000.0, + "grad_norm": 2.6381957487892067, + "learning_rate": 3.877781617102053e-06, + "loss": 0.8557, + "num_input_tokens_seen": 50120915, + "step": 1536 + }, + { + "epoch": 0.13861207557379268, + "flos": 26100919414080.0, + "grad_norm": 1.792810902680418, + "learning_rate": 3.877580442861389e-06, + "loss": 0.8318, + "num_input_tokens_seen": 50150900, + "step": 1537 + }, + { + "epoch": 0.13870225909726294, + "flos": 16989103498560.0, + "grad_norm": 1.8477523165804297, + "learning_rate": 3.877379108415209e-06, + "loss": 0.8204, + "num_input_tokens_seen": 50176545, + "step": 1538 + }, + { + "epoch": 0.1387924426207332, + "flos": 21659288499360.0, + "grad_norm": 1.6234134117887413, + "learning_rate": 3.8771776137806915e-06, + "loss": 0.8254, + "num_input_tokens_seen": 50205975, + "step": 1539 + }, + { + "epoch": 0.13888262614420346, + "flos": 24427087224480.0, + "grad_norm": 1.970527391566145, + "learning_rate": 3.8769759589750295e-06, + "loss": 0.7401, + "num_input_tokens_seen": 50235030, + "step": 1540 + }, + { + "epoch": 0.13897280966767372, + "flos": 21440472890400.0, + "grad_norm": 1.6945457306626117, + "learning_rate": 3.876774144015429e-06, + "loss": 0.8039, + "num_input_tokens_seen": 50262780, + "step": 1541 + }, + { + "epoch": 0.13906299319114399, + "flos": 26977594300800.0, + "grad_norm": 11.496350580082197, + "learning_rate": 3.87657216891911e-06, + "loss": 0.8337, + "num_input_tokens_seen": 50289590, + "step": 1542 + }, + { + "epoch": 0.13915317671461425, + "flos": 19215289986240.0, + "grad_norm": 1.9497956242198327, + "learning_rate": 3.876370033703307e-06, + "loss": 0.7961, + "num_input_tokens_seen": 50315610, + "step": 1543 + }, + { + "epoch": 0.1392433602380845, + "flos": 29784241158720.0, + "grad_norm": 1.5388645877518417, + "learning_rate": 3.876167738385265e-06, + "loss": 0.7025, + "num_input_tokens_seen": 50347750, + "step": 1544 + }, + { + "epoch": 0.13933354376155477, + "flos": 20346171827040.0, + "grad_norm": 1.8260508451287334, + "learning_rate": 3.875965282982247e-06, + "loss": 0.8452, + "num_input_tokens_seen": 50373415, + "step": 1545 + }, + { + "epoch": 0.13942372728502503, + "flos": 27997470037440.0, + "grad_norm": 1.950188038391698, + "learning_rate": 3.875762667511528e-06, + "loss": 0.8144, + "num_input_tokens_seen": 50404205, + "step": 1546 + }, + { + "epoch": 0.1395139108084953, + "flos": 21659065480800.0, + "grad_norm": 1.834226354655863, + "learning_rate": 3.875559891990394e-06, + "loss": 0.804, + "num_input_tokens_seen": 50434295, + "step": 1547 + }, + { + "epoch": 0.13960409433196555, + "flos": 32114668854240.0, + "grad_norm": 3.4114179691324296, + "learning_rate": 3.875356956436149e-06, + "loss": 0.7693, + "num_input_tokens_seen": 50464675, + "step": 1548 + }, + { + "epoch": 0.1396942778554358, + "flos": 19909804380960.0, + "grad_norm": 1.8293593062655265, + "learning_rate": 3.875153860866108e-06, + "loss": 0.7838, + "num_input_tokens_seen": 50490755, + "step": 1549 + }, + { + "epoch": 0.13978446137890607, + "flos": 25120003319520.0, + "grad_norm": 2.337931416728509, + "learning_rate": 3.8749506052976e-06, + "loss": 0.8439, + "num_input_tokens_seen": 50520470, + "step": 1550 + }, + { + "epoch": 0.13987464490237633, + "flos": 19145213521440.0, + "grad_norm": 1.5270733606169857, + "learning_rate": 3.874747189747968e-06, + "loss": 0.8245, + "num_input_tokens_seen": 50548165, + "step": 1551 + }, + { + "epoch": 0.1399648284258466, + "flos": 27013989229440.0, + "grad_norm": 1.9827855871131153, + "learning_rate": 3.874543614234568e-06, + "loss": 0.8079, + "num_input_tokens_seen": 50577295, + "step": 1552 + }, + { + "epoch": 0.14005501194931685, + "flos": 29452709636640.0, + "grad_norm": 1.8833873077130185, + "learning_rate": 3.874339878774771e-06, + "loss": 0.7802, + "num_input_tokens_seen": 50606795, + "step": 1553 + }, + { + "epoch": 0.14014519547278712, + "flos": 23622607478880.0, + "grad_norm": 1.923999325643092, + "learning_rate": 3.874135983385961e-06, + "loss": 0.8062, + "num_input_tokens_seen": 50633510, + "step": 1554 + }, + { + "epoch": 0.14023537899625738, + "flos": 15531856732320.0, + "grad_norm": 2.0993673779507387, + "learning_rate": 3.873931928085535e-06, + "loss": 0.7958, + "num_input_tokens_seen": 50659690, + "step": 1555 + }, + { + "epoch": 0.14032556251972764, + "flos": 19977910848480.0, + "grad_norm": 1.6778093350132466, + "learning_rate": 3.873727712890904e-06, + "loss": 0.8847, + "num_input_tokens_seen": 50685570, + "step": 1556 + }, + { + "epoch": 0.1404157460431979, + "flos": 29379473742240.0, + "grad_norm": 1.7142086025901193, + "learning_rate": 3.873523337819493e-06, + "loss": 0.7699, + "num_input_tokens_seen": 50718265, + "step": 1557 + }, + { + "epoch": 0.14050592956666816, + "flos": 23844917045280.0, + "grad_norm": 1.452907587009711, + "learning_rate": 3.873318802888739e-06, + "loss": 0.7307, + "num_input_tokens_seen": 50747365, + "step": 1558 + }, + { + "epoch": 0.14059611309013842, + "flos": 64104647399040.0, + "grad_norm": 0.5644455483531962, + "learning_rate": 3.873114108116097e-06, + "loss": 0.5305, + "num_input_tokens_seen": 50849675, + "step": 1559 + }, + { + "epoch": 0.14068629661360868, + "flos": 21330581879040.0, + "grad_norm": 1.989936604184899, + "learning_rate": 3.872909253519031e-06, + "loss": 0.8901, + "num_input_tokens_seen": 50877875, + "step": 1560 + }, + { + "epoch": 0.14077648013707897, + "flos": 21840556917120.0, + "grad_norm": 2.9434594920141044, + "learning_rate": 3.8727042391150195e-06, + "loss": 0.872, + "num_input_tokens_seen": 50906595, + "step": 1561 + }, + { + "epoch": 0.14086666366054923, + "flos": 20818153636800.0, + "grad_norm": 1.868740696708461, + "learning_rate": 3.872499064921556e-06, + "loss": 0.7389, + "num_input_tokens_seen": 50933155, + "step": 1562 + }, + { + "epoch": 0.1409568471840195, + "flos": 53937607305120.0, + "grad_norm": 1.8824858955979575, + "learning_rate": 3.872293730956149e-06, + "loss": 0.814, + "num_input_tokens_seen": 50964875, + "step": 1563 + }, + { + "epoch": 0.14104703070748975, + "flos": 23188433048640.0, + "grad_norm": 2.1625440207173883, + "learning_rate": 3.872088237236316e-06, + "loss": 0.7937, + "num_input_tokens_seen": 50992010, + "step": 1564 + }, + { + "epoch": 0.14113721423096, + "flos": 18124780238400.0, + "grad_norm": 1.8788268772419152, + "learning_rate": 3.871882583779592e-06, + "loss": 0.7045, + "num_input_tokens_seen": 51019285, + "step": 1565 + }, + { + "epoch": 0.14122739775443027, + "flos": 24683078327040.0, + "grad_norm": 1.6097690136991147, + "learning_rate": 3.871676770603525e-06, + "loss": 0.8035, + "num_input_tokens_seen": 51047655, + "step": 1566 + }, + { + "epoch": 0.14131758127790053, + "flos": 17067654668640.0, + "grad_norm": 2.0013196882781976, + "learning_rate": 3.871470797725676e-06, + "loss": 0.7904, + "num_input_tokens_seen": 51075115, + "step": 1567 + }, + { + "epoch": 0.1414077648013708, + "flos": 21075817378560.0, + "grad_norm": 2.641658512989381, + "learning_rate": 3.8712646651636185e-06, + "loss": 0.8714, + "num_input_tokens_seen": 51103525, + "step": 1568 + }, + { + "epoch": 0.14149794832484106, + "flos": 23845771949760.0, + "grad_norm": 2.1687268830312343, + "learning_rate": 3.871058372934942e-06, + "loss": 0.7988, + "num_input_tokens_seen": 51138215, + "step": 1569 + }, + { + "epoch": 0.14158813184831132, + "flos": 26393156766240.0, + "grad_norm": 1.7971577960846148, + "learning_rate": 3.8708519210572485e-06, + "loss": 0.6286, + "num_input_tokens_seen": 51166555, + "step": 1570 + }, + { + "epoch": 0.14167831537178158, + "flos": 21184741976160.0, + "grad_norm": 5.5051936856462635, + "learning_rate": 3.870645309548153e-06, + "loss": 0.7378, + "num_input_tokens_seen": 51194230, + "step": 1571 + }, + { + "epoch": 0.14176849889525184, + "flos": 59779523712960.0, + "grad_norm": 0.6220008655631001, + "learning_rate": 3.870438538425284e-06, + "loss": 0.5439, + "num_input_tokens_seen": 51288520, + "step": 1572 + }, + { + "epoch": 0.1418586824187221, + "flos": 19654779692160.0, + "grad_norm": 2.530149244045409, + "learning_rate": 3.870231607706287e-06, + "loss": 0.7454, + "num_input_tokens_seen": 51316505, + "step": 1573 + }, + { + "epoch": 0.14194886594219236, + "flos": 21002321295840.0, + "grad_norm": 1.926043138628972, + "learning_rate": 3.870024517408817e-06, + "loss": 0.8029, + "num_input_tokens_seen": 51346240, + "step": 1574 + }, + { + "epoch": 0.14203904946566262, + "flos": 20310334444800.0, + "grad_norm": 2.618850953698585, + "learning_rate": 3.8698172675505425e-06, + "loss": 0.7226, + "num_input_tokens_seen": 51372375, + "step": 1575 + }, + { + "epoch": 0.14212923298913288, + "flos": 61259115238560.0, + "grad_norm": 0.7621038392695422, + "learning_rate": 3.86960985814915e-06, + "loss": 0.5583, + "num_input_tokens_seen": 51466075, + "step": 1576 + }, + { + "epoch": 0.14221941651260314, + "flos": 40854283796640.0, + "grad_norm": 1.6903650898438747, + "learning_rate": 3.869402289222335e-06, + "loss": 0.7289, + "num_input_tokens_seen": 51499360, + "step": 1577 + }, + { + "epoch": 0.1423096000360734, + "flos": 18961900766880.0, + "grad_norm": 1.697558905544361, + "learning_rate": 3.869194560787808e-06, + "loss": 0.7805, + "num_input_tokens_seen": 51528405, + "step": 1578 + }, + { + "epoch": 0.14239978355954366, + "flos": 23513980239360.0, + "grad_norm": 1.9810246095702961, + "learning_rate": 3.868986672863296e-06, + "loss": 0.7263, + "num_input_tokens_seen": 51559115, + "step": 1579 + }, + { + "epoch": 0.14248996708301392, + "flos": 29163631714080.0, + "grad_norm": 1.6242419455423818, + "learning_rate": 3.868778625466535e-06, + "loss": 0.8998, + "num_input_tokens_seen": 51590345, + "step": 1580 + }, + { + "epoch": 0.14258015060648419, + "flos": 37360928193600.0, + "grad_norm": 2.2927212687197804, + "learning_rate": 3.868570418615278e-06, + "loss": 0.715, + "num_input_tokens_seen": 51619485, + "step": 1581 + }, + { + "epoch": 0.14267033412995445, + "flos": 67961952288480.0, + "grad_norm": 0.7299901444932537, + "learning_rate": 3.8683620523272885e-06, + "loss": 0.5666, + "num_input_tokens_seen": 51709500, + "step": 1582 + }, + { + "epoch": 0.1427605176534247, + "flos": 19545186038880.0, + "grad_norm": 1.9483153671249915, + "learning_rate": 3.8681535266203464e-06, + "loss": 0.9021, + "num_input_tokens_seen": 51736835, + "step": 1583 + }, + { + "epoch": 0.14285070117689497, + "flos": 22788051663840.0, + "grad_norm": 1.7841950547966416, + "learning_rate": 3.867944841512246e-06, + "loss": 0.8211, + "num_input_tokens_seen": 51764895, + "step": 1584 + }, + { + "epoch": 0.14294088470036526, + "flos": 22787456947680.0, + "grad_norm": 1.9093337209082233, + "learning_rate": 3.867735997020791e-06, + "loss": 0.8018, + "num_input_tokens_seen": 51789935, + "step": 1585 + }, + { + "epoch": 0.14303106822383552, + "flos": 37541081518560.0, + "grad_norm": 1.8371944588667122, + "learning_rate": 3.867526993163802e-06, + "loss": 0.6991, + "num_input_tokens_seen": 51819810, + "step": 1586 + }, + { + "epoch": 0.14312125174730578, + "flos": 33601173955200.0, + "grad_norm": 1.6310304106608462, + "learning_rate": 3.867317829959113e-06, + "loss": 0.6696, + "num_input_tokens_seen": 51852015, + "step": 1587 + }, + { + "epoch": 0.14321143527077604, + "flos": 24682632289920.0, + "grad_norm": 1.936334382417044, + "learning_rate": 3.8671085074245704e-06, + "loss": 0.594, + "num_input_tokens_seen": 51880075, + "step": 1588 + }, + { + "epoch": 0.1433016187942463, + "flos": 21767729890080.0, + "grad_norm": 2.163948248866685, + "learning_rate": 3.866899025578035e-06, + "loss": 0.8658, + "num_input_tokens_seen": 51907720, + "step": 1589 + }, + { + "epoch": 0.14339180231771656, + "flos": 22788126003360.0, + "grad_norm": 1.537479729247407, + "learning_rate": 3.86668938443738e-06, + "loss": 0.7646, + "num_input_tokens_seen": 51938620, + "step": 1590 + }, + { + "epoch": 0.14348198584118682, + "flos": 25447297488960.0, + "grad_norm": 2.468275773096115, + "learning_rate": 3.866479584020495e-06, + "loss": 0.8105, + "num_input_tokens_seen": 51965805, + "step": 1591 + }, + { + "epoch": 0.14357216936465708, + "flos": 13351580631840.0, + "grad_norm": 1.989012338345915, + "learning_rate": 3.866269624345279e-06, + "loss": 0.831, + "num_input_tokens_seen": 51991185, + "step": 1592 + }, + { + "epoch": 0.14366235288812734, + "flos": 25147626184800.0, + "grad_norm": 2.1712514956594884, + "learning_rate": 3.866059505429649e-06, + "loss": 0.8775, + "num_input_tokens_seen": 52017740, + "step": 1593 + }, + { + "epoch": 0.1437525364115976, + "flos": 66984939018720.0, + "grad_norm": 0.6573796313670877, + "learning_rate": 3.865849227291532e-06, + "loss": 0.5941, + "num_input_tokens_seen": 52105025, + "step": 1594 + }, + { + "epoch": 0.14384271993506786, + "flos": 25003310242080.0, + "grad_norm": 4.324236393369607, + "learning_rate": 3.865638789948872e-06, + "loss": 0.7909, + "num_input_tokens_seen": 52131855, + "step": 1595 + }, + { + "epoch": 0.14393290345853813, + "flos": 24281507509920.0, + "grad_norm": 2.987997841952973, + "learning_rate": 3.865428193419622e-06, + "loss": 0.747, + "num_input_tokens_seen": 52161530, + "step": 1596 + }, + { + "epoch": 0.14402308698200839, + "flos": 23334087102720.0, + "grad_norm": 2.1088738870550525, + "learning_rate": 3.865217437721753e-06, + "loss": 0.7844, + "num_input_tokens_seen": 52191135, + "step": 1597 + }, + { + "epoch": 0.14411327050547865, + "flos": 32619514465440.0, + "grad_norm": 13.852729588981, + "learning_rate": 3.865006522873249e-06, + "loss": 0.728, + "num_input_tokens_seen": 52223170, + "step": 1598 + }, + { + "epoch": 0.1442034540289489, + "flos": 20310408784320.0, + "grad_norm": 1.8355733307374698, + "learning_rate": 3.864795448892103e-06, + "loss": 0.7938, + "num_input_tokens_seen": 52251555, + "step": 1599 + }, + { + "epoch": 0.14429363755241917, + "flos": 23984289409920.0, + "grad_norm": 1.9326832486775032, + "learning_rate": 3.864584215796327e-06, + "loss": 0.8499, + "num_input_tokens_seen": 52281630, + "step": 1600 + }, + { + "epoch": 0.14438382107588943, + "flos": 15529886735040.0, + "grad_norm": 2.3351361485391715, + "learning_rate": 3.8643728236039455e-06, + "loss": 0.8178, + "num_input_tokens_seen": 52306450, + "step": 1601 + }, + { + "epoch": 0.1444740045993597, + "flos": 25374396122400.0, + "grad_norm": 1.726440287273568, + "learning_rate": 3.864161272332994e-06, + "loss": 0.7796, + "num_input_tokens_seen": 52338315, + "step": 1602 + }, + { + "epoch": 0.14456418812282995, + "flos": 25519641309120.0, + "grad_norm": 1.6971008431498902, + "learning_rate": 3.863949562001524e-06, + "loss": 0.8414, + "num_input_tokens_seen": 52369660, + "step": 1603 + }, + { + "epoch": 0.1446543716463002, + "flos": 26794578904320.0, + "grad_norm": 3.341530749695365, + "learning_rate": 3.8637376926276005e-06, + "loss": 0.7695, + "num_input_tokens_seen": 52398445, + "step": 1604 + }, + { + "epoch": 0.14474455516977047, + "flos": 19035248170560.0, + "grad_norm": 2.896441407248308, + "learning_rate": 3.8635256642293e-06, + "loss": 0.802, + "num_input_tokens_seen": 52424810, + "step": 1605 + }, + { + "epoch": 0.14483473869324073, + "flos": 18962718501600.0, + "grad_norm": 1.810597440636934, + "learning_rate": 3.863313476824714e-06, + "loss": 0.7786, + "num_input_tokens_seen": 52453320, + "step": 1606 + }, + { + "epoch": 0.144924922216711, + "flos": 14845519684800.0, + "grad_norm": 3.4685682644361653, + "learning_rate": 3.863101130431948e-06, + "loss": 0.7747, + "num_input_tokens_seen": 52479095, + "step": 1607 + }, + { + "epoch": 0.14501510574018128, + "flos": 24566868456480.0, + "grad_norm": 2.2374856277089474, + "learning_rate": 3.862888625069121e-06, + "loss": 0.84, + "num_input_tokens_seen": 52508685, + "step": 1608 + }, + { + "epoch": 0.14510528926365154, + "flos": 62864028959520.0, + "grad_norm": 1.7167268810390404, + "learning_rate": 3.8626759607543645e-06, + "loss": 0.7594, + "num_input_tokens_seen": 52544770, + "step": 1609 + }, + { + "epoch": 0.1451954727871218, + "flos": 18192626517600.0, + "grad_norm": 2.466126702035389, + "learning_rate": 3.862463137505825e-06, + "loss": 0.7515, + "num_input_tokens_seen": 52572100, + "step": 1610 + }, + { + "epoch": 0.14528565631059207, + "flos": 37543386043680.0, + "grad_norm": 1.7938003772214182, + "learning_rate": 3.862250155341659e-06, + "loss": 0.7599, + "num_input_tokens_seen": 52602850, + "step": 1611 + }, + { + "epoch": 0.14537583983406233, + "flos": 19727941247040.0, + "grad_norm": 1.6502245576813284, + "learning_rate": 3.862037014280043e-06, + "loss": 0.9248, + "num_input_tokens_seen": 52628930, + "step": 1612 + }, + { + "epoch": 0.1454660233575326, + "flos": 20381340153600.0, + "grad_norm": 1.9277942342846683, + "learning_rate": 3.861823714339162e-06, + "loss": 0.7927, + "num_input_tokens_seen": 52658555, + "step": 1613 + }, + { + "epoch": 0.14555620688100285, + "flos": 28579751725920.0, + "grad_norm": 3.7964097982115987, + "learning_rate": 3.861610255537215e-06, + "loss": 0.8017, + "num_input_tokens_seen": 52691495, + "step": 1614 + }, + { + "epoch": 0.1456463904044731, + "flos": 30256631835840.0, + "grad_norm": 1.6041756501710036, + "learning_rate": 3.8613966378924165e-06, + "loss": 0.8237, + "num_input_tokens_seen": 52720405, + "step": 1615 + }, + { + "epoch": 0.14573657392794337, + "flos": 22818202072800.0, + "grad_norm": 3.814177222364734, + "learning_rate": 3.861182861422993e-06, + "loss": 0.7422, + "num_input_tokens_seen": 52746565, + "step": 1616 + }, + { + "epoch": 0.14582675745141363, + "flos": 20419928098080.0, + "grad_norm": 1.6859290263305067, + "learning_rate": 3.860968926147185e-06, + "loss": 0.8664, + "num_input_tokens_seen": 52774735, + "step": 1617 + }, + { + "epoch": 0.1459169409748839, + "flos": 35065780333920.0, + "grad_norm": 2.0574984160888805, + "learning_rate": 3.860754832083247e-06, + "loss": 0.6292, + "num_input_tokens_seen": 52807480, + "step": 1618 + }, + { + "epoch": 0.14600712449835415, + "flos": 21440584399680.0, + "grad_norm": 2.3462348481442525, + "learning_rate": 3.8605405792494475e-06, + "loss": 0.8628, + "num_input_tokens_seen": 52836705, + "step": 1619 + }, + { + "epoch": 0.1460973080218244, + "flos": 23912688984960.0, + "grad_norm": 1.9971652206589265, + "learning_rate": 3.860326167664066e-06, + "loss": 0.81, + "num_input_tokens_seen": 52865255, + "step": 1620 + }, + { + "epoch": 0.14618749154529467, + "flos": 17140035658560.0, + "grad_norm": 2.409454589795808, + "learning_rate": 3.860111597345399e-06, + "loss": 0.7771, + "num_input_tokens_seen": 52891985, + "step": 1621 + }, + { + "epoch": 0.14627767506876493, + "flos": 24026408481600.0, + "grad_norm": 21.88648624218371, + "learning_rate": 3.859896868311753e-06, + "loss": 0.6489, + "num_input_tokens_seen": 52919770, + "step": 1622 + }, + { + "epoch": 0.1463678585922352, + "flos": 23330184277920.0, + "grad_norm": 1.655478459717241, + "learning_rate": 3.859681980581452e-06, + "loss": 0.8857, + "num_input_tokens_seen": 52947440, + "step": 1623 + }, + { + "epoch": 0.14645804211570546, + "flos": 21367051147200.0, + "grad_norm": 3.033238338897331, + "learning_rate": 3.859466934172829e-06, + "loss": 0.7713, + "num_input_tokens_seen": 52973855, + "step": 1624 + }, + { + "epoch": 0.14654822563917572, + "flos": 26578662536640.0, + "grad_norm": 1.5692414384079338, + "learning_rate": 3.859251729104235e-06, + "loss": 0.7253, + "num_input_tokens_seen": 53006805, + "step": 1625 + }, + { + "epoch": 0.14663840916264598, + "flos": 24204814827840.0, + "grad_norm": 1.7547040114810575, + "learning_rate": 3.859036365394031e-06, + "loss": 0.7415, + "num_input_tokens_seen": 53035070, + "step": 1626 + }, + { + "epoch": 0.14672859268611624, + "flos": 22273133047680.0, + "grad_norm": 2.636152463784463, + "learning_rate": 3.858820843060594e-06, + "loss": 0.7768, + "num_input_tokens_seen": 53062400, + "step": 1627 + }, + { + "epoch": 0.1468187762095865, + "flos": 20127876594720.0, + "grad_norm": 1.9977114219640324, + "learning_rate": 3.858605162122314e-06, + "loss": 0.7681, + "num_input_tokens_seen": 53091875, + "step": 1628 + }, + { + "epoch": 0.14690895973305676, + "flos": 16922706840000.0, + "grad_norm": 1.751899982101429, + "learning_rate": 3.858389322597592e-06, + "loss": 0.748, + "num_input_tokens_seen": 53118745, + "step": 1629 + }, + { + "epoch": 0.14699914325652702, + "flos": 20674915617120.0, + "grad_norm": 1.827896234267556, + "learning_rate": 3.858173324504847e-06, + "loss": 0.8067, + "num_input_tokens_seen": 53147440, + "step": 1630 + }, + { + "epoch": 0.14708932677999728, + "flos": 22569050206080.0, + "grad_norm": 1.5467075295777843, + "learning_rate": 3.857957167862508e-06, + "loss": 0.7492, + "num_input_tokens_seen": 53174255, + "step": 1631 + }, + { + "epoch": 0.14717951030346757, + "flos": 66288082929120.0, + "grad_norm": 0.6729480690808594, + "learning_rate": 3.857740852689018e-06, + "loss": 0.5788, + "num_input_tokens_seen": 53264685, + "step": 1632 + }, + { + "epoch": 0.14726969382693783, + "flos": 20490747958080.0, + "grad_norm": 2.30874132436662, + "learning_rate": 3.857524379002835e-06, + "loss": 0.7978, + "num_input_tokens_seen": 53289870, + "step": 1633 + }, + { + "epoch": 0.1473598773504081, + "flos": 24857581848480.0, + "grad_norm": 1.6477856544986023, + "learning_rate": 3.85730774682243e-06, + "loss": 0.8094, + "num_input_tokens_seen": 53318045, + "step": 1634 + }, + { + "epoch": 0.14745006087387835, + "flos": 24245484278880.0, + "grad_norm": 1.9861917664602018, + "learning_rate": 3.8570909561662875e-06, + "loss": 0.853, + "num_input_tokens_seen": 53348835, + "step": 1635 + }, + { + "epoch": 0.1475402443973486, + "flos": 23839899127680.0, + "grad_norm": 1.590938706555843, + "learning_rate": 3.8568740070529045e-06, + "loss": 0.8576, + "num_input_tokens_seen": 53378130, + "step": 1636 + }, + { + "epoch": 0.14763042792081887, + "flos": 16412954820480.0, + "grad_norm": 1.867444902498381, + "learning_rate": 3.856656899500792e-06, + "loss": 0.7554, + "num_input_tokens_seen": 53404455, + "step": 1637 + }, + { + "epoch": 0.14772061144428913, + "flos": 22022197032480.0, + "grad_norm": 1.774859089698194, + "learning_rate": 3.856439633528476e-06, + "loss": 0.7233, + "num_input_tokens_seen": 53429210, + "step": 1638 + }, + { + "epoch": 0.1478107949677594, + "flos": 25702396517280.0, + "grad_norm": 1.8631202100342221, + "learning_rate": 3.856222209154494e-06, + "loss": 0.8417, + "num_input_tokens_seen": 53458845, + "step": 1639 + }, + { + "epoch": 0.14790097849122966, + "flos": 20781275501280.0, + "grad_norm": 1.6582330312379716, + "learning_rate": 3.856004626397397e-06, + "loss": 0.8135, + "num_input_tokens_seen": 53487325, + "step": 1640 + }, + { + "epoch": 0.14799116201469992, + "flos": 25045838181120.0, + "grad_norm": 3.2347047794305435, + "learning_rate": 3.855786885275753e-06, + "loss": 0.7642, + "num_input_tokens_seen": 53515975, + "step": 1641 + }, + { + "epoch": 0.14808134553817018, + "flos": 26322485585280.0, + "grad_norm": 2.3051157408015785, + "learning_rate": 3.855568985808138e-06, + "loss": 0.7041, + "num_input_tokens_seen": 53544725, + "step": 1642 + }, + { + "epoch": 0.14817152906164044, + "flos": 33205215771840.0, + "grad_norm": 1.7027595197864878, + "learning_rate": 3.855350928013145e-06, + "loss": 0.7585, + "num_input_tokens_seen": 53576340, + "step": 1643 + }, + { + "epoch": 0.1482617125851107, + "flos": 20201967393600.0, + "grad_norm": 2.0764522847591147, + "learning_rate": 3.8551327119093825e-06, + "loss": 0.7948, + "num_input_tokens_seen": 53602735, + "step": 1644 + }, + { + "epoch": 0.14835189610858096, + "flos": 22387410090720.0, + "grad_norm": 2.285288123865792, + "learning_rate": 3.854914337515467e-06, + "loss": 0.7157, + "num_input_tokens_seen": 53631795, + "step": 1645 + }, + { + "epoch": 0.14844207963205122, + "flos": 27706050420000.0, + "grad_norm": 1.9230313427731973, + "learning_rate": 3.8546958048500324e-06, + "loss": 0.6844, + "num_input_tokens_seen": 53660425, + "step": 1646 + }, + { + "epoch": 0.14853226315552148, + "flos": 68402780105760.0, + "grad_norm": 1.2924983618189099, + "learning_rate": 3.854477113931725e-06, + "loss": 0.631, + "num_input_tokens_seen": 53750730, + "step": 1647 + }, + { + "epoch": 0.14862244667899174, + "flos": 20784583609920.0, + "grad_norm": 1.8127857509165821, + "learning_rate": 3.854258264779205e-06, + "loss": 0.8109, + "num_input_tokens_seen": 53779495, + "step": 1648 + }, + { + "epoch": 0.148712630202462, + "flos": 21108012124320.0, + "grad_norm": 1.7538896657588703, + "learning_rate": 3.854039257411145e-06, + "loss": 0.7679, + "num_input_tokens_seen": 53808350, + "step": 1649 + }, + { + "epoch": 0.14880281372593226, + "flos": 25848645287520.0, + "grad_norm": 2.068080634767865, + "learning_rate": 3.853820091846232e-06, + "loss": 0.7385, + "num_input_tokens_seen": 53838135, + "step": 1650 + }, + { + "epoch": 0.14889299724940253, + "flos": 18265304865600.0, + "grad_norm": 1.909373616469361, + "learning_rate": 3.853600768103169e-06, + "loss": 0.7885, + "num_input_tokens_seen": 53863855, + "step": 1651 + }, + { + "epoch": 0.1489831807728728, + "flos": 29815506660480.0, + "grad_norm": 2.9713104449885486, + "learning_rate": 3.853381286200667e-06, + "loss": 0.7757, + "num_input_tokens_seen": 53886890, + "step": 1652 + }, + { + "epoch": 0.14907336429634305, + "flos": 23407062808800.0, + "grad_norm": 2.080691824630859, + "learning_rate": 3.853161646157453e-06, + "loss": 0.8777, + "num_input_tokens_seen": 53914125, + "step": 1653 + }, + { + "epoch": 0.1491635478198133, + "flos": 21074925304320.0, + "grad_norm": 1.9994566298500924, + "learning_rate": 3.852941847992269e-06, + "loss": 0.7504, + "num_input_tokens_seen": 53942375, + "step": 1654 + }, + { + "epoch": 0.14925373134328357, + "flos": 19035545528640.0, + "grad_norm": 2.34601615703772, + "learning_rate": 3.852721891723871e-06, + "loss": 0.8034, + "num_input_tokens_seen": 53968880, + "step": 1655 + }, + { + "epoch": 0.14934391486675386, + "flos": 17905778780640.0, + "grad_norm": 2.691189739227692, + "learning_rate": 3.852501777371025e-06, + "loss": 0.7733, + "num_input_tokens_seen": 53995265, + "step": 1656 + }, + { + "epoch": 0.14943409839022412, + "flos": 24167341976160.0, + "grad_norm": 4.635673990143775, + "learning_rate": 3.8522815049525125e-06, + "loss": 0.8259, + "num_input_tokens_seen": 54025315, + "step": 1657 + }, + { + "epoch": 0.14952428191369438, + "flos": 22860692842080.0, + "grad_norm": 1.9023946328378591, + "learning_rate": 3.852061074487129e-06, + "loss": 0.8434, + "num_input_tokens_seen": 54052600, + "step": 1658 + }, + { + "epoch": 0.14961446543716464, + "flos": 33863818444800.0, + "grad_norm": 1.8918332739350467, + "learning_rate": 3.851840485993682e-06, + "loss": 0.6955, + "num_input_tokens_seen": 54085930, + "step": 1659 + }, + { + "epoch": 0.1497046489606349, + "flos": 19182574863840.0, + "grad_norm": 1.7512595548512053, + "learning_rate": 3.851619739490994e-06, + "loss": 0.7989, + "num_input_tokens_seen": 54114805, + "step": 1660 + }, + { + "epoch": 0.14979483248410516, + "flos": 21216230496480.0, + "grad_norm": 1.7343082586970209, + "learning_rate": 3.8513988349978996e-06, + "loss": 0.8245, + "num_input_tokens_seen": 54140960, + "step": 1661 + }, + { + "epoch": 0.14988501600757542, + "flos": 21293926762080.0, + "grad_norm": 1.671498764343444, + "learning_rate": 3.851177772533249e-06, + "loss": 0.8151, + "num_input_tokens_seen": 54170145, + "step": 1662 + }, + { + "epoch": 0.14997519953104568, + "flos": 19944377991360.0, + "grad_norm": 2.706755652992812, + "learning_rate": 3.850956552115903e-06, + "loss": 0.9631, + "num_input_tokens_seen": 54193415, + "step": 1663 + }, + { + "epoch": 0.15006538305451594, + "flos": 66503032883040.0, + "grad_norm": 0.7019559678671304, + "learning_rate": 3.850735173764738e-06, + "loss": 0.5824, + "num_input_tokens_seen": 54283935, + "step": 1664 + }, + { + "epoch": 0.1501555665779862, + "flos": 23297655004320.0, + "grad_norm": 1.7341830860529543, + "learning_rate": 3.850513637498642e-06, + "loss": 0.8071, + "num_input_tokens_seen": 54314130, + "step": 1665 + }, + { + "epoch": 0.15024575010145647, + "flos": 25192979025600.0, + "grad_norm": 2.219135089821032, + "learning_rate": 3.850291943336521e-06, + "loss": 0.8067, + "num_input_tokens_seen": 54343600, + "step": 1666 + }, + { + "epoch": 0.15033593362492673, + "flos": 23989679025120.0, + "grad_norm": 2.7988240361100627, + "learning_rate": 3.850070091297287e-06, + "loss": 0.8432, + "num_input_tokens_seen": 54371740, + "step": 1667 + }, + { + "epoch": 0.150426117148397, + "flos": 66642330908160.0, + "grad_norm": 0.7118551598815976, + "learning_rate": 3.8498480813998735e-06, + "loss": 0.6735, + "num_input_tokens_seen": 54458715, + "step": 1668 + }, + { + "epoch": 0.15051630067186725, + "flos": 24896281302240.0, + "grad_norm": 1.8529151450304537, + "learning_rate": 3.84962591366322e-06, + "loss": 0.7886, + "num_input_tokens_seen": 54488225, + "step": 1669 + }, + { + "epoch": 0.1506064841953375, + "flos": 24317530740960.0, + "grad_norm": 2.037219270428307, + "learning_rate": 3.8494035881062855e-06, + "loss": 0.8405, + "num_input_tokens_seen": 54516800, + "step": 1670 + }, + { + "epoch": 0.15069666771880777, + "flos": 14989798457760.0, + "grad_norm": 2.06318526010414, + "learning_rate": 3.84918110474804e-06, + "loss": 0.7728, + "num_input_tokens_seen": 54542755, + "step": 1671 + }, + { + "epoch": 0.15078685124227803, + "flos": 30397937028000.0, + "grad_norm": 1.5909417037593623, + "learning_rate": 3.8489584636074655e-06, + "loss": 0.7767, + "num_input_tokens_seen": 54575190, + "step": 1672 + }, + { + "epoch": 0.1508770347657483, + "flos": 69190013082720.0, + "grad_norm": 0.6860684749888342, + "learning_rate": 3.848735664703561e-06, + "loss": 0.5648, + "num_input_tokens_seen": 54670680, + "step": 1673 + }, + { + "epoch": 0.15096721828921855, + "flos": 25884854367360.0, + "grad_norm": 1.7963256659661182, + "learning_rate": 3.8485127080553346e-06, + "loss": 0.7864, + "num_input_tokens_seen": 54699840, + "step": 1674 + }, + { + "epoch": 0.1510574018126888, + "flos": 22350717804000.0, + "grad_norm": 1.8624218892353064, + "learning_rate": 3.8482895936818115e-06, + "loss": 0.7879, + "num_input_tokens_seen": 54730210, + "step": 1675 + }, + { + "epoch": 0.15114758533615907, + "flos": 21731929677600.0, + "grad_norm": 1.8752541687673296, + "learning_rate": 3.848066321602029e-06, + "loss": 0.808, + "num_input_tokens_seen": 54755285, + "step": 1676 + }, + { + "epoch": 0.15123776885962933, + "flos": 21038902073280.0, + "grad_norm": 2.914658554520738, + "learning_rate": 3.847842891835038e-06, + "loss": 0.7521, + "num_input_tokens_seen": 54782910, + "step": 1677 + }, + { + "epoch": 0.1513279523830996, + "flos": 64509451985280.0, + "grad_norm": 0.6263607708775077, + "learning_rate": 3.847619304399902e-06, + "loss": 0.5803, + "num_input_tokens_seen": 54884585, + "step": 1678 + }, + { + "epoch": 0.15141813590656986, + "flos": 34183976020320.0, + "grad_norm": 2.4584204986981355, + "learning_rate": 3.8473955593157e-06, + "loss": 0.8168, + "num_input_tokens_seen": 54913825, + "step": 1679 + }, + { + "epoch": 0.15150831943004014, + "flos": 33462693664800.0, + "grad_norm": 1.777920271043839, + "learning_rate": 3.847171656601522e-06, + "loss": 0.7535, + "num_input_tokens_seen": 54945305, + "step": 1680 + }, + { + "epoch": 0.1515985029535104, + "flos": 22315437968160.0, + "grad_norm": 2.067343717339664, + "learning_rate": 3.846947596276473e-06, + "loss": 0.8188, + "num_input_tokens_seen": 54972050, + "step": 1681 + }, + { + "epoch": 0.15168868647698067, + "flos": 32333075595840.0, + "grad_norm": 2.429116234056812, + "learning_rate": 3.846723378359672e-06, + "loss": 0.8404, + "num_input_tokens_seen": 55006755, + "step": 1682 + }, + { + "epoch": 0.15177887000045093, + "flos": 20339444100480.0, + "grad_norm": 2.746655250760153, + "learning_rate": 3.846499002870249e-06, + "loss": 0.7963, + "num_input_tokens_seen": 55030520, + "step": 1683 + }, + { + "epoch": 0.1518690535239212, + "flos": 20198845133760.0, + "grad_norm": 2.04391428435897, + "learning_rate": 3.846274469827349e-06, + "loss": 0.8114, + "num_input_tokens_seen": 55059115, + "step": 1684 + }, + { + "epoch": 0.15195923704739145, + "flos": 27050607176640.0, + "grad_norm": 1.86879906478313, + "learning_rate": 3.846049779250132e-06, + "loss": 0.8019, + "num_input_tokens_seen": 55089370, + "step": 1685 + }, + { + "epoch": 0.1520494205708617, + "flos": 65118167106720.0, + "grad_norm": 0.7861753085005617, + "learning_rate": 3.845824931157769e-06, + "loss": 0.6235, + "num_input_tokens_seen": 55185430, + "step": 1686 + }, + { + "epoch": 0.15213960409433197, + "flos": 15136307416320.0, + "grad_norm": 3.0315573624977405, + "learning_rate": 3.845599925569444e-06, + "loss": 0.7778, + "num_input_tokens_seen": 55211520, + "step": 1687 + }, + { + "epoch": 0.15222978761780223, + "flos": 26394829405440.0, + "grad_norm": 2.06671140324974, + "learning_rate": 3.845374762504357e-06, + "loss": 0.7532, + "num_input_tokens_seen": 55241565, + "step": 1688 + }, + { + "epoch": 0.1523199711412725, + "flos": 26978932412160.0, + "grad_norm": 4.261886131852763, + "learning_rate": 3.8451494419817204e-06, + "loss": 0.8409, + "num_input_tokens_seen": 55271170, + "step": 1689 + }, + { + "epoch": 0.15241015466474275, + "flos": 30110940612000.0, + "grad_norm": 2.5973589504742396, + "learning_rate": 3.8449239640207594e-06, + "loss": 0.7199, + "num_input_tokens_seen": 55302305, + "step": 1690 + }, + { + "epoch": 0.152500338188213, + "flos": 21804310667520.0, + "grad_norm": 2.3058168909138965, + "learning_rate": 3.844698328640713e-06, + "loss": 0.8362, + "num_input_tokens_seen": 55330125, + "step": 1691 + }, + { + "epoch": 0.15259052171168327, + "flos": 24316192629600.0, + "grad_norm": 1.4974296322753644, + "learning_rate": 3.844472535860833e-06, + "loss": 0.8076, + "num_input_tokens_seen": 55359825, + "step": 1692 + }, + { + "epoch": 0.15268070523515354, + "flos": 20018989166880.0, + "grad_norm": 2.9099451999813555, + "learning_rate": 3.8442465857003864e-06, + "loss": 0.8418, + "num_input_tokens_seen": 55385745, + "step": 1693 + }, + { + "epoch": 0.1527708887586238, + "flos": 22057365359040.0, + "grad_norm": 1.5327670309339785, + "learning_rate": 3.844020478178653e-06, + "loss": 0.7362, + "num_input_tokens_seen": 55414740, + "step": 1694 + }, + { + "epoch": 0.15286107228209406, + "flos": 23333975593440.0, + "grad_norm": 1.7239420756546324, + "learning_rate": 3.843794213314923e-06, + "loss": 0.8161, + "num_input_tokens_seen": 55442695, + "step": 1695 + }, + { + "epoch": 0.15295125580556432, + "flos": 24792411792000.0, + "grad_norm": 1.7445547673732533, + "learning_rate": 3.843567791128505e-06, + "loss": 0.8213, + "num_input_tokens_seen": 55471135, + "step": 1696 + }, + { + "epoch": 0.15304143932903458, + "flos": 30142391962560.0, + "grad_norm": 1.7687368062204605, + "learning_rate": 3.843341211638717e-06, + "loss": 0.7802, + "num_input_tokens_seen": 55500775, + "step": 1697 + }, + { + "epoch": 0.15313162285250484, + "flos": 23223229677600.0, + "grad_norm": 2.19272012473324, + "learning_rate": 3.843114474864894e-06, + "loss": 0.752, + "num_input_tokens_seen": 55530590, + "step": 1698 + }, + { + "epoch": 0.1532218063759751, + "flos": 25847864722560.0, + "grad_norm": 1.6444968901825472, + "learning_rate": 3.84288758082638e-06, + "loss": 0.7913, + "num_input_tokens_seen": 55561065, + "step": 1699 + }, + { + "epoch": 0.15331198989944536, + "flos": 22422875775360.0, + "grad_norm": 1.822587977249508, + "learning_rate": 3.842660529542536e-06, + "loss": 0.8759, + "num_input_tokens_seen": 55590950, + "step": 1700 + }, + { + "epoch": 0.15340217342291562, + "flos": 27046964540160.0, + "grad_norm": 1.6259504619714453, + "learning_rate": 3.842433321032736e-06, + "loss": 0.8047, + "num_input_tokens_seen": 55619075, + "step": 1701 + }, + { + "epoch": 0.15349235694638588, + "flos": 31638635540640.0, + "grad_norm": 1.7983175706688173, + "learning_rate": 3.842205955316365e-06, + "loss": 0.7161, + "num_input_tokens_seen": 55649765, + "step": 1702 + }, + { + "epoch": 0.15358254046985614, + "flos": 29273113858080.0, + "grad_norm": 2.0669328851383297, + "learning_rate": 3.8419784324128256e-06, + "loss": 0.6991, + "num_input_tokens_seen": 55680895, + "step": 1703 + }, + { + "epoch": 0.15367272399332643, + "flos": 21401624757600.0, + "grad_norm": 1.836955336208957, + "learning_rate": 3.841750752341529e-06, + "loss": 0.7549, + "num_input_tokens_seen": 55702530, + "step": 1704 + }, + { + "epoch": 0.1537629075167967, + "flos": 20966669762400.0, + "grad_norm": 1.80912578493817, + "learning_rate": 3.841522915121902e-06, + "loss": 0.7968, + "num_input_tokens_seen": 55730990, + "step": 1705 + }, + { + "epoch": 0.15385309104026695, + "flos": 23443978114080.0, + "grad_norm": 2.08183450824958, + "learning_rate": 3.841294920773387e-06, + "loss": 0.7202, + "num_input_tokens_seen": 55759710, + "step": 1706 + }, + { + "epoch": 0.15394327456373721, + "flos": 17141373769920.0, + "grad_norm": 2.2058322586654544, + "learning_rate": 3.841066769315436e-06, + "loss": 0.8049, + "num_input_tokens_seen": 55787030, + "step": 1707 + }, + { + "epoch": 0.15403345808720748, + "flos": 19109376139200.0, + "grad_norm": 1.6190655257199158, + "learning_rate": 3.840838460767518e-06, + "loss": 0.8018, + "num_input_tokens_seen": 55812810, + "step": 1708 + }, + { + "epoch": 0.15412364161067774, + "flos": 23949827308800.0, + "grad_norm": 2.095264171978898, + "learning_rate": 3.840609995149111e-06, + "loss": 0.7626, + "num_input_tokens_seen": 55839260, + "step": 1709 + }, + { + "epoch": 0.154213825134148, + "flos": 30545672588640.0, + "grad_norm": 1.7164406954851523, + "learning_rate": 3.84038137247971e-06, + "loss": 0.7947, + "num_input_tokens_seen": 55869000, + "step": 1710 + }, + { + "epoch": 0.15430400865761826, + "flos": 25149596182080.0, + "grad_norm": 1.6880965414688498, + "learning_rate": 3.840152592778823e-06, + "loss": 0.7778, + "num_input_tokens_seen": 55897900, + "step": 1711 + }, + { + "epoch": 0.15439419218108852, + "flos": 27414519293280.0, + "grad_norm": 1.3460613598859237, + "learning_rate": 3.83992365606597e-06, + "loss": 0.8376, + "num_input_tokens_seen": 55929790, + "step": 1712 + }, + { + "epoch": 0.15448437570455878, + "flos": 26246461958880.0, + "grad_norm": 2.0304572098841813, + "learning_rate": 3.8396945623606855e-06, + "loss": 0.744, + "num_input_tokens_seen": 55960250, + "step": 1713 + }, + { + "epoch": 0.15457455922802904, + "flos": 23079731469600.0, + "grad_norm": 1.5875336831191462, + "learning_rate": 3.8394653116825174e-06, + "loss": 0.7463, + "num_input_tokens_seen": 55989420, + "step": 1714 + }, + { + "epoch": 0.1546647427514993, + "flos": 24317605080480.0, + "grad_norm": 2.4739238749998824, + "learning_rate": 3.839235904051026e-06, + "loss": 0.7754, + "num_input_tokens_seen": 56018435, + "step": 1715 + }, + { + "epoch": 0.15475492627496956, + "flos": 21694754184000.0, + "grad_norm": 5.833739166674743, + "learning_rate": 3.8390063394857855e-06, + "loss": 0.7691, + "num_input_tokens_seen": 56047860, + "step": 1716 + }, + { + "epoch": 0.15484510979843982, + "flos": 26172296820480.0, + "grad_norm": 1.4896572690456624, + "learning_rate": 3.838776618006385e-06, + "loss": 0.8728, + "num_input_tokens_seen": 56078085, + "step": 1717 + }, + { + "epoch": 0.15493529332191008, + "flos": 20889902740800.0, + "grad_norm": 1.8821455537978864, + "learning_rate": 3.838546739632423e-06, + "loss": 0.7961, + "num_input_tokens_seen": 56107225, + "step": 1718 + }, + { + "epoch": 0.15502547684538034, + "flos": 22711582000320.0, + "grad_norm": 1.5560155209304125, + "learning_rate": 3.838316704383517e-06, + "loss": 0.8395, + "num_input_tokens_seen": 56134090, + "step": 1719 + }, + { + "epoch": 0.1551156603688506, + "flos": 21439952513760.0, + "grad_norm": 2.0589220628480738, + "learning_rate": 3.838086512279292e-06, + "loss": 0.7958, + "num_input_tokens_seen": 56163200, + "step": 1720 + }, + { + "epoch": 0.15520584389232087, + "flos": 28173943556160.0, + "grad_norm": 1.9738754281891528, + "learning_rate": 3.837856163339391e-06, + "loss": 0.7933, + "num_input_tokens_seen": 56191785, + "step": 1721 + }, + { + "epoch": 0.15529602741579113, + "flos": 60094929292800.0, + "grad_norm": 1.493343843922638, + "learning_rate": 3.837625657583469e-06, + "loss": 0.7189, + "num_input_tokens_seen": 56227075, + "step": 1722 + }, + { + "epoch": 0.1553862109392614, + "flos": 34409036148960.0, + "grad_norm": 1.8451733671575563, + "learning_rate": 3.837394995031193e-06, + "loss": 0.7504, + "num_input_tokens_seen": 56260525, + "step": 1723 + }, + { + "epoch": 0.15547639446273165, + "flos": 31604582306880.0, + "grad_norm": 1.6694341650061175, + "learning_rate": 3.837164175702245e-06, + "loss": 0.7906, + "num_input_tokens_seen": 56291015, + "step": 1724 + }, + { + "epoch": 0.1555665779862019, + "flos": 17651274468480.0, + "grad_norm": 2.080092452452124, + "learning_rate": 3.836933199616319e-06, + "loss": 0.71, + "num_input_tokens_seen": 56317555, + "step": 1725 + }, + { + "epoch": 0.15565676150967217, + "flos": 19144618805280.0, + "grad_norm": 1.7297620858890672, + "learning_rate": 3.836702066793124e-06, + "loss": 0.8607, + "num_input_tokens_seen": 56344365, + "step": 1726 + }, + { + "epoch": 0.15574694503314246, + "flos": 21032806232640.0, + "grad_norm": 1.8857031827554356, + "learning_rate": 3.836470777252381e-06, + "loss": 0.8527, + "num_input_tokens_seen": 56374115, + "step": 1727 + }, + { + "epoch": 0.15583712855661272, + "flos": 25483915436160.0, + "grad_norm": 2.576565542502657, + "learning_rate": 3.836239331013825e-06, + "loss": 0.8052, + "num_input_tokens_seen": 56404570, + "step": 1728 + }, + { + "epoch": 0.15592731208008298, + "flos": 24245186920800.0, + "grad_norm": 2.206597371622023, + "learning_rate": 3.836007728097205e-06, + "loss": 0.7539, + "num_input_tokens_seen": 56435110, + "step": 1729 + }, + { + "epoch": 0.15601749560355324, + "flos": 24973531530720.0, + "grad_norm": 1.624839187881253, + "learning_rate": 3.835775968522282e-06, + "loss": 0.7451, + "num_input_tokens_seen": 56465365, + "step": 1730 + }, + { + "epoch": 0.1561076791270235, + "flos": 20633093903520.0, + "grad_norm": 3.7993865321028744, + "learning_rate": 3.83554405230883e-06, + "loss": 0.8099, + "num_input_tokens_seen": 56492400, + "step": 1731 + }, + { + "epoch": 0.15619786265049376, + "flos": 31093343496960.0, + "grad_norm": 1.5720514945412118, + "learning_rate": 3.835311979476639e-06, + "loss": 0.7746, + "num_input_tokens_seen": 56526885, + "step": 1732 + }, + { + "epoch": 0.15628804617396402, + "flos": 22752251451360.0, + "grad_norm": 1.5851470029513275, + "learning_rate": 3.83507975004551e-06, + "loss": 0.6787, + "num_input_tokens_seen": 56556980, + "step": 1733 + }, + { + "epoch": 0.15637822969743428, + "flos": 16340313642240.0, + "grad_norm": 1.7236764529487758, + "learning_rate": 3.834847364035258e-06, + "loss": 0.7687, + "num_input_tokens_seen": 56584505, + "step": 1734 + }, + { + "epoch": 0.15646841322090455, + "flos": 24572964297120.0, + "grad_norm": 2.670956370058413, + "learning_rate": 3.834614821465712e-06, + "loss": 0.7342, + "num_input_tokens_seen": 56612825, + "step": 1735 + }, + { + "epoch": 0.1565585967443748, + "flos": 23116237907520.0, + "grad_norm": 4.434992336881569, + "learning_rate": 3.834382122356713e-06, + "loss": 0.8285, + "num_input_tokens_seen": 56639550, + "step": 1736 + }, + { + "epoch": 0.15664878026784507, + "flos": 25514363203200.0, + "grad_norm": 1.7110730121893112, + "learning_rate": 3.834149266728117e-06, + "loss": 0.7482, + "num_input_tokens_seen": 56667720, + "step": 1737 + }, + { + "epoch": 0.15673896379131533, + "flos": 54336384656640.0, + "grad_norm": 0.6804695234502345, + "learning_rate": 3.833916254599792e-06, + "loss": 0.6195, + "num_input_tokens_seen": 56752940, + "step": 1738 + }, + { + "epoch": 0.1568291473147856, + "flos": 67274388638880.0, + "grad_norm": 0.6344275410509088, + "learning_rate": 3.83368308599162e-06, + "loss": 0.5656, + "num_input_tokens_seen": 56844040, + "step": 1739 + }, + { + "epoch": 0.15691933083825585, + "flos": 71285822287680.0, + "grad_norm": 0.6188170974648506, + "learning_rate": 3.833449760923498e-06, + "loss": 0.5374, + "num_input_tokens_seen": 56944100, + "step": 1740 + }, + { + "epoch": 0.1570095143617261, + "flos": 19252353970560.0, + "grad_norm": 1.7024706362126008, + "learning_rate": 3.83321627941533e-06, + "loss": 0.8283, + "num_input_tokens_seen": 56970600, + "step": 1741 + }, + { + "epoch": 0.15709969788519637, + "flos": 17139924149280.0, + "grad_norm": 1.9380754416144166, + "learning_rate": 3.832982641487043e-06, + "loss": 0.8453, + "num_input_tokens_seen": 56996985, + "step": 1742 + }, + { + "epoch": 0.15718988140866663, + "flos": 19654593843360.0, + "grad_norm": 1.9631643227832207, + "learning_rate": 3.832748847158568e-06, + "loss": 0.8787, + "num_input_tokens_seen": 57025385, + "step": 1743 + }, + { + "epoch": 0.1572800649321369, + "flos": 31460266364160.0, + "grad_norm": 1.602372053278606, + "learning_rate": 3.832514896449858e-06, + "loss": 0.8124, + "num_input_tokens_seen": 57058730, + "step": 1744 + }, + { + "epoch": 0.15737024845560715, + "flos": 23841943464480.0, + "grad_norm": 1.5324945647001724, + "learning_rate": 3.832280789380871e-06, + "loss": 0.7255, + "num_input_tokens_seen": 57087090, + "step": 1745 + }, + { + "epoch": 0.15746043197907741, + "flos": 21730963263840.0, + "grad_norm": 1.6246004346971636, + "learning_rate": 3.832046525971584e-06, + "loss": 0.69, + "num_input_tokens_seen": 57115765, + "step": 1746 + }, + { + "epoch": 0.15755061550254768, + "flos": 24166375562400.0, + "grad_norm": 1.8873571101790418, + "learning_rate": 3.831812106241987e-06, + "loss": 0.8474, + "num_input_tokens_seen": 57143965, + "step": 1747 + }, + { + "epoch": 0.15764079902601794, + "flos": 18306829221120.0, + "grad_norm": 1.7668421103375191, + "learning_rate": 3.8315775302120796e-06, + "loss": 0.6943, + "num_input_tokens_seen": 57171795, + "step": 1748 + }, + { + "epoch": 0.1577309825494882, + "flos": 25082716316640.0, + "grad_norm": 1.6253973401734707, + "learning_rate": 3.831342797901878e-06, + "loss": 0.8019, + "num_input_tokens_seen": 57204295, + "step": 1749 + }, + { + "epoch": 0.15782116607295846, + "flos": 68529769206720.0, + "grad_norm": 0.6711548560925058, + "learning_rate": 3.831107909331411e-06, + "loss": 0.5785, + "num_input_tokens_seen": 57307375, + "step": 1750 + }, + { + "epoch": 0.15791134959642875, + "flos": 22674257827680.0, + "grad_norm": 1.8641333720826827, + "learning_rate": 3.830872864520721e-06, + "loss": 0.829, + "num_input_tokens_seen": 57333160, + "step": 1751 + }, + { + "epoch": 0.158001533119899, + "flos": 64132753471200.0, + "grad_norm": 0.7166601565161441, + "learning_rate": 3.830637663489862e-06, + "loss": 0.6582, + "num_input_tokens_seen": 57422165, + "step": 1752 + }, + { + "epoch": 0.15809171664336927, + "flos": 65856659061120.0, + "grad_norm": 0.7789512766657617, + "learning_rate": 3.830402306258904e-06, + "loss": 0.6886, + "num_input_tokens_seen": 57511710, + "step": 1753 + }, + { + "epoch": 0.15818190016683953, + "flos": 18598471857120.0, + "grad_norm": 3.097366992967339, + "learning_rate": 3.830166792847929e-06, + "loss": 0.7741, + "num_input_tokens_seen": 57539580, + "step": 1754 + }, + { + "epoch": 0.1582720836903098, + "flos": 32986474502400.0, + "grad_norm": 1.5305719236655244, + "learning_rate": 3.829931123277031e-06, + "loss": 0.7251, + "num_input_tokens_seen": 57572850, + "step": 1755 + }, + { + "epoch": 0.15836226721378005, + "flos": 37360853854080.0, + "grad_norm": 2.2941413233759365, + "learning_rate": 3.8296952975663204e-06, + "loss": 0.7746, + "num_input_tokens_seen": 57603575, + "step": 1756 + }, + { + "epoch": 0.1584524507372503, + "flos": 25520198855520.0, + "grad_norm": 1.554119830751072, + "learning_rate": 3.829459315735918e-06, + "loss": 0.8543, + "num_input_tokens_seen": 57632335, + "step": 1757 + }, + { + "epoch": 0.15854263426072057, + "flos": 22569087375840.0, + "grad_norm": 2.038159634583347, + "learning_rate": 3.829223177805959e-06, + "loss": 0.7105, + "num_input_tokens_seen": 57661335, + "step": 1758 + }, + { + "epoch": 0.15863281778419083, + "flos": 26825955915360.0, + "grad_norm": 1.740945073667454, + "learning_rate": 3.828986883796591e-06, + "loss": 0.8272, + "num_input_tokens_seen": 57687270, + "step": 1759 + }, + { + "epoch": 0.1587230013076611, + "flos": 32515719294720.0, + "grad_norm": 1.7331565494740344, + "learning_rate": 3.828750433727979e-06, + "loss": 0.7783, + "num_input_tokens_seen": 57715520, + "step": 1760 + }, + { + "epoch": 0.15881318483113135, + "flos": 23698965633120.0, + "grad_norm": 2.0123076826812127, + "learning_rate": 3.828513827620296e-06, + "loss": 0.8342, + "num_input_tokens_seen": 57745540, + "step": 1761 + }, + { + "epoch": 0.15890336835460162, + "flos": 24457869519360.0, + "grad_norm": 2.111563444477506, + "learning_rate": 3.82827706549373e-06, + "loss": 0.8347, + "num_input_tokens_seen": 57769845, + "step": 1762 + }, + { + "epoch": 0.15899355187807188, + "flos": 19399420475520.0, + "grad_norm": 2.0917861610515773, + "learning_rate": 3.828040147368484e-06, + "loss": 0.7768, + "num_input_tokens_seen": 57797950, + "step": 1763 + }, + { + "epoch": 0.15908373540154214, + "flos": 24755607996000.0, + "grad_norm": 1.9055409598030961, + "learning_rate": 3.827803073264774e-06, + "loss": 0.7836, + "num_input_tokens_seen": 57826685, + "step": 1764 + }, + { + "epoch": 0.1591739189250124, + "flos": 15465571583040.0, + "grad_norm": 2.2905342656545113, + "learning_rate": 3.827565843202826e-06, + "loss": 0.7847, + "num_input_tokens_seen": 57853435, + "step": 1765 + }, + { + "epoch": 0.15926410244848266, + "flos": 37069657255200.0, + "grad_norm": 1.976436264168463, + "learning_rate": 3.827328457202884e-06, + "loss": 0.6577, + "num_input_tokens_seen": 57881570, + "step": 1766 + }, + { + "epoch": 0.15935428597195292, + "flos": 34517291690880.0, + "grad_norm": 1.4550401925896725, + "learning_rate": 3.8270909152852014e-06, + "loss": 0.7166, + "num_input_tokens_seen": 57913730, + "step": 1767 + }, + { + "epoch": 0.15944446949542318, + "flos": 26759113219680.0, + "grad_norm": 1.9649306291366462, + "learning_rate": 3.826853217470048e-06, + "loss": 0.7813, + "num_input_tokens_seen": 57939690, + "step": 1768 + }, + { + "epoch": 0.15953465301889344, + "flos": 24973754549280.0, + "grad_norm": 1.7952083653725992, + "learning_rate": 3.826615363777705e-06, + "loss": 0.865, + "num_input_tokens_seen": 57969415, + "step": 1769 + }, + { + "epoch": 0.1596248365423637, + "flos": 22676525183040.0, + "grad_norm": 1.89089478388539, + "learning_rate": 3.826377354228468e-06, + "loss": 0.7646, + "num_input_tokens_seen": 57998770, + "step": 1770 + }, + { + "epoch": 0.15971502006583396, + "flos": 63914532578400.0, + "grad_norm": 0.6536775761124655, + "learning_rate": 3.826139188842643e-06, + "loss": 0.5912, + "num_input_tokens_seen": 58089690, + "step": 1771 + }, + { + "epoch": 0.15980520358930422, + "flos": 28613210243520.0, + "grad_norm": 1.5869301381397296, + "learning_rate": 3.825900867640554e-06, + "loss": 0.8397, + "num_input_tokens_seen": 58118960, + "step": 1772 + }, + { + "epoch": 0.15989538711277448, + "flos": 23766774742560.0, + "grad_norm": 1.461378224851649, + "learning_rate": 3.825662390642535e-06, + "loss": 0.8247, + "num_input_tokens_seen": 58150000, + "step": 1773 + }, + { + "epoch": 0.15998557063624474, + "flos": 20638743707040.0, + "grad_norm": 1.9200387261379352, + "learning_rate": 3.825423757868934e-06, + "loss": 0.8883, + "num_input_tokens_seen": 58177120, + "step": 1774 + }, + { + "epoch": 0.16007575415971503, + "flos": 19614556278240.0, + "grad_norm": 1.6950460479467908, + "learning_rate": 3.825184969340114e-06, + "loss": 0.8107, + "num_input_tokens_seen": 58205390, + "step": 1775 + }, + { + "epoch": 0.1601659376831853, + "flos": 28762841461920.0, + "grad_norm": 1.6508389738635787, + "learning_rate": 3.824946025076447e-06, + "loss": 0.8138, + "num_input_tokens_seen": 58238950, + "step": 1776 + }, + { + "epoch": 0.16025612120665556, + "flos": 23552977051200.0, + "grad_norm": 1.914482849052241, + "learning_rate": 3.824706925098323e-06, + "loss": 0.734, + "num_input_tokens_seen": 58267035, + "step": 1777 + }, + { + "epoch": 0.16034630473012582, + "flos": 13896649656960.0, + "grad_norm": 2.373704657420375, + "learning_rate": 3.824467669426143e-06, + "loss": 0.8448, + "num_input_tokens_seen": 58290725, + "step": 1778 + }, + { + "epoch": 0.16043648825359608, + "flos": 22606374378720.0, + "grad_norm": 1.8260407263025653, + "learning_rate": 3.824228258080321e-06, + "loss": 0.7526, + "num_input_tokens_seen": 58320300, + "step": 1779 + }, + { + "epoch": 0.16052667177706634, + "flos": 25265768882880.0, + "grad_norm": 1.761523589089693, + "learning_rate": 3.823988691081285e-06, + "loss": 0.6994, + "num_input_tokens_seen": 58348095, + "step": 1780 + }, + { + "epoch": 0.1606168553005366, + "flos": 23366133169440.0, + "grad_norm": 1.663788822216955, + "learning_rate": 3.823748968449478e-06, + "loss": 0.8172, + "num_input_tokens_seen": 58376335, + "step": 1781 + }, + { + "epoch": 0.16070703882400686, + "flos": 16047964780800.0, + "grad_norm": 1.935930970360287, + "learning_rate": 3.823509090205352e-06, + "loss": 0.6782, + "num_input_tokens_seen": 58403230, + "step": 1782 + }, + { + "epoch": 0.16079722234747712, + "flos": 27450691203360.0, + "grad_norm": 2.201563396683802, + "learning_rate": 3.823269056369376e-06, + "loss": 0.7407, + "num_input_tokens_seen": 58433080, + "step": 1783 + }, + { + "epoch": 0.16088740587094738, + "flos": 21330842067360.0, + "grad_norm": 1.8652285590755204, + "learning_rate": 3.8230288669620295e-06, + "loss": 0.8196, + "num_input_tokens_seen": 58461295, + "step": 1784 + }, + { + "epoch": 0.16097758939441764, + "flos": 20602385948160.0, + "grad_norm": 1.4895846048985457, + "learning_rate": 3.822788522003809e-06, + "loss": 0.7899, + "num_input_tokens_seen": 58492595, + "step": 1785 + }, + { + "epoch": 0.1610677729178879, + "flos": 14007432742560.0, + "grad_norm": 1.9671510530932887, + "learning_rate": 3.822548021515221e-06, + "loss": 0.8596, + "num_input_tokens_seen": 58515410, + "step": 1786 + }, + { + "epoch": 0.16115795644135816, + "flos": 22161606566880.0, + "grad_norm": 1.661723037090071, + "learning_rate": 3.822307365516787e-06, + "loss": 0.7608, + "num_input_tokens_seen": 58542965, + "step": 1787 + }, + { + "epoch": 0.16124813996482842, + "flos": 24607017530880.0, + "grad_norm": 1.905137910231365, + "learning_rate": 3.8220665540290395e-06, + "loss": 0.8035, + "num_input_tokens_seen": 58571885, + "step": 1788 + }, + { + "epoch": 0.16133832348829868, + "flos": 27087559651680.0, + "grad_norm": 1.4636196871377722, + "learning_rate": 3.8218255870725265e-06, + "loss": 0.8505, + "num_input_tokens_seen": 58601375, + "step": 1789 + }, + { + "epoch": 0.16142850701176895, + "flos": 28718975411520.0, + "grad_norm": 1.681437233067874, + "learning_rate": 3.82158446466781e-06, + "loss": 0.7353, + "num_input_tokens_seen": 58630505, + "step": 1790 + }, + { + "epoch": 0.1615186905352392, + "flos": 27231801254880.0, + "grad_norm": 1.5864328447057279, + "learning_rate": 3.821343186835462e-06, + "loss": 0.8001, + "num_input_tokens_seen": 58660550, + "step": 1791 + }, + { + "epoch": 0.16160887405870947, + "flos": 18999856825440.0, + "grad_norm": 1.5816955210174508, + "learning_rate": 3.821101753596072e-06, + "loss": 0.8531, + "num_input_tokens_seen": 58688055, + "step": 1792 + }, + { + "epoch": 0.16169905758217973, + "flos": 13971037813920.0, + "grad_norm": 2.127361556142199, + "learning_rate": 3.820860164970237e-06, + "loss": 0.7725, + "num_input_tokens_seen": 58712410, + "step": 1793 + }, + { + "epoch": 0.16178924110565, + "flos": 27268382032320.0, + "grad_norm": 1.528173899111984, + "learning_rate": 3.820618420978574e-06, + "loss": 0.8538, + "num_input_tokens_seen": 58742390, + "step": 1794 + }, + { + "epoch": 0.16187942462912025, + "flos": 20383570339200.0, + "grad_norm": 2.058694855091005, + "learning_rate": 3.820376521641708e-06, + "loss": 0.7832, + "num_input_tokens_seen": 58770865, + "step": 1795 + }, + { + "epoch": 0.1619696081525905, + "flos": 30110829102720.0, + "grad_norm": 4.480677126734093, + "learning_rate": 3.82013446698028e-06, + "loss": 0.7917, + "num_input_tokens_seen": 58800985, + "step": 1796 + }, + { + "epoch": 0.16205979167606077, + "flos": 33065508879360.0, + "grad_norm": 3.2355043954903393, + "learning_rate": 3.819892257014943e-06, + "loss": 0.7529, + "num_input_tokens_seen": 58834935, + "step": 1797 + }, + { + "epoch": 0.16214997519953103, + "flos": 25593248901120.0, + "grad_norm": 1.7215673071350466, + "learning_rate": 3.819649891766364e-06, + "loss": 0.8111, + "num_input_tokens_seen": 58864825, + "step": 1798 + }, + { + "epoch": 0.16224015872300132, + "flos": 21622521873120.0, + "grad_norm": 1.528090506634363, + "learning_rate": 3.819407371255222e-06, + "loss": 0.7798, + "num_input_tokens_seen": 58893880, + "step": 1799 + }, + { + "epoch": 0.16233034224647158, + "flos": 27852187680960.0, + "grad_norm": 1.7339393824725342, + "learning_rate": 3.819164695502212e-06, + "loss": 0.7979, + "num_input_tokens_seen": 58923280, + "step": 1800 + }, + { + "epoch": 0.16242052576994184, + "flos": 23005677840480.0, + "grad_norm": 1.6209344624773478, + "learning_rate": 3.818921864528039e-06, + "loss": 0.7286, + "num_input_tokens_seen": 58949505, + "step": 1801 + }, + { + "epoch": 0.1625107092934121, + "flos": 21439841004480.0, + "grad_norm": 2.544725224588786, + "learning_rate": 3.818678878353423e-06, + "loss": 0.8383, + "num_input_tokens_seen": 58977185, + "step": 1802 + }, + { + "epoch": 0.16260089281688236, + "flos": 20638223330400.0, + "grad_norm": 2.199257885870976, + "learning_rate": 3.818435736999097e-06, + "loss": 0.7297, + "num_input_tokens_seen": 59004615, + "step": 1803 + }, + { + "epoch": 0.16269107634035262, + "flos": 24536532198720.0, + "grad_norm": 1.913994425906014, + "learning_rate": 3.818192440485807e-06, + "loss": 0.7811, + "num_input_tokens_seen": 59033345, + "step": 1804 + }, + { + "epoch": 0.16278125986382289, + "flos": 19210457917440.0, + "grad_norm": 2.4502367644834586, + "learning_rate": 3.817948988834314e-06, + "loss": 0.7185, + "num_input_tokens_seen": 59060955, + "step": 1805 + }, + { + "epoch": 0.16287144338729315, + "flos": 21727097608800.0, + "grad_norm": 1.5498516842423353, + "learning_rate": 3.817705382065388e-06, + "loss": 0.782, + "num_input_tokens_seen": 59090975, + "step": 1806 + }, + { + "epoch": 0.1629616269107634, + "flos": 28653210638880.0, + "grad_norm": 1.2469030640008507, + "learning_rate": 3.8174616201998155e-06, + "loss": 0.7503, + "num_input_tokens_seen": 59123635, + "step": 1807 + }, + { + "epoch": 0.16305181043423367, + "flos": 17541011759520.0, + "grad_norm": 3.2547897842897613, + "learning_rate": 3.817217703258397e-06, + "loss": 0.8345, + "num_input_tokens_seen": 59148125, + "step": 1808 + }, + { + "epoch": 0.16314199395770393, + "flos": 27809622572160.0, + "grad_norm": 1.5811772054569857, + "learning_rate": 3.816973631261943e-06, + "loss": 0.7722, + "num_input_tokens_seen": 59175160, + "step": 1809 + }, + { + "epoch": 0.1632321774811742, + "flos": 19982928766080.0, + "grad_norm": 2.7323742331331804, + "learning_rate": 3.816729404231281e-06, + "loss": 0.8159, + "num_input_tokens_seen": 59203015, + "step": 1810 + }, + { + "epoch": 0.16332236100464445, + "flos": 23583350478720.0, + "grad_norm": 2.624478333726617, + "learning_rate": 3.816485022187249e-06, + "loss": 0.8527, + "num_input_tokens_seen": 59225565, + "step": 1811 + }, + { + "epoch": 0.1634125445281147, + "flos": 31234871707680.0, + "grad_norm": 2.881099864167698, + "learning_rate": 3.816240485150698e-06, + "loss": 0.7405, + "num_input_tokens_seen": 59254680, + "step": 1812 + }, + { + "epoch": 0.16350272805158497, + "flos": 22351052331840.0, + "grad_norm": 1.6347363832836843, + "learning_rate": 3.815995793142495e-06, + "loss": 0.7852, + "num_input_tokens_seen": 59284710, + "step": 1813 + }, + { + "epoch": 0.16359291157505523, + "flos": 25374544801440.0, + "grad_norm": 1.8287771643018056, + "learning_rate": 3.815750946183518e-06, + "loss": 0.7239, + "num_input_tokens_seen": 59313870, + "step": 1814 + }, + { + "epoch": 0.1636830950985255, + "flos": 15100618713120.0, + "grad_norm": 1.8654897465072624, + "learning_rate": 3.815505944294658e-06, + "loss": 0.8407, + "num_input_tokens_seen": 59340225, + "step": 1815 + }, + { + "epoch": 0.16377327862199575, + "flos": 27598017896640.0, + "grad_norm": 1.7792602937289912, + "learning_rate": 3.81526078749682e-06, + "loss": 0.7406, + "num_input_tokens_seen": 59372750, + "step": 1816 + }, + { + "epoch": 0.16386346214546602, + "flos": 20784100403040.0, + "grad_norm": 1.9189710309583807, + "learning_rate": 3.8150154758109225e-06, + "loss": 0.7232, + "num_input_tokens_seen": 59402290, + "step": 1817 + }, + { + "epoch": 0.16395364566893628, + "flos": 32332964086560.0, + "grad_norm": 1.59341023133207, + "learning_rate": 3.814770009257896e-06, + "loss": 0.7589, + "num_input_tokens_seen": 59433280, + "step": 1818 + }, + { + "epoch": 0.16404382919240654, + "flos": 28323426095520.0, + "grad_norm": 1.8282091822356161, + "learning_rate": 3.814524387858687e-06, + "loss": 0.7976, + "num_input_tokens_seen": 59463530, + "step": 1819 + }, + { + "epoch": 0.1641340127158768, + "flos": 19035694207680.0, + "grad_norm": 1.7860943022049587, + "learning_rate": 3.814278611634251e-06, + "loss": 0.8092, + "num_input_tokens_seen": 59492110, + "step": 1820 + }, + { + "epoch": 0.16422419623934706, + "flos": 23438068122240.0, + "grad_norm": 1.7809791358997937, + "learning_rate": 3.8140326806055606e-06, + "loss": 0.7585, + "num_input_tokens_seen": 59520610, + "step": 1821 + }, + { + "epoch": 0.16431437976281732, + "flos": 19581729646560.0, + "grad_norm": 3.3943796771485597, + "learning_rate": 3.8137865947935992e-06, + "loss": 0.8442, + "num_input_tokens_seen": 59548775, + "step": 1822 + }, + { + "epoch": 0.1644045632862876, + "flos": 24281470340160.0, + "grad_norm": 2.5817813723525207, + "learning_rate": 3.8135403542193646e-06, + "loss": 0.8014, + "num_input_tokens_seen": 59577515, + "step": 1823 + }, + { + "epoch": 0.16449474680975787, + "flos": 16959807994080.0, + "grad_norm": 2.0894997572103424, + "learning_rate": 3.813293958903867e-06, + "loss": 0.7201, + "num_input_tokens_seen": 59603105, + "step": 1824 + }, + { + "epoch": 0.16458493033322813, + "flos": 44137112647200.0, + "grad_norm": 1.9092371314887449, + "learning_rate": 3.8130474088681306e-06, + "loss": 0.7498, + "num_input_tokens_seen": 59634675, + "step": 1825 + }, + { + "epoch": 0.1646751138566984, + "flos": 25629272132160.0, + "grad_norm": 1.8162741336151134, + "learning_rate": 3.8128007041331927e-06, + "loss": 0.7146, + "num_input_tokens_seen": 59662955, + "step": 1826 + }, + { + "epoch": 0.16476529738016865, + "flos": 20638148990880.0, + "grad_norm": 1.581718432735331, + "learning_rate": 3.812553844720102e-06, + "loss": 0.7007, + "num_input_tokens_seen": 59693975, + "step": 1827 + }, + { + "epoch": 0.1648554809036389, + "flos": 35095596215040.0, + "grad_norm": 1.8179985040673206, + "learning_rate": 3.8123068306499236e-06, + "loss": 0.8012, + "num_input_tokens_seen": 59721815, + "step": 1828 + }, + { + "epoch": 0.16494566442710917, + "flos": 31197324516480.0, + "grad_norm": 4.879623958920273, + "learning_rate": 3.812059661943733e-06, + "loss": 0.781, + "num_input_tokens_seen": 59750325, + "step": 1829 + }, + { + "epoch": 0.16503584795057943, + "flos": 23948563536960.0, + "grad_norm": 1.4331232696935337, + "learning_rate": 3.811812338622621e-06, + "loss": 0.754, + "num_input_tokens_seen": 59777725, + "step": 1830 + }, + { + "epoch": 0.1651260314740497, + "flos": 20784360591360.0, + "grad_norm": 1.6851337231512915, + "learning_rate": 3.81156486070769e-06, + "loss": 0.7763, + "num_input_tokens_seen": 59806440, + "step": 1831 + }, + { + "epoch": 0.16521621499751996, + "flos": 21002953181760.0, + "grad_norm": 1.8034820494026826, + "learning_rate": 3.811317228220056e-06, + "loss": 0.8281, + "num_input_tokens_seen": 59835600, + "step": 1832 + }, + { + "epoch": 0.16530639852099022, + "flos": 23990013552960.0, + "grad_norm": 7.407194053261719, + "learning_rate": 3.811069441180849e-06, + "loss": 0.8291, + "num_input_tokens_seen": 59865540, + "step": 1833 + }, + { + "epoch": 0.16539658204446048, + "flos": 18307238088480.0, + "grad_norm": 2.179682030893116, + "learning_rate": 3.8108214996112107e-06, + "loss": 0.7873, + "num_input_tokens_seen": 59891930, + "step": 1834 + }, + { + "epoch": 0.16548676556793074, + "flos": 21002172616800.0, + "grad_norm": 1.9842740130822483, + "learning_rate": 3.810573403532297e-06, + "loss": 0.8224, + "num_input_tokens_seen": 59916140, + "step": 1835 + }, + { + "epoch": 0.165576949091401, + "flos": 22457932592640.0, + "grad_norm": 1.9799717386714135, + "learning_rate": 3.8103251529652774e-06, + "loss": 0.7651, + "num_input_tokens_seen": 59944890, + "step": 1836 + }, + { + "epoch": 0.16566713261487126, + "flos": 23371039577760.0, + "grad_norm": 1.6782987309323163, + "learning_rate": 3.810076747931334e-06, + "loss": 0.7613, + "num_input_tokens_seen": 59973880, + "step": 1837 + }, + { + "epoch": 0.16575731613834152, + "flos": 23625878417760.0, + "grad_norm": 2.7642461928101363, + "learning_rate": 3.809828188451662e-06, + "loss": 0.7403, + "num_input_tokens_seen": 60003025, + "step": 1838 + }, + { + "epoch": 0.16584749966181178, + "flos": 28435138425120.0, + "grad_norm": 1.6728161453443944, + "learning_rate": 3.809579474547469e-06, + "loss": 0.7793, + "num_input_tokens_seen": 60033335, + "step": 1839 + }, + { + "epoch": 0.16593768318528204, + "flos": 18010354516320.0, + "grad_norm": 1.8384356445635133, + "learning_rate": 3.809330606239977e-06, + "loss": 0.8375, + "num_input_tokens_seen": 60060295, + "step": 1840 + }, + { + "epoch": 0.1660278667087523, + "flos": 25442948627040.0, + "grad_norm": 1.693492633367689, + "learning_rate": 3.809081583550422e-06, + "loss": 0.8492, + "num_input_tokens_seen": 60091155, + "step": 1841 + }, + { + "epoch": 0.16611805023222256, + "flos": 19181199582720.0, + "grad_norm": 2.209063055524326, + "learning_rate": 3.808832406500051e-06, + "loss": 0.863, + "num_input_tokens_seen": 60119515, + "step": 1842 + }, + { + "epoch": 0.16620823375569282, + "flos": 17939200128480.0, + "grad_norm": 1.662480493302024, + "learning_rate": 3.8085830751101253e-06, + "loss": 0.8554, + "num_input_tokens_seen": 60143810, + "step": 1843 + }, + { + "epoch": 0.16629841727916309, + "flos": 18086787010080.0, + "grad_norm": 1.750588386258991, + "learning_rate": 3.808333589401919e-06, + "loss": 0.8195, + "num_input_tokens_seen": 60172245, + "step": 1844 + }, + { + "epoch": 0.16638860080263335, + "flos": 18197644435200.0, + "grad_norm": 1.7911951272549782, + "learning_rate": 3.8080839493967194e-06, + "loss": 0.714, + "num_input_tokens_seen": 60199395, + "step": 1845 + }, + { + "epoch": 0.1664787843261036, + "flos": 67325688641280.0, + "grad_norm": 0.6620265267451007, + "learning_rate": 3.807834155115828e-06, + "loss": 0.5447, + "num_input_tokens_seen": 60298835, + "step": 1846 + }, + { + "epoch": 0.1665689678495739, + "flos": 70305092041920.0, + "grad_norm": 0.8505212559945216, + "learning_rate": 3.8075842065805584e-06, + "loss": 0.6496, + "num_input_tokens_seen": 60404460, + "step": 1847 + }, + { + "epoch": 0.16665915137304416, + "flos": 23151852271200.0, + "grad_norm": 1.8952895701657446, + "learning_rate": 3.8073341038122374e-06, + "loss": 0.7701, + "num_input_tokens_seen": 60434200, + "step": 1848 + }, + { + "epoch": 0.16674933489651442, + "flos": 25662024424320.0, + "grad_norm": 1.697481657443154, + "learning_rate": 3.8070838468322048e-06, + "loss": 0.8504, + "num_input_tokens_seen": 60463630, + "step": 1849 + }, + { + "epoch": 0.16683951841998468, + "flos": 20565210454560.0, + "grad_norm": 1.8741804239060174, + "learning_rate": 3.8068334356618143e-06, + "loss": 0.7206, + "num_input_tokens_seen": 60490845, + "step": 1850 + }, + { + "epoch": 0.16692970194345494, + "flos": 23516396273760.0, + "grad_norm": 2.798524593870195, + "learning_rate": 3.8065828703224324e-06, + "loss": 0.8155, + "num_input_tokens_seen": 60519075, + "step": 1851 + }, + { + "epoch": 0.1670198854669252, + "flos": 18962644162080.0, + "grad_norm": 1.9839246343941734, + "learning_rate": 3.8063321508354386e-06, + "loss": 0.7818, + "num_input_tokens_seen": 60542685, + "step": 1852 + }, + { + "epoch": 0.16711006899039546, + "flos": 19361204228640.0, + "grad_norm": 1.6698595537301812, + "learning_rate": 3.8060812772222255e-06, + "loss": 0.7638, + "num_input_tokens_seen": 60570970, + "step": 1853 + }, + { + "epoch": 0.16720025251386572, + "flos": 21732004017120.0, + "grad_norm": 1.5614341152044384, + "learning_rate": 3.8058302495041993e-06, + "loss": 0.8514, + "num_input_tokens_seen": 60598765, + "step": 1854 + }, + { + "epoch": 0.16729043603733598, + "flos": 18051990381120.0, + "grad_norm": 3.422954789590004, + "learning_rate": 3.805579067702779e-06, + "loss": 0.8013, + "num_input_tokens_seen": 60627300, + "step": 1855 + }, + { + "epoch": 0.16738061956080624, + "flos": 21658247746080.0, + "grad_norm": 1.7981240616664353, + "learning_rate": 3.8053277318393967e-06, + "loss": 0.7826, + "num_input_tokens_seen": 60654780, + "step": 1856 + }, + { + "epoch": 0.1674708030842765, + "flos": 23551118563200.0, + "grad_norm": 4.684581203897335, + "learning_rate": 3.805076241935498e-06, + "loss": 0.7993, + "num_input_tokens_seen": 60683710, + "step": 1857 + }, + { + "epoch": 0.16756098660774676, + "flos": 20056164660480.0, + "grad_norm": 6.617151757946565, + "learning_rate": 3.804824598012541e-06, + "loss": 0.662, + "num_input_tokens_seen": 60711935, + "step": 1858 + }, + { + "epoch": 0.16765117013121703, + "flos": 29602824061920.0, + "grad_norm": 1.9954306141091143, + "learning_rate": 3.8045728000919975e-06, + "loss": 0.8248, + "num_input_tokens_seen": 60743415, + "step": 1859 + }, + { + "epoch": 0.1677413536546873, + "flos": 25008848536320.0, + "grad_norm": 2.448625848727834, + "learning_rate": 3.8043208481953524e-06, + "loss": 0.8206, + "num_input_tokens_seen": 60774485, + "step": 1860 + }, + { + "epoch": 0.16783153717815755, + "flos": 37579520784000.0, + "grad_norm": 2.392931654365795, + "learning_rate": 3.804068742344104e-06, + "loss": 0.7107, + "num_input_tokens_seen": 60806475, + "step": 1861 + }, + { + "epoch": 0.1679217207016278, + "flos": 30401802683040.0, + "grad_norm": 2.156052843008938, + "learning_rate": 3.8038164825597628e-06, + "loss": 0.7551, + "num_input_tokens_seen": 60833545, + "step": 1862 + }, + { + "epoch": 0.16801190422509807, + "flos": 21112137967680.0, + "grad_norm": 2.1588516020596638, + "learning_rate": 3.8035640688638537e-06, + "loss": 0.8655, + "num_input_tokens_seen": 60858915, + "step": 1863 + }, + { + "epoch": 0.16810208774856833, + "flos": 22752028432800.0, + "grad_norm": 1.9923059764862603, + "learning_rate": 3.8033115012779125e-06, + "loss": 0.8598, + "num_input_tokens_seen": 60885940, + "step": 1864 + }, + { + "epoch": 0.1681922712720386, + "flos": 25954744983360.0, + "grad_norm": 1.6366608456950318, + "learning_rate": 3.8030587798234915e-06, + "loss": 0.8273, + "num_input_tokens_seen": 60917100, + "step": 1865 + }, + { + "epoch": 0.16828245479550885, + "flos": 23472641732640.0, + "grad_norm": 1.7586640054573206, + "learning_rate": 3.802805904522153e-06, + "loss": 0.7183, + "num_input_tokens_seen": 60941995, + "step": 1866 + }, + { + "epoch": 0.1683726383189791, + "flos": 30693556828320.0, + "grad_norm": 1.4281382687409276, + "learning_rate": 3.8025528753954742e-06, + "loss": 0.775, + "num_input_tokens_seen": 60973690, + "step": 1867 + }, + { + "epoch": 0.16846282184244937, + "flos": 23881460652960.0, + "grad_norm": 1.967221287201624, + "learning_rate": 3.802299692465045e-06, + "loss": 0.7264, + "num_input_tokens_seen": 61001880, + "step": 1868 + }, + { + "epoch": 0.16855300536591963, + "flos": 26066011275840.0, + "grad_norm": 1.9211966705366486, + "learning_rate": 3.802046355752468e-06, + "loss": 0.7725, + "num_input_tokens_seen": 61029600, + "step": 1869 + }, + { + "epoch": 0.16864318888938992, + "flos": 23479927005600.0, + "grad_norm": 2.0928871312589825, + "learning_rate": 3.80179286527936e-06, + "loss": 0.8408, + "num_input_tokens_seen": 61055300, + "step": 1870 + }, + { + "epoch": 0.16873337241286018, + "flos": 21692189470560.0, + "grad_norm": 1.6763576414406927, + "learning_rate": 3.801539221067349e-06, + "loss": 0.6955, + "num_input_tokens_seen": 61084325, + "step": 1871 + }, + { + "epoch": 0.16882355593633044, + "flos": 22605779662560.0, + "grad_norm": 1.888590829682398, + "learning_rate": 3.801285423138079e-06, + "loss": 0.7382, + "num_input_tokens_seen": 61113205, + "step": 1872 + }, + { + "epoch": 0.1689137394598007, + "flos": 27997618716480.0, + "grad_norm": 1.9312408002615755, + "learning_rate": 3.8010314715132037e-06, + "loss": 0.7359, + "num_input_tokens_seen": 61143825, + "step": 1873 + }, + { + "epoch": 0.16900392298327097, + "flos": 31896001924320.0, + "grad_norm": 1.9817856843636699, + "learning_rate": 3.800777366214393e-06, + "loss": 0.7349, + "num_input_tokens_seen": 61175815, + "step": 1874 + }, + { + "epoch": 0.16909410650674123, + "flos": 20711905261920.0, + "grad_norm": 2.038844539057198, + "learning_rate": 3.800523107263328e-06, + "loss": 0.761, + "num_input_tokens_seen": 61204830, + "step": 1875 + }, + { + "epoch": 0.1691842900302115, + "flos": 26096236024320.0, + "grad_norm": 2.148592523703892, + "learning_rate": 3.800268694681703e-06, + "loss": 0.8383, + "num_input_tokens_seen": 61233600, + "step": 1876 + }, + { + "epoch": 0.16927447355368175, + "flos": 69869319312000.0, + "grad_norm": 0.7475490656668472, + "learning_rate": 3.800014128491227e-06, + "loss": 0.5989, + "num_input_tokens_seen": 61329760, + "step": 1877 + }, + { + "epoch": 0.169364657077152, + "flos": 23404795453440.0, + "grad_norm": 2.3078768112323447, + "learning_rate": 3.79975940871362e-06, + "loss": 0.8359, + "num_input_tokens_seen": 61359785, + "step": 1878 + }, + { + "epoch": 0.16945484060062227, + "flos": 25301531925600.0, + "grad_norm": 2.077863404294011, + "learning_rate": 3.799504535370617e-06, + "loss": 0.7877, + "num_input_tokens_seen": 61386115, + "step": 1879 + }, + { + "epoch": 0.16954502412409253, + "flos": 19325032318560.0, + "grad_norm": 5.5702275808744455, + "learning_rate": 3.799249508483964e-06, + "loss": 0.7238, + "num_input_tokens_seen": 61412340, + "step": 1880 + }, + { + "epoch": 0.1696352076475628, + "flos": 61931099025120.0, + "grad_norm": 0.7199270313428082, + "learning_rate": 3.798994328075422e-06, + "loss": 0.6588, + "num_input_tokens_seen": 61511010, + "step": 1881 + }, + { + "epoch": 0.16972539117103305, + "flos": 20383124302080.0, + "grad_norm": 2.5027921343396464, + "learning_rate": 3.798738994166765e-06, + "loss": 0.7151, + "num_input_tokens_seen": 61538640, + "step": 1882 + }, + { + "epoch": 0.1698155746945033, + "flos": 21184816315680.0, + "grad_norm": 1.7012420885068873, + "learning_rate": 3.7984835067797788e-06, + "loss": 0.7868, + "num_input_tokens_seen": 61566995, + "step": 1883 + }, + { + "epoch": 0.16990575821797357, + "flos": 27558203350080.0, + "grad_norm": 1.7015181494503477, + "learning_rate": 3.798227865936263e-06, + "loss": 0.8757, + "num_input_tokens_seen": 61597830, + "step": 1884 + }, + { + "epoch": 0.16999594174144383, + "flos": 39144614224800.0, + "grad_norm": 1.9524509945433046, + "learning_rate": 3.7979720716580297e-06, + "loss": 0.6534, + "num_input_tokens_seen": 61629550, + "step": 1885 + }, + { + "epoch": 0.1700861252649141, + "flos": 23552642523360.0, + "grad_norm": 1.6221083074502896, + "learning_rate": 3.7977161239669057e-06, + "loss": 0.7752, + "num_input_tokens_seen": 61658775, + "step": 1886 + }, + { + "epoch": 0.17017630878838436, + "flos": 24463556492640.0, + "grad_norm": 2.7076909305954144, + "learning_rate": 3.7974600228847294e-06, + "loss": 0.7625, + "num_input_tokens_seen": 61686145, + "step": 1887 + }, + { + "epoch": 0.17026649231185462, + "flos": 14512910239680.0, + "grad_norm": 2.4257488156421627, + "learning_rate": 3.7972037684333534e-06, + "loss": 0.7255, + "num_input_tokens_seen": 61710575, + "step": 1888 + }, + { + "epoch": 0.17035667583532488, + "flos": 36486632171520.0, + "grad_norm": 1.6118976118403174, + "learning_rate": 3.796947360634642e-06, + "loss": 0.7093, + "num_input_tokens_seen": 61741415, + "step": 1889 + }, + { + "epoch": 0.17044685935879514, + "flos": 25411274257920.0, + "grad_norm": 1.9382447072850744, + "learning_rate": 3.796690799510473e-06, + "loss": 0.8304, + "num_input_tokens_seen": 61770485, + "step": 1890 + }, + { + "epoch": 0.1705370428822654, + "flos": 21182771978880.0, + "grad_norm": 1.80361715175183, + "learning_rate": 3.7964340850827387e-06, + "loss": 0.7477, + "num_input_tokens_seen": 61799915, + "step": 1891 + }, + { + "epoch": 0.17062722640573566, + "flos": 59498591701440.0, + "grad_norm": 0.677707132896801, + "learning_rate": 3.7961772173733425e-06, + "loss": 0.5658, + "num_input_tokens_seen": 61885820, + "step": 1892 + }, + { + "epoch": 0.17071740992920592, + "flos": 28029999311040.0, + "grad_norm": 1.4454901508746556, + "learning_rate": 3.7959201964042024e-06, + "loss": 0.7698, + "num_input_tokens_seen": 61915885, + "step": 1893 + }, + { + "epoch": 0.1708075934526762, + "flos": 20126352634560.0, + "grad_norm": 2.311115262698251, + "learning_rate": 3.795663022197248e-06, + "loss": 0.8098, + "num_input_tokens_seen": 61945140, + "step": 1894 + }, + { + "epoch": 0.17089777697614647, + "flos": 22751805414240.0, + "grad_norm": 2.4463345765164686, + "learning_rate": 3.7954056947744242e-06, + "loss": 0.7831, + "num_input_tokens_seen": 61973650, + "step": 1895 + }, + { + "epoch": 0.17098796049961673, + "flos": 18961937936640.0, + "grad_norm": 1.8851204809849633, + "learning_rate": 3.7951482141576863e-06, + "loss": 0.7581, + "num_input_tokens_seen": 62001945, + "step": 1896 + }, + { + "epoch": 0.171078144023087, + "flos": 22788237512640.0, + "grad_norm": 1.4612034846229054, + "learning_rate": 3.794890580369004e-06, + "loss": 0.8343, + "num_input_tokens_seen": 62028950, + "step": 1897 + }, + { + "epoch": 0.17116832754655725, + "flos": 21366344921760.0, + "grad_norm": 1.840400109533627, + "learning_rate": 3.7946327934303612e-06, + "loss": 0.8132, + "num_input_tokens_seen": 62052230, + "step": 1898 + }, + { + "epoch": 0.1712585110700275, + "flos": 22967387254080.0, + "grad_norm": 2.4698011840310867, + "learning_rate": 3.794374853363752e-06, + "loss": 0.8786, + "num_input_tokens_seen": 62079035, + "step": 1899 + }, + { + "epoch": 0.17134869459349777, + "flos": 21476050084320.0, + "grad_norm": 1.7404839778733634, + "learning_rate": 3.794116760191187e-06, + "loss": 0.8066, + "num_input_tokens_seen": 62107755, + "step": 1900 + }, + { + "epoch": 0.17143887811696804, + "flos": 18999150600000.0, + "grad_norm": 2.134261425118655, + "learning_rate": 3.7938585139346877e-06, + "loss": 0.8739, + "num_input_tokens_seen": 62134880, + "step": 1901 + }, + { + "epoch": 0.1715290616404383, + "flos": 19399457645280.0, + "grad_norm": 2.0151471894983297, + "learning_rate": 3.793600114616288e-06, + "loss": 0.7635, + "num_input_tokens_seen": 62163280, + "step": 1902 + }, + { + "epoch": 0.17161924516390856, + "flos": 56823208466880.0, + "grad_norm": 0.73886997223384, + "learning_rate": 3.793341562258037e-06, + "loss": 0.5556, + "num_input_tokens_seen": 62249405, + "step": 1903 + }, + { + "epoch": 0.17170942868737882, + "flos": 24206598976320.0, + "grad_norm": 4.035514763916747, + "learning_rate": 3.7930828568819953e-06, + "loss": 0.7909, + "num_input_tokens_seen": 62279925, + "step": 1904 + }, + { + "epoch": 0.17179961221084908, + "flos": 25591427582880.0, + "grad_norm": 1.804730303368209, + "learning_rate": 3.7928239985102378e-06, + "loss": 0.7805, + "num_input_tokens_seen": 62307350, + "step": 1905 + }, + { + "epoch": 0.17188979573431934, + "flos": 19290867575520.0, + "grad_norm": 1.624832144650953, + "learning_rate": 3.7925649871648505e-06, + "loss": 0.7844, + "num_input_tokens_seen": 62335225, + "step": 1906 + }, + { + "epoch": 0.1719799792577896, + "flos": 21076337755200.0, + "grad_norm": 1.555395990468121, + "learning_rate": 3.792305822867935e-06, + "loss": 0.7026, + "num_input_tokens_seen": 62364270, + "step": 1907 + }, + { + "epoch": 0.17207016278125986, + "flos": 19983151784640.0, + "grad_norm": 1.9582202215873272, + "learning_rate": 3.792046505641604e-06, + "loss": 0.7454, + "num_input_tokens_seen": 62391465, + "step": 1908 + }, + { + "epoch": 0.17216034630473012, + "flos": 25077624059520.0, + "grad_norm": 2.7900124008832865, + "learning_rate": 3.791787035507984e-06, + "loss": 0.7435, + "num_input_tokens_seen": 62418735, + "step": 1909 + }, + { + "epoch": 0.17225052982820038, + "flos": 25151231651520.0, + "grad_norm": 1.9366488279067133, + "learning_rate": 3.7915274124892136e-06, + "loss": 0.7412, + "num_input_tokens_seen": 62445460, + "step": 1910 + }, + { + "epoch": 0.17234071335167064, + "flos": 28870948324800.0, + "grad_norm": 1.726817944992733, + "learning_rate": 3.7912676366074466e-06, + "loss": 0.7625, + "num_input_tokens_seen": 62477840, + "step": 1911 + }, + { + "epoch": 0.1724308968751409, + "flos": 24427273073280.0, + "grad_norm": 3.2727921243465143, + "learning_rate": 3.7910077078848478e-06, + "loss": 0.748, + "num_input_tokens_seen": 62506695, + "step": 1912 + }, + { + "epoch": 0.17252108039861117, + "flos": 30213174652800.0, + "grad_norm": 4.912386415833135, + "learning_rate": 3.7907476263435957e-06, + "loss": 0.7465, + "num_input_tokens_seen": 62537105, + "step": 1913 + }, + { + "epoch": 0.17261126392208143, + "flos": 39619346596800.0, + "grad_norm": 2.682873225205245, + "learning_rate": 3.7904873920058826e-06, + "loss": 0.751, + "num_input_tokens_seen": 62566330, + "step": 1914 + }, + { + "epoch": 0.1727014474455517, + "flos": 21799590108000.0, + "grad_norm": 2.8472399413841956, + "learning_rate": 3.7902270048939114e-06, + "loss": 0.6608, + "num_input_tokens_seen": 62592530, + "step": 1915 + }, + { + "epoch": 0.17279163096902195, + "flos": 27924977538240.0, + "grad_norm": 2.5551519409210535, + "learning_rate": 3.7899664650299023e-06, + "loss": 0.7456, + "num_input_tokens_seen": 62620975, + "step": 1916 + }, + { + "epoch": 0.1728818144924922, + "flos": 21403111548000.0, + "grad_norm": 1.2892842419958974, + "learning_rate": 3.7897057724360836e-06, + "loss": 0.7911, + "num_input_tokens_seen": 62650770, + "step": 1917 + }, + { + "epoch": 0.1729719980159625, + "flos": 21804831044160.0, + "grad_norm": 2.2774417391547344, + "learning_rate": 3.7894449271347004e-06, + "loss": 0.7959, + "num_input_tokens_seen": 62678350, + "step": 1918 + }, + { + "epoch": 0.17306218153943276, + "flos": 22641654214560.0, + "grad_norm": 1.6602360246228378, + "learning_rate": 3.789183929148009e-06, + "loss": 0.8415, + "num_input_tokens_seen": 62708455, + "step": 1919 + }, + { + "epoch": 0.17315236506290302, + "flos": 23659820142240.0, + "grad_norm": 1.6801454851190034, + "learning_rate": 3.7889227784982795e-06, + "loss": 0.8094, + "num_input_tokens_seen": 62738210, + "step": 1920 + }, + { + "epoch": 0.17324254858637328, + "flos": 23916666149280.0, + "grad_norm": 1.491778412510132, + "learning_rate": 3.7886614752077945e-06, + "loss": 0.7776, + "num_input_tokens_seen": 62768190, + "step": 1921 + }, + { + "epoch": 0.17333273210984354, + "flos": 20784472100640.0, + "grad_norm": 5.91326036140864, + "learning_rate": 3.7884000192988495e-06, + "loss": 0.7781, + "num_input_tokens_seen": 62795620, + "step": 1922 + }, + { + "epoch": 0.1734229156333138, + "flos": 23043150692160.0, + "grad_norm": 1.8827747008775444, + "learning_rate": 3.7881384107937546e-06, + "loss": 0.8311, + "num_input_tokens_seen": 62823835, + "step": 1923 + }, + { + "epoch": 0.17351309915678406, + "flos": 37907446839360.0, + "grad_norm": 2.5964247918096772, + "learning_rate": 3.78787664971483e-06, + "loss": 0.7713, + "num_input_tokens_seen": 62855790, + "step": 1924 + }, + { + "epoch": 0.17360328268025432, + "flos": 27705418534080.0, + "grad_norm": 1.3692027471056212, + "learning_rate": 3.7876147360844115e-06, + "loss": 0.7502, + "num_input_tokens_seen": 62888265, + "step": 1925 + }, + { + "epoch": 0.17369346620372458, + "flos": 18816246712800.0, + "grad_norm": 2.321039875731636, + "learning_rate": 3.7873526699248474e-06, + "loss": 0.8456, + "num_input_tokens_seen": 62915345, + "step": 1926 + }, + { + "epoch": 0.17378364972719484, + "flos": 19800173557920.0, + "grad_norm": 1.650069238192588, + "learning_rate": 3.7870904512584974e-06, + "loss": 0.7896, + "num_input_tokens_seen": 62945825, + "step": 1927 + }, + { + "epoch": 0.1738738332506651, + "flos": 26643609574560.0, + "grad_norm": 1.6303081380337168, + "learning_rate": 3.7868280801077368e-06, + "loss": 0.7808, + "num_input_tokens_seen": 62976500, + "step": 1928 + }, + { + "epoch": 0.17396401677413537, + "flos": 27525339548640.0, + "grad_norm": 1.87683117775025, + "learning_rate": 3.7865655564949517e-06, + "loss": 0.7725, + "num_input_tokens_seen": 63010575, + "step": 1929 + }, + { + "epoch": 0.17405420029760563, + "flos": 24935612641920.0, + "grad_norm": 1.5191601297583517, + "learning_rate": 3.786302880442542e-06, + "loss": 0.7499, + "num_input_tokens_seen": 63040185, + "step": 1930 + }, + { + "epoch": 0.1741443838210759, + "flos": 18777212731200.0, + "grad_norm": 2.221959117042899, + "learning_rate": 3.7860400519729215e-06, + "loss": 0.7779, + "num_input_tokens_seen": 63067030, + "step": 1931 + }, + { + "epoch": 0.17423456734454615, + "flos": 21038567545440.0, + "grad_norm": 2.3059289724324383, + "learning_rate": 3.7857770711085157e-06, + "loss": 0.8011, + "num_input_tokens_seen": 63092535, + "step": 1932 + }, + { + "epoch": 0.1743247508680164, + "flos": 22417783518240.0, + "grad_norm": 2.1025524431326432, + "learning_rate": 3.785513937871763e-06, + "loss": 0.8287, + "num_input_tokens_seen": 63121065, + "step": 1933 + }, + { + "epoch": 0.17441493439148667, + "flos": 24751742340960.0, + "grad_norm": 2.3623228541571004, + "learning_rate": 3.785250652285116e-06, + "loss": 0.7498, + "num_input_tokens_seen": 63149705, + "step": 1934 + }, + { + "epoch": 0.17450511791495693, + "flos": 27231280878240.0, + "grad_norm": 1.9202932035564575, + "learning_rate": 3.78498721437104e-06, + "loss": 0.8598, + "num_input_tokens_seen": 63176065, + "step": 1935 + }, + { + "epoch": 0.1745953014384272, + "flos": 14846040061440.0, + "grad_norm": 2.54056736879568, + "learning_rate": 3.784723624152012e-06, + "loss": 0.7798, + "num_input_tokens_seen": 63199935, + "step": 1936 + }, + { + "epoch": 0.17468548496189745, + "flos": 26357430893280.0, + "grad_norm": 2.574583268007902, + "learning_rate": 3.784459881650524e-06, + "loss": 0.7255, + "num_input_tokens_seen": 63229370, + "step": 1937 + }, + { + "epoch": 0.1747756684853677, + "flos": 23443978114080.0, + "grad_norm": 1.708493023673649, + "learning_rate": 3.784195986889079e-06, + "loss": 0.7501, + "num_input_tokens_seen": 63259265, + "step": 1938 + }, + { + "epoch": 0.17486585200883797, + "flos": 28690125944160.0, + "grad_norm": 1.7146379699283234, + "learning_rate": 3.7839319398901946e-06, + "loss": 0.7058, + "num_input_tokens_seen": 63287150, + "step": 1939 + }, + { + "epoch": 0.17495603553230824, + "flos": 26868409514880.0, + "grad_norm": 1.6987891344695267, + "learning_rate": 3.7836677406764013e-06, + "loss": 0.85, + "num_input_tokens_seen": 63317275, + "step": 1940 + }, + { + "epoch": 0.1750462190557785, + "flos": 25411014069600.0, + "grad_norm": 1.6291985192400569, + "learning_rate": 3.7834033892702407e-06, + "loss": 0.8245, + "num_input_tokens_seen": 63345935, + "step": 1941 + }, + { + "epoch": 0.17513640257924878, + "flos": 25593248901120.0, + "grad_norm": 1.9125582019202505, + "learning_rate": 3.783138885694269e-06, + "loss": 0.7367, + "num_input_tokens_seen": 63375375, + "step": 1942 + }, + { + "epoch": 0.17522658610271905, + "flos": 18015706961760.0, + "grad_norm": 1.98603595096437, + "learning_rate": 3.7828742299710558e-06, + "loss": 0.8561, + "num_input_tokens_seen": 63402590, + "step": 1943 + }, + { + "epoch": 0.1753167696261893, + "flos": 23876479905120.0, + "grad_norm": 1.6833537698518655, + "learning_rate": 3.782609422123183e-06, + "loss": 0.7574, + "num_input_tokens_seen": 63429540, + "step": 1944 + }, + { + "epoch": 0.17540695314965957, + "flos": 18488618015520.0, + "grad_norm": 2.224143980140893, + "learning_rate": 3.7823444621732444e-06, + "loss": 0.7794, + "num_input_tokens_seen": 63457955, + "step": 1945 + }, + { + "epoch": 0.17549713667312983, + "flos": 24318831682560.0, + "grad_norm": 1.8471284076065235, + "learning_rate": 3.782079350143849e-06, + "loss": 0.7944, + "num_input_tokens_seen": 63486700, + "step": 1946 + }, + { + "epoch": 0.1755873201966001, + "flos": 22090563688320.0, + "grad_norm": 2.1145231366789385, + "learning_rate": 3.781814086057617e-06, + "loss": 0.8791, + "num_input_tokens_seen": 63512930, + "step": 1947 + }, + { + "epoch": 0.17567750372007035, + "flos": 17140593204960.0, + "grad_norm": 2.023720032986683, + "learning_rate": 3.7815486699371826e-06, + "loss": 0.7707, + "num_input_tokens_seen": 63540100, + "step": 1948 + }, + { + "epoch": 0.1757676872435406, + "flos": 42133607423520.0, + "grad_norm": 2.032565585834261, + "learning_rate": 3.7812831018051918e-06, + "loss": 0.7526, + "num_input_tokens_seen": 63572245, + "step": 1949 + }, + { + "epoch": 0.17585787076701087, + "flos": 19763369761920.0, + "grad_norm": 1.7674367605763774, + "learning_rate": 3.7810173816843058e-06, + "loss": 0.786, + "num_input_tokens_seen": 63600060, + "step": 1950 + }, + { + "epoch": 0.17594805429048113, + "flos": 23405018472000.0, + "grad_norm": 1.9463473711787507, + "learning_rate": 3.7807515095971955e-06, + "loss": 0.8531, + "num_input_tokens_seen": 63631395, + "step": 1951 + }, + { + "epoch": 0.1760382378139514, + "flos": 21690479661600.0, + "grad_norm": 1.7890880199761123, + "learning_rate": 3.7804854855665475e-06, + "loss": 0.7787, + "num_input_tokens_seen": 63659145, + "step": 1952 + }, + { + "epoch": 0.17612842133742165, + "flos": 20344462018080.0, + "grad_norm": 1.8189466206297977, + "learning_rate": 3.7802193096150606e-06, + "loss": 0.7894, + "num_input_tokens_seen": 63686775, + "step": 1953 + }, + { + "epoch": 0.17621860486089191, + "flos": 21986285310720.0, + "grad_norm": 1.5203980471780385, + "learning_rate": 3.779952981765446e-06, + "loss": 0.7321, + "num_input_tokens_seen": 63715910, + "step": 1954 + }, + { + "epoch": 0.17630878838436218, + "flos": 28799533748640.0, + "grad_norm": 1.5504166104730959, + "learning_rate": 3.779686502040429e-06, + "loss": 0.797, + "num_input_tokens_seen": 63747305, + "step": 1955 + }, + { + "epoch": 0.17639897190783244, + "flos": 69078486601920.0, + "grad_norm": 0.7086865769119636, + "learning_rate": 3.779419870462746e-06, + "loss": 0.5741, + "num_input_tokens_seen": 63828405, + "step": 1956 + }, + { + "epoch": 0.1764891554313027, + "flos": 24354780574080.0, + "grad_norm": 2.2101120772790055, + "learning_rate": 3.779153087055148e-06, + "loss": 0.7598, + "num_input_tokens_seen": 63857380, + "step": 1957 + }, + { + "epoch": 0.17657933895477296, + "flos": 25083385372320.0, + "grad_norm": 2.516240394145673, + "learning_rate": 3.7788861518403988e-06, + "loss": 0.782, + "num_input_tokens_seen": 63884510, + "step": 1958 + }, + { + "epoch": 0.17666952247824322, + "flos": 21257048626560.0, + "grad_norm": 2.429734698754032, + "learning_rate": 3.7786190648412742e-06, + "loss": 0.6778, + "num_input_tokens_seen": 63913325, + "step": 1959 + }, + { + "epoch": 0.17675970600171348, + "flos": 60886505398080.0, + "grad_norm": 0.6353366337595725, + "learning_rate": 3.778351826080564e-06, + "loss": 0.5867, + "num_input_tokens_seen": 64000150, + "step": 1960 + }, + { + "epoch": 0.17684988952518374, + "flos": 21876580148160.0, + "grad_norm": 1.7516301600568682, + "learning_rate": 3.7780844355810704e-06, + "loss": 0.8853, + "num_input_tokens_seen": 64029480, + "step": 1961 + }, + { + "epoch": 0.176940073048654, + "flos": 19800433746240.0, + "grad_norm": 1.7306399293023462, + "learning_rate": 3.777816893365608e-06, + "loss": 0.7938, + "num_input_tokens_seen": 64059845, + "step": 1962 + }, + { + "epoch": 0.17703025657212426, + "flos": 21403929282720.0, + "grad_norm": 2.7115719627188, + "learning_rate": 3.7775491994570057e-06, + "loss": 0.8566, + "num_input_tokens_seen": 64086260, + "step": 1963 + }, + { + "epoch": 0.17712044009559452, + "flos": 20124791504640.0, + "grad_norm": 2.2444448771124863, + "learning_rate": 3.777281353878105e-06, + "loss": 0.7596, + "num_input_tokens_seen": 64114610, + "step": 1964 + }, + { + "epoch": 0.17721062361906478, + "flos": 59601117366720.0, + "grad_norm": 0.629112304671583, + "learning_rate": 3.777013356651758e-06, + "loss": 0.5923, + "num_input_tokens_seen": 64213585, + "step": 1965 + }, + { + "epoch": 0.17730080714253507, + "flos": 17214721173600.0, + "grad_norm": 2.1011336218689762, + "learning_rate": 3.776745207800834e-06, + "loss": 0.7996, + "num_input_tokens_seen": 64239565, + "step": 1966 + }, + { + "epoch": 0.17739099066600533, + "flos": 22789055247360.0, + "grad_norm": 1.9624706217516743, + "learning_rate": 3.7764769073482122e-06, + "loss": 0.8083, + "num_input_tokens_seen": 64269440, + "step": 1967 + }, + { + "epoch": 0.1774811741894756, + "flos": 23219475531840.0, + "grad_norm": 1.3808212167088931, + "learning_rate": 3.7762084553167846e-06, + "loss": 0.8467, + "num_input_tokens_seen": 64299410, + "step": 1968 + }, + { + "epoch": 0.17757135771294585, + "flos": 24382477778880.0, + "grad_norm": 1.7567389028320026, + "learning_rate": 3.775939851729458e-06, + "loss": 0.8746, + "num_input_tokens_seen": 64323710, + "step": 1969 + }, + { + "epoch": 0.17766154123641612, + "flos": 19251276047520.0, + "grad_norm": 2.3969556971447568, + "learning_rate": 3.775671096609151e-06, + "loss": 0.7401, + "num_input_tokens_seen": 64350025, + "step": 1970 + }, + { + "epoch": 0.17775172475988638, + "flos": 39542876933280.0, + "grad_norm": 1.7490608556377458, + "learning_rate": 3.775402189978795e-06, + "loss": 0.7647, + "num_input_tokens_seen": 64381915, + "step": 1971 + }, + { + "epoch": 0.17784190828335664, + "flos": 23442268305120.0, + "grad_norm": 2.452857519254966, + "learning_rate": 3.7751331318613343e-06, + "loss": 0.7645, + "num_input_tokens_seen": 64413000, + "step": 1972 + }, + { + "epoch": 0.1779320918068269, + "flos": 29812495909920.0, + "grad_norm": 1.8365639125214241, + "learning_rate": 3.774863922279727e-06, + "loss": 0.8021, + "num_input_tokens_seen": 64444550, + "step": 1973 + }, + { + "epoch": 0.17802227533029716, + "flos": 22093946136480.0, + "grad_norm": 3.2670953534201974, + "learning_rate": 3.7745945612569435e-06, + "loss": 0.8417, + "num_input_tokens_seen": 64472795, + "step": 1974 + }, + { + "epoch": 0.17811245885376742, + "flos": 27446156492640.0, + "grad_norm": 1.4645538450105897, + "learning_rate": 3.7743250488159674e-06, + "loss": 0.6604, + "num_input_tokens_seen": 64503580, + "step": 1975 + }, + { + "epoch": 0.17820264237723768, + "flos": 17869495361280.0, + "grad_norm": 2.2699270510610408, + "learning_rate": 3.774055384979794e-06, + "loss": 0.6812, + "num_input_tokens_seen": 64529685, + "step": 1976 + }, + { + "epoch": 0.17829282590070794, + "flos": 26427358679040.0, + "grad_norm": 2.2043459059220543, + "learning_rate": 3.773785569771433e-06, + "loss": 0.7621, + "num_input_tokens_seen": 64559705, + "step": 1977 + }, + { + "epoch": 0.1783830094241782, + "flos": 24209015010720.0, + "grad_norm": 1.6297245303049637, + "learning_rate": 3.7735156032139066e-06, + "loss": 0.7663, + "num_input_tokens_seen": 64590010, + "step": 1978 + }, + { + "epoch": 0.17847319294764846, + "flos": 26717700373440.0, + "grad_norm": 1.829237249079963, + "learning_rate": 3.773245485330251e-06, + "loss": 0.8813, + "num_input_tokens_seen": 64620495, + "step": 1979 + }, + { + "epoch": 0.17856337647111872, + "flos": 26430889806240.0, + "grad_norm": 2.1007011378438225, + "learning_rate": 3.7729752161435115e-06, + "loss": 0.7503, + "num_input_tokens_seen": 64649120, + "step": 1980 + }, + { + "epoch": 0.17865355999458898, + "flos": 24828323513760.0, + "grad_norm": 1.6405966620041472, + "learning_rate": 3.7727047956767514e-06, + "loss": 0.8475, + "num_input_tokens_seen": 64677705, + "step": 1981 + }, + { + "epoch": 0.17874374351805924, + "flos": 20532941369280.0, + "grad_norm": 1.6136928522819134, + "learning_rate": 3.7724342239530436e-06, + "loss": 0.9103, + "num_input_tokens_seen": 64708850, + "step": 1982 + }, + { + "epoch": 0.1788339270415295, + "flos": 34665956495520.0, + "grad_norm": 1.9856305092941378, + "learning_rate": 3.772163500995474e-06, + "loss": 0.7306, + "num_input_tokens_seen": 64741055, + "step": 1983 + }, + { + "epoch": 0.17892411056499977, + "flos": 27415039669920.0, + "grad_norm": 2.068212516549216, + "learning_rate": 3.7718926268271437e-06, + "loss": 0.8755, + "num_input_tokens_seen": 64768330, + "step": 1984 + }, + { + "epoch": 0.17901429408847003, + "flos": 22350940822560.0, + "grad_norm": 2.0858002235303474, + "learning_rate": 3.771621601471164e-06, + "loss": 0.8027, + "num_input_tokens_seen": 64795535, + "step": 1985 + }, + { + "epoch": 0.1791044776119403, + "flos": 31386026886240.0, + "grad_norm": 3.5152004409442497, + "learning_rate": 3.771350424950661e-06, + "loss": 0.8304, + "num_input_tokens_seen": 64826275, + "step": 1986 + }, + { + "epoch": 0.17919466113541055, + "flos": 26612975958720.0, + "grad_norm": 3.6821963429646294, + "learning_rate": 3.771079097288772e-06, + "loss": 0.7269, + "num_input_tokens_seen": 64853120, + "step": 1987 + }, + { + "epoch": 0.1792848446588808, + "flos": 22933482699360.0, + "grad_norm": 1.466210869204366, + "learning_rate": 3.770807618508649e-06, + "loss": 0.7985, + "num_input_tokens_seen": 64884505, + "step": 1988 + }, + { + "epoch": 0.17937502818235107, + "flos": 19467378264000.0, + "grad_norm": 1.6606390755846137, + "learning_rate": 3.7705359886334555e-06, + "loss": 0.7706, + "num_input_tokens_seen": 64911680, + "step": 1989 + }, + { + "epoch": 0.17946521170582136, + "flos": 23552679693120.0, + "grad_norm": 2.0101718918287776, + "learning_rate": 3.7702642076863694e-06, + "loss": 0.7609, + "num_input_tokens_seen": 64938505, + "step": 1990 + }, + { + "epoch": 0.17955539522929162, + "flos": 21184704806400.0, + "grad_norm": 1.6468047241041255, + "learning_rate": 3.7699922756905795e-06, + "loss": 0.8401, + "num_input_tokens_seen": 64966570, + "step": 1991 + }, + { + "epoch": 0.17964557875276188, + "flos": 27596828464320.0, + "grad_norm": 1.7262530157351548, + "learning_rate": 3.7697201926692895e-06, + "loss": 0.7333, + "num_input_tokens_seen": 64995445, + "step": 1992 + }, + { + "epoch": 0.17973576227623214, + "flos": 25847864722560.0, + "grad_norm": 2.081111167882369, + "learning_rate": 3.7694479586457144e-06, + "loss": 0.7953, + "num_input_tokens_seen": 65020950, + "step": 1993 + }, + { + "epoch": 0.1798259457997024, + "flos": 18269765236800.0, + "grad_norm": 2.723688835588219, + "learning_rate": 3.7691755736430827e-06, + "loss": 0.7983, + "num_input_tokens_seen": 65046690, + "step": 1994 + }, + { + "epoch": 0.17991612932317266, + "flos": 31131039367200.0, + "grad_norm": 2.026973036313361, + "learning_rate": 3.768903037684636e-06, + "loss": 0.8255, + "num_input_tokens_seen": 65078190, + "step": 1995 + }, + { + "epoch": 0.18000631284664292, + "flos": 25593620598720.0, + "grad_norm": 1.8319064425187892, + "learning_rate": 3.7686303507936284e-06, + "loss": 0.8335, + "num_input_tokens_seen": 65106295, + "step": 1996 + }, + { + "epoch": 0.18009649637011318, + "flos": 14481087191520.0, + "grad_norm": 2.2858638352618805, + "learning_rate": 3.7683575129933272e-06, + "loss": 0.8262, + "num_input_tokens_seen": 65131360, + "step": 1997 + }, + { + "epoch": 0.18018667989358345, + "flos": 23079285432480.0, + "grad_norm": 1.7351941073228352, + "learning_rate": 3.7680845243070128e-06, + "loss": 0.7909, + "num_input_tokens_seen": 65162910, + "step": 1998 + }, + { + "epoch": 0.1802768634170537, + "flos": 61678304521920.0, + "grad_norm": 0.7156958998827706, + "learning_rate": 3.7678113847579767e-06, + "loss": 0.5539, + "num_input_tokens_seen": 65246220, + "step": 1999 + }, + { + "epoch": 0.18036704694052397, + "flos": 23735434901280.0, + "grad_norm": 2.349162857613707, + "learning_rate": 3.7675380943695264e-06, + "loss": 0.854, + "num_input_tokens_seen": 65273650, + "step": 2000 + }, + { + "epoch": 0.18045723046399423, + "flos": 23735546410560.0, + "grad_norm": 1.9490833426555756, + "learning_rate": 3.7672646531649795e-06, + "loss": 0.7485, + "num_input_tokens_seen": 65302860, + "step": 2001 + }, + { + "epoch": 0.1805474139874645, + "flos": 64140707799840.0, + "grad_norm": 0.5768015378918298, + "learning_rate": 3.7669910611676682e-06, + "loss": 0.576, + "num_input_tokens_seen": 65400480, + "step": 2002 + }, + { + "epoch": 0.18063759751093475, + "flos": 30179344437600.0, + "grad_norm": 1.8260907976236482, + "learning_rate": 3.7667173184009356e-06, + "loss": 0.7702, + "num_input_tokens_seen": 65432500, + "step": 2003 + }, + { + "epoch": 0.180727781034405, + "flos": 22021490807040.0, + "grad_norm": 2.045087310307559, + "learning_rate": 3.7664434248881403e-06, + "loss": 0.8484, + "num_input_tokens_seen": 65462375, + "step": 2004 + }, + { + "epoch": 0.18081796455787527, + "flos": 19327150994880.0, + "grad_norm": 2.2886059406638455, + "learning_rate": 3.766169380652652e-06, + "loss": 0.7934, + "num_input_tokens_seen": 65490375, + "step": 2005 + }, + { + "epoch": 0.18090814808134553, + "flos": 19836828674880.0, + "grad_norm": 1.6870354613470362, + "learning_rate": 3.7658951857178537e-06, + "loss": 0.8281, + "num_input_tokens_seen": 65518395, + "step": 2006 + }, + { + "epoch": 0.1809983316048158, + "flos": 22203205261920.0, + "grad_norm": 1.6153888374839354, + "learning_rate": 3.7656208401071414e-06, + "loss": 0.7938, + "num_input_tokens_seen": 65546695, + "step": 2007 + }, + { + "epoch": 0.18108851512828605, + "flos": 23407099978560.0, + "grad_norm": 5.146680795587572, + "learning_rate": 3.7653463438439225e-06, + "loss": 0.7244, + "num_input_tokens_seen": 65574690, + "step": 2008 + }, + { + "epoch": 0.18117869865175631, + "flos": 20090775440640.0, + "grad_norm": 2.105440893325112, + "learning_rate": 3.7650716969516203e-06, + "loss": 0.7972, + "num_input_tokens_seen": 65600755, + "step": 2009 + }, + { + "epoch": 0.18126888217522658, + "flos": 29965472406720.0, + "grad_norm": 1.9203989206582397, + "learning_rate": 3.764796899453668e-06, + "loss": 0.6863, + "num_input_tokens_seen": 65631750, + "step": 2010 + }, + { + "epoch": 0.18135906569869684, + "flos": 31203122999040.0, + "grad_norm": 1.9053169218152506, + "learning_rate": 3.7645219513735134e-06, + "loss": 0.8253, + "num_input_tokens_seen": 65661435, + "step": 2011 + }, + { + "epoch": 0.1814492492221671, + "flos": 32625275778240.0, + "grad_norm": 1.6551847811112186, + "learning_rate": 3.764246852734617e-06, + "loss": 0.8276, + "num_input_tokens_seen": 65693150, + "step": 2012 + }, + { + "epoch": 0.18153943274563739, + "flos": 18816841428960.0, + "grad_norm": 2.1424005203565515, + "learning_rate": 3.7639716035604502e-06, + "loss": 0.76, + "num_input_tokens_seen": 65720205, + "step": 2013 + }, + { + "epoch": 0.18162961626910765, + "flos": 71752563161280.0, + "grad_norm": 0.6405210410644208, + "learning_rate": 3.7636962038745e-06, + "loss": 0.548, + "num_input_tokens_seen": 65817730, + "step": 2014 + }, + { + "epoch": 0.1817197997925779, + "flos": 28435287104160.0, + "grad_norm": 1.9120070027737885, + "learning_rate": 3.763420653700265e-06, + "loss": 0.7483, + "num_input_tokens_seen": 65846500, + "step": 2015 + }, + { + "epoch": 0.18180998331604817, + "flos": 25734479753760.0, + "grad_norm": 1.8443930199679035, + "learning_rate": 3.7631449530612565e-06, + "loss": 0.7431, + "num_input_tokens_seen": 65874775, + "step": 2016 + }, + { + "epoch": 0.18190016683951843, + "flos": 25151529009600.0, + "grad_norm": 1.5593042116782392, + "learning_rate": 3.762869101980999e-06, + "loss": 0.8748, + "num_input_tokens_seen": 65901380, + "step": 2017 + }, + { + "epoch": 0.1819903503629887, + "flos": 22496186009280.0, + "grad_norm": 1.9391835804283475, + "learning_rate": 3.7625931004830287e-06, + "loss": 0.7829, + "num_input_tokens_seen": 65931360, + "step": 2018 + }, + { + "epoch": 0.18208053388645895, + "flos": 25993444437120.0, + "grad_norm": 1.8686261241260536, + "learning_rate": 3.7623169485908966e-06, + "loss": 0.7416, + "num_input_tokens_seen": 65962435, + "step": 2019 + }, + { + "epoch": 0.1821707174099292, + "flos": 67376734188960.0, + "grad_norm": 0.7182289390941662, + "learning_rate": 3.7620406463281647e-06, + "loss": 0.5948, + "num_input_tokens_seen": 66039400, + "step": 2020 + }, + { + "epoch": 0.18226090093339947, + "flos": 28324950055680.0, + "grad_norm": 2.7171327889250034, + "learning_rate": 3.7617641937184095e-06, + "loss": 0.8084, + "num_input_tokens_seen": 66069585, + "step": 2021 + }, + { + "epoch": 0.18235108445686973, + "flos": 24026780179200.0, + "grad_norm": 5.21052167568781, + "learning_rate": 3.761487590785219e-06, + "loss": 0.8137, + "num_input_tokens_seen": 66096660, + "step": 2022 + }, + { + "epoch": 0.18244126798034, + "flos": 22489272433920.0, + "grad_norm": 2.1123593165034786, + "learning_rate": 3.7612108375521942e-06, + "loss": 0.7573, + "num_input_tokens_seen": 66125090, + "step": 2023 + }, + { + "epoch": 0.18253145150381025, + "flos": 23479964175360.0, + "grad_norm": 1.5788503811562609, + "learning_rate": 3.76093393404295e-06, + "loss": 0.8337, + "num_input_tokens_seen": 66153680, + "step": 2024 + }, + { + "epoch": 0.18262163502728052, + "flos": 22533026975040.0, + "grad_norm": 2.1687688777222913, + "learning_rate": 3.7606568802811126e-06, + "loss": 0.7777, + "num_input_tokens_seen": 66182060, + "step": 2025 + }, + { + "epoch": 0.18271181855075078, + "flos": 17206432317120.0, + "grad_norm": 5.931093313159413, + "learning_rate": 3.760379676290322e-06, + "loss": 0.7981, + "num_input_tokens_seen": 66204435, + "step": 2026 + }, + { + "epoch": 0.18280200207422104, + "flos": 20524169305920.0, + "grad_norm": 5.138355809419423, + "learning_rate": 3.760102322094231e-06, + "loss": 0.7158, + "num_input_tokens_seen": 66232095, + "step": 2027 + }, + { + "epoch": 0.1828921855976913, + "flos": 25739088804000.0, + "grad_norm": 1.8692139677927075, + "learning_rate": 3.759824817716504e-06, + "loss": 0.7483, + "num_input_tokens_seen": 66263480, + "step": 2028 + }, + { + "epoch": 0.18298236912116156, + "flos": 26430406599360.0, + "grad_norm": 1.2606168960968973, + "learning_rate": 3.759547163180821e-06, + "loss": 0.8375, + "num_input_tokens_seen": 66293645, + "step": 2029 + }, + { + "epoch": 0.18307255264463182, + "flos": 27925423575360.0, + "grad_norm": 1.7289113082156717, + "learning_rate": 3.759269358510871e-06, + "loss": 0.5874, + "num_input_tokens_seen": 66323555, + "step": 2030 + }, + { + "epoch": 0.18316273616810208, + "flos": 27670213037760.0, + "grad_norm": 1.8497667577989547, + "learning_rate": 3.75899140373036e-06, + "loss": 0.6913, + "num_input_tokens_seen": 66351760, + "step": 2031 + }, + { + "epoch": 0.18325291969157234, + "flos": 18087716254080.0, + "grad_norm": 2.7227298940060343, + "learning_rate": 3.7587132988630028e-06, + "loss": 0.8439, + "num_input_tokens_seen": 66379650, + "step": 2032 + }, + { + "epoch": 0.1833431032150426, + "flos": 14773175864640.0, + "grad_norm": 1.9724276268546133, + "learning_rate": 3.7584350439325295e-06, + "loss": 0.8192, + "num_input_tokens_seen": 66404565, + "step": 2033 + }, + { + "epoch": 0.18343328673851286, + "flos": 17322865206240.0, + "grad_norm": 1.92662837526138, + "learning_rate": 3.758156638962682e-06, + "loss": 0.7741, + "num_input_tokens_seen": 66432025, + "step": 2034 + }, + { + "epoch": 0.18352347026198312, + "flos": 27156037816800.0, + "grad_norm": 1.7891838332431897, + "learning_rate": 3.757878083977216e-06, + "loss": 0.7464, + "num_input_tokens_seen": 66462635, + "step": 2035 + }, + { + "epoch": 0.18361365378545338, + "flos": 26140510942080.0, + "grad_norm": 1.670654207681433, + "learning_rate": 3.7575993789999e-06, + "loss": 0.7742, + "num_input_tokens_seen": 66493385, + "step": 2036 + }, + { + "epoch": 0.18370383730892367, + "flos": 22824409422720.0, + "grad_norm": 1.852559654109333, + "learning_rate": 3.757320524054512e-06, + "loss": 0.8211, + "num_input_tokens_seen": 66520660, + "step": 2037 + }, + { + "epoch": 0.18379402083239393, + "flos": 22168445802720.0, + "grad_norm": 1.84867904774541, + "learning_rate": 3.757041519164848e-06, + "loss": 0.7166, + "num_input_tokens_seen": 66551500, + "step": 2038 + }, + { + "epoch": 0.1838842043558642, + "flos": 25077809908320.0, + "grad_norm": 2.385222670039203, + "learning_rate": 3.7567623643547133e-06, + "loss": 0.7929, + "num_input_tokens_seen": 66578560, + "step": 2039 + }, + { + "epoch": 0.18397438787933446, + "flos": 25338558740160.0, + "grad_norm": 1.749066103654986, + "learning_rate": 3.756483059647927e-06, + "loss": 0.784, + "num_input_tokens_seen": 66605945, + "step": 2040 + }, + { + "epoch": 0.18406457140280472, + "flos": 34914364967040.0, + "grad_norm": 1.54842999838543, + "learning_rate": 3.756203605068321e-06, + "loss": 0.7588, + "num_input_tokens_seen": 66636750, + "step": 2041 + }, + { + "epoch": 0.18415475492627498, + "flos": 25302386830080.0, + "grad_norm": 1.793821443736605, + "learning_rate": 3.7559240006397396e-06, + "loss": 0.774, + "num_input_tokens_seen": 66667395, + "step": 2042 + }, + { + "epoch": 0.18424493844974524, + "flos": 23691903378720.0, + "grad_norm": 1.8515832312344809, + "learning_rate": 3.7556442463860406e-06, + "loss": 0.9032, + "num_input_tokens_seen": 66696920, + "step": 2043 + }, + { + "epoch": 0.1843351219732155, + "flos": 23699262991200.0, + "grad_norm": 2.228404250441883, + "learning_rate": 3.7553643423310934e-06, + "loss": 0.8948, + "num_input_tokens_seen": 66724955, + "step": 2044 + }, + { + "epoch": 0.18442530549668576, + "flos": 26939155035360.0, + "grad_norm": 2.482332937836909, + "learning_rate": 3.755084288498782e-06, + "loss": 0.6392, + "num_input_tokens_seen": 66755910, + "step": 2045 + }, + { + "epoch": 0.18451548902015602, + "flos": 26358062779200.0, + "grad_norm": 1.798163441342245, + "learning_rate": 3.754804084913002e-06, + "loss": 0.7902, + "num_input_tokens_seen": 66784685, + "step": 2046 + }, + { + "epoch": 0.18460567254362628, + "flos": 21729439303680.0, + "grad_norm": 1.6807385372172141, + "learning_rate": 3.754523731597661e-06, + "loss": 0.8509, + "num_input_tokens_seen": 66813810, + "step": 2047 + }, + { + "epoch": 0.18469585606709654, + "flos": 20601754062240.0, + "grad_norm": 2.2522775990288495, + "learning_rate": 3.754243228576681e-06, + "loss": 0.6812, + "num_input_tokens_seen": 66842535, + "step": 2048 + }, + { + "epoch": 0.1847860395905668, + "flos": 27415225518720.0, + "grad_norm": 1.933837998501804, + "learning_rate": 3.753962575873996e-06, + "loss": 0.8384, + "num_input_tokens_seen": 66869350, + "step": 2049 + }, + { + "epoch": 0.18487622311403706, + "flos": 22679312915040.0, + "grad_norm": 1.716675425325043, + "learning_rate": 3.7536817735135527e-06, + "loss": 0.7463, + "num_input_tokens_seen": 66898125, + "step": 2050 + }, + { + "epoch": 0.18496640663750732, + "flos": 27955499644800.0, + "grad_norm": 1.5009343892751652, + "learning_rate": 3.753400821519311e-06, + "loss": 0.8273, + "num_input_tokens_seen": 66927860, + "step": 2051 + }, + { + "epoch": 0.18505659016097759, + "flos": 43481000348160.0, + "grad_norm": 2.881307069291864, + "learning_rate": 3.7531197199152426e-06, + "loss": 0.7161, + "num_input_tokens_seen": 66964005, + "step": 2052 + }, + { + "epoch": 0.18514677368444785, + "flos": 64781171629920.0, + "grad_norm": 0.7119367514763669, + "learning_rate": 3.7528384687253335e-06, + "loss": 0.6174, + "num_input_tokens_seen": 67053575, + "step": 2053 + }, + { + "epoch": 0.1852369572079181, + "flos": 27560024668320.0, + "grad_norm": 2.1386072157748504, + "learning_rate": 3.7525570679735815e-06, + "loss": 0.8915, + "num_input_tokens_seen": 67082830, + "step": 2054 + }, + { + "epoch": 0.18532714073138837, + "flos": 21689141550240.0, + "grad_norm": 1.7584012550264485, + "learning_rate": 3.7522755176839965e-06, + "loss": 0.8005, + "num_input_tokens_seen": 67109515, + "step": 2055 + }, + { + "epoch": 0.18541732425485863, + "flos": 21140801586240.0, + "grad_norm": 2.3835370072344078, + "learning_rate": 3.7519938178806027e-06, + "loss": 0.7967, + "num_input_tokens_seen": 67136990, + "step": 2056 + }, + { + "epoch": 0.1855075077783289, + "flos": 23807667212160.0, + "grad_norm": 2.245218282168244, + "learning_rate": 3.7517119685874358e-06, + "loss": 0.7364, + "num_input_tokens_seen": 67165250, + "step": 2057 + }, + { + "epoch": 0.18559769130179915, + "flos": 23079508451040.0, + "grad_norm": 1.7584150849753097, + "learning_rate": 3.7514299698285447e-06, + "loss": 0.7764, + "num_input_tokens_seen": 67195875, + "step": 2058 + }, + { + "epoch": 0.1856878748252694, + "flos": 18925431498720.0, + "grad_norm": 1.3917989597524425, + "learning_rate": 3.751147821627991e-06, + "loss": 0.7652, + "num_input_tokens_seen": 67224965, + "step": 2059 + }, + { + "epoch": 0.18577805834873967, + "flos": 20347175410560.0, + "grad_norm": 2.586823464373582, + "learning_rate": 3.75086552400985e-06, + "loss": 0.8046, + "num_input_tokens_seen": 67252155, + "step": 2060 + }, + { + "epoch": 0.18586824187220996, + "flos": 21840556917120.0, + "grad_norm": 2.0448937764828736, + "learning_rate": 3.750583076998208e-06, + "loss": 0.7633, + "num_input_tokens_seen": 67280115, + "step": 2061 + }, + { + "epoch": 0.18595842539568022, + "flos": 24500211609600.0, + "grad_norm": 1.8682777701199742, + "learning_rate": 3.7503004806171655e-06, + "loss": 0.764, + "num_input_tokens_seen": 67309320, + "step": 2062 + }, + { + "epoch": 0.18604860891915048, + "flos": 24095853060480.0, + "grad_norm": 2.0508284499776654, + "learning_rate": 3.7500177348908354e-06, + "loss": 0.717, + "num_input_tokens_seen": 67338045, + "step": 2063 + }, + { + "epoch": 0.18613879244262074, + "flos": 23043894087360.0, + "grad_norm": 1.7215521424944968, + "learning_rate": 3.749734839843342e-06, + "loss": 0.8073, + "num_input_tokens_seen": 67365425, + "step": 2064 + }, + { + "epoch": 0.186228975966091, + "flos": 23148246804480.0, + "grad_norm": 2.542385333240857, + "learning_rate": 3.7494517954988245e-06, + "loss": 0.785, + "num_input_tokens_seen": 67394410, + "step": 2065 + }, + { + "epoch": 0.18631915948956126, + "flos": 26832126095520.0, + "grad_norm": 1.8753739302022612, + "learning_rate": 3.749168601881433e-06, + "loss": 0.774, + "num_input_tokens_seen": 67423550, + "step": 2066 + }, + { + "epoch": 0.18640934301303153, + "flos": 20347547108160.0, + "grad_norm": 2.085243938455986, + "learning_rate": 3.7488852590153315e-06, + "loss": 0.7678, + "num_input_tokens_seen": 67449005, + "step": 2067 + }, + { + "epoch": 0.1864995265365018, + "flos": 20529373072320.0, + "grad_norm": 1.739578346047324, + "learning_rate": 3.748601766924697e-06, + "loss": 0.7839, + "num_input_tokens_seen": 67476265, + "step": 2068 + }, + { + "epoch": 0.18658971005997205, + "flos": 20092113552000.0, + "grad_norm": 1.9932245285529113, + "learning_rate": 3.7483181256337176e-06, + "loss": 0.7542, + "num_input_tokens_seen": 67504095, + "step": 2069 + }, + { + "epoch": 0.1866798935834423, + "flos": 23402862625920.0, + "grad_norm": 1.54087652973524, + "learning_rate": 3.7480343351665962e-06, + "loss": 0.7897, + "num_input_tokens_seen": 67532990, + "step": 2070 + }, + { + "epoch": 0.18677007710691257, + "flos": 18633863202240.0, + "grad_norm": 1.7004498111056119, + "learning_rate": 3.747750395547546e-06, + "loss": 0.7701, + "num_input_tokens_seen": 67562055, + "step": 2071 + }, + { + "epoch": 0.18686026063038283, + "flos": 22095024059520.0, + "grad_norm": 1.53791526210779, + "learning_rate": 3.7474663068007956e-06, + "loss": 0.7701, + "num_input_tokens_seen": 67593605, + "step": 2072 + }, + { + "epoch": 0.1869504441538531, + "flos": 35721372256320.0, + "grad_norm": 1.2817526040866412, + "learning_rate": 3.747182068950584e-06, + "loss": 0.7706, + "num_input_tokens_seen": 67626235, + "step": 2073 + }, + { + "epoch": 0.18704062767732335, + "flos": 25409378600160.0, + "grad_norm": 1.8671520022286, + "learning_rate": 3.7468976820211643e-06, + "loss": 0.7834, + "num_input_tokens_seen": 67655335, + "step": 2074 + }, + { + "epoch": 0.1871308112007936, + "flos": 20383124302080.0, + "grad_norm": 1.8310289060684974, + "learning_rate": 3.746613146036803e-06, + "loss": 0.7831, + "num_input_tokens_seen": 67683135, + "step": 2075 + }, + { + "epoch": 0.18722099472426387, + "flos": 20019398034240.0, + "grad_norm": 1.7102949773670428, + "learning_rate": 3.7463284610217766e-06, + "loss": 0.8216, + "num_input_tokens_seen": 67710625, + "step": 2076 + }, + { + "epoch": 0.18731117824773413, + "flos": 19363174225920.0, + "grad_norm": 2.0484495440492014, + "learning_rate": 3.746043627000377e-06, + "loss": 0.7175, + "num_input_tokens_seen": 67739275, + "step": 2077 + }, + { + "epoch": 0.1874013617712044, + "flos": 52585010614080.0, + "grad_norm": 1.899567507157526, + "learning_rate": 3.7457586439969076e-06, + "loss": 0.73, + "num_input_tokens_seen": 67771950, + "step": 2078 + }, + { + "epoch": 0.18749154529467466, + "flos": 26429997732000.0, + "grad_norm": 1.5353579789034766, + "learning_rate": 3.7454735120356842e-06, + "loss": 0.7703, + "num_input_tokens_seen": 67802930, + "step": 2079 + }, + { + "epoch": 0.18758172881814492, + "flos": 22896864752160.0, + "grad_norm": 1.518254152837206, + "learning_rate": 3.7451882311410373e-06, + "loss": 0.8487, + "num_input_tokens_seen": 67831940, + "step": 2080 + }, + { + "epoch": 0.18767191234161518, + "flos": 20310631802880.0, + "grad_norm": 1.5662629402034047, + "learning_rate": 3.7449028013373074e-06, + "loss": 0.8206, + "num_input_tokens_seen": 67860115, + "step": 2081 + }, + { + "epoch": 0.18776209586508544, + "flos": 26027980877760.0, + "grad_norm": 1.8229193623840745, + "learning_rate": 3.7446172226488485e-06, + "loss": 0.783, + "num_input_tokens_seen": 67889555, + "step": 2082 + }, + { + "epoch": 0.1878522793885557, + "flos": 26576581030080.0, + "grad_norm": 1.999498125526139, + "learning_rate": 3.7443314951000285e-06, + "loss": 0.8438, + "num_input_tokens_seen": 67916025, + "step": 2083 + }, + { + "epoch": 0.18794246291202596, + "flos": 24244963902240.0, + "grad_norm": 2.653613049445326, + "learning_rate": 3.7440456187152276e-06, + "loss": 0.7675, + "num_input_tokens_seen": 67945160, + "step": 2084 + }, + { + "epoch": 0.18803264643549625, + "flos": 20638074651360.0, + "grad_norm": 2.1095576454490774, + "learning_rate": 3.7437595935188377e-06, + "loss": 0.8638, + "num_input_tokens_seen": 67973500, + "step": 2085 + }, + { + "epoch": 0.1881228299589665, + "flos": 21840742765920.0, + "grad_norm": 2.521354169608525, + "learning_rate": 3.7434734195352647e-06, + "loss": 0.7807, + "num_input_tokens_seen": 68002565, + "step": 2086 + }, + { + "epoch": 0.18821301348243677, + "flos": 57444078099840.0, + "grad_norm": 0.6536846200040857, + "learning_rate": 3.743187096788926e-06, + "loss": 0.58, + "num_input_tokens_seen": 68086760, + "step": 2087 + }, + { + "epoch": 0.18830319700590703, + "flos": 14190262290240.0, + "grad_norm": 1.8028130643291254, + "learning_rate": 3.7429006253042524e-06, + "loss": 0.7812, + "num_input_tokens_seen": 68113305, + "step": 2088 + }, + { + "epoch": 0.1883933805293773, + "flos": 22350680634240.0, + "grad_norm": 1.7361061460635925, + "learning_rate": 3.7426140051056867e-06, + "loss": 0.7293, + "num_input_tokens_seen": 68140515, + "step": 2089 + }, + { + "epoch": 0.18848356405284755, + "flos": 17687632227360.0, + "grad_norm": 2.1652385944928896, + "learning_rate": 3.7423272362176856e-06, + "loss": 0.7236, + "num_input_tokens_seen": 68168615, + "step": 2090 + }, + { + "epoch": 0.1885737475763178, + "flos": 21984352483200.0, + "grad_norm": 1.8062876423540013, + "learning_rate": 3.742040318664718e-06, + "loss": 0.8269, + "num_input_tokens_seen": 68198435, + "step": 2091 + }, + { + "epoch": 0.18866393109978807, + "flos": 21911859984000.0, + "grad_norm": 1.813208605927183, + "learning_rate": 3.7417532524712643e-06, + "loss": 0.84, + "num_input_tokens_seen": 68226070, + "step": 2092 + }, + { + "epoch": 0.18875411462325833, + "flos": 22532580937920.0, + "grad_norm": 1.841406674682413, + "learning_rate": 3.7414660376618195e-06, + "loss": 0.8109, + "num_input_tokens_seen": 68253255, + "step": 2093 + }, + { + "epoch": 0.1888442981467286, + "flos": 70312302975360.0, + "grad_norm": 0.6113262237421763, + "learning_rate": 3.74117867426089e-06, + "loss": 0.5901, + "num_input_tokens_seen": 68355435, + "step": 2094 + }, + { + "epoch": 0.18893448167019886, + "flos": 25447966544640.0, + "grad_norm": 1.8333670945382354, + "learning_rate": 3.7408911622929954e-06, + "loss": 0.7589, + "num_input_tokens_seen": 68385225, + "step": 2095 + }, + { + "epoch": 0.18902466519366912, + "flos": 25483952605920.0, + "grad_norm": 2.6885206369264574, + "learning_rate": 3.740603501782668e-06, + "loss": 0.8159, + "num_input_tokens_seen": 68415700, + "step": 2096 + }, + { + "epoch": 0.18911484871713938, + "flos": 24609545074560.0, + "grad_norm": 2.124241637499101, + "learning_rate": 3.7403156927544516e-06, + "loss": 0.7219, + "num_input_tokens_seen": 68446280, + "step": 2097 + }, + { + "epoch": 0.18920503224060964, + "flos": 25044648748800.0, + "grad_norm": 2.9907972328074255, + "learning_rate": 3.740027735232904e-06, + "loss": 0.7168, + "num_input_tokens_seen": 68477125, + "step": 2098 + }, + { + "epoch": 0.1892952157640799, + "flos": 18707879661600.0, + "grad_norm": 1.9303194316327643, + "learning_rate": 3.7397396292425966e-06, + "loss": 0.7666, + "num_input_tokens_seen": 68506295, + "step": 2099 + }, + { + "epoch": 0.18938539928755016, + "flos": 21543636175200.0, + "grad_norm": 1.883136642222558, + "learning_rate": 3.7394513748081105e-06, + "loss": 0.7571, + "num_input_tokens_seen": 68533755, + "step": 2100 + }, + { + "epoch": 0.18947558281102042, + "flos": 22417746348480.0, + "grad_norm": 1.726097009662124, + "learning_rate": 3.7391629719540418e-06, + "loss": 0.8138, + "num_input_tokens_seen": 68561705, + "step": 2101 + }, + { + "epoch": 0.18956576633449068, + "flos": 21621927156960.0, + "grad_norm": 6.755592290181677, + "learning_rate": 3.7388744207049998e-06, + "loss": 0.7721, + "num_input_tokens_seen": 68590100, + "step": 2102 + }, + { + "epoch": 0.18965594985796094, + "flos": 17250521386080.0, + "grad_norm": 3.130798301172484, + "learning_rate": 3.7385857210856027e-06, + "loss": 0.7533, + "num_input_tokens_seen": 68615490, + "step": 2103 + }, + { + "epoch": 0.1897461333814312, + "flos": 28581015497760.0, + "grad_norm": 2.018048035071685, + "learning_rate": 3.738296873120486e-06, + "loss": 0.7659, + "num_input_tokens_seen": 68644280, + "step": 2104 + }, + { + "epoch": 0.18983631690490146, + "flos": 26721937726080.0, + "grad_norm": 2.0843648518511158, + "learning_rate": 3.7380078768342955e-06, + "loss": 0.7495, + "num_input_tokens_seen": 68672495, + "step": 2105 + }, + { + "epoch": 0.18992650042837173, + "flos": 25802697730560.0, + "grad_norm": 1.8152879779700648, + "learning_rate": 3.7377187322516895e-06, + "loss": 0.7982, + "num_input_tokens_seen": 68699315, + "step": 2106 + }, + { + "epoch": 0.19001668395184199, + "flos": 16267300766400.0, + "grad_norm": 1.8193728923883683, + "learning_rate": 3.7374294393973395e-06, + "loss": 0.731, + "num_input_tokens_seen": 68727325, + "step": 2107 + }, + { + "epoch": 0.19010686747531225, + "flos": 28761131652960.0, + "grad_norm": 1.6904660754857412, + "learning_rate": 3.7371399982959294e-06, + "loss": 0.8344, + "num_input_tokens_seen": 68756015, + "step": 2108 + }, + { + "epoch": 0.19019705099878254, + "flos": 41213363844480.0, + "grad_norm": 1.9429535561869804, + "learning_rate": 3.7368504089721565e-06, + "loss": 0.6975, + "num_input_tokens_seen": 68785010, + "step": 2109 + }, + { + "epoch": 0.1902872345222528, + "flos": 28690311792960.0, + "grad_norm": 1.6736705087099624, + "learning_rate": 3.73656067145073e-06, + "loss": 0.7337, + "num_input_tokens_seen": 68815970, + "step": 2110 + }, + { + "epoch": 0.19037741804572306, + "flos": 22861176048960.0, + "grad_norm": 1.9465999402842507, + "learning_rate": 3.736270785756371e-06, + "loss": 0.8288, + "num_input_tokens_seen": 68846650, + "step": 2111 + }, + { + "epoch": 0.19046760156919332, + "flos": 23216390441760.0, + "grad_norm": 3.083249391667541, + "learning_rate": 3.7359807519138156e-06, + "loss": 0.7852, + "num_input_tokens_seen": 68871895, + "step": 2112 + }, + { + "epoch": 0.19055778509266358, + "flos": 25265583034080.0, + "grad_norm": 1.6714804353108148, + "learning_rate": 3.73569056994781e-06, + "loss": 0.7792, + "num_input_tokens_seen": 68900420, + "step": 2113 + }, + { + "epoch": 0.19064796861613384, + "flos": 20492606446080.0, + "grad_norm": 2.5401455960634305, + "learning_rate": 3.7354002398831144e-06, + "loss": 0.82, + "num_input_tokens_seen": 68926745, + "step": 2114 + }, + { + "epoch": 0.1907381521396041, + "flos": 20274088195200.0, + "grad_norm": 4.02238803250119, + "learning_rate": 3.7351097617445015e-06, + "loss": 0.7776, + "num_input_tokens_seen": 68955430, + "step": 2115 + }, + { + "epoch": 0.19082833566307436, + "flos": 63737464343520.0, + "grad_norm": 0.690237641772067, + "learning_rate": 3.7348191355567567e-06, + "loss": 0.6111, + "num_input_tokens_seen": 69036905, + "step": 2116 + }, + { + "epoch": 0.19091851918654462, + "flos": 22751842584000.0, + "grad_norm": 2.112982483790777, + "learning_rate": 3.734528361344677e-06, + "loss": 0.8088, + "num_input_tokens_seen": 69065455, + "step": 2117 + }, + { + "epoch": 0.19100870271001488, + "flos": 20164531711680.0, + "grad_norm": 1.9229363092953244, + "learning_rate": 3.734237439133074e-06, + "loss": 0.7771, + "num_input_tokens_seen": 69091640, + "step": 2118 + }, + { + "epoch": 0.19109888623348514, + "flos": 25006432501920.0, + "grad_norm": 1.739994876020933, + "learning_rate": 3.7339463689467702e-06, + "loss": 0.8268, + "num_input_tokens_seen": 69118820, + "step": 2119 + }, + { + "epoch": 0.1911890697569554, + "flos": 15898891108800.0, + "grad_norm": 1.6513064760748186, + "learning_rate": 3.733655150810601e-06, + "loss": 0.7508, + "num_input_tokens_seen": 69145075, + "step": 2120 + }, + { + "epoch": 0.19127925328042567, + "flos": 24020461320000.0, + "grad_norm": 2.0081262009465997, + "learning_rate": 3.7333637847494154e-06, + "loss": 0.9123, + "num_input_tokens_seen": 69173290, + "step": 2121 + }, + { + "epoch": 0.19136943680389593, + "flos": 29015450116320.0, + "grad_norm": 1.5812051005265035, + "learning_rate": 3.7330722707880734e-06, + "loss": 0.9353, + "num_input_tokens_seen": 69202675, + "step": 2122 + }, + { + "epoch": 0.1914596203273662, + "flos": 23079062413920.0, + "grad_norm": 1.800435395034861, + "learning_rate": 3.7327806089514497e-06, + "loss": 0.6881, + "num_input_tokens_seen": 69231465, + "step": 2123 + }, + { + "epoch": 0.19154980385083645, + "flos": 16814376958560.0, + "grad_norm": 1.5874305800030235, + "learning_rate": 3.7324887992644297e-06, + "loss": 0.8767, + "num_input_tokens_seen": 69258005, + "step": 2124 + }, + { + "epoch": 0.1916399873743067, + "flos": 65553679648320.0, + "grad_norm": 0.7232654787207793, + "learning_rate": 3.7321968417519123e-06, + "loss": 0.5851, + "num_input_tokens_seen": 69350730, + "step": 2125 + }, + { + "epoch": 0.19173017089777697, + "flos": 23990162232000.0, + "grad_norm": 1.8489707480716755, + "learning_rate": 3.7319047364388097e-06, + "loss": 0.7706, + "num_input_tokens_seen": 69379525, + "step": 2126 + }, + { + "epoch": 0.19182035442124723, + "flos": 52182765007680.0, + "grad_norm": 2.874159814552316, + "learning_rate": 3.7316124833500453e-06, + "loss": 0.5711, + "num_input_tokens_seen": 69467920, + "step": 2127 + }, + { + "epoch": 0.1919105379447175, + "flos": 22424474075040.0, + "grad_norm": 1.611783124088371, + "learning_rate": 3.731320082510556e-06, + "loss": 0.766, + "num_input_tokens_seen": 69499000, + "step": 2128 + }, + { + "epoch": 0.19200072146818775, + "flos": 23080920901920.0, + "grad_norm": 2.0858254849078843, + "learning_rate": 3.7310275339452906e-06, + "loss": 0.7917, + "num_input_tokens_seen": 69529175, + "step": 2129 + }, + { + "epoch": 0.192090904991658, + "flos": 22708125212640.0, + "grad_norm": 2.074697780431394, + "learning_rate": 3.7307348376792113e-06, + "loss": 0.767, + "num_input_tokens_seen": 69558385, + "step": 2130 + }, + { + "epoch": 0.19218108851512827, + "flos": 23043002013120.0, + "grad_norm": 1.6511341474205905, + "learning_rate": 3.730441993737292e-06, + "loss": 0.7735, + "num_input_tokens_seen": 69585870, + "step": 2131 + }, + { + "epoch": 0.19227127203859856, + "flos": 22747493722080.0, + "grad_norm": 2.681043187415008, + "learning_rate": 3.7301490021445205e-06, + "loss": 0.7649, + "num_input_tokens_seen": 69612405, + "step": 2132 + }, + { + "epoch": 0.19236145556206882, + "flos": 22933185341280.0, + "grad_norm": 2.6286893077742453, + "learning_rate": 3.7298558629258966e-06, + "loss": 0.7176, + "num_input_tokens_seen": 69638520, + "step": 2133 + }, + { + "epoch": 0.19245163908553908, + "flos": 22132571250720.0, + "grad_norm": 2.0134451263960624, + "learning_rate": 3.7295625761064314e-06, + "loss": 0.8129, + "num_input_tokens_seen": 69665665, + "step": 2134 + }, + { + "epoch": 0.19254182260900934, + "flos": 28580680969920.0, + "grad_norm": 2.061064017969707, + "learning_rate": 3.7292691417111504e-06, + "loss": 0.7442, + "num_input_tokens_seen": 69696275, + "step": 2135 + }, + { + "epoch": 0.1926320061324796, + "flos": 18560701647360.0, + "grad_norm": 2.3936910477592557, + "learning_rate": 3.728975559765092e-06, + "loss": 0.7349, + "num_input_tokens_seen": 69722995, + "step": 2136 + }, + { + "epoch": 0.19272218965594987, + "flos": 23298063871680.0, + "grad_norm": 2.84090725846665, + "learning_rate": 3.728681830293305e-06, + "loss": 0.8077, + "num_input_tokens_seen": 69751170, + "step": 2137 + }, + { + "epoch": 0.19281237317942013, + "flos": 31021185525600.0, + "grad_norm": 1.789035646995015, + "learning_rate": 3.7283879533208523e-06, + "loss": 0.7345, + "num_input_tokens_seen": 69782060, + "step": 2138 + }, + { + "epoch": 0.1929025567028904, + "flos": 25338372891360.0, + "grad_norm": 1.603021840723116, + "learning_rate": 3.7280939288728094e-06, + "loss": 0.7673, + "num_input_tokens_seen": 69812015, + "step": 2139 + }, + { + "epoch": 0.19299274022636065, + "flos": 20966260895040.0, + "grad_norm": 2.0056609860737558, + "learning_rate": 3.7277997569742637e-06, + "loss": 0.7389, + "num_input_tokens_seen": 69841400, + "step": 2140 + }, + { + "epoch": 0.1930829237498309, + "flos": 19435332197280.0, + "grad_norm": 2.32714536396933, + "learning_rate": 3.7275054376503155e-06, + "loss": 0.7427, + "num_input_tokens_seen": 69868315, + "step": 2141 + }, + { + "epoch": 0.19317310727330117, + "flos": 22824037725120.0, + "grad_norm": 7.850443746088373, + "learning_rate": 3.7272109709260783e-06, + "loss": 0.7173, + "num_input_tokens_seen": 69894875, + "step": 2142 + }, + { + "epoch": 0.19326329079677143, + "flos": 65482376581440.0, + "grad_norm": 0.7735593295236183, + "learning_rate": 3.7269163568266774e-06, + "loss": 0.5869, + "num_input_tokens_seen": 69982140, + "step": 2143 + }, + { + "epoch": 0.1933534743202417, + "flos": 21877026185280.0, + "grad_norm": 1.666898147084733, + "learning_rate": 3.7266215953772512e-06, + "loss": 0.763, + "num_input_tokens_seen": 70012060, + "step": 2144 + }, + { + "epoch": 0.19344365784371195, + "flos": 26358471646560.0, + "grad_norm": 1.9279268085627121, + "learning_rate": 3.7263266866029492e-06, + "loss": 0.8111, + "num_input_tokens_seen": 70041300, + "step": 2145 + }, + { + "epoch": 0.1935338413671822, + "flos": 34408106904960.0, + "grad_norm": 1.6362115989241297, + "learning_rate": 3.726031630528936e-06, + "loss": 0.7447, + "num_input_tokens_seen": 70071265, + "step": 2146 + }, + { + "epoch": 0.19362402489065247, + "flos": 26649519566400.0, + "grad_norm": 1.4606380869810038, + "learning_rate": 3.7257364271803865e-06, + "loss": 0.6706, + "num_input_tokens_seen": 70101305, + "step": 2147 + }, + { + "epoch": 0.19371420841412273, + "flos": 21731186282400.0, + "grad_norm": 3.375683609288212, + "learning_rate": 3.7254410765824896e-06, + "loss": 0.6981, + "num_input_tokens_seen": 70127355, + "step": 2148 + }, + { + "epoch": 0.193804391937593, + "flos": 22970249325600.0, + "grad_norm": 1.8420957482601323, + "learning_rate": 3.725145578760446e-06, + "loss": 0.6936, + "num_input_tokens_seen": 70154760, + "step": 2149 + }, + { + "epoch": 0.19389457546106326, + "flos": 24864755612160.0, + "grad_norm": 1.8982572744715687, + "learning_rate": 3.7248499337394696e-06, + "loss": 0.6916, + "num_input_tokens_seen": 70183715, + "step": 2150 + }, + { + "epoch": 0.19398475898453352, + "flos": 19327485522720.0, + "grad_norm": 2.398190762132902, + "learning_rate": 3.7245541415447848e-06, + "loss": 0.8662, + "num_input_tokens_seen": 70208465, + "step": 2151 + }, + { + "epoch": 0.19407494250800378, + "flos": 18962309634240.0, + "grad_norm": 2.215256425468305, + "learning_rate": 3.724258202201633e-06, + "loss": 0.7968, + "num_input_tokens_seen": 70236095, + "step": 2152 + }, + { + "epoch": 0.19416512603147404, + "flos": 24166449901920.0, + "grad_norm": 1.7380293417144954, + "learning_rate": 3.7239621157352633e-06, + "loss": 0.7908, + "num_input_tokens_seen": 70264240, + "step": 2153 + }, + { + "epoch": 0.1942553095549443, + "flos": 29052365421600.0, + "grad_norm": 2.6313211535621406, + "learning_rate": 3.7236658821709403e-06, + "loss": 0.7637, + "num_input_tokens_seen": 70293985, + "step": 2154 + }, + { + "epoch": 0.19434549307841456, + "flos": 19361204228640.0, + "grad_norm": 1.5660133534979004, + "learning_rate": 3.7233695015339404e-06, + "loss": 0.8137, + "num_input_tokens_seen": 70321025, + "step": 2155 + }, + { + "epoch": 0.19443567660188485, + "flos": 26209063446720.0, + "grad_norm": 3.3907278018959723, + "learning_rate": 3.7230729738495513e-06, + "loss": 0.7684, + "num_input_tokens_seen": 70348205, + "step": 2156 + }, + { + "epoch": 0.1945258601253551, + "flos": 26028872952000.0, + "grad_norm": 1.5850510398276274, + "learning_rate": 3.722776299143075e-06, + "loss": 0.769, + "num_input_tokens_seen": 70378530, + "step": 2157 + }, + { + "epoch": 0.19461604364882537, + "flos": 68795058482880.0, + "grad_norm": 0.6681177073581178, + "learning_rate": 3.722479477439826e-06, + "loss": 0.586, + "num_input_tokens_seen": 70468185, + "step": 2158 + }, + { + "epoch": 0.19470622717229563, + "flos": 22459791080640.0, + "grad_norm": 1.4100439391493182, + "learning_rate": 3.7221825087651306e-06, + "loss": 0.7545, + "num_input_tokens_seen": 70497460, + "step": 2159 + }, + { + "epoch": 0.1947964106957659, + "flos": 25957495545600.0, + "grad_norm": 1.8929153944959094, + "learning_rate": 3.7218853931443274e-06, + "loss": 0.7123, + "num_input_tokens_seen": 70527010, + "step": 2160 + }, + { + "epoch": 0.19488659421923615, + "flos": 24354557555520.0, + "grad_norm": 1.6888289693330971, + "learning_rate": 3.721588130602768e-06, + "loss": 0.7137, + "num_input_tokens_seen": 70556970, + "step": 2161 + }, + { + "epoch": 0.19497677774270641, + "flos": 13024360801920.0, + "grad_norm": 2.0683245633431455, + "learning_rate": 3.7212907211658164e-06, + "loss": 0.7496, + "num_input_tokens_seen": 70584560, + "step": 2162 + }, + { + "epoch": 0.19506696126617667, + "flos": 44206557226080.0, + "grad_norm": 1.6135849932309885, + "learning_rate": 3.72099316485885e-06, + "loss": 0.743, + "num_input_tokens_seen": 70616805, + "step": 2163 + }, + { + "epoch": 0.19515714478964694, + "flos": 30141834416160.0, + "grad_norm": 1.9977268175200944, + "learning_rate": 3.720695461707256e-06, + "loss": 0.8058, + "num_input_tokens_seen": 70647695, + "step": 2164 + }, + { + "epoch": 0.1952473283131172, + "flos": 21221434262880.0, + "grad_norm": 1.6610913324755625, + "learning_rate": 3.7203976117364383e-06, + "loss": 0.7898, + "num_input_tokens_seen": 70676265, + "step": 2165 + }, + { + "epoch": 0.19533751183658746, + "flos": 23480001345120.0, + "grad_norm": 1.9284015648831354, + "learning_rate": 3.7200996149718105e-06, + "loss": 0.7585, + "num_input_tokens_seen": 70703605, + "step": 2166 + }, + { + "epoch": 0.19542769536005772, + "flos": 21840891444960.0, + "grad_norm": 1.4153852651348984, + "learning_rate": 3.7198014714387985e-06, + "loss": 0.7624, + "num_input_tokens_seen": 70734655, + "step": 2167 + }, + { + "epoch": 0.19551787888352798, + "flos": 28144462202880.0, + "grad_norm": 1.916910660458341, + "learning_rate": 3.7195031811628422e-06, + "loss": 0.7078, + "num_input_tokens_seen": 70764355, + "step": 2168 + }, + { + "epoch": 0.19560806240699824, + "flos": 22675149901920.0, + "grad_norm": 2.5872402079171417, + "learning_rate": 3.719204744169393e-06, + "loss": 0.808, + "num_input_tokens_seen": 70790265, + "step": 2169 + }, + { + "epoch": 0.1956982459304685, + "flos": 24063175107840.0, + "grad_norm": 2.0294367987055386, + "learning_rate": 3.718906160483916e-06, + "loss": 0.7482, + "num_input_tokens_seen": 70817650, + "step": 2170 + }, + { + "epoch": 0.19578842945393876, + "flos": 22676636692320.0, + "grad_norm": 1.6486944206303962, + "learning_rate": 3.7186074301318868e-06, + "loss": 0.7474, + "num_input_tokens_seen": 70846580, + "step": 2171 + }, + { + "epoch": 0.19587861297740902, + "flos": 21980932865280.0, + "grad_norm": 1.6982945526806577, + "learning_rate": 3.7183085531387957e-06, + "loss": 0.8487, + "num_input_tokens_seen": 70874225, + "step": 2172 + }, + { + "epoch": 0.19596879650087928, + "flos": 24063918503040.0, + "grad_norm": 2.111618092657449, + "learning_rate": 3.7180095295301443e-06, + "loss": 0.7446, + "num_input_tokens_seen": 70903750, + "step": 2173 + }, + { + "epoch": 0.19605898002434954, + "flos": 21146451389760.0, + "grad_norm": 1.3932991034046338, + "learning_rate": 3.7177103593314465e-06, + "loss": 0.8362, + "num_input_tokens_seen": 70934720, + "step": 2174 + }, + { + "epoch": 0.1961491635478198, + "flos": 18597802801440.0, + "grad_norm": 1.8623119848195389, + "learning_rate": 3.7174110425682297e-06, + "loss": 0.8119, + "num_input_tokens_seen": 70963300, + "step": 2175 + }, + { + "epoch": 0.19623934707129007, + "flos": 21913383944160.0, + "grad_norm": 1.7447706154839173, + "learning_rate": 3.7171115792660333e-06, + "loss": 0.8977, + "num_input_tokens_seen": 70989975, + "step": 2176 + }, + { + "epoch": 0.19632953059476033, + "flos": 18156751965600.0, + "grad_norm": 2.3354139474880733, + "learning_rate": 3.7168119694504083e-06, + "loss": 0.8395, + "num_input_tokens_seen": 71016630, + "step": 2177 + }, + { + "epoch": 0.1964197141182306, + "flos": 19979880845760.0, + "grad_norm": 1.8194002804560852, + "learning_rate": 3.71651221314692e-06, + "loss": 0.7661, + "num_input_tokens_seen": 71045885, + "step": 2178 + }, + { + "epoch": 0.19650989764170085, + "flos": 23116200737760.0, + "grad_norm": 2.531670523668389, + "learning_rate": 3.716212310381145e-06, + "loss": 0.7711, + "num_input_tokens_seen": 71074790, + "step": 2179 + }, + { + "epoch": 0.19660008116517114, + "flos": 23808782304960.0, + "grad_norm": 2.030867085677923, + "learning_rate": 3.7159122611786725e-06, + "loss": 0.8176, + "num_input_tokens_seen": 71102115, + "step": 2180 + }, + { + "epoch": 0.1966902646886414, + "flos": 26321965208640.0, + "grad_norm": 2.05380554997085, + "learning_rate": 3.7156120655651045e-06, + "loss": 0.7573, + "num_input_tokens_seen": 71131885, + "step": 2181 + }, + { + "epoch": 0.19678044821211166, + "flos": 27195703684320.0, + "grad_norm": 1.718215096207849, + "learning_rate": 3.7153117235660553e-06, + "loss": 0.756, + "num_input_tokens_seen": 71162975, + "step": 2182 + }, + { + "epoch": 0.19687063173558192, + "flos": 68602936495200.0, + "grad_norm": 0.6937140189110509, + "learning_rate": 3.7150112352071514e-06, + "loss": 0.5848, + "num_input_tokens_seen": 71257295, + "step": 2183 + }, + { + "epoch": 0.19696081525905218, + "flos": 22707939363840.0, + "grad_norm": 1.9594533706883293, + "learning_rate": 3.7147106005140326e-06, + "loss": 0.7429, + "num_input_tokens_seen": 71282405, + "step": 2184 + }, + { + "epoch": 0.19705099878252244, + "flos": 67665254565120.0, + "grad_norm": 0.6843424375617193, + "learning_rate": 3.714409819512351e-06, + "loss": 0.5937, + "num_input_tokens_seen": 71374380, + "step": 2185 + }, + { + "epoch": 0.1971411823059927, + "flos": 20668373739360.0, + "grad_norm": 2.209417352985364, + "learning_rate": 3.7141088922277695e-06, + "loss": 0.7975, + "num_input_tokens_seen": 71399425, + "step": 2186 + }, + { + "epoch": 0.19723136582946296, + "flos": 23916777658560.0, + "grad_norm": 2.042240866538031, + "learning_rate": 3.7138078186859664e-06, + "loss": 0.8177, + "num_input_tokens_seen": 71428220, + "step": 2187 + }, + { + "epoch": 0.19732154935293322, + "flos": 21106859861760.0, + "grad_norm": 1.713712071194506, + "learning_rate": 3.7135065989126303e-06, + "loss": 0.7876, + "num_input_tokens_seen": 71456060, + "step": 2188 + }, + { + "epoch": 0.19741173287640348, + "flos": 64680424379520.0, + "grad_norm": 0.735235562236923, + "learning_rate": 3.713205232933463e-06, + "loss": 0.649, + "num_input_tokens_seen": 71541775, + "step": 2189 + }, + { + "epoch": 0.19750191639987374, + "flos": 17724213004800.0, + "grad_norm": 1.7287262351829213, + "learning_rate": 3.7129037207741792e-06, + "loss": 0.7836, + "num_input_tokens_seen": 71568035, + "step": 2190 + }, + { + "epoch": 0.197592099923344, + "flos": 20851128947520.0, + "grad_norm": 2.8455970550646277, + "learning_rate": 3.7126020624605046e-06, + "loss": 0.7866, + "num_input_tokens_seen": 71595480, + "step": 2191 + }, + { + "epoch": 0.19768228344681427, + "flos": 31750459379520.0, + "grad_norm": 1.6248581164114047, + "learning_rate": 3.7123002580181785e-06, + "loss": 0.7257, + "num_input_tokens_seen": 71628200, + "step": 2192 + }, + { + "epoch": 0.19777246697028453, + "flos": 23332897670400.0, + "grad_norm": 1.3672592580219929, + "learning_rate": 3.7119983074729532e-06, + "loss": 0.8478, + "num_input_tokens_seen": 71656245, + "step": 2193 + }, + { + "epoch": 0.1978626504937548, + "flos": 20963733351360.0, + "grad_norm": 2.3916180112809706, + "learning_rate": 3.7116962108505926e-06, + "loss": 0.8289, + "num_input_tokens_seen": 71683470, + "step": 2194 + }, + { + "epoch": 0.19795283401722505, + "flos": 22641951572640.0, + "grad_norm": 1.6076218769152006, + "learning_rate": 3.711393968176873e-06, + "loss": 0.8722, + "num_input_tokens_seen": 71713115, + "step": 2195 + }, + { + "epoch": 0.1980430175406953, + "flos": 20961949202880.0, + "grad_norm": 1.4748346624004491, + "learning_rate": 3.711091579477584e-06, + "loss": 0.7845, + "num_input_tokens_seen": 71742395, + "step": 2196 + }, + { + "epoch": 0.19813320106416557, + "flos": 17613206900640.0, + "grad_norm": 2.1182617615204746, + "learning_rate": 3.7107890447785255e-06, + "loss": 0.8041, + "num_input_tokens_seen": 71768260, + "step": 2197 + }, + { + "epoch": 0.19822338458763583, + "flos": 22018405716960.0, + "grad_norm": 2.971889562400655, + "learning_rate": 3.710486364105513e-06, + "loss": 0.8299, + "num_input_tokens_seen": 71797475, + "step": 2198 + }, + { + "epoch": 0.1983135681111061, + "flos": 18124074012960.0, + "grad_norm": 2.0720350476986633, + "learning_rate": 3.7101835374843728e-06, + "loss": 0.7729, + "num_input_tokens_seen": 71823320, + "step": 2199 + }, + { + "epoch": 0.19840375163457635, + "flos": 26066754671040.0, + "grad_norm": 1.9599237139122199, + "learning_rate": 3.7098805649409427e-06, + "loss": 0.7965, + "num_input_tokens_seen": 71853625, + "step": 2200 + }, + { + "epoch": 0.1984939351580466, + "flos": 19909098155520.0, + "grad_norm": 2.381952749768586, + "learning_rate": 3.7095774465010748e-06, + "loss": 0.8437, + "num_input_tokens_seen": 71878475, + "step": 2201 + }, + { + "epoch": 0.19858411868151687, + "flos": 25883553425760.0, + "grad_norm": 1.9733487264037104, + "learning_rate": 3.7092741821906328e-06, + "loss": 0.7711, + "num_input_tokens_seen": 71908090, + "step": 2202 + }, + { + "epoch": 0.19867430220498714, + "flos": 22819279995840.0, + "grad_norm": 2.0082180934571503, + "learning_rate": 3.708970772035493e-06, + "loss": 0.7583, + "num_input_tokens_seen": 71933445, + "step": 2203 + }, + { + "epoch": 0.19876448572845742, + "flos": 24203625395520.0, + "grad_norm": 1.7461831840124007, + "learning_rate": 3.7086672160615427e-06, + "loss": 0.7441, + "num_input_tokens_seen": 71963165, + "step": 2204 + }, + { + "epoch": 0.19885466925192768, + "flos": 23479592477760.0, + "grad_norm": 1.7488400838688862, + "learning_rate": 3.7083635142946852e-06, + "loss": 0.7614, + "num_input_tokens_seen": 71992855, + "step": 2205 + }, + { + "epoch": 0.19894485277539795, + "flos": 27450839882400.0, + "grad_norm": 1.7325125652259412, + "learning_rate": 3.7080596667608327e-06, + "loss": 0.8153, + "num_input_tokens_seen": 72019340, + "step": 2206 + }, + { + "epoch": 0.1990350362988682, + "flos": 23007090291360.0, + "grad_norm": 1.8583010609423236, + "learning_rate": 3.707755673485911e-06, + "loss": 0.8215, + "num_input_tokens_seen": 72049205, + "step": 2207 + }, + { + "epoch": 0.19912521982233847, + "flos": 17504170793760.0, + "grad_norm": 2.0631029808166015, + "learning_rate": 3.7074515344958584e-06, + "loss": 0.6887, + "num_input_tokens_seen": 72073915, + "step": 2208 + }, + { + "epoch": 0.19921540334580873, + "flos": 67508481019200.0, + "grad_norm": 0.6657422083428448, + "learning_rate": 3.707147249816627e-06, + "loss": 0.6133, + "num_input_tokens_seen": 72167200, + "step": 2209 + }, + { + "epoch": 0.199305586869279, + "flos": 25520459043840.0, + "grad_norm": 1.659054379280753, + "learning_rate": 3.706842819474178e-06, + "loss": 0.7999, + "num_input_tokens_seen": 72197430, + "step": 2210 + }, + { + "epoch": 0.19939577039274925, + "flos": 13533666784320.0, + "grad_norm": 1.8121503295349137, + "learning_rate": 3.706538243494489e-06, + "loss": 0.7133, + "num_input_tokens_seen": 72225420, + "step": 2211 + }, + { + "epoch": 0.1994859539162195, + "flos": 35028790689120.0, + "grad_norm": 2.125433709258798, + "learning_rate": 3.706233521903547e-06, + "loss": 0.75, + "num_input_tokens_seen": 72252415, + "step": 2212 + }, + { + "epoch": 0.19957613743968977, + "flos": 16595152482240.0, + "grad_norm": 1.9001534952100283, + "learning_rate": 3.705928654727353e-06, + "loss": 0.8589, + "num_input_tokens_seen": 72278785, + "step": 2213 + }, + { + "epoch": 0.19966632096316003, + "flos": 22606337208960.0, + "grad_norm": 1.5033333616122788, + "learning_rate": 3.7056236419919195e-06, + "loss": 0.7532, + "num_input_tokens_seen": 72309810, + "step": 2214 + }, + { + "epoch": 0.1997565044866303, + "flos": 28795742433120.0, + "grad_norm": 2.0150985621728275, + "learning_rate": 3.705318483723273e-06, + "loss": 0.829, + "num_input_tokens_seen": 72341445, + "step": 2215 + }, + { + "epoch": 0.19984668801010055, + "flos": 23147577748800.0, + "grad_norm": 1.9099631494692044, + "learning_rate": 3.7050131799474493e-06, + "loss": 0.7527, + "num_input_tokens_seen": 72366075, + "step": 2216 + }, + { + "epoch": 0.19993687153357081, + "flos": 22895192112960.0, + "grad_norm": 1.537367841421511, + "learning_rate": 3.7047077306905e-06, + "loss": 0.7778, + "num_input_tokens_seen": 72395585, + "step": 2217 + }, + { + "epoch": 0.20002705505704108, + "flos": 26423790382080.0, + "grad_norm": 1.6358677511275406, + "learning_rate": 3.704402135978488e-06, + "loss": 0.8119, + "num_input_tokens_seen": 72425340, + "step": 2218 + }, + { + "epoch": 0.20011723858051134, + "flos": 17869309512480.0, + "grad_norm": 3.372346187898739, + "learning_rate": 3.7040963958374877e-06, + "loss": 0.7987, + "num_input_tokens_seen": 72448705, + "step": 2219 + }, + { + "epoch": 0.2002074221039816, + "flos": 46689292362720.0, + "grad_norm": 9.478947653004443, + "learning_rate": 3.7037905102935864e-06, + "loss": 0.8085, + "num_input_tokens_seen": 72483840, + "step": 2220 + }, + { + "epoch": 0.20029760562745186, + "flos": 20598445953600.0, + "grad_norm": 1.6352877312958067, + "learning_rate": 3.7034844793728837e-06, + "loss": 0.7856, + "num_input_tokens_seen": 72509365, + "step": 2221 + }, + { + "epoch": 0.20038778915092212, + "flos": 32150246048160.0, + "grad_norm": 1.5705582156025932, + "learning_rate": 3.7031783031014933e-06, + "loss": 0.6674, + "num_input_tokens_seen": 72542380, + "step": 2222 + }, + { + "epoch": 0.20047797267439238, + "flos": 21185634050400.0, + "grad_norm": 1.925137549601234, + "learning_rate": 3.702871981505538e-06, + "loss": 0.7536, + "num_input_tokens_seen": 72570270, + "step": 2223 + }, + { + "epoch": 0.20056815619786264, + "flos": 23735509240800.0, + "grad_norm": 2.114425152153667, + "learning_rate": 3.7025655146111563e-06, + "loss": 0.7191, + "num_input_tokens_seen": 72601020, + "step": 2224 + }, + { + "epoch": 0.2006583397213329, + "flos": 71027192132160.0, + "grad_norm": 0.6923068156858339, + "learning_rate": 3.702258902444497e-06, + "loss": 0.6642, + "num_input_tokens_seen": 72701640, + "step": 2225 + }, + { + "epoch": 0.20074852324480316, + "flos": 25192756007040.0, + "grad_norm": 1.711778547251461, + "learning_rate": 3.701952145031722e-06, + "loss": 0.8317, + "num_input_tokens_seen": 72732765, + "step": 2226 + }, + { + "epoch": 0.20083870676827342, + "flos": 24354929253120.0, + "grad_norm": 1.5466207430176746, + "learning_rate": 3.701645242399005e-06, + "loss": 0.8604, + "num_input_tokens_seen": 72762785, + "step": 2227 + }, + { + "epoch": 0.2009288902917437, + "flos": 19400015191680.0, + "grad_norm": 10.86723032544581, + "learning_rate": 3.701338194572533e-06, + "loss": 0.8429, + "num_input_tokens_seen": 72790735, + "step": 2228 + }, + { + "epoch": 0.20101907381521397, + "flos": 15792605564160.0, + "grad_norm": 2.074116907295583, + "learning_rate": 3.7010310015785056e-06, + "loss": 0.7586, + "num_input_tokens_seen": 72816545, + "step": 2229 + }, + { + "epoch": 0.20110925733868423, + "flos": 23954250510240.0, + "grad_norm": 1.6460801032124823, + "learning_rate": 3.700723663443134e-06, + "loss": 0.72, + "num_input_tokens_seen": 72845970, + "step": 2230 + }, + { + "epoch": 0.2011994408621545, + "flos": 55317560939520.0, + "grad_norm": 0.7142487848476479, + "learning_rate": 3.7004161801926416e-06, + "loss": 0.5907, + "num_input_tokens_seen": 72931035, + "step": 2231 + }, + { + "epoch": 0.20128962438562475, + "flos": 19394365388160.0, + "grad_norm": 1.6911318514564795, + "learning_rate": 3.7001085518532643e-06, + "loss": 0.8017, + "num_input_tokens_seen": 72957565, + "step": 2232 + }, + { + "epoch": 0.20137980790909502, + "flos": 22314620233440.0, + "grad_norm": 2.7200076822565085, + "learning_rate": 3.6998007784512515e-06, + "loss": 0.7723, + "num_input_tokens_seen": 72986075, + "step": 2233 + }, + { + "epoch": 0.20146999143256528, + "flos": 37356579331680.0, + "grad_norm": 1.623292539437781, + "learning_rate": 3.6994928600128637e-06, + "loss": 0.6545, + "num_input_tokens_seen": 73019705, + "step": 2234 + }, + { + "epoch": 0.20156017495603554, + "flos": 23804173254720.0, + "grad_norm": 1.566627619980469, + "learning_rate": 3.6991847965643742e-06, + "loss": 0.7433, + "num_input_tokens_seen": 73047230, + "step": 2235 + }, + { + "epoch": 0.2016503584795058, + "flos": 22678829708160.0, + "grad_norm": 2.2565714979553655, + "learning_rate": 3.698876588132068e-06, + "loss": 0.7348, + "num_input_tokens_seen": 73074905, + "step": 2236 + }, + { + "epoch": 0.20174054200297606, + "flos": 25738196729760.0, + "grad_norm": 1.9584928677555393, + "learning_rate": 3.6985682347422446e-06, + "loss": 0.739, + "num_input_tokens_seen": 73105740, + "step": 2237 + }, + { + "epoch": 0.20183072552644632, + "flos": 18342889621920.0, + "grad_norm": 1.9515095858581606, + "learning_rate": 3.698259736421213e-06, + "loss": 0.7906, + "num_input_tokens_seen": 73131660, + "step": 2238 + }, + { + "epoch": 0.20192090904991658, + "flos": 23694505261920.0, + "grad_norm": 1.9163915049642315, + "learning_rate": 3.697951093195297e-06, + "loss": 0.7599, + "num_input_tokens_seen": 73160475, + "step": 2239 + }, + { + "epoch": 0.20201109257338684, + "flos": 25809313947840.0, + "grad_norm": 1.7076549656308433, + "learning_rate": 3.6976423050908307e-06, + "loss": 0.7529, + "num_input_tokens_seen": 73188450, + "step": 2240 + }, + { + "epoch": 0.2021012760968571, + "flos": 27524298795360.0, + "grad_norm": 1.5769803267602198, + "learning_rate": 3.697333372134163e-06, + "loss": 0.7658, + "num_input_tokens_seen": 73216910, + "step": 2241 + }, + { + "epoch": 0.20219145962032736, + "flos": 22897682486880.0, + "grad_norm": 2.4792826542582453, + "learning_rate": 3.697024294351653e-06, + "loss": 0.8714, + "num_input_tokens_seen": 73245200, + "step": 2242 + }, + { + "epoch": 0.20228164314379762, + "flos": 24386083245600.0, + "grad_norm": 1.7709569850460234, + "learning_rate": 3.696715071769672e-06, + "loss": 0.7381, + "num_input_tokens_seen": 73271940, + "step": 2243 + }, + { + "epoch": 0.20237182666726788, + "flos": 24026631500160.0, + "grad_norm": 2.1643250233858704, + "learning_rate": 3.696405704414606e-06, + "loss": 0.7756, + "num_input_tokens_seen": 73301895, + "step": 2244 + }, + { + "epoch": 0.20246201019073815, + "flos": 27304776960960.0, + "grad_norm": 1.7952834668978388, + "learning_rate": 3.6960961923128514e-06, + "loss": 0.8801, + "num_input_tokens_seen": 73332650, + "step": 2245 + }, + { + "epoch": 0.2025521937142084, + "flos": 22351052331840.0, + "grad_norm": 2.774635077873963, + "learning_rate": 3.6957865354908177e-06, + "loss": 0.7073, + "num_input_tokens_seen": 73360300, + "step": 2246 + }, + { + "epoch": 0.20264237723767867, + "flos": 28543802834400.0, + "grad_norm": 1.7924550635798384, + "learning_rate": 3.6954767339749262e-06, + "loss": 0.7985, + "num_input_tokens_seen": 73389690, + "step": 2247 + }, + { + "epoch": 0.20273256076114893, + "flos": 34300483248960.0, + "grad_norm": 2.348350282675134, + "learning_rate": 3.6951667877916113e-06, + "loss": 0.7523, + "num_input_tokens_seen": 73419375, + "step": 2248 + }, + { + "epoch": 0.2028227442846192, + "flos": 30474443861280.0, + "grad_norm": 2.0932840996113717, + "learning_rate": 3.694856696967319e-06, + "loss": 0.784, + "num_input_tokens_seen": 73453025, + "step": 2249 + }, + { + "epoch": 0.20291292780808945, + "flos": 22277742097920.0, + "grad_norm": 1.9070996126504276, + "learning_rate": 3.6945464615285077e-06, + "loss": 0.7647, + "num_input_tokens_seen": 73482890, + "step": 2250 + }, + { + "epoch": 0.2030031113315597, + "flos": 71213404128000.0, + "grad_norm": 0.6995691668075952, + "learning_rate": 3.694236081501648e-06, + "loss": 0.6912, + "num_input_tokens_seen": 73582080, + "step": 2251 + }, + { + "epoch": 0.20309329485503, + "flos": 25627302134880.0, + "grad_norm": 1.5717301683080351, + "learning_rate": 3.6939255569132246e-06, + "loss": 0.7387, + "num_input_tokens_seen": 73612510, + "step": 2252 + }, + { + "epoch": 0.20318347837850026, + "flos": 70070516454720.0, + "grad_norm": 0.8386482068414678, + "learning_rate": 3.693614887789733e-06, + "loss": 0.6199, + "num_input_tokens_seen": 73703070, + "step": 2253 + }, + { + "epoch": 0.20327366190197052, + "flos": 68231844881280.0, + "grad_norm": 0.7810926278875646, + "learning_rate": 3.69330407415768e-06, + "loss": 0.6205, + "num_input_tokens_seen": 73793915, + "step": 2254 + }, + { + "epoch": 0.20336384542544078, + "flos": 31853436815520.0, + "grad_norm": 1.717617979925507, + "learning_rate": 3.6929931160435867e-06, + "loss": 0.7815, + "num_input_tokens_seen": 73823780, + "step": 2255 + }, + { + "epoch": 0.20345402894891104, + "flos": 18525496151040.0, + "grad_norm": 2.230511933825613, + "learning_rate": 3.6926820134739858e-06, + "loss": 0.8326, + "num_input_tokens_seen": 73851930, + "step": 2256 + }, + { + "epoch": 0.2035442124723813, + "flos": 21621666968640.0, + "grad_norm": 3.1596673273667446, + "learning_rate": 3.692370766475422e-06, + "loss": 0.7137, + "num_input_tokens_seen": 73880420, + "step": 2257 + }, + { + "epoch": 0.20363439599585156, + "flos": 19580317195680.0, + "grad_norm": 2.3840758929749604, + "learning_rate": 3.692059375074453e-06, + "loss": 0.863, + "num_input_tokens_seen": 73905050, + "step": 2258 + }, + { + "epoch": 0.20372457951932182, + "flos": 36012903383040.0, + "grad_norm": 1.596257201128484, + "learning_rate": 3.6917478392976475e-06, + "loss": 0.8432, + "num_input_tokens_seen": 73935025, + "step": 2259 + }, + { + "epoch": 0.20381476304279209, + "flos": 29490405506880.0, + "grad_norm": 1.746013255472106, + "learning_rate": 3.691436159171589e-06, + "loss": 0.8378, + "num_input_tokens_seen": 73967070, + "step": 2260 + }, + { + "epoch": 0.20390494656626235, + "flos": 16412025576480.0, + "grad_norm": 1.9555181261596353, + "learning_rate": 3.6911243347228703e-06, + "loss": 0.7557, + "num_input_tokens_seen": 73994860, + "step": 2261 + }, + { + "epoch": 0.2039951300897326, + "flos": 62942165528640.0, + "grad_norm": 0.6361243158697706, + "learning_rate": 3.690812365978099e-06, + "loss": 0.533, + "num_input_tokens_seen": 74092495, + "step": 2262 + }, + { + "epoch": 0.20408531361320287, + "flos": 23442825851520.0, + "grad_norm": 1.8455586371809798, + "learning_rate": 3.690500252963893e-06, + "loss": 0.7325, + "num_input_tokens_seen": 74122160, + "step": 2263 + }, + { + "epoch": 0.20417549713667313, + "flos": 21252365236800.0, + "grad_norm": 1.8493485231557265, + "learning_rate": 3.6901879957068846e-06, + "loss": 0.7661, + "num_input_tokens_seen": 74151625, + "step": 2264 + }, + { + "epoch": 0.2042656806601434, + "flos": 25768421478240.0, + "grad_norm": 1.8744403003741539, + "learning_rate": 3.689875594233717e-06, + "loss": 0.8652, + "num_input_tokens_seen": 74182435, + "step": 2265 + }, + { + "epoch": 0.20435586418361365, + "flos": 27851146927680.0, + "grad_norm": 1.6278283374270062, + "learning_rate": 3.689563048571046e-06, + "loss": 0.7575, + "num_input_tokens_seen": 74213550, + "step": 2266 + }, + { + "epoch": 0.2044460477070839, + "flos": 20194384762560.0, + "grad_norm": 2.392967223622442, + "learning_rate": 3.6892503587455395e-06, + "loss": 0.7708, + "num_input_tokens_seen": 74241450, + "step": 2267 + }, + { + "epoch": 0.20453623123055417, + "flos": 42353538125280.0, + "grad_norm": 1.613572507710881, + "learning_rate": 3.6889375247838766e-06, + "loss": 0.7891, + "num_input_tokens_seen": 74275920, + "step": 2268 + }, + { + "epoch": 0.20462641475402443, + "flos": 21222214827840.0, + "grad_norm": 1.5461123349237498, + "learning_rate": 3.688624546712753e-06, + "loss": 0.7783, + "num_input_tokens_seen": 74304425, + "step": 2269 + }, + { + "epoch": 0.2047165982774947, + "flos": 24823305596160.0, + "grad_norm": 1.593183603806423, + "learning_rate": 3.688311424558871e-06, + "loss": 0.7295, + "num_input_tokens_seen": 74334450, + "step": 2270 + }, + { + "epoch": 0.20480678180096495, + "flos": 22605407964960.0, + "grad_norm": 2.549884539215468, + "learning_rate": 3.6879981583489496e-06, + "loss": 0.675, + "num_input_tokens_seen": 74357840, + "step": 2271 + }, + { + "epoch": 0.20489696532443522, + "flos": 38745979818720.0, + "grad_norm": 1.7324246274287605, + "learning_rate": 3.687684748109718e-06, + "loss": 0.7919, + "num_input_tokens_seen": 74391895, + "step": 2272 + }, + { + "epoch": 0.20498714884790548, + "flos": 15137125151040.0, + "grad_norm": 2.1321661671982035, + "learning_rate": 3.6873711938679174e-06, + "loss": 0.838, + "num_input_tokens_seen": 74418670, + "step": 2273 + }, + { + "epoch": 0.20507733237137574, + "flos": 19796642430720.0, + "grad_norm": 1.5698329515941003, + "learning_rate": 3.6870574956503027e-06, + "loss": 0.7669, + "num_input_tokens_seen": 74447740, + "step": 2274 + }, + { + "epoch": 0.20516751589484603, + "flos": 20055458435040.0, + "grad_norm": 2.0536308327369768, + "learning_rate": 3.68674365348364e-06, + "loss": 0.7666, + "num_input_tokens_seen": 74471610, + "step": 2275 + }, + { + "epoch": 0.2052576994183163, + "flos": 19865789651520.0, + "grad_norm": 3.3225558253435583, + "learning_rate": 3.6864296673947086e-06, + "loss": 0.7802, + "num_input_tokens_seen": 74498440, + "step": 2276 + }, + { + "epoch": 0.20534788294178655, + "flos": 20561679327360.0, + "grad_norm": 2.058939432439988, + "learning_rate": 3.686115537410298e-06, + "loss": 0.7573, + "num_input_tokens_seen": 74525500, + "step": 2277 + }, + { + "epoch": 0.2054380664652568, + "flos": 39146398373280.0, + "grad_norm": 1.9084844782025447, + "learning_rate": 3.685801263557214e-06, + "loss": 0.7258, + "num_input_tokens_seen": 74558120, + "step": 2278 + }, + { + "epoch": 0.20552824998872707, + "flos": 18561928249440.0, + "grad_norm": 1.8884198909268675, + "learning_rate": 3.68548684586227e-06, + "loss": 0.8689, + "num_input_tokens_seen": 74582930, + "step": 2279 + }, + { + "epoch": 0.20561843351219733, + "flos": 30985348143360.0, + "grad_norm": 2.6038198214658963, + "learning_rate": 3.685172284352295e-06, + "loss": 0.7718, + "num_input_tokens_seen": 74611805, + "step": 2280 + }, + { + "epoch": 0.2057086170356676, + "flos": 28438781061600.0, + "grad_norm": 1.5797160975191038, + "learning_rate": 3.684857579054128e-06, + "loss": 0.7086, + "num_input_tokens_seen": 74643685, + "step": 2281 + }, + { + "epoch": 0.20579880055913785, + "flos": 26169063051360.0, + "grad_norm": 2.1707587358052267, + "learning_rate": 3.6845427299946233e-06, + "loss": 0.8365, + "num_input_tokens_seen": 74669935, + "step": 2282 + }, + { + "epoch": 0.2058889840826081, + "flos": 23327582394720.0, + "grad_norm": 1.8555367226447586, + "learning_rate": 3.6842277372006434e-06, + "loss": 0.8733, + "num_input_tokens_seen": 74696445, + "step": 2283 + }, + { + "epoch": 0.20597916760607837, + "flos": 26429142827520.0, + "grad_norm": 2.5099353777405007, + "learning_rate": 3.6839126006990664e-06, + "loss": 0.7467, + "num_input_tokens_seen": 74729850, + "step": 2284 + }, + { + "epoch": 0.20606935112954863, + "flos": 31490937149760.0, + "grad_norm": 1.498295928663578, + "learning_rate": 3.6835973205167818e-06, + "loss": 0.8297, + "num_input_tokens_seen": 74761885, + "step": 2285 + }, + { + "epoch": 0.2061595346530189, + "flos": 18847958251680.0, + "grad_norm": 2.0431399446899485, + "learning_rate": 3.6832818966806904e-06, + "loss": 0.8118, + "num_input_tokens_seen": 74787415, + "step": 2286 + }, + { + "epoch": 0.20624971817648916, + "flos": 24425042887680.0, + "grad_norm": 1.8540438350610486, + "learning_rate": 3.682966329217706e-06, + "loss": 0.8559, + "num_input_tokens_seen": 74816905, + "step": 2287 + }, + { + "epoch": 0.20633990169995942, + "flos": 20785029647040.0, + "grad_norm": 1.7909556537933513, + "learning_rate": 3.6826506181547543e-06, + "loss": 0.8105, + "num_input_tokens_seen": 74844105, + "step": 2288 + }, + { + "epoch": 0.20643008522342968, + "flos": 21148309877760.0, + "grad_norm": 3.362905944371034, + "learning_rate": 3.682334763518774e-06, + "loss": 0.8011, + "num_input_tokens_seen": 74871905, + "step": 2289 + }, + { + "epoch": 0.20652026874689994, + "flos": 23480447382240.0, + "grad_norm": 1.9222421683058775, + "learning_rate": 3.6820187653367158e-06, + "loss": 0.7449, + "num_input_tokens_seen": 74900935, + "step": 2290 + }, + { + "epoch": 0.2066104522703702, + "flos": 45591683190720.0, + "grad_norm": 1.6727530767218373, + "learning_rate": 3.6817026236355412e-06, + "loss": 0.6684, + "num_input_tokens_seen": 74934750, + "step": 2291 + }, + { + "epoch": 0.20670063579384046, + "flos": 28690014434880.0, + "grad_norm": 3.43134325244696, + "learning_rate": 3.681386338442227e-06, + "loss": 0.7551, + "num_input_tokens_seen": 74964445, + "step": 2292 + }, + { + "epoch": 0.20679081931731072, + "flos": 25701764631360.0, + "grad_norm": 1.5542986206136922, + "learning_rate": 3.681069909783758e-06, + "loss": 0.7417, + "num_input_tokens_seen": 74995965, + "step": 2293 + }, + { + "epoch": 0.20688100284078098, + "flos": 20492680785600.0, + "grad_norm": 1.978594673889759, + "learning_rate": 3.680753337687136e-06, + "loss": 0.7848, + "num_input_tokens_seen": 75024900, + "step": 2294 + }, + { + "epoch": 0.20697118636425124, + "flos": 24209275199040.0, + "grad_norm": 2.3435830629024843, + "learning_rate": 3.680436622179371e-06, + "loss": 0.7026, + "num_input_tokens_seen": 75052235, + "step": 2295 + }, + { + "epoch": 0.2070613698877215, + "flos": 20383756188000.0, + "grad_norm": 1.976280584321854, + "learning_rate": 3.680119763287488e-06, + "loss": 0.7616, + "num_input_tokens_seen": 75079610, + "step": 2296 + }, + { + "epoch": 0.20715155341119176, + "flos": 17979460712160.0, + "grad_norm": 2.129722396777384, + "learning_rate": 3.6798027610385227e-06, + "loss": 0.7141, + "num_input_tokens_seen": 75107910, + "step": 2297 + }, + { + "epoch": 0.20724173693466202, + "flos": 51783207091200.0, + "grad_norm": 2.3073132056844767, + "learning_rate": 3.6794856154595235e-06, + "loss": 0.8243, + "num_input_tokens_seen": 75137155, + "step": 2298 + }, + { + "epoch": 0.2073319204581323, + "flos": 19800322236960.0, + "grad_norm": 2.2538056508163278, + "learning_rate": 3.6791683265775506e-06, + "loss": 0.7827, + "num_input_tokens_seen": 75164050, + "step": 2299 + }, + { + "epoch": 0.20742210398160257, + "flos": 26321816529600.0, + "grad_norm": 2.659007088911772, + "learning_rate": 3.6788508944196773e-06, + "loss": 0.737, + "num_input_tokens_seen": 75193140, + "step": 2300 + }, + { + "epoch": 0.20751228750507283, + "flos": 25336774591680.0, + "grad_norm": 1.4234983797590055, + "learning_rate": 3.678533319012989e-06, + "loss": 0.8511, + "num_input_tokens_seen": 75224165, + "step": 2301 + }, + { + "epoch": 0.2076024710285431, + "flos": 25483469399040.0, + "grad_norm": 1.787862316838449, + "learning_rate": 3.6782156003845826e-06, + "loss": 0.7258, + "num_input_tokens_seen": 75251440, + "step": 2302 + }, + { + "epoch": 0.20769265455201336, + "flos": 65628699691200.0, + "grad_norm": 0.6422542927109643, + "learning_rate": 3.6778977385615676e-06, + "loss": 0.5952, + "num_input_tokens_seen": 75344795, + "step": 2303 + }, + { + "epoch": 0.20778283807548362, + "flos": 30037816226880.0, + "grad_norm": 1.7777716823645575, + "learning_rate": 3.6775797335710656e-06, + "loss": 0.8275, + "num_input_tokens_seen": 75372065, + "step": 2304 + }, + { + "epoch": 0.20787302159895388, + "flos": 21364597943040.0, + "grad_norm": 1.553797097877152, + "learning_rate": 3.6772615854402105e-06, + "loss": 0.725, + "num_input_tokens_seen": 75401405, + "step": 2305 + }, + { + "epoch": 0.20796320512242414, + "flos": 48144829320000.0, + "grad_norm": 1.354886284381476, + "learning_rate": 3.6769432941961487e-06, + "loss": 0.6548, + "num_input_tokens_seen": 75434465, + "step": 2306 + }, + { + "epoch": 0.2080533886458944, + "flos": 39801618598080.0, + "grad_norm": 1.8049618070793794, + "learning_rate": 3.676624859866038e-06, + "loss": 0.7742, + "num_input_tokens_seen": 75465960, + "step": 2307 + }, + { + "epoch": 0.20814357216936466, + "flos": 15646840000800.0, + "grad_norm": 2.813709253383022, + "learning_rate": 3.67630628247705e-06, + "loss": 0.7031, + "num_input_tokens_seen": 75490965, + "step": 2308 + }, + { + "epoch": 0.20823375569283492, + "flos": 21221620111680.0, + "grad_norm": 2.029494715602674, + "learning_rate": 3.675987562056367e-06, + "loss": 0.8397, + "num_input_tokens_seen": 75517715, + "step": 2309 + }, + { + "epoch": 0.20832393921630518, + "flos": 22642286100480.0, + "grad_norm": 1.8199706186291416, + "learning_rate": 3.675668698631184e-06, + "loss": 0.8113, + "num_input_tokens_seen": 75546450, + "step": 2310 + }, + { + "epoch": 0.20841412273977544, + "flos": 16151128065600.0, + "grad_norm": 2.2765052259332004, + "learning_rate": 3.675349692228708e-06, + "loss": 0.8376, + "num_input_tokens_seen": 75573640, + "step": 2311 + }, + { + "epoch": 0.2085043062632457, + "flos": 27560730893760.0, + "grad_norm": 2.1340396690292516, + "learning_rate": 3.6750305428761578e-06, + "loss": 0.7468, + "num_input_tokens_seen": 75602875, + "step": 2312 + }, + { + "epoch": 0.20859448978671596, + "flos": 14517593629440.0, + "grad_norm": 2.1029119874002595, + "learning_rate": 3.674711250600766e-06, + "loss": 0.7836, + "num_input_tokens_seen": 75627655, + "step": 2313 + }, + { + "epoch": 0.20868467331018623, + "flos": 70923805828800.0, + "grad_norm": 0.6770462368315097, + "learning_rate": 3.6743918154297765e-06, + "loss": 0.5766, + "num_input_tokens_seen": 75717175, + "step": 2314 + }, + { + "epoch": 0.20877485683365649, + "flos": 24827505779040.0, + "grad_norm": 2.022052069668022, + "learning_rate": 3.6740722373904446e-06, + "loss": 0.7652, + "num_input_tokens_seen": 75743895, + "step": 2315 + }, + { + "epoch": 0.20886504035712675, + "flos": 32295862932480.0, + "grad_norm": 1.7628759402562204, + "learning_rate": 3.6737525165100383e-06, + "loss": 0.7519, + "num_input_tokens_seen": 75776265, + "step": 2316 + }, + { + "epoch": 0.208955223880597, + "flos": 26941050693120.0, + "grad_norm": 1.760580096662437, + "learning_rate": 3.6734326528158385e-06, + "loss": 0.7928, + "num_input_tokens_seen": 75806740, + "step": 2317 + }, + { + "epoch": 0.20904540740406727, + "flos": 36376034934720.0, + "grad_norm": 1.789125799533503, + "learning_rate": 3.673112646335138e-06, + "loss": 0.764, + "num_input_tokens_seen": 75836270, + "step": 2318 + }, + { + "epoch": 0.20913559092753753, + "flos": 36741359502240.0, + "grad_norm": 2.38007374945745, + "learning_rate": 3.672792497095241e-06, + "loss": 0.7235, + "num_input_tokens_seen": 75869595, + "step": 2319 + }, + { + "epoch": 0.2092257744510078, + "flos": 17978940335520.0, + "grad_norm": 2.229446845109445, + "learning_rate": 3.672472205123464e-06, + "loss": 0.7725, + "num_input_tokens_seen": 75896240, + "step": 2320 + }, + { + "epoch": 0.20931595797447805, + "flos": 22752214281600.0, + "grad_norm": 1.8454957229907825, + "learning_rate": 3.6721517704471363e-06, + "loss": 0.7045, + "num_input_tokens_seen": 75927475, + "step": 2321 + }, + { + "epoch": 0.2094061414979483, + "flos": 22165249203360.0, + "grad_norm": 1.7544977426892183, + "learning_rate": 3.6718311930936e-06, + "loss": 0.8097, + "num_input_tokens_seen": 75956990, + "step": 2322 + }, + { + "epoch": 0.2094963250214186, + "flos": 16332024785760.0, + "grad_norm": 2.213828309891047, + "learning_rate": 3.6715104730902074e-06, + "loss": 0.7285, + "num_input_tokens_seen": 75979860, + "step": 2323 + }, + { + "epoch": 0.20958650854488886, + "flos": 19436224271520.0, + "grad_norm": 2.0696876590618687, + "learning_rate": 3.671189610464325e-06, + "loss": 0.8121, + "num_input_tokens_seen": 76005250, + "step": 2324 + }, + { + "epoch": 0.20967669206835912, + "flos": 19905604198080.0, + "grad_norm": 2.15997722083902, + "learning_rate": 3.6708686052433303e-06, + "loss": 0.7233, + "num_input_tokens_seen": 76031485, + "step": 2325 + }, + { + "epoch": 0.20976687559182938, + "flos": 27743337422880.0, + "grad_norm": 1.7769748179079534, + "learning_rate": 3.6705474574546127e-06, + "loss": 0.7898, + "num_input_tokens_seen": 76060710, + "step": 2326 + }, + { + "epoch": 0.20985705911529964, + "flos": 26357653911840.0, + "grad_norm": 3.62845972638004, + "learning_rate": 3.670226167125575e-06, + "loss": 0.7931, + "num_input_tokens_seen": 76091085, + "step": 2327 + }, + { + "epoch": 0.2099472426387699, + "flos": 17758786615200.0, + "grad_norm": 1.915621189032742, + "learning_rate": 3.6699047342836313e-06, + "loss": 0.7773, + "num_input_tokens_seen": 76118165, + "step": 2328 + }, + { + "epoch": 0.21003742616224017, + "flos": 28032601194240.0, + "grad_norm": 1.450692550749654, + "learning_rate": 3.669583158956208e-06, + "loss": 0.8097, + "num_input_tokens_seen": 76149710, + "step": 2329 + }, + { + "epoch": 0.21012760968571043, + "flos": 62034819856320.0, + "grad_norm": 0.6745523734603679, + "learning_rate": 3.669261441170743e-06, + "loss": 0.6066, + "num_input_tokens_seen": 76244345, + "step": 2330 + }, + { + "epoch": 0.2102177932091807, + "flos": 21440138362560.0, + "grad_norm": 1.7694660368433162, + "learning_rate": 3.668939580954688e-06, + "loss": 0.7852, + "num_input_tokens_seen": 76272695, + "step": 2331 + }, + { + "epoch": 0.21030797673265095, + "flos": 27268382032320.0, + "grad_norm": 2.054318471049015, + "learning_rate": 3.668617578335506e-06, + "loss": 0.7504, + "num_input_tokens_seen": 76304785, + "step": 2332 + }, + { + "epoch": 0.2103981602561212, + "flos": 27997693056000.0, + "grad_norm": 1.7774038451378846, + "learning_rate": 3.6682954333406707e-06, + "loss": 0.8044, + "num_input_tokens_seen": 76335020, + "step": 2333 + }, + { + "epoch": 0.21048834377959147, + "flos": 16376820080160.0, + "grad_norm": 2.2559592484993813, + "learning_rate": 3.6679731459976707e-06, + "loss": 0.9117, + "num_input_tokens_seen": 76358245, + "step": 2334 + }, + { + "epoch": 0.21057852730306173, + "flos": 21295450722240.0, + "grad_norm": 1.5845074901072744, + "learning_rate": 3.6676507163340046e-06, + "loss": 0.7699, + "num_input_tokens_seen": 76387960, + "step": 2335 + }, + { + "epoch": 0.210668710826532, + "flos": 23626621812960.0, + "grad_norm": 2.604699154137236, + "learning_rate": 3.6673281443771842e-06, + "loss": 0.8086, + "num_input_tokens_seen": 76416620, + "step": 2336 + }, + { + "epoch": 0.21075889435000225, + "flos": 22350643464480.0, + "grad_norm": 1.768754733992533, + "learning_rate": 3.667005430154733e-06, + "loss": 0.6689, + "num_input_tokens_seen": 76446370, + "step": 2337 + }, + { + "epoch": 0.2108490778734725, + "flos": 70805700300480.0, + "grad_norm": 0.671515346343217, + "learning_rate": 3.666682573694186e-06, + "loss": 0.6803, + "num_input_tokens_seen": 76536445, + "step": 2338 + }, + { + "epoch": 0.21093926139694277, + "flos": 65665243298880.0, + "grad_norm": 0.655764650067036, + "learning_rate": 3.6663595750230924e-06, + "loss": 0.5501, + "num_input_tokens_seen": 76626735, + "step": 2339 + }, + { + "epoch": 0.21102944492041303, + "flos": 21579882424800.0, + "grad_norm": 2.142887340060406, + "learning_rate": 3.666036434169012e-06, + "loss": 0.8277, + "num_input_tokens_seen": 76654320, + "step": 2340 + }, + { + "epoch": 0.2111196284438833, + "flos": 22711693509600.0, + "grad_norm": 1.7062937105609424, + "learning_rate": 3.665713151159516e-06, + "loss": 0.7557, + "num_input_tokens_seen": 76684755, + "step": 2341 + }, + { + "epoch": 0.21120981196735356, + "flos": 26975401284960.0, + "grad_norm": 1.652642117158832, + "learning_rate": 3.665389726022189e-06, + "loss": 0.8762, + "num_input_tokens_seen": 76714570, + "step": 2342 + }, + { + "epoch": 0.21129999549082382, + "flos": 25885040216160.0, + "grad_norm": 1.9331555419048378, + "learning_rate": 3.6650661587846283e-06, + "loss": 0.7504, + "num_input_tokens_seen": 76743680, + "step": 2343 + }, + { + "epoch": 0.21139017901429408, + "flos": 22750169944800.0, + "grad_norm": 2.7539312699716714, + "learning_rate": 3.6647424494744418e-06, + "loss": 0.7399, + "num_input_tokens_seen": 76769165, + "step": 2344 + }, + { + "epoch": 0.21148036253776434, + "flos": 21582372798720.0, + "grad_norm": 1.6288165975731164, + "learning_rate": 3.6644185981192503e-06, + "loss": 0.8257, + "num_input_tokens_seen": 76797150, + "step": 2345 + }, + { + "epoch": 0.2115705460612346, + "flos": 23225162505120.0, + "grad_norm": 1.9160672646503814, + "learning_rate": 3.6640946047466868e-06, + "loss": 0.7061, + "num_input_tokens_seen": 76826720, + "step": 2346 + }, + { + "epoch": 0.2116607295847049, + "flos": 61545039465120.0, + "grad_norm": 1.9537707672675044, + "learning_rate": 3.6637704693843953e-06, + "loss": 0.7462, + "num_input_tokens_seen": 76860430, + "step": 2347 + }, + { + "epoch": 0.21175091310817515, + "flos": 24427347412800.0, + "grad_norm": 2.4920665747504183, + "learning_rate": 3.6634461920600337e-06, + "loss": 0.7092, + "num_input_tokens_seen": 76891070, + "step": 2348 + }, + { + "epoch": 0.2118410966316454, + "flos": 11020372371360.0, + "grad_norm": 2.340017576559193, + "learning_rate": 3.66312177280127e-06, + "loss": 0.8036, + "num_input_tokens_seen": 76918800, + "step": 2349 + }, + { + "epoch": 0.21193128015511567, + "flos": 20565544982400.0, + "grad_norm": 1.4124195396665253, + "learning_rate": 3.6627972116357872e-06, + "loss": 0.816, + "num_input_tokens_seen": 76947610, + "step": 2350 + }, + { + "epoch": 0.21202146367858593, + "flos": 20018580299520.0, + "grad_norm": 1.9526862910608547, + "learning_rate": 3.662472508591278e-06, + "loss": 0.8396, + "num_input_tokens_seen": 76976230, + "step": 2351 + }, + { + "epoch": 0.2121116472020562, + "flos": 29528473074720.0, + "grad_norm": 1.855546616376003, + "learning_rate": 3.662147663695447e-06, + "loss": 0.7642, + "num_input_tokens_seen": 77006810, + "step": 2352 + }, + { + "epoch": 0.21220183072552645, + "flos": 26139693207360.0, + "grad_norm": 3.0599834456650483, + "learning_rate": 3.6618226769760127e-06, + "loss": 0.7672, + "num_input_tokens_seen": 77038590, + "step": 2353 + }, + { + "epoch": 0.2122920142489967, + "flos": 29236718929440.0, + "grad_norm": 1.7868462707622372, + "learning_rate": 3.661497548460704e-06, + "loss": 0.655, + "num_input_tokens_seen": 77069240, + "step": 2354 + }, + { + "epoch": 0.21238219777246697, + "flos": 17796891352800.0, + "grad_norm": 2.2254147432375264, + "learning_rate": 3.6611722781772635e-06, + "loss": 0.7959, + "num_input_tokens_seen": 77095135, + "step": 2355 + }, + { + "epoch": 0.21247238129593723, + "flos": 63237041933760.0, + "grad_norm": 0.8331218252386228, + "learning_rate": 3.6608468661534444e-06, + "loss": 0.6447, + "num_input_tokens_seen": 77176585, + "step": 2356 + }, + { + "epoch": 0.2125625648194075, + "flos": 15647732075040.0, + "grad_norm": 2.727310908657183, + "learning_rate": 3.660521312417013e-06, + "loss": 0.699, + "num_input_tokens_seen": 77202880, + "step": 2357 + }, + { + "epoch": 0.21265274834287776, + "flos": 22311237785280.0, + "grad_norm": 2.037624392778547, + "learning_rate": 3.660195616995747e-06, + "loss": 0.7478, + "num_input_tokens_seen": 77231305, + "step": 2358 + }, + { + "epoch": 0.21274293186634802, + "flos": 22060115921280.0, + "grad_norm": 3.2179902300357393, + "learning_rate": 3.6598697799174367e-06, + "loss": 0.8021, + "num_input_tokens_seen": 77261945, + "step": 2359 + }, + { + "epoch": 0.21283311538981828, + "flos": 17723655458400.0, + "grad_norm": 2.09858895988676, + "learning_rate": 3.6595438012098844e-06, + "loss": 0.8353, + "num_input_tokens_seen": 77288890, + "step": 2360 + }, + { + "epoch": 0.21292329891328854, + "flos": 33820955977920.0, + "grad_norm": 2.8707516005738403, + "learning_rate": 3.6592176809009045e-06, + "loss": 0.748, + "num_input_tokens_seen": 77318215, + "step": 2361 + }, + { + "epoch": 0.2130134824367588, + "flos": 24973568700480.0, + "grad_norm": 1.8414621544542347, + "learning_rate": 3.6588914190183227e-06, + "loss": 0.8145, + "num_input_tokens_seen": 77347905, + "step": 2362 + }, + { + "epoch": 0.21310366596022906, + "flos": 18816544070880.0, + "grad_norm": 1.9732978608298195, + "learning_rate": 3.658565015589978e-06, + "loss": 0.649, + "num_input_tokens_seen": 77376140, + "step": 2363 + }, + { + "epoch": 0.21319384948369932, + "flos": 20456062838400.0, + "grad_norm": 4.224022403455335, + "learning_rate": 3.6582384706437217e-06, + "loss": 0.7855, + "num_input_tokens_seen": 77403985, + "step": 2364 + }, + { + "epoch": 0.21328403300716958, + "flos": 25047324971520.0, + "grad_norm": 2.0449896129913303, + "learning_rate": 3.6579117842074156e-06, + "loss": 0.7941, + "num_input_tokens_seen": 77433395, + "step": 2365 + }, + { + "epoch": 0.21337421653063984, + "flos": 24500211609600.0, + "grad_norm": 1.7374615397416135, + "learning_rate": 3.657584956308934e-06, + "loss": 0.7158, + "num_input_tokens_seen": 77463225, + "step": 2366 + }, + { + "epoch": 0.2134644000541101, + "flos": 16556118500640.0, + "grad_norm": 2.302977053317177, + "learning_rate": 3.6572579869761648e-06, + "loss": 0.8415, + "num_input_tokens_seen": 77486300, + "step": 2367 + }, + { + "epoch": 0.21355458357758036, + "flos": 16224661318080.0, + "grad_norm": 1.906222108730701, + "learning_rate": 3.6569308762370056e-06, + "loss": 0.8325, + "num_input_tokens_seen": 77511825, + "step": 2368 + }, + { + "epoch": 0.21364476710105063, + "flos": 25483506568800.0, + "grad_norm": 2.3671569822251572, + "learning_rate": 3.6566036241193676e-06, + "loss": 0.8181, + "num_input_tokens_seen": 77540430, + "step": 2369 + }, + { + "epoch": 0.2137349506245209, + "flos": 69132388487520.0, + "grad_norm": 0.6199000421440035, + "learning_rate": 3.656276230651174e-06, + "loss": 0.608, + "num_input_tokens_seen": 77636350, + "step": 2370 + }, + { + "epoch": 0.21382513414799117, + "flos": 56710120856160.0, + "grad_norm": 0.7058945892450317, + "learning_rate": 3.65594869586036e-06, + "loss": 0.6413, + "num_input_tokens_seen": 77722835, + "step": 2371 + }, + { + "epoch": 0.21391531767146144, + "flos": 29126456220480.0, + "grad_norm": 1.7753908935815252, + "learning_rate": 3.6556210197748724e-06, + "loss": 0.6958, + "num_input_tokens_seen": 77753315, + "step": 2372 + }, + { + "epoch": 0.2140055011949317, + "flos": 20783840214720.0, + "grad_norm": 1.8801137490050963, + "learning_rate": 3.655293202422671e-06, + "loss": 0.7421, + "num_input_tokens_seen": 77781725, + "step": 2373 + }, + { + "epoch": 0.21409568471840196, + "flos": 26831308360800.0, + "grad_norm": 1.5317880113510323, + "learning_rate": 3.654965243831725e-06, + "loss": 0.7881, + "num_input_tokens_seen": 77812390, + "step": 2374 + }, + { + "epoch": 0.21418586824187222, + "flos": 20892913491360.0, + "grad_norm": 1.614346279667113, + "learning_rate": 3.65463714403002e-06, + "loss": 0.8127, + "num_input_tokens_seen": 77840015, + "step": 2375 + }, + { + "epoch": 0.21427605176534248, + "flos": 29600110669440.0, + "grad_norm": 1.533232670396057, + "learning_rate": 3.65430890304555e-06, + "loss": 0.7309, + "num_input_tokens_seen": 77871790, + "step": 2376 + }, + { + "epoch": 0.21436623528881274, + "flos": 24828397853280.0, + "grad_norm": 1.4028052205631738, + "learning_rate": 3.653980520906323e-06, + "loss": 0.8036, + "num_input_tokens_seen": 77902070, + "step": 2377 + }, + { + "epoch": 0.214456418812283, + "flos": 28216657344000.0, + "grad_norm": 3.2379784896918373, + "learning_rate": 3.653651997640358e-06, + "loss": 0.7518, + "num_input_tokens_seen": 77934565, + "step": 2378 + }, + { + "epoch": 0.21454660233575326, + "flos": 23839230072000.0, + "grad_norm": 1.8421820099338717, + "learning_rate": 3.653323333275686e-06, + "loss": 0.7798, + "num_input_tokens_seen": 77961400, + "step": 2379 + }, + { + "epoch": 0.21463678585922352, + "flos": 18850114097760.0, + "grad_norm": 2.877121515778007, + "learning_rate": 3.652994527840351e-06, + "loss": 0.7581, + "num_input_tokens_seen": 77990095, + "step": 2380 + }, + { + "epoch": 0.21472696938269378, + "flos": 13497717892800.0, + "grad_norm": 2.207075108455259, + "learning_rate": 3.6526655813624087e-06, + "loss": 0.737, + "num_input_tokens_seen": 78015690, + "step": 2381 + }, + { + "epoch": 0.21481715290616404, + "flos": 67687630760640.0, + "grad_norm": 0.5812629390629989, + "learning_rate": 3.652336493869925e-06, + "loss": 0.601, + "num_input_tokens_seen": 78116660, + "step": 2382 + }, + { + "epoch": 0.2149073364296343, + "flos": 29819260806240.0, + "grad_norm": 1.9330349201337944, + "learning_rate": 3.6520072653909823e-06, + "loss": 0.7835, + "num_input_tokens_seen": 78145880, + "step": 2383 + }, + { + "epoch": 0.21499751995310457, + "flos": 20675027126400.0, + "grad_norm": 2.0981166341488127, + "learning_rate": 3.6516778959536702e-06, + "loss": 0.7717, + "num_input_tokens_seen": 78174475, + "step": 2384 + }, + { + "epoch": 0.21508770347657483, + "flos": 15752382150240.0, + "grad_norm": 2.111982528238565, + "learning_rate": 3.6513483855860923e-06, + "loss": 0.8229, + "num_input_tokens_seen": 78200470, + "step": 2385 + }, + { + "epoch": 0.2151778870000451, + "flos": 23151926610720.0, + "grad_norm": 2.204401282339566, + "learning_rate": 3.6510187343163654e-06, + "loss": 0.7182, + "num_input_tokens_seen": 78229180, + "step": 2386 + }, + { + "epoch": 0.21526807052351535, + "flos": 21038864903520.0, + "grad_norm": 1.74089358053716, + "learning_rate": 3.650688942172616e-06, + "loss": 0.8059, + "num_input_tokens_seen": 78254260, + "step": 2387 + }, + { + "epoch": 0.2153582540469856, + "flos": 20857187618400.0, + "grad_norm": 1.9737417727994104, + "learning_rate": 3.650359009182984e-06, + "loss": 0.7146, + "num_input_tokens_seen": 78281485, + "step": 2388 + }, + { + "epoch": 0.21544843757045587, + "flos": 21075371341440.0, + "grad_norm": 1.8487308272682355, + "learning_rate": 3.650028935375622e-06, + "loss": 0.8539, + "num_input_tokens_seen": 78307720, + "step": 2389 + }, + { + "epoch": 0.21553862109392613, + "flos": 28796634507360.0, + "grad_norm": 1.599975245999537, + "learning_rate": 3.6496987207786926e-06, + "loss": 0.7392, + "num_input_tokens_seen": 78340960, + "step": 2390 + }, + { + "epoch": 0.2156288046173964, + "flos": 20670232227360.0, + "grad_norm": 1.839344386145103, + "learning_rate": 3.6493683654203724e-06, + "loss": 0.7223, + "num_input_tokens_seen": 78368010, + "step": 2391 + }, + { + "epoch": 0.21571898814086665, + "flos": 24281507509920.0, + "grad_norm": 2.812950352534067, + "learning_rate": 3.6490378693288484e-06, + "loss": 0.7632, + "num_input_tokens_seen": 78395700, + "step": 2392 + }, + { + "epoch": 0.2158091716643369, + "flos": 21650925303360.0, + "grad_norm": 1.7222029939555692, + "learning_rate": 3.648707232532321e-06, + "loss": 0.8206, + "num_input_tokens_seen": 78421575, + "step": 2393 + }, + { + "epoch": 0.2158993551878072, + "flos": 23843132896800.0, + "grad_norm": 2.473676184911407, + "learning_rate": 3.6483764550590017e-06, + "loss": 0.7749, + "num_input_tokens_seen": 78448480, + "step": 2394 + }, + { + "epoch": 0.21598953871127746, + "flos": 20930126154720.0, + "grad_norm": 2.2316069629974233, + "learning_rate": 3.6480455369371133e-06, + "loss": 0.7922, + "num_input_tokens_seen": 78477190, + "step": 2395 + }, + { + "epoch": 0.21607972223474772, + "flos": 23223229677600.0, + "grad_norm": 2.56921669463296, + "learning_rate": 3.647714478194893e-06, + "loss": 0.7818, + "num_input_tokens_seen": 78506790, + "step": 2396 + }, + { + "epoch": 0.21616990575821798, + "flos": 20347063901280.0, + "grad_norm": 3.74350290135765, + "learning_rate": 3.647383278860588e-06, + "loss": 0.7199, + "num_input_tokens_seen": 78533690, + "step": 2397 + }, + { + "epoch": 0.21626008928168824, + "flos": 14518225515360.0, + "grad_norm": 1.6382459557569355, + "learning_rate": 3.6470519389624587e-06, + "loss": 0.7616, + "num_input_tokens_seen": 78559255, + "step": 2398 + }, + { + "epoch": 0.2163502728051585, + "flos": 27050086800000.0, + "grad_norm": 1.572866535861571, + "learning_rate": 3.646720458528776e-06, + "loss": 0.8354, + "num_input_tokens_seen": 78587180, + "step": 2399 + }, + { + "epoch": 0.21644045632862877, + "flos": 27733896303840.0, + "grad_norm": 2.469748765778506, + "learning_rate": 3.6463888375878235e-06, + "loss": 0.7452, + "num_input_tokens_seen": 78610015, + "step": 2400 + }, + { + "epoch": 0.21653063985209903, + "flos": 26977668640320.0, + "grad_norm": 2.1422640797222146, + "learning_rate": 3.646057076167897e-06, + "loss": 0.7897, + "num_input_tokens_seen": 78636875, + "step": 2401 + }, + { + "epoch": 0.2166208233755693, + "flos": 18671893600320.0, + "grad_norm": 1.7913153178197956, + "learning_rate": 3.645725174297305e-06, + "loss": 0.7796, + "num_input_tokens_seen": 78664360, + "step": 2402 + }, + { + "epoch": 0.21671100689903955, + "flos": 26868744042720.0, + "grad_norm": 2.67786570198451, + "learning_rate": 3.645393132004367e-06, + "loss": 0.7564, + "num_input_tokens_seen": 78693615, + "step": 2403 + }, + { + "epoch": 0.2168011904225098, + "flos": 37834285284480.0, + "grad_norm": 3.2333335489751476, + "learning_rate": 3.6450609493174135e-06, + "loss": 0.7316, + "num_input_tokens_seen": 78724450, + "step": 2404 + }, + { + "epoch": 0.21689137394598007, + "flos": 14881431406560.0, + "grad_norm": 1.8772424639597445, + "learning_rate": 3.6447286262647896e-06, + "loss": 0.7352, + "num_input_tokens_seen": 78751045, + "step": 2405 + }, + { + "epoch": 0.21698155746945033, + "flos": 14408446013280.0, + "grad_norm": 6.470612925722067, + "learning_rate": 3.64439616287485e-06, + "loss": 0.7405, + "num_input_tokens_seen": 78777625, + "step": 2406 + }, + { + "epoch": 0.2170717409929206, + "flos": 23804247594240.0, + "grad_norm": 3.362583916620294, + "learning_rate": 3.644063559175963e-06, + "loss": 0.7375, + "num_input_tokens_seen": 78805465, + "step": 2407 + }, + { + "epoch": 0.21716192451639085, + "flos": 28943924030880.0, + "grad_norm": 1.6534730195185028, + "learning_rate": 3.6437308151965074e-06, + "loss": 0.7544, + "num_input_tokens_seen": 78835680, + "step": 2408 + }, + { + "epoch": 0.2172521080398611, + "flos": 25627859681280.0, + "grad_norm": 1.6040120070904647, + "learning_rate": 3.643397930964876e-06, + "loss": 0.7871, + "num_input_tokens_seen": 78865160, + "step": 2409 + }, + { + "epoch": 0.21734229156333137, + "flos": 27924754519680.0, + "grad_norm": 2.6342342171660955, + "learning_rate": 3.6430649065094707e-06, + "loss": 0.8448, + "num_input_tokens_seen": 78893800, + "step": 2410 + }, + { + "epoch": 0.21743247508680164, + "flos": 18664868515680.0, + "grad_norm": 2.053308756833982, + "learning_rate": 3.6427317418587086e-06, + "loss": 0.8368, + "num_input_tokens_seen": 78914970, + "step": 2411 + }, + { + "epoch": 0.2175226586102719, + "flos": 22897682486880.0, + "grad_norm": 1.774055810095608, + "learning_rate": 3.6423984370410157e-06, + "loss": 0.7136, + "num_input_tokens_seen": 78945725, + "step": 2412 + }, + { + "epoch": 0.21761284213374216, + "flos": 24826836723360.0, + "grad_norm": 1.7660675169457738, + "learning_rate": 3.6420649920848324e-06, + "loss": 0.7394, + "num_input_tokens_seen": 78976095, + "step": 2413 + }, + { + "epoch": 0.21770302565721242, + "flos": 24206413127520.0, + "grad_norm": 2.04166629038697, + "learning_rate": 3.6417314070186096e-06, + "loss": 0.7822, + "num_input_tokens_seen": 79006305, + "step": 2414 + }, + { + "epoch": 0.21779320918068268, + "flos": 28326027978720.0, + "grad_norm": 1.5872405763851958, + "learning_rate": 3.641397681870811e-06, + "loss": 0.7816, + "num_input_tokens_seen": 79036300, + "step": 2415 + }, + { + "epoch": 0.21788339270415294, + "flos": 24901187710560.0, + "grad_norm": 1.6695029999373845, + "learning_rate": 3.641063816669911e-06, + "loss": 0.7938, + "num_input_tokens_seen": 79065165, + "step": 2416 + }, + { + "epoch": 0.2179735762276232, + "flos": 25265322845760.0, + "grad_norm": 2.3449458822419142, + "learning_rate": 3.640729811444398e-06, + "loss": 0.8338, + "num_input_tokens_seen": 79094330, + "step": 2417 + }, + { + "epoch": 0.2180637597510935, + "flos": 36886270161120.0, + "grad_norm": 1.7167260477604114, + "learning_rate": 3.6403956662227706e-06, + "loss": 0.7775, + "num_input_tokens_seen": 79126955, + "step": 2418 + }, + { + "epoch": 0.21815394327456375, + "flos": 22241607357600.0, + "grad_norm": 1.878091234865245, + "learning_rate": 3.6400613810335396e-06, + "loss": 0.8271, + "num_input_tokens_seen": 79154835, + "step": 2419 + }, + { + "epoch": 0.218244126798034, + "flos": 32551407997920.0, + "grad_norm": 1.742226980814036, + "learning_rate": 3.639726955905228e-06, + "loss": 0.6903, + "num_input_tokens_seen": 79183780, + "step": 2420 + }, + { + "epoch": 0.21833431032150427, + "flos": 27415002500160.0, + "grad_norm": 1.6115016774433057, + "learning_rate": 3.639392390866372e-06, + "loss": 0.7975, + "num_input_tokens_seen": 79212245, + "step": 2421 + }, + { + "epoch": 0.21842449384497453, + "flos": 23226463446720.0, + "grad_norm": 1.7110463286325144, + "learning_rate": 3.639057685945517e-06, + "loss": 0.7511, + "num_input_tokens_seen": 79241625, + "step": 2422 + }, + { + "epoch": 0.2185146773684448, + "flos": 41291580486720.0, + "grad_norm": 1.4936325505900878, + "learning_rate": 3.638722841171223e-06, + "loss": 0.662, + "num_input_tokens_seen": 79277685, + "step": 2423 + }, + { + "epoch": 0.21860486089191505, + "flos": 23516544952800.0, + "grad_norm": 1.8551080292865658, + "learning_rate": 3.638387856572061e-06, + "loss": 0.813, + "num_input_tokens_seen": 79307340, + "step": 2424 + }, + { + "epoch": 0.21869504441538531, + "flos": 67299149432640.0, + "grad_norm": 0.6564619539484459, + "learning_rate": 3.638052732176612e-06, + "loss": 0.6071, + "num_input_tokens_seen": 79393245, + "step": 2425 + }, + { + "epoch": 0.21878522793885558, + "flos": 21327013582080.0, + "grad_norm": 1.6482050582658148, + "learning_rate": 3.637717468013472e-06, + "loss": 0.7861, + "num_input_tokens_seen": 79420230, + "step": 2426 + }, + { + "epoch": 0.21887541146232584, + "flos": 65921345910720.0, + "grad_norm": 0.5957770379910771, + "learning_rate": 3.6373820641112475e-06, + "loss": 0.559, + "num_input_tokens_seen": 79512450, + "step": 2427 + }, + { + "epoch": 0.2189655949857961, + "flos": 20128768668960.0, + "grad_norm": 1.6207266886063074, + "learning_rate": 3.6370465204985567e-06, + "loss": 0.7903, + "num_input_tokens_seen": 79537380, + "step": 2428 + }, + { + "epoch": 0.21905577850926636, + "flos": 69205698721440.0, + "grad_norm": 0.6325084520989006, + "learning_rate": 3.6367108372040304e-06, + "loss": 0.6199, + "num_input_tokens_seen": 79632080, + "step": 2429 + }, + { + "epoch": 0.21914596203273662, + "flos": 23480187193920.0, + "grad_norm": 1.9115037904023775, + "learning_rate": 3.6363750142563107e-06, + "loss": 0.7766, + "num_input_tokens_seen": 79658990, + "step": 2430 + }, + { + "epoch": 0.21923614555620688, + "flos": 64481643271200.0, + "grad_norm": 0.7198329727826721, + "learning_rate": 3.636039051684052e-06, + "loss": 0.6562, + "num_input_tokens_seen": 79760370, + "step": 2431 + }, + { + "epoch": 0.21932632907967714, + "flos": 13351729310880.0, + "grad_norm": 2.0148126902262598, + "learning_rate": 3.6357029495159203e-06, + "loss": 0.7454, + "num_input_tokens_seen": 79788120, + "step": 2432 + }, + { + "epoch": 0.2194165126031474, + "flos": 23989902043680.0, + "grad_norm": 1.6008072723909783, + "learning_rate": 3.6353667077805934e-06, + "loss": 0.7497, + "num_input_tokens_seen": 79816810, + "step": 2433 + }, + { + "epoch": 0.21950669612661766, + "flos": 21440324211360.0, + "grad_norm": 1.8600472268907686, + "learning_rate": 3.6350303265067625e-06, + "loss": 0.8106, + "num_input_tokens_seen": 79845335, + "step": 2434 + }, + { + "epoch": 0.21959687965008792, + "flos": 27124809484800.0, + "grad_norm": 2.0010084947427673, + "learning_rate": 3.6346938057231285e-06, + "loss": 0.8042, + "num_input_tokens_seen": 79877745, + "step": 2435 + }, + { + "epoch": 0.21968706317355818, + "flos": 25265917561920.0, + "grad_norm": 1.7718946829212232, + "learning_rate": 3.6343571454584047e-06, + "loss": 0.7202, + "num_input_tokens_seen": 79909720, + "step": 2436 + }, + { + "epoch": 0.21977724669702844, + "flos": 21876877506240.0, + "grad_norm": 2.9798629765082567, + "learning_rate": 3.6340203457413176e-06, + "loss": 0.8144, + "num_input_tokens_seen": 79935930, + "step": 2437 + }, + { + "epoch": 0.2198674302204987, + "flos": 14809236265440.0, + "grad_norm": 2.1899619281194247, + "learning_rate": 3.633683406600605e-06, + "loss": 0.7994, + "num_input_tokens_seen": 79961820, + "step": 2438 + }, + { + "epoch": 0.21995761374396897, + "flos": 19722328613280.0, + "grad_norm": 1.4825562427524273, + "learning_rate": 3.6333463280650165e-06, + "loss": 0.8384, + "num_input_tokens_seen": 79989320, + "step": 2439 + }, + { + "epoch": 0.22004779726743923, + "flos": 56815551496320.0, + "grad_norm": 0.703788867778278, + "learning_rate": 3.6330091101633126e-06, + "loss": 0.5694, + "num_input_tokens_seen": 80077595, + "step": 2440 + }, + { + "epoch": 0.2201379807909095, + "flos": 24646237361280.0, + "grad_norm": 2.6385863069316486, + "learning_rate": 3.632671752924267e-06, + "loss": 0.8158, + "num_input_tokens_seen": 80109225, + "step": 2441 + }, + { + "epoch": 0.22022816431437978, + "flos": 15605464324320.0, + "grad_norm": 2.2637815144460616, + "learning_rate": 3.632334256376665e-06, + "loss": 0.7839, + "num_input_tokens_seen": 80134485, + "step": 2442 + }, + { + "epoch": 0.22031834783785004, + "flos": 22897496638080.0, + "grad_norm": 1.8108054164742267, + "learning_rate": 3.6319966205493044e-06, + "loss": 0.8327, + "num_input_tokens_seen": 80163165, + "step": 2443 + }, + { + "epoch": 0.2204085313613203, + "flos": 56010737222880.0, + "grad_norm": 0.6925428498779028, + "learning_rate": 3.6316588454709922e-06, + "loss": 0.6527, + "num_input_tokens_seen": 80253000, + "step": 2444 + }, + { + "epoch": 0.22049871488479056, + "flos": 22095841794240.0, + "grad_norm": 1.8431924526513737, + "learning_rate": 3.6313209311705514e-06, + "loss": 0.7287, + "num_input_tokens_seen": 80280705, + "step": 2445 + }, + { + "epoch": 0.22058889840826082, + "flos": 16412099916000.0, + "grad_norm": 1.945476050185114, + "learning_rate": 3.6309828776768133e-06, + "loss": 0.8198, + "num_input_tokens_seen": 80306955, + "step": 2446 + }, + { + "epoch": 0.22067908193173108, + "flos": 28106468974560.0, + "grad_norm": 2.8200545681753346, + "learning_rate": 3.630644685018623e-06, + "loss": 0.7414, + "num_input_tokens_seen": 80336840, + "step": 2447 + }, + { + "epoch": 0.22076926545520134, + "flos": 22241124150720.0, + "grad_norm": 2.1085285890453402, + "learning_rate": 3.6303063532248367e-06, + "loss": 0.8565, + "num_input_tokens_seen": 80363195, + "step": 2448 + }, + { + "epoch": 0.2208594489786716, + "flos": 18668325303360.0, + "grad_norm": 2.0002405498802522, + "learning_rate": 3.6299678823243236e-06, + "loss": 0.8599, + "num_input_tokens_seen": 80390515, + "step": 2449 + }, + { + "epoch": 0.22094963250214186, + "flos": 22746564478080.0, + "grad_norm": 2.0485134410384607, + "learning_rate": 3.629629272345963e-06, + "loss": 0.6814, + "num_input_tokens_seen": 80418680, + "step": 2450 + }, + { + "epoch": 0.22103981602561212, + "flos": 20345391262080.0, + "grad_norm": 3.1976136965964375, + "learning_rate": 3.6292905233186468e-06, + "loss": 0.8991, + "num_input_tokens_seen": 80446210, + "step": 2451 + }, + { + "epoch": 0.22112999954908238, + "flos": 20923807295520.0, + "grad_norm": 2.1408638925278147, + "learning_rate": 3.6289516352712796e-06, + "loss": 0.739, + "num_input_tokens_seen": 80474460, + "step": 2452 + }, + { + "epoch": 0.22122018307255265, + "flos": 64454360667360.0, + "grad_norm": 0.7096299944095932, + "learning_rate": 3.6286126082327764e-06, + "loss": 0.6028, + "num_input_tokens_seen": 80560095, + "step": 2453 + }, + { + "epoch": 0.2213103665960229, + "flos": 19217854699680.0, + "grad_norm": 1.6084909325635504, + "learning_rate": 3.628273442232066e-06, + "loss": 0.8089, + "num_input_tokens_seen": 80588150, + "step": 2454 + }, + { + "epoch": 0.22140055011949317, + "flos": 27082913431680.0, + "grad_norm": 1.3389055620811174, + "learning_rate": 3.627934137298087e-06, + "loss": 0.7693, + "num_input_tokens_seen": 80618015, + "step": 2455 + }, + { + "epoch": 0.22149073364296343, + "flos": 24827840306880.0, + "grad_norm": 2.0823409832420583, + "learning_rate": 3.627594693459792e-06, + "loss": 0.7533, + "num_input_tokens_seen": 80646110, + "step": 2456 + }, + { + "epoch": 0.2215809171664337, + "flos": 17942694085920.0, + "grad_norm": 1.7908302469110082, + "learning_rate": 3.6272551107461424e-06, + "loss": 0.7366, + "num_input_tokens_seen": 80673570, + "step": 2457 + }, + { + "epoch": 0.22167110068990395, + "flos": 36523250118720.0, + "grad_norm": 1.7279932644456704, + "learning_rate": 3.6269153891861137e-06, + "loss": 0.7676, + "num_input_tokens_seen": 80707080, + "step": 2458 + }, + { + "epoch": 0.2217612842133742, + "flos": 34482346382880.0, + "grad_norm": 1.4557968515164237, + "learning_rate": 3.6265755288086944e-06, + "loss": 0.7473, + "num_input_tokens_seen": 80742820, + "step": 2459 + }, + { + "epoch": 0.22185146773684447, + "flos": 22168334293440.0, + "grad_norm": 2.4998884176148546, + "learning_rate": 3.626235529642881e-06, + "loss": 0.7548, + "num_input_tokens_seen": 80769715, + "step": 2460 + }, + { + "epoch": 0.22194165126031473, + "flos": 26103521297280.0, + "grad_norm": 1.8533692377486752, + "learning_rate": 3.625895391717686e-06, + "loss": 0.7117, + "num_input_tokens_seen": 80795405, + "step": 2461 + }, + { + "epoch": 0.222031834783785, + "flos": 43771899588960.0, + "grad_norm": 1.990086184029964, + "learning_rate": 3.625555115062131e-06, + "loss": 0.7457, + "num_input_tokens_seen": 80829900, + "step": 2462 + }, + { + "epoch": 0.22212201830725525, + "flos": 17099180358720.0, + "grad_norm": 7.5252190339410765, + "learning_rate": 3.6252146997052507e-06, + "loss": 0.8425, + "num_input_tokens_seen": 80856295, + "step": 2463 + }, + { + "epoch": 0.22221220183072551, + "flos": 70601752595520.0, + "grad_norm": 0.6672912865346953, + "learning_rate": 3.6248741456760898e-06, + "loss": 0.5596, + "num_input_tokens_seen": 80956960, + "step": 2464 + }, + { + "epoch": 0.22230238535419578, + "flos": 25663213856640.0, + "grad_norm": 2.326609585543673, + "learning_rate": 3.624533453003708e-06, + "loss": 0.7615, + "num_input_tokens_seen": 80988420, + "step": 2465 + }, + { + "epoch": 0.22239256887766606, + "flos": 18597877140960.0, + "grad_norm": 1.99063977725594, + "learning_rate": 3.6241926217171745e-06, + "loss": 0.7832, + "num_input_tokens_seen": 81013995, + "step": 2466 + }, + { + "epoch": 0.22248275240113632, + "flos": 30617830560000.0, + "grad_norm": 1.7763457126588134, + "learning_rate": 3.6238516518455703e-06, + "loss": 0.6948, + "num_input_tokens_seen": 81046330, + "step": 2467 + }, + { + "epoch": 0.22257293592460659, + "flos": 17249815160640.0, + "grad_norm": 3.730522179345784, + "learning_rate": 3.62351054341799e-06, + "loss": 0.899, + "num_input_tokens_seen": 81071930, + "step": 2468 + }, + { + "epoch": 0.22266311944807685, + "flos": 22749575228640.0, + "grad_norm": 1.919722553512512, + "learning_rate": 3.623169296463538e-06, + "loss": 0.8039, + "num_input_tokens_seen": 81100245, + "step": 2469 + }, + { + "epoch": 0.2227533029715471, + "flos": 22242350752800.0, + "grad_norm": 1.719662580106794, + "learning_rate": 3.6228279110113316e-06, + "loss": 0.7764, + "num_input_tokens_seen": 81128360, + "step": 2470 + }, + { + "epoch": 0.22284348649501737, + "flos": 18524678416320.0, + "grad_norm": 2.103667499014788, + "learning_rate": 3.6224863870904994e-06, + "loss": 0.7861, + "num_input_tokens_seen": 81155815, + "step": 2471 + }, + { + "epoch": 0.22293367001848763, + "flos": 25739274652800.0, + "grad_norm": 1.7532456137225478, + "learning_rate": 3.6221447247301827e-06, + "loss": 0.7922, + "num_input_tokens_seen": 81184585, + "step": 2472 + }, + { + "epoch": 0.2230238535419579, + "flos": 28872137757120.0, + "grad_norm": 2.270754354894205, + "learning_rate": 3.6218029239595332e-06, + "loss": 0.7275, + "num_input_tokens_seen": 81213565, + "step": 2473 + }, + { + "epoch": 0.22311403706542815, + "flos": 27300353759520.0, + "grad_norm": 2.128986098165712, + "learning_rate": 3.621460984807716e-06, + "loss": 0.8071, + "num_input_tokens_seen": 81241370, + "step": 2474 + }, + { + "epoch": 0.2232042205888984, + "flos": 20165237937120.0, + "grad_norm": 2.876579016907955, + "learning_rate": 3.621118907303907e-06, + "loss": 0.748, + "num_input_tokens_seen": 81270715, + "step": 2475 + }, + { + "epoch": 0.22329440411236867, + "flos": 18270768820320.0, + "grad_norm": 1.919115644277233, + "learning_rate": 3.620776691477294e-06, + "loss": 0.7721, + "num_input_tokens_seen": 81299060, + "step": 2476 + }, + { + "epoch": 0.22338458763583893, + "flos": 54218948184000.0, + "grad_norm": 0.6815797182870237, + "learning_rate": 3.6204343373570765e-06, + "loss": 0.5507, + "num_input_tokens_seen": 81382890, + "step": 2477 + }, + { + "epoch": 0.2234747711593092, + "flos": 31677966880320.0, + "grad_norm": 2.0686287420871614, + "learning_rate": 3.620091844972467e-06, + "loss": 0.8015, + "num_input_tokens_seen": 81411185, + "step": 2478 + }, + { + "epoch": 0.22356495468277945, + "flos": 19909878720480.0, + "grad_norm": 2.434852140038006, + "learning_rate": 3.619749214352688e-06, + "loss": 0.7985, + "num_input_tokens_seen": 81437975, + "step": 2479 + }, + { + "epoch": 0.22365513820624972, + "flos": 17833026093120.0, + "grad_norm": 2.1723880299114247, + "learning_rate": 3.6194064455269744e-06, + "loss": 0.8817, + "num_input_tokens_seen": 81464290, + "step": 2480 + }, + { + "epoch": 0.22374532172971998, + "flos": 22605779662560.0, + "grad_norm": 2.2487244699247135, + "learning_rate": 3.6190635385245737e-06, + "loss": 0.6995, + "num_input_tokens_seen": 81493750, + "step": 2481 + }, + { + "epoch": 0.22383550525319024, + "flos": 21475789896000.0, + "grad_norm": 1.9633947531651885, + "learning_rate": 3.618720493374745e-06, + "loss": 0.7109, + "num_input_tokens_seen": 81523705, + "step": 2482 + }, + { + "epoch": 0.2239256887766605, + "flos": 27009268669920.0, + "grad_norm": 1.9854315400608893, + "learning_rate": 3.6183773101067575e-06, + "loss": 0.6378, + "num_input_tokens_seen": 81552960, + "step": 2483 + }, + { + "epoch": 0.22401587230013076, + "flos": 28981619901120.0, + "grad_norm": 1.7373916921545853, + "learning_rate": 3.6180339887498948e-06, + "loss": 0.6992, + "num_input_tokens_seen": 81584510, + "step": 2484 + }, + { + "epoch": 0.22410605582360102, + "flos": 20995110362400.0, + "grad_norm": 2.0860279077947133, + "learning_rate": 3.61769052933345e-06, + "loss": 0.781, + "num_input_tokens_seen": 81610535, + "step": 2485 + }, + { + "epoch": 0.22419623934707128, + "flos": 12185381785440.0, + "grad_norm": 2.074460462721636, + "learning_rate": 3.6173469318867297e-06, + "loss": 0.9323, + "num_input_tokens_seen": 81632275, + "step": 2486 + }, + { + "epoch": 0.22428642287054154, + "flos": 21877509392160.0, + "grad_norm": 2.193456224730357, + "learning_rate": 3.617003196439051e-06, + "loss": 0.7959, + "num_input_tokens_seen": 81660160, + "step": 2487 + }, + { + "epoch": 0.2243766063940118, + "flos": 28690014434880.0, + "grad_norm": 1.975023597866032, + "learning_rate": 3.616659323019744e-06, + "loss": 0.7902, + "num_input_tokens_seen": 81692085, + "step": 2488 + }, + { + "epoch": 0.22446678991748206, + "flos": 23329180694400.0, + "grad_norm": 2.1804352164893332, + "learning_rate": 3.616315311658149e-06, + "loss": 0.8062, + "num_input_tokens_seen": 81720185, + "step": 2489 + }, + { + "epoch": 0.22455697344095235, + "flos": 21695125881600.0, + "grad_norm": 1.9115771292238457, + "learning_rate": 3.6159711623836195e-06, + "loss": 0.8329, + "num_input_tokens_seen": 81750440, + "step": 2490 + }, + { + "epoch": 0.2246471569644226, + "flos": 71032804765920.0, + "grad_norm": 0.6475962961974562, + "learning_rate": 3.6156268752255203e-06, + "loss": 0.5644, + "num_input_tokens_seen": 81848550, + "step": 2491 + }, + { + "epoch": 0.22473734048789287, + "flos": 29197982305920.0, + "grad_norm": 1.9271251020973992, + "learning_rate": 3.615282450213227e-06, + "loss": 0.7418, + "num_input_tokens_seen": 81878200, + "step": 2492 + }, + { + "epoch": 0.22482752401136313, + "flos": 20966186555520.0, + "grad_norm": 1.6310152082172733, + "learning_rate": 3.614937887376128e-06, + "loss": 0.792, + "num_input_tokens_seen": 81906495, + "step": 2493 + }, + { + "epoch": 0.2249177075348334, + "flos": 26248952332800.0, + "grad_norm": 1.9692872540518174, + "learning_rate": 3.614593186743625e-06, + "loss": 0.793, + "num_input_tokens_seen": 81933830, + "step": 2494 + }, + { + "epoch": 0.22500789105830366, + "flos": 19217222813760.0, + "grad_norm": 2.37882657223016, + "learning_rate": 3.614248348345128e-06, + "loss": 0.7113, + "num_input_tokens_seen": 81959565, + "step": 2495 + }, + { + "epoch": 0.22509807458177392, + "flos": 25991660288640.0, + "grad_norm": 2.261113166443653, + "learning_rate": 3.6139033722100614e-06, + "loss": 0.8277, + "num_input_tokens_seen": 81988275, + "step": 2496 + }, + { + "epoch": 0.22518825810524418, + "flos": 18634049051040.0, + "grad_norm": 2.766767535228455, + "learning_rate": 3.6135582583678596e-06, + "loss": 0.8558, + "num_input_tokens_seen": 82013115, + "step": 2497 + }, + { + "epoch": 0.22527844162871444, + "flos": 12950641700640.0, + "grad_norm": 2.0793810038914184, + "learning_rate": 3.61321300684797e-06, + "loss": 0.796, + "num_input_tokens_seen": 82039355, + "step": 2498 + }, + { + "epoch": 0.2253686251521847, + "flos": 25193164874400.0, + "grad_norm": 1.477706466036139, + "learning_rate": 3.6128676176798527e-06, + "loss": 0.8023, + "num_input_tokens_seen": 82071555, + "step": 2499 + }, + { + "epoch": 0.22545880867565496, + "flos": 27813265208640.0, + "grad_norm": 1.8855910317178148, + "learning_rate": 3.612522090892976e-06, + "loss": 0.7969, + "num_input_tokens_seen": 82102790, + "step": 2500 + }, + { + "epoch": 0.22554899219912522, + "flos": 47889024066240.0, + "grad_norm": 1.4652166680262975, + "learning_rate": 3.6121764265168232e-06, + "loss": 0.7459, + "num_input_tokens_seen": 82135155, + "step": 2501 + }, + { + "epoch": 0.22563917572259548, + "flos": 22970063476800.0, + "grad_norm": 1.534517499906577, + "learning_rate": 3.611830624580888e-06, + "loss": 0.8042, + "num_input_tokens_seen": 82164565, + "step": 2502 + }, + { + "epoch": 0.22572935924606574, + "flos": 17286321598560.0, + "grad_norm": 1.7685796271450405, + "learning_rate": 3.6114846851146767e-06, + "loss": 0.7911, + "num_input_tokens_seen": 82191345, + "step": 2503 + }, + { + "epoch": 0.225819542769536, + "flos": 37359404233440.0, + "grad_norm": 1.9313067233449615, + "learning_rate": 3.6111386081477068e-06, + "loss": 0.6855, + "num_input_tokens_seen": 82222865, + "step": 2504 + }, + { + "epoch": 0.22590972629300626, + "flos": 30981519658080.0, + "grad_norm": 1.805663336283825, + "learning_rate": 3.6107923937095066e-06, + "loss": 0.7987, + "num_input_tokens_seen": 82253950, + "step": 2505 + }, + { + "epoch": 0.22599990981647652, + "flos": 19107963688320.0, + "grad_norm": 2.422938845383436, + "learning_rate": 3.6104460418296173e-06, + "loss": 0.7003, + "num_input_tokens_seen": 82280645, + "step": 2506 + }, + { + "epoch": 0.22609009333994678, + "flos": 24244517865120.0, + "grad_norm": 1.8906985812410215, + "learning_rate": 3.6100995525375924e-06, + "loss": 0.8293, + "num_input_tokens_seen": 82306575, + "step": 2507 + }, + { + "epoch": 0.22618027686341705, + "flos": 30435632898240.0, + "grad_norm": 1.5632449091052578, + "learning_rate": 3.6097529258629952e-06, + "loss": 0.7243, + "num_input_tokens_seen": 82334005, + "step": 2508 + }, + { + "epoch": 0.2262704603868873, + "flos": 21695794937280.0, + "grad_norm": 2.341398050868558, + "learning_rate": 3.6094061618354027e-06, + "loss": 0.8072, + "num_input_tokens_seen": 82362865, + "step": 2509 + }, + { + "epoch": 0.22636064391035757, + "flos": 22855080208320.0, + "grad_norm": 1.884887122156681, + "learning_rate": 3.609059260484402e-06, + "loss": 0.7771, + "num_input_tokens_seen": 82391875, + "step": 2510 + }, + { + "epoch": 0.22645082743382783, + "flos": 21402107964480.0, + "grad_norm": 3.5680425260002426, + "learning_rate": 3.6087122218395935e-06, + "loss": 0.72, + "num_input_tokens_seen": 82414240, + "step": 2511 + }, + { + "epoch": 0.2265410109572981, + "flos": 23221111001280.0, + "grad_norm": 1.5955440670755097, + "learning_rate": 3.608365045930587e-06, + "loss": 0.7479, + "num_input_tokens_seen": 82442350, + "step": 2512 + }, + { + "epoch": 0.22663119448076835, + "flos": 22600538726400.0, + "grad_norm": 1.9895780340260605, + "learning_rate": 3.608017732787007e-06, + "loss": 0.7686, + "num_input_tokens_seen": 82469720, + "step": 2513 + }, + { + "epoch": 0.22672137800423864, + "flos": 17828454212640.0, + "grad_norm": 1.6583964995553835, + "learning_rate": 3.6076702824384875e-06, + "loss": 0.7118, + "num_input_tokens_seen": 82496260, + "step": 2514 + }, + { + "epoch": 0.2268115615277089, + "flos": 31131002197440.0, + "grad_norm": 1.7265841797723296, + "learning_rate": 3.607322694914675e-06, + "loss": 0.7965, + "num_input_tokens_seen": 82525600, + "step": 2515 + }, + { + "epoch": 0.22690174505117916, + "flos": 19509311486880.0, + "grad_norm": 1.9839922318037295, + "learning_rate": 3.606974970245227e-06, + "loss": 0.7509, + "num_input_tokens_seen": 82554965, + "step": 2516 + }, + { + "epoch": 0.22699192857464942, + "flos": 31167694484160.0, + "grad_norm": 1.8153294150335524, + "learning_rate": 3.606627108459814e-06, + "loss": 0.732, + "num_input_tokens_seen": 82584915, + "step": 2517 + }, + { + "epoch": 0.22708211209811968, + "flos": 20637925972320.0, + "grad_norm": 1.716801388401526, + "learning_rate": 3.6062791095881174e-06, + "loss": 0.7059, + "num_input_tokens_seen": 82613635, + "step": 2518 + }, + { + "epoch": 0.22717229562158994, + "flos": 26504200040160.0, + "grad_norm": 2.0548021407835244, + "learning_rate": 3.6059309736598303e-06, + "loss": 0.8519, + "num_input_tokens_seen": 82641115, + "step": 2519 + }, + { + "epoch": 0.2272624791450602, + "flos": 13605378718560.0, + "grad_norm": 2.1454701409485764, + "learning_rate": 3.605582700704657e-06, + "loss": 0.7587, + "num_input_tokens_seen": 82666580, + "step": 2520 + }, + { + "epoch": 0.22735266266853046, + "flos": 18887215251840.0, + "grad_norm": 1.7133586596436625, + "learning_rate": 3.6052342907523146e-06, + "loss": 0.7832, + "num_input_tokens_seen": 82695210, + "step": 2521 + }, + { + "epoch": 0.22744284619200072, + "flos": 22781398276800.0, + "grad_norm": 2.1812079264345856, + "learning_rate": 3.604885743832532e-06, + "loss": 0.708, + "num_input_tokens_seen": 82721340, + "step": 2522 + }, + { + "epoch": 0.22753302971547099, + "flos": 57713530389120.0, + "grad_norm": 0.6929387495735485, + "learning_rate": 3.6045370599750482e-06, + "loss": 0.58, + "num_input_tokens_seen": 82806030, + "step": 2523 + }, + { + "epoch": 0.22762321323894125, + "flos": 21002655823680.0, + "grad_norm": 1.732044682087174, + "learning_rate": 3.604188239209615e-06, + "loss": 0.7577, + "num_input_tokens_seen": 82833900, + "step": 2524 + }, + { + "epoch": 0.2277133967624115, + "flos": 49711112193120.0, + "grad_norm": 2.451650869066631, + "learning_rate": 3.603839281565996e-06, + "loss": 0.7009, + "num_input_tokens_seen": 82867690, + "step": 2525 + }, + { + "epoch": 0.22780358028588177, + "flos": 16702924817280.0, + "grad_norm": 2.1629469437214035, + "learning_rate": 3.603490187073966e-06, + "loss": 0.7964, + "num_input_tokens_seen": 82894820, + "step": 2526 + }, + { + "epoch": 0.22789376380935203, + "flos": 26030731440000.0, + "grad_norm": 2.509131159164295, + "learning_rate": 3.6031409557633117e-06, + "loss": 0.8366, + "num_input_tokens_seen": 82924450, + "step": 2527 + }, + { + "epoch": 0.2279839473328223, + "flos": 20710455641280.0, + "grad_norm": 2.892045945094088, + "learning_rate": 3.602791587663831e-06, + "loss": 0.818, + "num_input_tokens_seen": 82952500, + "step": 2528 + }, + { + "epoch": 0.22807413085629255, + "flos": 21183255185760.0, + "grad_norm": 1.6190343069549724, + "learning_rate": 3.6024420828053348e-06, + "loss": 0.8164, + "num_input_tokens_seen": 82981570, + "step": 2529 + }, + { + "epoch": 0.2281643143797628, + "flos": 16267635294240.0, + "grad_norm": 2.238565423578156, + "learning_rate": 3.6020924412176445e-06, + "loss": 0.849, + "num_input_tokens_seen": 83009265, + "step": 2530 + }, + { + "epoch": 0.22825449790323307, + "flos": 25993407267360.0, + "grad_norm": 2.922647792189045, + "learning_rate": 3.601742662930593e-06, + "loss": 0.7485, + "num_input_tokens_seen": 83040210, + "step": 2531 + }, + { + "epoch": 0.22834468142670333, + "flos": 21841448991360.0, + "grad_norm": 1.7352659669115944, + "learning_rate": 3.6013927479740248e-06, + "loss": 0.7429, + "num_input_tokens_seen": 83069035, + "step": 2532 + }, + { + "epoch": 0.2284348649501736, + "flos": 16886237571840.0, + "grad_norm": 1.9989558713048656, + "learning_rate": 3.6010426963777985e-06, + "loss": 0.8573, + "num_input_tokens_seen": 83095710, + "step": 2533 + }, + { + "epoch": 0.22852504847364385, + "flos": 17793062867520.0, + "grad_norm": 1.9783472740798393, + "learning_rate": 3.6006925081717804e-06, + "loss": 0.8237, + "num_input_tokens_seen": 83121875, + "step": 2534 + }, + { + "epoch": 0.22861523199711412, + "flos": 26976999584640.0, + "grad_norm": 1.8012249249017405, + "learning_rate": 3.600342183385852e-06, + "loss": 0.7485, + "num_input_tokens_seen": 83152885, + "step": 2535 + }, + { + "epoch": 0.22870541552058438, + "flos": 20857076109120.0, + "grad_norm": 1.9397228779431206, + "learning_rate": 3.5999917220499043e-06, + "loss": 0.7175, + "num_input_tokens_seen": 83180340, + "step": 2536 + }, + { + "epoch": 0.22879559904405466, + "flos": 23042370127200.0, + "grad_norm": 1.9560364652366713, + "learning_rate": 3.5996411241938404e-06, + "loss": 0.7816, + "num_input_tokens_seen": 83207955, + "step": 2537 + }, + { + "epoch": 0.22888578256752493, + "flos": 21401996455200.0, + "grad_norm": 1.7629147548871038, + "learning_rate": 3.5992903898475752e-06, + "loss": 0.7343, + "num_input_tokens_seen": 83237670, + "step": 2538 + }, + { + "epoch": 0.2289759660909952, + "flos": 24169683671040.0, + "grad_norm": 1.502132568751711, + "learning_rate": 3.5989395190410365e-06, + "loss": 0.8744, + "num_input_tokens_seen": 83266525, + "step": 2539 + }, + { + "epoch": 0.22906614961446545, + "flos": 24313553576640.0, + "grad_norm": 1.8400116214177873, + "learning_rate": 3.598588511804161e-06, + "loss": 0.7843, + "num_input_tokens_seen": 83296455, + "step": 2540 + }, + { + "epoch": 0.2291563331379357, + "flos": 26902462748640.0, + "grad_norm": 2.4780369169304, + "learning_rate": 3.5982373681668987e-06, + "loss": 0.7856, + "num_input_tokens_seen": 83320205, + "step": 2541 + }, + { + "epoch": 0.22924651666140597, + "flos": 14517890987520.0, + "grad_norm": 1.9724879429089712, + "learning_rate": 3.597886088159212e-06, + "loss": 0.787, + "num_input_tokens_seen": 83348550, + "step": 2542 + }, + { + "epoch": 0.22933670018487623, + "flos": 45667558138080.0, + "grad_norm": 1.6212819920826735, + "learning_rate": 3.597534671811074e-06, + "loss": 0.7094, + "num_input_tokens_seen": 83382335, + "step": 2543 + }, + { + "epoch": 0.2294268837083465, + "flos": 29017903320480.0, + "grad_norm": 13.16435459616588, + "learning_rate": 3.5971831191524684e-06, + "loss": 0.7583, + "num_input_tokens_seen": 83412010, + "step": 2544 + }, + { + "epoch": 0.22951706723181675, + "flos": 25297889289120.0, + "grad_norm": 2.1959362297874154, + "learning_rate": 3.5968314302133925e-06, + "loss": 0.8287, + "num_input_tokens_seen": 83439585, + "step": 2545 + }, + { + "epoch": 0.229607250755287, + "flos": 16958692901280.0, + "grad_norm": 2.095792542724855, + "learning_rate": 3.596479605023854e-06, + "loss": 0.7773, + "num_input_tokens_seen": 83463805, + "step": 2546 + }, + { + "epoch": 0.22969743427875727, + "flos": 11165468879040.0, + "grad_norm": 1.6464244971920892, + "learning_rate": 3.596127643613873e-06, + "loss": 0.7928, + "num_input_tokens_seen": 83488390, + "step": 2547 + }, + { + "epoch": 0.22978761780222753, + "flos": 24866205232800.0, + "grad_norm": 1.7735953239761264, + "learning_rate": 3.59577554601348e-06, + "loss": 0.7993, + "num_input_tokens_seen": 83518010, + "step": 2548 + }, + { + "epoch": 0.2298778013256978, + "flos": 21980970035040.0, + "grad_norm": 2.506291674032028, + "learning_rate": 3.595423312252719e-06, + "loss": 0.8296, + "num_input_tokens_seen": 83547990, + "step": 2549 + }, + { + "epoch": 0.22996798484916806, + "flos": 22678792538400.0, + "grad_norm": 1.7549717306919659, + "learning_rate": 3.5950709423616436e-06, + "loss": 0.7616, + "num_input_tokens_seen": 83574635, + "step": 2550 + }, + { + "epoch": 0.23005816837263832, + "flos": 29017568792640.0, + "grad_norm": 2.7321994519182264, + "learning_rate": 3.5947184363703203e-06, + "loss": 0.7572, + "num_input_tokens_seen": 83603800, + "step": 2551 + }, + { + "epoch": 0.23014835189610858, + "flos": 31162156189920.0, + "grad_norm": 2.8594266959765005, + "learning_rate": 3.5943657943088274e-06, + "loss": 0.7292, + "num_input_tokens_seen": 83635100, + "step": 2552 + }, + { + "epoch": 0.23023853541957884, + "flos": 29164263600000.0, + "grad_norm": 1.6362067175051744, + "learning_rate": 3.5940130162072525e-06, + "loss": 0.704, + "num_input_tokens_seen": 83668160, + "step": 2553 + }, + { + "epoch": 0.2303287189430491, + "flos": 20893508207520.0, + "grad_norm": 3.937930129036303, + "learning_rate": 3.5936601020956985e-06, + "loss": 0.7814, + "num_input_tokens_seen": 83697125, + "step": 2554 + }, + { + "epoch": 0.23041890246651936, + "flos": 20745846986400.0, + "grad_norm": 1.85246744052624, + "learning_rate": 3.5933070520042772e-06, + "loss": 0.8118, + "num_input_tokens_seen": 83725065, + "step": 2555 + }, + { + "epoch": 0.23050908598998962, + "flos": 52823669141280.0, + "grad_norm": 0.775512147696392, + "learning_rate": 3.5929538659631133e-06, + "loss": 0.63, + "num_input_tokens_seen": 83814390, + "step": 2556 + }, + { + "epoch": 0.23059926951345988, + "flos": 15355680571680.0, + "grad_norm": 1.8942646900909832, + "learning_rate": 3.592600544002341e-06, + "loss": 0.8108, + "num_input_tokens_seen": 83841525, + "step": 2557 + }, + { + "epoch": 0.23068945303693014, + "flos": 21331027916160.0, + "grad_norm": 1.8207646410079876, + "learning_rate": 3.5922470861521098e-06, + "loss": 0.7822, + "num_input_tokens_seen": 83868545, + "step": 2558 + }, + { + "epoch": 0.2307796365604004, + "flos": 19763852968800.0, + "grad_norm": 6.080393389975312, + "learning_rate": 3.591893492442577e-06, + "loss": 0.7486, + "num_input_tokens_seen": 83894695, + "step": 2559 + }, + { + "epoch": 0.23086982008387066, + "flos": 21871599400320.0, + "grad_norm": 1.7929594417820744, + "learning_rate": 3.591539762903914e-06, + "loss": 0.8357, + "num_input_tokens_seen": 83920430, + "step": 2560 + }, + { + "epoch": 0.23096000360734095, + "flos": 23808261928320.0, + "grad_norm": 1.7920024320637002, + "learning_rate": 3.591185897566303e-06, + "loss": 0.7077, + "num_input_tokens_seen": 83950765, + "step": 2561 + }, + { + "epoch": 0.2310501871308112, + "flos": 27524075776800.0, + "grad_norm": 1.827819016109597, + "learning_rate": 3.590831896459937e-06, + "loss": 0.7769, + "num_input_tokens_seen": 83980115, + "step": 2562 + }, + { + "epoch": 0.23114037065428147, + "flos": 26868558193920.0, + "grad_norm": 1.9344302439157763, + "learning_rate": 3.5904777596150222e-06, + "loss": 0.7832, + "num_input_tokens_seen": 84009970, + "step": 2563 + }, + { + "epoch": 0.23123055417775173, + "flos": 20601977080800.0, + "grad_norm": 2.189282154550639, + "learning_rate": 3.590123487061775e-06, + "loss": 0.7761, + "num_input_tokens_seen": 84037430, + "step": 2564 + }, + { + "epoch": 0.231320737701222, + "flos": 26977631470560.0, + "grad_norm": 2.383897778111636, + "learning_rate": 3.589769078830424e-06, + "loss": 0.8745, + "num_input_tokens_seen": 84066705, + "step": 2565 + }, + { + "epoch": 0.23141092122469226, + "flos": 35757804354720.0, + "grad_norm": 1.9537322664340782, + "learning_rate": 3.58941453495121e-06, + "loss": 0.7581, + "num_input_tokens_seen": 84096680, + "step": 2566 + }, + { + "epoch": 0.23150110474816252, + "flos": 21330321690720.0, + "grad_norm": 2.8009217039507903, + "learning_rate": 3.5890598554543834e-06, + "loss": 0.7894, + "num_input_tokens_seen": 84122735, + "step": 2567 + }, + { + "epoch": 0.23159128827163278, + "flos": 21877323543360.0, + "grad_norm": 3.0092402252725585, + "learning_rate": 3.5887050403702073e-06, + "loss": 0.8279, + "num_input_tokens_seen": 84149035, + "step": 2568 + }, + { + "epoch": 0.23168147179510304, + "flos": 18598137329280.0, + "grad_norm": 1.7531483123371676, + "learning_rate": 3.588350089728958e-06, + "loss": 0.7543, + "num_input_tokens_seen": 84176605, + "step": 2569 + }, + { + "epoch": 0.2317716553185733, + "flos": 18562114098240.0, + "grad_norm": 2.3344843500511083, + "learning_rate": 3.5879950035609204e-06, + "loss": 0.8034, + "num_input_tokens_seen": 84202780, + "step": 2570 + }, + { + "epoch": 0.23186183884204356, + "flos": 26212891932000.0, + "grad_norm": 2.2578871559551197, + "learning_rate": 3.5876397818963933e-06, + "loss": 0.777, + "num_input_tokens_seen": 84231545, + "step": 2571 + }, + { + "epoch": 0.23195202236551382, + "flos": 21658247746080.0, + "grad_norm": 1.5873083194143343, + "learning_rate": 3.5872844247656858e-06, + "loss": 0.8352, + "num_input_tokens_seen": 84261015, + "step": 2572 + }, + { + "epoch": 0.23204220588898408, + "flos": 23516024576160.0, + "grad_norm": 1.6095958669268138, + "learning_rate": 3.5869289321991195e-06, + "loss": 0.7883, + "num_input_tokens_seen": 84291405, + "step": 2573 + }, + { + "epoch": 0.23213238941245434, + "flos": 24973680209760.0, + "grad_norm": 1.9506771770271496, + "learning_rate": 3.5865733042270263e-06, + "loss": 0.7722, + "num_input_tokens_seen": 84322565, + "step": 2574 + }, + { + "epoch": 0.2322225729359246, + "flos": 23225943070080.0, + "grad_norm": 2.0579757776271443, + "learning_rate": 3.5862175408797498e-06, + "loss": 0.8075, + "num_input_tokens_seen": 84351910, + "step": 2575 + }, + { + "epoch": 0.23231275645939486, + "flos": 27087299463360.0, + "grad_norm": 1.6915034382937693, + "learning_rate": 3.585861642187647e-06, + "loss": 0.7174, + "num_input_tokens_seen": 84381620, + "step": 2576 + }, + { + "epoch": 0.23240293998286513, + "flos": 26759410577760.0, + "grad_norm": 1.5122337548343028, + "learning_rate": 3.5855056081810845e-06, + "loss": 0.7755, + "num_input_tokens_seen": 84409905, + "step": 2577 + }, + { + "epoch": 0.2324931235063354, + "flos": 24388610789280.0, + "grad_norm": 1.5254273550615474, + "learning_rate": 3.5851494388904406e-06, + "loss": 0.78, + "num_input_tokens_seen": 84440315, + "step": 2578 + }, + { + "epoch": 0.23258330702980565, + "flos": 28325693450880.0, + "grad_norm": 2.292116178873491, + "learning_rate": 3.5847931343461064e-06, + "loss": 0.7163, + "num_input_tokens_seen": 84469750, + "step": 2579 + }, + { + "epoch": 0.2326734905532759, + "flos": 17796742673760.0, + "grad_norm": 1.8721781795589985, + "learning_rate": 3.5844366945784835e-06, + "loss": 0.7055, + "num_input_tokens_seen": 84496140, + "step": 2580 + }, + { + "epoch": 0.23276367407674617, + "flos": 18196975379520.0, + "grad_norm": 2.3698854085525762, + "learning_rate": 3.5840801196179856e-06, + "loss": 0.8352, + "num_input_tokens_seen": 84520645, + "step": 2581 + }, + { + "epoch": 0.23285385760021643, + "flos": 25184578659840.0, + "grad_norm": 1.891603201026975, + "learning_rate": 3.583723409495037e-06, + "loss": 0.8339, + "num_input_tokens_seen": 84546225, + "step": 2582 + }, + { + "epoch": 0.2329440411236867, + "flos": 30146406296640.0, + "grad_norm": 2.2059214274367656, + "learning_rate": 3.5833665642400747e-06, + "loss": 0.8331, + "num_input_tokens_seen": 84574960, + "step": 2583 + }, + { + "epoch": 0.23303422464715695, + "flos": 21765759892800.0, + "grad_norm": 2.155363370525417, + "learning_rate": 3.5830095838835472e-06, + "loss": 0.705, + "num_input_tokens_seen": 84602840, + "step": 2584 + }, + { + "epoch": 0.23312440817062724, + "flos": 20963250144480.0, + "grad_norm": 1.7433624411147206, + "learning_rate": 3.5826524684559125e-06, + "loss": 0.7491, + "num_input_tokens_seen": 84631400, + "step": 2585 + }, + { + "epoch": 0.2332145916940975, + "flos": 21913383944160.0, + "grad_norm": 2.1757455271556894, + "learning_rate": 3.5822952179876433e-06, + "loss": 0.8598, + "num_input_tokens_seen": 84658845, + "step": 2586 + }, + { + "epoch": 0.23330477521756776, + "flos": 30801886709760.0, + "grad_norm": 1.5288929606831345, + "learning_rate": 3.5819378325092205e-06, + "loss": 0.749, + "num_input_tokens_seen": 84690730, + "step": 2587 + }, + { + "epoch": 0.23339495874103802, + "flos": 38266229529120.0, + "grad_norm": 4.666195482354015, + "learning_rate": 3.581580312051139e-06, + "loss": 0.8004, + "num_input_tokens_seen": 84720750, + "step": 2588 + }, + { + "epoch": 0.23348514226450828, + "flos": 17687706566880.0, + "grad_norm": 1.7255053844119181, + "learning_rate": 3.5812226566439057e-06, + "loss": 0.705, + "num_input_tokens_seen": 84748225, + "step": 2589 + }, + { + "epoch": 0.23357532578797854, + "flos": 21658730952960.0, + "grad_norm": 1.9089257893433078, + "learning_rate": 3.580864866318036e-06, + "loss": 0.7555, + "num_input_tokens_seen": 84777725, + "step": 2590 + }, + { + "epoch": 0.2336655093114488, + "flos": 21367162656480.0, + "grad_norm": 2.14428787379103, + "learning_rate": 3.580506941104059e-06, + "loss": 0.7845, + "num_input_tokens_seen": 84805715, + "step": 2591 + }, + { + "epoch": 0.23375569283491907, + "flos": 30912112248960.0, + "grad_norm": 1.744022036426332, + "learning_rate": 3.580148881032515e-06, + "loss": 0.7664, + "num_input_tokens_seen": 84835485, + "step": 2592 + }, + { + "epoch": 0.23384587635838933, + "flos": 24171616498560.0, + "grad_norm": 1.8804626363314942, + "learning_rate": 3.5797906861339556e-06, + "loss": 0.7829, + "num_input_tokens_seen": 84865865, + "step": 2593 + }, + { + "epoch": 0.2339360598818596, + "flos": 43989637274880.0, + "grad_norm": 1.9002215632702446, + "learning_rate": 3.5794323564389435e-06, + "loss": 0.8037, + "num_input_tokens_seen": 84897965, + "step": 2594 + }, + { + "epoch": 0.23402624340532985, + "flos": 22933334020320.0, + "grad_norm": 2.438689210760754, + "learning_rate": 3.579073891978055e-06, + "loss": 0.7919, + "num_input_tokens_seen": 84923815, + "step": 2595 + }, + { + "epoch": 0.2341164269288001, + "flos": 20857001769600.0, + "grad_norm": 2.8545464775777196, + "learning_rate": 3.5787152927818746e-06, + "loss": 0.7496, + "num_input_tokens_seen": 84953215, + "step": 2596 + }, + { + "epoch": 0.23420661045227037, + "flos": 26758778691840.0, + "grad_norm": 1.62557885057178, + "learning_rate": 3.5783565588810003e-06, + "loss": 0.8433, + "num_input_tokens_seen": 84983335, + "step": 2597 + }, + { + "epoch": 0.23429679397574063, + "flos": 20747482455840.0, + "grad_norm": 1.7594774965070517, + "learning_rate": 3.5779976903060412e-06, + "loss": 0.7504, + "num_input_tokens_seen": 85012275, + "step": 2598 + }, + { + "epoch": 0.2343869774992109, + "flos": 25338595909920.0, + "grad_norm": 1.679156410179651, + "learning_rate": 3.577638687087619e-06, + "loss": 0.8075, + "num_input_tokens_seen": 85041080, + "step": 2599 + }, + { + "epoch": 0.23447716102268115, + "flos": 23152298308320.0, + "grad_norm": 1.732611988736247, + "learning_rate": 3.577279549256364e-06, + "loss": 0.786, + "num_input_tokens_seen": 85069570, + "step": 2600 + }, + { + "epoch": 0.2345673445461514, + "flos": 24172359893760.0, + "grad_norm": 2.1022786901890873, + "learning_rate": 3.5769202768429213e-06, + "loss": 0.8247, + "num_input_tokens_seen": 85098260, + "step": 2601 + }, + { + "epoch": 0.23465752806962167, + "flos": 28981062354720.0, + "grad_norm": 2.365150204459442, + "learning_rate": 3.5765608698779454e-06, + "loss": 0.7609, + "num_input_tokens_seen": 85126080, + "step": 2602 + }, + { + "epoch": 0.23474771159309193, + "flos": 20893805565600.0, + "grad_norm": 1.8407189622091429, + "learning_rate": 3.5762013283921033e-06, + "loss": 0.7881, + "num_input_tokens_seen": 85155110, + "step": 2603 + }, + { + "epoch": 0.2348378951165622, + "flos": 22131567667200.0, + "grad_norm": 4.584550846110368, + "learning_rate": 3.5758416524160728e-06, + "loss": 0.7825, + "num_input_tokens_seen": 85183410, + "step": 2604 + }, + { + "epoch": 0.23492807864003246, + "flos": 21181099339680.0, + "grad_norm": 1.8725965990887363, + "learning_rate": 3.5754818419805427e-06, + "loss": 0.7723, + "num_input_tokens_seen": 85210435, + "step": 2605 + }, + { + "epoch": 0.23501826216350272, + "flos": 28580383611840.0, + "grad_norm": 1.489216428087902, + "learning_rate": 3.575121897116216e-06, + "loss": 0.7601, + "num_input_tokens_seen": 85243790, + "step": 2606 + }, + { + "epoch": 0.23510844568697298, + "flos": 30038225094240.0, + "grad_norm": 1.5135268800262154, + "learning_rate": 3.574761817853803e-06, + "loss": 0.8172, + "num_input_tokens_seen": 85271920, + "step": 2607 + }, + { + "epoch": 0.23519862921044324, + "flos": 31713544074240.0, + "grad_norm": 1.4513914370882013, + "learning_rate": 3.5744016042240287e-06, + "loss": 0.7158, + "num_input_tokens_seen": 85304635, + "step": 2608 + }, + { + "epoch": 0.23528881273391353, + "flos": 24609024697920.0, + "grad_norm": 2.125986319130977, + "learning_rate": 3.5740412562576286e-06, + "loss": 0.7154, + "num_input_tokens_seen": 85332630, + "step": 2609 + }, + { + "epoch": 0.2353789962573838, + "flos": 24425823452640.0, + "grad_norm": 1.9294731855108396, + "learning_rate": 3.573680773985349e-06, + "loss": 0.7749, + "num_input_tokens_seen": 85360660, + "step": 2610 + }, + { + "epoch": 0.23546917978085405, + "flos": 32256754611360.0, + "grad_norm": 2.636876803711941, + "learning_rate": 3.5733201574379486e-06, + "loss": 0.7741, + "num_input_tokens_seen": 85389915, + "step": 2611 + }, + { + "epoch": 0.2355593633043243, + "flos": 23735026033920.0, + "grad_norm": 1.8514578101584251, + "learning_rate": 3.5729594066461975e-06, + "loss": 0.7162, + "num_input_tokens_seen": 85419890, + "step": 2612 + }, + { + "epoch": 0.23564954682779457, + "flos": 28652950450560.0, + "grad_norm": 1.6047663950631565, + "learning_rate": 3.572598521640876e-06, + "loss": 0.7952, + "num_input_tokens_seen": 85450310, + "step": 2613 + }, + { + "epoch": 0.23573973035126483, + "flos": 61968466101120.0, + "grad_norm": 0.8007067017629348, + "learning_rate": 3.5722375024527782e-06, + "loss": 0.6342, + "num_input_tokens_seen": 85537770, + "step": 2614 + }, + { + "epoch": 0.2358299138747351, + "flos": 23622533139360.0, + "grad_norm": 2.591571766464059, + "learning_rate": 3.571876349112707e-06, + "loss": 0.8004, + "num_input_tokens_seen": 85566035, + "step": 2615 + }, + { + "epoch": 0.23592009739820535, + "flos": 27047707935360.0, + "grad_norm": 1.858172757291733, + "learning_rate": 3.5715150616514784e-06, + "loss": 0.7783, + "num_input_tokens_seen": 85594945, + "step": 2616 + }, + { + "epoch": 0.2360102809216756, + "flos": 16776532409280.0, + "grad_norm": 2.0938851611792817, + "learning_rate": 3.5711536400999196e-06, + "loss": 0.833, + "num_input_tokens_seen": 85621530, + "step": 2617 + }, + { + "epoch": 0.23610046444514587, + "flos": 57687951860640.0, + "grad_norm": 0.7103904044330204, + "learning_rate": 3.570792084488869e-06, + "loss": 0.637, + "num_input_tokens_seen": 85710805, + "step": 2618 + }, + { + "epoch": 0.23619064796861614, + "flos": 66674637163200.0, + "grad_norm": 0.733282941112767, + "learning_rate": 3.5704303948491764e-06, + "loss": 0.649, + "num_input_tokens_seen": 85791420, + "step": 2619 + }, + { + "epoch": 0.2362808314920864, + "flos": 28653879694560.0, + "grad_norm": 1.7067370128414066, + "learning_rate": 3.5700685712117035e-06, + "loss": 0.8001, + "num_input_tokens_seen": 85824235, + "step": 2620 + }, + { + "epoch": 0.23637101501555666, + "flos": 22460237117760.0, + "grad_norm": 2.1249713895779063, + "learning_rate": 3.5697066136073227e-06, + "loss": 0.7645, + "num_input_tokens_seen": 85853310, + "step": 2621 + }, + { + "epoch": 0.23646119853902692, + "flos": 20419928098080.0, + "grad_norm": 1.5249064870210682, + "learning_rate": 3.5693445220669184e-06, + "loss": 0.79, + "num_input_tokens_seen": 85882255, + "step": 2622 + }, + { + "epoch": 0.23655138206249718, + "flos": 26103707146080.0, + "grad_norm": 3.0445901764689722, + "learning_rate": 3.568982296621386e-06, + "loss": 0.7731, + "num_input_tokens_seen": 85914380, + "step": 2623 + }, + { + "epoch": 0.23664156558596744, + "flos": 21251361653280.0, + "grad_norm": 2.2669763152538933, + "learning_rate": 3.5686199373016325e-06, + "loss": 0.8559, + "num_input_tokens_seen": 85942210, + "step": 2624 + }, + { + "epoch": 0.2367317491094377, + "flos": 26358657495360.0, + "grad_norm": 1.5699472334674793, + "learning_rate": 3.568257444138577e-06, + "loss": 0.7262, + "num_input_tokens_seen": 85971970, + "step": 2625 + }, + { + "epoch": 0.23682193263290796, + "flos": 18707247775680.0, + "grad_norm": 2.514015393618947, + "learning_rate": 3.5678948171631495e-06, + "loss": 0.7794, + "num_input_tokens_seen": 85998995, + "step": 2626 + }, + { + "epoch": 0.23691211615637822, + "flos": 17323274073600.0, + "grad_norm": 1.9778840242141542, + "learning_rate": 3.5675320564062908e-06, + "loss": 0.7849, + "num_input_tokens_seen": 86025860, + "step": 2627 + }, + { + "epoch": 0.23700229967984848, + "flos": 34700827464000.0, + "grad_norm": 1.7976188778463649, + "learning_rate": 3.5671691618989533e-06, + "loss": 0.6922, + "num_input_tokens_seen": 86054355, + "step": 2628 + }, + { + "epoch": 0.23709248320331874, + "flos": 21766874985600.0, + "grad_norm": 2.3280856750034458, + "learning_rate": 3.5668061336721024e-06, + "loss": 0.7852, + "num_input_tokens_seen": 86082515, + "step": 2629 + }, + { + "epoch": 0.237182666726789, + "flos": 20274422723040.0, + "grad_norm": 2.0923669146985424, + "learning_rate": 3.5664429717567117e-06, + "loss": 0.7735, + "num_input_tokens_seen": 86108520, + "step": 2630 + }, + { + "epoch": 0.23727285025025927, + "flos": 25010781363840.0, + "grad_norm": 1.6147951130770142, + "learning_rate": 3.56607967618377e-06, + "loss": 0.8106, + "num_input_tokens_seen": 86139330, + "step": 2631 + }, + { + "epoch": 0.23736303377372953, + "flos": 22606300039200.0, + "grad_norm": 1.8503166546329601, + "learning_rate": 3.5657162469842754e-06, + "loss": 0.7931, + "num_input_tokens_seen": 86166835, + "step": 2632 + }, + { + "epoch": 0.23745321729719981, + "flos": 20485618531200.0, + "grad_norm": 1.9336451051616101, + "learning_rate": 3.5653526841892374e-06, + "loss": 0.7821, + "num_input_tokens_seen": 86195095, + "step": 2633 + }, + { + "epoch": 0.23754340082067008, + "flos": 24861521843040.0, + "grad_norm": 1.8441963302127344, + "learning_rate": 3.564988987829676e-06, + "loss": 0.8104, + "num_input_tokens_seen": 86222910, + "step": 2634 + }, + { + "epoch": 0.23763358434414034, + "flos": 24860815617600.0, + "grad_norm": 1.6318129258784804, + "learning_rate": 3.564625157936626e-06, + "loss": 0.7221, + "num_input_tokens_seen": 86252515, + "step": 2635 + }, + { + "epoch": 0.2377237678676106, + "flos": 25266103410720.0, + "grad_norm": 1.757341855939209, + "learning_rate": 3.56426119454113e-06, + "loss": 0.8181, + "num_input_tokens_seen": 86281205, + "step": 2636 + }, + { + "epoch": 0.23781395139108086, + "flos": 21621852817440.0, + "grad_norm": 6.686178772083919, + "learning_rate": 3.5638970976742436e-06, + "loss": 0.8547, + "num_input_tokens_seen": 86305480, + "step": 2637 + }, + { + "epoch": 0.23790413491455112, + "flos": 26684316195360.0, + "grad_norm": 2.5350408542508633, + "learning_rate": 3.5635328673670335e-06, + "loss": 0.8082, + "num_input_tokens_seen": 86330295, + "step": 2638 + }, + { + "epoch": 0.23799431843802138, + "flos": 29596430863200.0, + "grad_norm": 6.428931598564682, + "learning_rate": 3.5631685036505783e-06, + "loss": 0.8705, + "num_input_tokens_seen": 86360130, + "step": 2639 + }, + { + "epoch": 0.23808450196149164, + "flos": 25408932563040.0, + "grad_norm": 1.477720180184145, + "learning_rate": 3.562804006555966e-06, + "loss": 0.7934, + "num_input_tokens_seen": 86392060, + "step": 2640 + }, + { + "epoch": 0.2381746854849619, + "flos": 27633595090560.0, + "grad_norm": 1.6732314836932938, + "learning_rate": 3.5624393761143e-06, + "loss": 0.7196, + "num_input_tokens_seen": 86421660, + "step": 2641 + }, + { + "epoch": 0.23826486900843216, + "flos": 21040574712480.0, + "grad_norm": 1.940726516206126, + "learning_rate": 3.5620746123566906e-06, + "loss": 0.693, + "num_input_tokens_seen": 86449740, + "step": 2642 + }, + { + "epoch": 0.23835505253190242, + "flos": 23839899127680.0, + "grad_norm": 1.7190131939597468, + "learning_rate": 3.5617097153142623e-06, + "loss": 0.7275, + "num_input_tokens_seen": 86478340, + "step": 2643 + }, + { + "epoch": 0.23844523605537268, + "flos": 23589780847200.0, + "grad_norm": 1.703679685646031, + "learning_rate": 3.5613446850181497e-06, + "loss": 0.8218, + "num_input_tokens_seen": 86506370, + "step": 2644 + }, + { + "epoch": 0.23853541957884294, + "flos": 57928251590880.0, + "grad_norm": 0.606968235000135, + "learning_rate": 3.5609795214994996e-06, + "loss": 0.5373, + "num_input_tokens_seen": 86599100, + "step": 2645 + }, + { + "epoch": 0.2386256031023132, + "flos": 25338781758720.0, + "grad_norm": 2.0252959800391386, + "learning_rate": 3.560614224789469e-06, + "loss": 0.8384, + "num_input_tokens_seen": 86629490, + "step": 2646 + }, + { + "epoch": 0.23871578662578347, + "flos": 24136039304640.0, + "grad_norm": 1.624135227450377, + "learning_rate": 3.5602487949192285e-06, + "loss": 0.7293, + "num_input_tokens_seen": 86657010, + "step": 2647 + }, + { + "epoch": 0.23880597014925373, + "flos": 40785433933920.0, + "grad_norm": 2.062254226375027, + "learning_rate": 3.559883231919957e-06, + "loss": 0.7227, + "num_input_tokens_seen": 86688180, + "step": 2648 + }, + { + "epoch": 0.238896153672724, + "flos": 24317716589760.0, + "grad_norm": 1.590001728728692, + "learning_rate": 3.5595175358228473e-06, + "loss": 0.7536, + "num_input_tokens_seen": 86718105, + "step": 2649 + }, + { + "epoch": 0.23898633719619425, + "flos": 28325656281120.0, + "grad_norm": 1.9083250664722118, + "learning_rate": 3.5591517066591027e-06, + "loss": 0.7773, + "num_input_tokens_seen": 86747935, + "step": 2650 + }, + { + "epoch": 0.2390765207196645, + "flos": 14955299186880.0, + "grad_norm": 3.9317564475845344, + "learning_rate": 3.5587857444599364e-06, + "loss": 0.7716, + "num_input_tokens_seen": 86772965, + "step": 2651 + }, + { + "epoch": 0.23916670424313477, + "flos": 16958804410560.0, + "grad_norm": 1.5878517361954534, + "learning_rate": 3.5584196492565766e-06, + "loss": 0.7463, + "num_input_tokens_seen": 86800465, + "step": 2652 + }, + { + "epoch": 0.23925688776660503, + "flos": 17286990654240.0, + "grad_norm": 1.8112986299187679, + "learning_rate": 3.5580534210802587e-06, + "loss": 0.8428, + "num_input_tokens_seen": 86825060, + "step": 2653 + }, + { + "epoch": 0.2393470712900753, + "flos": 26394866575200.0, + "grad_norm": 1.5410422357889324, + "learning_rate": 3.557687059962232e-06, + "loss": 0.8265, + "num_input_tokens_seen": 86855275, + "step": 2654 + }, + { + "epoch": 0.23943725481354555, + "flos": 23328994845600.0, + "grad_norm": 2.4394054301282235, + "learning_rate": 3.5573205659337558e-06, + "loss": 0.7732, + "num_input_tokens_seen": 86883930, + "step": 2655 + }, + { + "epoch": 0.23952743833701584, + "flos": 21038902073280.0, + "grad_norm": 1.5625987212465513, + "learning_rate": 3.5569539390261025e-06, + "loss": 0.7681, + "num_input_tokens_seen": 86911735, + "step": 2656 + }, + { + "epoch": 0.2396176218604861, + "flos": 24245224090560.0, + "grad_norm": 1.3190431531161955, + "learning_rate": 3.5565871792705543e-06, + "loss": 0.7949, + "num_input_tokens_seen": 86942270, + "step": 2657 + }, + { + "epoch": 0.23970780538395636, + "flos": 23552791202400.0, + "grad_norm": 1.5046647970322902, + "learning_rate": 3.5562202866984045e-06, + "loss": 0.7727, + "num_input_tokens_seen": 86970120, + "step": 2658 + }, + { + "epoch": 0.23979798890742662, + "flos": 23297952362400.0, + "grad_norm": 2.2678246704969283, + "learning_rate": 3.5558532613409594e-06, + "loss": 0.7946, + "num_input_tokens_seen": 86997255, + "step": 2659 + }, + { + "epoch": 0.23988817243089688, + "flos": 29746991325600.0, + "grad_norm": 1.8342463719567497, + "learning_rate": 3.555486103229535e-06, + "loss": 0.7268, + "num_input_tokens_seen": 87028125, + "step": 2660 + }, + { + "epoch": 0.23997835595436715, + "flos": 20019360864480.0, + "grad_norm": 32.042107534827316, + "learning_rate": 3.5551188123954595e-06, + "loss": 0.8022, + "num_input_tokens_seen": 87053760, + "step": 2661 + }, + { + "epoch": 0.2400685394778374, + "flos": 20019509543520.0, + "grad_norm": 1.7787557840688868, + "learning_rate": 3.5547513888700715e-06, + "loss": 0.6763, + "num_input_tokens_seen": 87080710, + "step": 2662 + }, + { + "epoch": 0.24015872300130767, + "flos": 25952551967520.0, + "grad_norm": 2.1796796244817402, + "learning_rate": 3.5543838326847224e-06, + "loss": 0.8129, + "num_input_tokens_seen": 87108845, + "step": 2663 + }, + { + "epoch": 0.24024890652477793, + "flos": 33936050755680.0, + "grad_norm": 1.8048270695869706, + "learning_rate": 3.5540161438707744e-06, + "loss": 0.7588, + "num_input_tokens_seen": 87138675, + "step": 2664 + }, + { + "epoch": 0.2403390900482482, + "flos": 32187198523200.0, + "grad_norm": 1.7644345481625994, + "learning_rate": 3.5536483224596e-06, + "loss": 0.9394, + "num_input_tokens_seen": 87170780, + "step": 2665 + }, + { + "epoch": 0.24042927357171845, + "flos": 29054261079360.0, + "grad_norm": 3.1455102379880433, + "learning_rate": 3.553280368482584e-06, + "loss": 0.7589, + "num_input_tokens_seen": 87201265, + "step": 2666 + }, + { + "epoch": 0.2405194570951887, + "flos": 49708175782080.0, + "grad_norm": 2.0883145736050635, + "learning_rate": 3.5529122819711227e-06, + "loss": 0.6008, + "num_input_tokens_seen": 87232860, + "step": 2667 + }, + { + "epoch": 0.24060964061865897, + "flos": 24828360683520.0, + "grad_norm": 2.039117571207645, + "learning_rate": 3.5525440629566223e-06, + "loss": 0.7263, + "num_input_tokens_seen": 87263515, + "step": 2668 + }, + { + "epoch": 0.24069982414212923, + "flos": 19254323967840.0, + "grad_norm": 2.9924676430338377, + "learning_rate": 3.552175711470502e-06, + "loss": 0.8009, + "num_input_tokens_seen": 87291460, + "step": 2669 + }, + { + "epoch": 0.2407900076655995, + "flos": 24135741946560.0, + "grad_norm": 2.061395407152079, + "learning_rate": 3.5518072275441912e-06, + "loss": 0.8232, + "num_input_tokens_seen": 87319175, + "step": 2670 + }, + { + "epoch": 0.24088019118906975, + "flos": 24605456400960.0, + "grad_norm": 2.090623143399131, + "learning_rate": 3.551438611209131e-06, + "loss": 0.8057, + "num_input_tokens_seen": 87349435, + "step": 2671 + }, + { + "epoch": 0.24097037471254001, + "flos": 27810960683520.0, + "grad_norm": 1.8914797403459118, + "learning_rate": 3.551069862496774e-06, + "loss": 0.8762, + "num_input_tokens_seen": 87379125, + "step": 2672 + }, + { + "epoch": 0.24106055823601027, + "flos": 70851982385280.0, + "grad_norm": 0.654462332230996, + "learning_rate": 3.5507009814385846e-06, + "loss": 0.5981, + "num_input_tokens_seen": 87469815, + "step": 2673 + }, + { + "epoch": 0.24115074175948054, + "flos": 56562831332640.0, + "grad_norm": 0.8047028697499666, + "learning_rate": 3.550331968066036e-06, + "loss": 0.5891, + "num_input_tokens_seen": 87556830, + "step": 2674 + }, + { + "epoch": 0.2412409252829508, + "flos": 17067729008160.0, + "grad_norm": 2.7996583014846688, + "learning_rate": 3.549962822410616e-06, + "loss": 0.7955, + "num_input_tokens_seen": 87584840, + "step": 2675 + }, + { + "epoch": 0.24133110880642106, + "flos": 70807298600160.0, + "grad_norm": 0.6912672474246008, + "learning_rate": 3.5495935445038217e-06, + "loss": 0.5862, + "num_input_tokens_seen": 87686395, + "step": 2676 + }, + { + "epoch": 0.24142129232989132, + "flos": 21469062169440.0, + "grad_norm": 1.7805271355034953, + "learning_rate": 3.5492241343771612e-06, + "loss": 0.8193, + "num_input_tokens_seen": 87714030, + "step": 2677 + }, + { + "epoch": 0.24151147585336158, + "flos": 26540557799040.0, + "grad_norm": 2.210417993121479, + "learning_rate": 3.548854592062156e-06, + "loss": 0.7789, + "num_input_tokens_seen": 87740225, + "step": 2678 + }, + { + "epoch": 0.24160165937683184, + "flos": 17248179691200.0, + "grad_norm": 2.5839047466531366, + "learning_rate": 3.548484917590336e-06, + "loss": 0.7595, + "num_input_tokens_seen": 87767835, + "step": 2679 + }, + { + "epoch": 0.24169184290030213, + "flos": 43079541040320.0, + "grad_norm": 2.465867979891988, + "learning_rate": 3.5481151109932447e-06, + "loss": 0.6985, + "num_input_tokens_seen": 87795650, + "step": 2680 + }, + { + "epoch": 0.2417820264237724, + "flos": 21695274560640.0, + "grad_norm": 1.6199675911145868, + "learning_rate": 3.5477451723024364e-06, + "loss": 0.8056, + "num_input_tokens_seen": 87823575, + "step": 2681 + }, + { + "epoch": 0.24187220994724265, + "flos": 70631642816160.0, + "grad_norm": 0.5913849457666616, + "learning_rate": 3.5473751015494757e-06, + "loss": 0.5705, + "num_input_tokens_seen": 87923415, + "step": 2682 + }, + { + "epoch": 0.2419623934707129, + "flos": 21257643342720.0, + "grad_norm": 1.701760681773376, + "learning_rate": 3.547004898765939e-06, + "loss": 0.7949, + "num_input_tokens_seen": 87952655, + "step": 2683 + }, + { + "epoch": 0.24205257699418317, + "flos": 23585803682880.0, + "grad_norm": 1.3922986877772772, + "learning_rate": 3.546634563983414e-06, + "loss": 0.7979, + "num_input_tokens_seen": 87981425, + "step": 2684 + }, + { + "epoch": 0.24214276051765343, + "flos": 24859291657440.0, + "grad_norm": 1.545506266758634, + "learning_rate": 3.5462640972335002e-06, + "loss": 0.7418, + "num_input_tokens_seen": 88012340, + "step": 2685 + }, + { + "epoch": 0.2422329440411237, + "flos": 26721714707520.0, + "grad_norm": 1.633036109908179, + "learning_rate": 3.5458934985478077e-06, + "loss": 0.7992, + "num_input_tokens_seen": 88042740, + "step": 2686 + }, + { + "epoch": 0.24232312756459395, + "flos": 21912603379200.0, + "grad_norm": 2.1404805378304217, + "learning_rate": 3.5455227679579577e-06, + "loss": 0.7605, + "num_input_tokens_seen": 88069785, + "step": 2687 + }, + { + "epoch": 0.24241331108806422, + "flos": 22968576686400.0, + "grad_norm": 2.2758360070746337, + "learning_rate": 3.545151905495584e-06, + "loss": 0.8878, + "num_input_tokens_seen": 88097970, + "step": 2688 + }, + { + "epoch": 0.24250349461153448, + "flos": 21075520020480.0, + "grad_norm": 6.629203182801624, + "learning_rate": 3.544780911192329e-06, + "loss": 0.7653, + "num_input_tokens_seen": 88125615, + "step": 2689 + }, + { + "epoch": 0.24259367813500474, + "flos": 24169534992000.0, + "grad_norm": 1.6705977650372956, + "learning_rate": 3.544409785079849e-06, + "loss": 0.7306, + "num_input_tokens_seen": 88156005, + "step": 2690 + }, + { + "epoch": 0.242683861658475, + "flos": 19545409057440.0, + "grad_norm": 1.5633915455736347, + "learning_rate": 3.5440385271898103e-06, + "loss": 0.7275, + "num_input_tokens_seen": 88184795, + "step": 2691 + }, + { + "epoch": 0.24277404518194526, + "flos": 24274036388160.0, + "grad_norm": 1.7627515789149886, + "learning_rate": 3.5436671375538903e-06, + "loss": 0.7069, + "num_input_tokens_seen": 88209645, + "step": 2692 + }, + { + "epoch": 0.24286422870541552, + "flos": 24172471403040.0, + "grad_norm": 2.3097774872393795, + "learning_rate": 3.543295616203779e-06, + "loss": 0.7781, + "num_input_tokens_seen": 88237560, + "step": 2693 + }, + { + "epoch": 0.24295441222888578, + "flos": 29012625214560.0, + "grad_norm": 1.8232817438673985, + "learning_rate": 3.542923963171176e-06, + "loss": 0.8769, + "num_input_tokens_seen": 88267220, + "step": 2694 + }, + { + "epoch": 0.24304459575235604, + "flos": 29416463387040.0, + "grad_norm": 1.5172407838243254, + "learning_rate": 3.542552178487793e-06, + "loss": 0.6999, + "num_input_tokens_seen": 88299895, + "step": 2695 + }, + { + "epoch": 0.2431347792758263, + "flos": 24785646895680.0, + "grad_norm": 2.1325504575173415, + "learning_rate": 3.5421802621853523e-06, + "loss": 0.7733, + "num_input_tokens_seen": 88328440, + "step": 2696 + }, + { + "epoch": 0.24322496279929656, + "flos": 27812150115840.0, + "grad_norm": 1.892307510161257, + "learning_rate": 3.5418082142955887e-06, + "loss": 0.7765, + "num_input_tokens_seen": 88357575, + "step": 2697 + }, + { + "epoch": 0.24331514632276682, + "flos": 59079433854240.0, + "grad_norm": 0.6672907505948149, + "learning_rate": 3.5414360348502463e-06, + "loss": 0.6145, + "num_input_tokens_seen": 88440900, + "step": 2698 + }, + { + "epoch": 0.24340532984623708, + "flos": 28107063690720.0, + "grad_norm": 1.937729606559493, + "learning_rate": 3.5410637238810825e-06, + "loss": 0.6831, + "num_input_tokens_seen": 88470770, + "step": 2699 + }, + { + "epoch": 0.24349551336970734, + "flos": 27706050420000.0, + "grad_norm": 6.794530284329298, + "learning_rate": 3.5406912814198635e-06, + "loss": 0.7344, + "num_input_tokens_seen": 88499645, + "step": 2700 + }, + { + "epoch": 0.2435856968931776, + "flos": 28507445075520.0, + "grad_norm": 4.217121177171604, + "learning_rate": 3.54031870749837e-06, + "loss": 0.7464, + "num_input_tokens_seen": 88529765, + "step": 2701 + }, + { + "epoch": 0.24367588041664787, + "flos": 22569124545600.0, + "grad_norm": 1.5482693947578685, + "learning_rate": 3.539946002148391e-06, + "loss": 0.7897, + "num_input_tokens_seen": 88556965, + "step": 2702 + }, + { + "epoch": 0.24376606394011813, + "flos": 21985876443360.0, + "grad_norm": 2.638535524991901, + "learning_rate": 3.5395731654017277e-06, + "loss": 0.7843, + "num_input_tokens_seen": 88580470, + "step": 2703 + }, + { + "epoch": 0.24385624746358842, + "flos": 19325329676640.0, + "grad_norm": 2.0876703386992683, + "learning_rate": 3.5392001972901923e-06, + "loss": 0.822, + "num_input_tokens_seen": 88603955, + "step": 2704 + }, + { + "epoch": 0.24394643098705868, + "flos": 22315029100800.0, + "grad_norm": 1.5311357883755634, + "learning_rate": 3.5388270978456098e-06, + "loss": 0.742, + "num_input_tokens_seen": 88632560, + "step": 2705 + }, + { + "epoch": 0.24403661451052894, + "flos": 24389502863520.0, + "grad_norm": 4.322049169554628, + "learning_rate": 3.5384538670998137e-06, + "loss": 0.7532, + "num_input_tokens_seen": 88659795, + "step": 2706 + }, + { + "epoch": 0.2441267980339992, + "flos": 19617975896160.0, + "grad_norm": 2.0035200137192377, + "learning_rate": 3.538080505084651e-06, + "loss": 0.8022, + "num_input_tokens_seen": 88686220, + "step": 2707 + }, + { + "epoch": 0.24421698155746946, + "flos": 26066791840800.0, + "grad_norm": 1.550173623605086, + "learning_rate": 3.5377070118319788e-06, + "loss": 0.7941, + "num_input_tokens_seen": 88715565, + "step": 2708 + }, + { + "epoch": 0.24430716508093972, + "flos": 40487806966560.0, + "grad_norm": 2.6840996633238174, + "learning_rate": 3.5373333873736657e-06, + "loss": 0.7401, + "num_input_tokens_seen": 88744725, + "step": 2709 + }, + { + "epoch": 0.24439734860440998, + "flos": 25593620598720.0, + "grad_norm": 1.530729555707763, + "learning_rate": 3.536959631741591e-06, + "loss": 0.7866, + "num_input_tokens_seen": 88774395, + "step": 2710 + }, + { + "epoch": 0.24448753212788024, + "flos": 26139284340000.0, + "grad_norm": 1.4181403274346684, + "learning_rate": 3.536585744967646e-06, + "loss": 0.8046, + "num_input_tokens_seen": 88807785, + "step": 2711 + }, + { + "epoch": 0.2445777156513505, + "flos": 16995459527520.0, + "grad_norm": 2.4995012739544173, + "learning_rate": 3.5362117270837326e-06, + "loss": 0.7603, + "num_input_tokens_seen": 88835400, + "step": 2712 + }, + { + "epoch": 0.24466789917482076, + "flos": 26468362657920.0, + "grad_norm": 3.740672400101727, + "learning_rate": 3.5358375781217634e-06, + "loss": 0.6846, + "num_input_tokens_seen": 88869270, + "step": 2713 + }, + { + "epoch": 0.24475808269829102, + "flos": 26613347656320.0, + "grad_norm": 2.0057178064772647, + "learning_rate": 3.535463298113664e-06, + "loss": 0.8111, + "num_input_tokens_seen": 88896510, + "step": 2714 + }, + { + "epoch": 0.24484826622176128, + "flos": 25885226064960.0, + "grad_norm": 1.7209702456538234, + "learning_rate": 3.5350888870913697e-06, + "loss": 0.741, + "num_input_tokens_seen": 88924465, + "step": 2715 + }, + { + "epoch": 0.24493844974523155, + "flos": 25294209482880.0, + "grad_norm": 2.23428557957268, + "learning_rate": 3.5347143450868273e-06, + "loss": 0.781, + "num_input_tokens_seen": 88952260, + "step": 2716 + }, + { + "epoch": 0.2450286332687018, + "flos": 20273753667360.0, + "grad_norm": 5.191566068716607, + "learning_rate": 3.534339672131994e-06, + "loss": 0.8004, + "num_input_tokens_seen": 88980090, + "step": 2717 + }, + { + "epoch": 0.24511881679217207, + "flos": 13893304378560.0, + "grad_norm": 2.6942868257632115, + "learning_rate": 3.5339648682588397e-06, + "loss": 0.7145, + "num_input_tokens_seen": 89006380, + "step": 2718 + }, + { + "epoch": 0.24520900031564233, + "flos": 19581841155840.0, + "grad_norm": 1.7833851515588715, + "learning_rate": 3.533589933499345e-06, + "loss": 0.819, + "num_input_tokens_seen": 89034925, + "step": 2719 + }, + { + "epoch": 0.2452991838391126, + "flos": 21950150570400.0, + "grad_norm": 2.718884958008793, + "learning_rate": 3.533214867885501e-06, + "loss": 0.7168, + "num_input_tokens_seen": 89063575, + "step": 2720 + }, + { + "epoch": 0.24538936736258285, + "flos": 25443803531520.0, + "grad_norm": 2.089682770705601, + "learning_rate": 3.53283967144931e-06, + "loss": 0.7232, + "num_input_tokens_seen": 89090860, + "step": 2721 + }, + { + "epoch": 0.2454795508860531, + "flos": 22131530497440.0, + "grad_norm": 1.826072192403884, + "learning_rate": 3.532464344222787e-06, + "loss": 0.7898, + "num_input_tokens_seen": 89118465, + "step": 2722 + }, + { + "epoch": 0.24556973440952337, + "flos": 16594148898720.0, + "grad_norm": 2.1130516505718973, + "learning_rate": 3.532088886237956e-06, + "loss": 0.8185, + "num_input_tokens_seen": 89146730, + "step": 2723 + }, + { + "epoch": 0.24565991793299363, + "flos": 71012881774560.0, + "grad_norm": 0.6526146514691693, + "learning_rate": 3.5317132975268535e-06, + "loss": 0.5268, + "num_input_tokens_seen": 89237950, + "step": 2724 + }, + { + "epoch": 0.2457501014564639, + "flos": 22895786829120.0, + "grad_norm": 2.817545201532178, + "learning_rate": 3.531337578121526e-06, + "loss": 0.807, + "num_input_tokens_seen": 89267360, + "step": 2725 + }, + { + "epoch": 0.24584028497993415, + "flos": 15100916071200.0, + "grad_norm": 1.6959260454383231, + "learning_rate": 3.530961728054033e-06, + "loss": 0.8158, + "num_input_tokens_seen": 89294350, + "step": 2726 + }, + { + "epoch": 0.24593046850340441, + "flos": 21768510455040.0, + "grad_norm": 1.6119232003680843, + "learning_rate": 3.5305857473564435e-06, + "loss": 0.7425, + "num_input_tokens_seen": 89321385, + "step": 2727 + }, + { + "epoch": 0.2460206520268747, + "flos": 16120345770720.0, + "grad_norm": 1.973240590182189, + "learning_rate": 3.5302096360608385e-06, + "loss": 0.7095, + "num_input_tokens_seen": 89346650, + "step": 2728 + }, + { + "epoch": 0.24611083555034496, + "flos": 20820644010720.0, + "grad_norm": 1.6526633239923618, + "learning_rate": 3.5298333941993105e-06, + "loss": 0.8343, + "num_input_tokens_seen": 89374175, + "step": 2729 + }, + { + "epoch": 0.24620101907381522, + "flos": 23042778994560.0, + "grad_norm": 2.017680221884331, + "learning_rate": 3.529457021803962e-06, + "loss": 0.7734, + "num_input_tokens_seen": 89399975, + "step": 2730 + }, + { + "epoch": 0.24629120259728549, + "flos": 14299595755200.0, + "grad_norm": 2.0802931410290006, + "learning_rate": 3.529080518906906e-06, + "loss": 0.7953, + "num_input_tokens_seen": 89426490, + "step": 2731 + }, + { + "epoch": 0.24638138612075575, + "flos": 25046841764640.0, + "grad_norm": 1.536759643320881, + "learning_rate": 3.5287038855402696e-06, + "loss": 0.7631, + "num_input_tokens_seen": 89456685, + "step": 2732 + }, + { + "epoch": 0.246471569644226, + "flos": 18634272069600.0, + "grad_norm": 1.9777154989771428, + "learning_rate": 3.528327121736188e-06, + "loss": 0.814, + "num_input_tokens_seen": 89484705, + "step": 2733 + }, + { + "epoch": 0.24656175316769627, + "flos": 21835315980960.0, + "grad_norm": 1.9188840589380924, + "learning_rate": 3.52795022752681e-06, + "loss": 0.8585, + "num_input_tokens_seen": 89511965, + "step": 2734 + }, + { + "epoch": 0.24665193669116653, + "flos": 21585495058560.0, + "grad_norm": 2.534374522104465, + "learning_rate": 3.5275732029442925e-06, + "loss": 0.8487, + "num_input_tokens_seen": 89539845, + "step": 2735 + }, + { + "epoch": 0.2467421202146368, + "flos": 24026928858240.0, + "grad_norm": 2.5772030193277424, + "learning_rate": 3.5271960480208077e-06, + "loss": 0.7209, + "num_input_tokens_seen": 89568965, + "step": 2736 + }, + { + "epoch": 0.24683230373810705, + "flos": 67412385722400.0, + "grad_norm": 0.7905618378510312, + "learning_rate": 3.526818762788534e-06, + "loss": 0.6273, + "num_input_tokens_seen": 89660490, + "step": 2737 + }, + { + "epoch": 0.2469224872615773, + "flos": 23406691111200.0, + "grad_norm": 1.6151140714815486, + "learning_rate": 3.5264413472796653e-06, + "loss": 0.8173, + "num_input_tokens_seen": 89689805, + "step": 2738 + }, + { + "epoch": 0.24701267078504757, + "flos": 21431998185120.0, + "grad_norm": 2.453223451230291, + "learning_rate": 3.5260638015264037e-06, + "loss": 0.7571, + "num_input_tokens_seen": 89714960, + "step": 2739 + }, + { + "epoch": 0.24710285430851783, + "flos": 27638724517440.0, + "grad_norm": 1.532441356961977, + "learning_rate": 3.5256861255609644e-06, + "loss": 0.7507, + "num_input_tokens_seen": 89749165, + "step": 2740 + }, + { + "epoch": 0.2471930378319881, + "flos": 18925877535840.0, + "grad_norm": 2.2610559101640852, + "learning_rate": 3.5253083194155723e-06, + "loss": 0.7953, + "num_input_tokens_seen": 89775515, + "step": 2741 + }, + { + "epoch": 0.24728322135545835, + "flos": 23657998824000.0, + "grad_norm": 1.3673863293619295, + "learning_rate": 3.5249303831224637e-06, + "loss": 0.7987, + "num_input_tokens_seen": 89805600, + "step": 2742 + }, + { + "epoch": 0.24737340487892862, + "flos": 22018888923840.0, + "grad_norm": 1.587013518518032, + "learning_rate": 3.524552316713887e-06, + "loss": 0.7669, + "num_input_tokens_seen": 89834000, + "step": 2743 + }, + { + "epoch": 0.24746358840239888, + "flos": 20854325546880.0, + "grad_norm": 1.8361901292070857, + "learning_rate": 3.5241741202220995e-06, + "loss": 0.8369, + "num_input_tokens_seen": 89862205, + "step": 2744 + }, + { + "epoch": 0.24755377192586914, + "flos": 25003533260640.0, + "grad_norm": 2.2312369551931788, + "learning_rate": 3.5237957936793724e-06, + "loss": 0.7652, + "num_input_tokens_seen": 89892730, + "step": 2745 + }, + { + "epoch": 0.2476439554493394, + "flos": 22312018350240.0, + "grad_norm": 1.4966075204293503, + "learning_rate": 3.523417337117986e-06, + "loss": 0.7425, + "num_input_tokens_seen": 89923345, + "step": 2746 + }, + { + "epoch": 0.24773413897280966, + "flos": 27596382427200.0, + "grad_norm": 1.486526375781323, + "learning_rate": 3.523038750570232e-06, + "loss": 0.7302, + "num_input_tokens_seen": 89954230, + "step": 2747 + }, + { + "epoch": 0.24782432249627992, + "flos": 24026334142080.0, + "grad_norm": 3.099730289170428, + "learning_rate": 3.522660034068414e-06, + "loss": 0.8095, + "num_input_tokens_seen": 89983700, + "step": 2748 + }, + { + "epoch": 0.24791450601975018, + "flos": 18999039090720.0, + "grad_norm": 1.893476495412071, + "learning_rate": 3.5222811876448464e-06, + "loss": 0.7915, + "num_input_tokens_seen": 90010340, + "step": 2749 + }, + { + "epoch": 0.24800468954322044, + "flos": 32077753548960.0, + "grad_norm": 1.5571182647752952, + "learning_rate": 3.521902211331854e-06, + "loss": 0.799, + "num_input_tokens_seen": 90044685, + "step": 2750 + }, + { + "epoch": 0.2480948730666907, + "flos": 24973531530720.0, + "grad_norm": 1.8834000932907142, + "learning_rate": 3.5215231051617726e-06, + "loss": 0.7279, + "num_input_tokens_seen": 90072695, + "step": 2751 + }, + { + "epoch": 0.248185056590161, + "flos": 31489636208160.0, + "grad_norm": 1.7447076881657786, + "learning_rate": 3.521143869166951e-06, + "loss": 0.6831, + "num_input_tokens_seen": 90100340, + "step": 2752 + }, + { + "epoch": 0.24827524011363125, + "flos": 14226508539840.0, + "grad_norm": 2.2001118718226937, + "learning_rate": 3.5207645033797464e-06, + "loss": 0.812, + "num_input_tokens_seen": 90124860, + "step": 2753 + }, + { + "epoch": 0.2483654236371015, + "flos": 31640568368160.0, + "grad_norm": 2.1593079927815833, + "learning_rate": 3.5203850078325293e-06, + "loss": 0.7017, + "num_input_tokens_seen": 90152900, + "step": 2754 + }, + { + "epoch": 0.24845560716057177, + "flos": 23259698945760.0, + "grad_norm": 2.188991583057322, + "learning_rate": 3.5200053825576797e-06, + "loss": 0.6928, + "num_input_tokens_seen": 90181965, + "step": 2755 + }, + { + "epoch": 0.24854579068404203, + "flos": 26468362657920.0, + "grad_norm": 1.6708654047043459, + "learning_rate": 3.51962562758759e-06, + "loss": 0.7812, + "num_input_tokens_seen": 90213030, + "step": 2756 + }, + { + "epoch": 0.2486359742075123, + "flos": 28435212764640.0, + "grad_norm": 3.648567148392383, + "learning_rate": 3.5192457429546627e-06, + "loss": 0.7562, + "num_input_tokens_seen": 90243065, + "step": 2757 + }, + { + "epoch": 0.24872615773098256, + "flos": 36848648630400.0, + "grad_norm": 3.4791812249219802, + "learning_rate": 3.5188657286913115e-06, + "loss": 0.7462, + "num_input_tokens_seen": 90273245, + "step": 2758 + }, + { + "epoch": 0.24881634125445282, + "flos": 24244852392960.0, + "grad_norm": 2.2498691308850876, + "learning_rate": 3.518485584829961e-06, + "loss": 0.7401, + "num_input_tokens_seen": 90301285, + "step": 2759 + }, + { + "epoch": 0.24890652477792308, + "flos": 19181199582720.0, + "grad_norm": 1.8321920946516774, + "learning_rate": 3.5181053114030485e-06, + "loss": 0.7876, + "num_input_tokens_seen": 90328645, + "step": 2760 + }, + { + "epoch": 0.24899670830139334, + "flos": 30507307662720.0, + "grad_norm": 2.319888606028027, + "learning_rate": 3.5177249084430198e-06, + "loss": 0.7731, + "num_input_tokens_seen": 90359130, + "step": 2761 + }, + { + "epoch": 0.2490868918248636, + "flos": 16660285368960.0, + "grad_norm": 2.2621681939555556, + "learning_rate": 3.517344375982333e-06, + "loss": 0.811, + "num_input_tokens_seen": 90385800, + "step": 2762 + }, + { + "epoch": 0.24917707534833386, + "flos": 23328288620160.0, + "grad_norm": 1.838499448758007, + "learning_rate": 3.5169637140534565e-06, + "loss": 0.8101, + "num_input_tokens_seen": 90412580, + "step": 2763 + }, + { + "epoch": 0.24926725887180412, + "flos": 25848236420160.0, + "grad_norm": 2.0310795639133117, + "learning_rate": 3.5165829226888733e-06, + "loss": 0.7496, + "num_input_tokens_seen": 90442355, + "step": 2764 + }, + { + "epoch": 0.24935744239527438, + "flos": 23261297245440.0, + "grad_norm": 1.6356893502427154, + "learning_rate": 3.516202001921072e-06, + "loss": 0.8035, + "num_input_tokens_seen": 90471780, + "step": 2765 + }, + { + "epoch": 0.24944762591874464, + "flos": 22929914402400.0, + "grad_norm": 3.790296593487317, + "learning_rate": 3.515820951782555e-06, + "loss": 0.8045, + "num_input_tokens_seen": 90501850, + "step": 2766 + }, + { + "epoch": 0.2495378094422149, + "flos": 23297766513600.0, + "grad_norm": 2.366038833763812, + "learning_rate": 3.5154397723058366e-06, + "loss": 0.7051, + "num_input_tokens_seen": 90530215, + "step": 2767 + }, + { + "epoch": 0.24962799296568516, + "flos": 18889519776960.0, + "grad_norm": 1.8935302606888889, + "learning_rate": 3.5150584635234416e-06, + "loss": 0.8321, + "num_input_tokens_seen": 90558900, + "step": 2768 + }, + { + "epoch": 0.24971817648915542, + "flos": 24062097184800.0, + "grad_norm": 1.4763951282403274, + "learning_rate": 3.5146770254679035e-06, + "loss": 0.8201, + "num_input_tokens_seen": 90590050, + "step": 2769 + }, + { + "epoch": 0.24980836001262569, + "flos": 21401290229760.0, + "grad_norm": 3.0575679067211943, + "learning_rate": 3.51429545817177e-06, + "loss": 0.6967, + "num_input_tokens_seen": 90619390, + "step": 2770 + }, + { + "epoch": 0.24989854353609595, + "flos": 22278002286240.0, + "grad_norm": 2.8294964410409285, + "learning_rate": 3.5139137616675985e-06, + "loss": 0.7875, + "num_input_tokens_seen": 90647425, + "step": 2771 + }, + { + "epoch": 0.2499887270595662, + "flos": 25442168062080.0, + "grad_norm": 1.9974490228672288, + "learning_rate": 3.513531935987957e-06, + "loss": 0.7214, + "num_input_tokens_seen": 90677430, + "step": 2772 + }, + { + "epoch": 0.2500789105830365, + "flos": 20817521750880.0, + "grad_norm": 1.7573845236379315, + "learning_rate": 3.5131499811654253e-06, + "loss": 0.7989, + "num_input_tokens_seen": 90706385, + "step": 2773 + }, + { + "epoch": 0.25016909410650673, + "flos": 25994076323040.0, + "grad_norm": 2.081127894030017, + "learning_rate": 3.512767897232594e-06, + "loss": 0.7454, + "num_input_tokens_seen": 90736525, + "step": 2774 + }, + { + "epoch": 0.250259277629977, + "flos": 19763629950240.0, + "grad_norm": 1.9868719656154026, + "learning_rate": 3.512385684222064e-06, + "loss": 0.7721, + "num_input_tokens_seen": 90762685, + "step": 2775 + }, + { + "epoch": 0.25034946115344725, + "flos": 15027940365120.0, + "grad_norm": 2.0182309124026387, + "learning_rate": 3.512003342166449e-06, + "loss": 0.8748, + "num_input_tokens_seen": 90785090, + "step": 2776 + }, + { + "epoch": 0.25043964467691754, + "flos": 22205249598720.0, + "grad_norm": 1.4685614444553374, + "learning_rate": 3.511620871098371e-06, + "loss": 0.8161, + "num_input_tokens_seen": 90815060, + "step": 2777 + }, + { + "epoch": 0.25052982820038777, + "flos": 22892590229760.0, + "grad_norm": 2.054302455084288, + "learning_rate": 3.511238271050465e-06, + "loss": 0.8094, + "num_input_tokens_seen": 90840700, + "step": 2778 + }, + { + "epoch": 0.25062001172385806, + "flos": 38340208818720.0, + "grad_norm": 1.7121588123450453, + "learning_rate": 3.5108555420553778e-06, + "loss": 0.6931, + "num_input_tokens_seen": 90871820, + "step": 2779 + }, + { + "epoch": 0.2507101952473283, + "flos": 25739237483040.0, + "grad_norm": 2.0234098101161, + "learning_rate": 3.510472684145764e-06, + "loss": 0.5965, + "num_input_tokens_seen": 90902860, + "step": 2780 + }, + { + "epoch": 0.2508003787707986, + "flos": 64279193823840.0, + "grad_norm": 1.8275696000441053, + "learning_rate": 3.5100896973542926e-06, + "loss": 0.6684, + "num_input_tokens_seen": 90937760, + "step": 2781 + }, + { + "epoch": 0.2508905622942688, + "flos": 23334087102720.0, + "grad_norm": 2.2492332211579127, + "learning_rate": 3.509706581713642e-06, + "loss": 0.8151, + "num_input_tokens_seen": 90964245, + "step": 2782 + }, + { + "epoch": 0.2509807458177391, + "flos": 27196558588800.0, + "grad_norm": 1.6103601495942972, + "learning_rate": 3.509323337256501e-06, + "loss": 0.8057, + "num_input_tokens_seen": 90994880, + "step": 2783 + }, + { + "epoch": 0.25107092934120934, + "flos": 19071494420160.0, + "grad_norm": 1.9721792720840043, + "learning_rate": 3.5089399640155703e-06, + "loss": 0.7289, + "num_input_tokens_seen": 91021160, + "step": 2784 + }, + { + "epoch": 0.2511611128646796, + "flos": 23990348080800.0, + "grad_norm": 1.6429427713978662, + "learning_rate": 3.508556462023562e-06, + "loss": 0.7719, + "num_input_tokens_seen": 91049295, + "step": 2785 + }, + { + "epoch": 0.25125129638814986, + "flos": 17935371643200.0, + "grad_norm": 2.119530201625881, + "learning_rate": 3.5081728313131984e-06, + "loss": 0.7618, + "num_input_tokens_seen": 91074775, + "step": 2786 + }, + { + "epoch": 0.25134147991162015, + "flos": 24792411792000.0, + "grad_norm": 2.617350558994617, + "learning_rate": 3.5077890719172125e-06, + "loss": 0.7714, + "num_input_tokens_seen": 91105370, + "step": 2787 + }, + { + "epoch": 0.25143166343509044, + "flos": 25374358952640.0, + "grad_norm": 2.3211894320883575, + "learning_rate": 3.5074051838683497e-06, + "loss": 0.7764, + "num_input_tokens_seen": 91133775, + "step": 2788 + }, + { + "epoch": 0.25152184695856067, + "flos": 17940426730560.0, + "grad_norm": 1.8409752737536667, + "learning_rate": 3.5070211671993643e-06, + "loss": 0.7004, + "num_input_tokens_seen": 91161350, + "step": 2789 + }, + { + "epoch": 0.25161203048203096, + "flos": 20456137177920.0, + "grad_norm": 1.7995666247109834, + "learning_rate": 3.5066370219430238e-06, + "loss": 0.8006, + "num_input_tokens_seen": 91188310, + "step": 2790 + }, + { + "epoch": 0.2517022140055012, + "flos": 21950485098240.0, + "grad_norm": 25.422759713016628, + "learning_rate": 3.5062527481321044e-06, + "loss": 0.7722, + "num_input_tokens_seen": 91218730, + "step": 2791 + }, + { + "epoch": 0.2517923975289715, + "flos": 21840854275200.0, + "grad_norm": 2.34460230271271, + "learning_rate": 3.5058683457993954e-06, + "loss": 0.7921, + "num_input_tokens_seen": 91245970, + "step": 2792 + }, + { + "epoch": 0.2518825810524417, + "flos": 66808168141920.0, + "grad_norm": 0.7622250170543703, + "learning_rate": 3.5054838149776963e-06, + "loss": 0.5354, + "num_input_tokens_seen": 91342655, + "step": 2793 + }, + { + "epoch": 0.251972764575912, + "flos": 22059000828480.0, + "grad_norm": 1.6651531102116068, + "learning_rate": 3.505099155699816e-06, + "loss": 0.8101, + "num_input_tokens_seen": 91372000, + "step": 2794 + }, + { + "epoch": 0.25206294809938223, + "flos": 35684419781280.0, + "grad_norm": 1.6922784895311294, + "learning_rate": 3.5047143679985775e-06, + "loss": 0.6498, + "num_input_tokens_seen": 91403125, + "step": 2795 + }, + { + "epoch": 0.2521531316228525, + "flos": 21475938575040.0, + "grad_norm": 1.580793258984861, + "learning_rate": 3.5043294519068126e-06, + "loss": 0.7931, + "num_input_tokens_seen": 91431860, + "step": 2796 + }, + { + "epoch": 0.25224331514632276, + "flos": 24500025760800.0, + "grad_norm": 1.641809598958286, + "learning_rate": 3.503944407457363e-06, + "loss": 0.6971, + "num_input_tokens_seen": 91459735, + "step": 2797 + }, + { + "epoch": 0.25233349866979304, + "flos": 24026557160640.0, + "grad_norm": 2.8015647228634255, + "learning_rate": 3.5035592346830846e-06, + "loss": 0.7888, + "num_input_tokens_seen": 91491015, + "step": 2798 + }, + { + "epoch": 0.2524236821932633, + "flos": 23624986343520.0, + "grad_norm": 2.4408412432148996, + "learning_rate": 3.503173933616841e-06, + "loss": 0.8479, + "num_input_tokens_seen": 91516510, + "step": 2799 + }, + { + "epoch": 0.25251386571673357, + "flos": 19944898368000.0, + "grad_norm": 4.925625779725802, + "learning_rate": 3.50278850429151e-06, + "loss": 0.8506, + "num_input_tokens_seen": 91543295, + "step": 2800 + }, + { + "epoch": 0.2526040492402038, + "flos": 19946310818880.0, + "grad_norm": 1.932740795930207, + "learning_rate": 3.502402946739977e-06, + "loss": 0.8013, + "num_input_tokens_seen": 91569100, + "step": 2801 + }, + { + "epoch": 0.2526942327636741, + "flos": 20856853090560.0, + "grad_norm": 2.5152650184619003, + "learning_rate": 3.5020172609951405e-06, + "loss": 0.7997, + "num_input_tokens_seen": 91596750, + "step": 2802 + }, + { + "epoch": 0.2527844162871443, + "flos": 27629692265760.0, + "grad_norm": 1.4671984654870946, + "learning_rate": 3.501631447089909e-06, + "loss": 0.793, + "num_input_tokens_seen": 91627960, + "step": 2803 + }, + { + "epoch": 0.2528745998106146, + "flos": 23553088560480.0, + "grad_norm": 6.860421276892225, + "learning_rate": 3.501245505057203e-06, + "loss": 0.7493, + "num_input_tokens_seen": 91656525, + "step": 2804 + }, + { + "epoch": 0.25296478333408484, + "flos": 25228927917120.0, + "grad_norm": 2.343564242313217, + "learning_rate": 3.5008594349299526e-06, + "loss": 0.8502, + "num_input_tokens_seen": 91684255, + "step": 2805 + }, + { + "epoch": 0.25305496685755513, + "flos": 36155137819200.0, + "grad_norm": 1.7668296957667946, + "learning_rate": 3.500473236741099e-06, + "loss": 0.7613, + "num_input_tokens_seen": 91713795, + "step": 2806 + }, + { + "epoch": 0.25314515038102536, + "flos": 20092410910080.0, + "grad_norm": 1.9511430373557137, + "learning_rate": 3.500086910523596e-06, + "loss": 0.7379, + "num_input_tokens_seen": 91741430, + "step": 2807 + }, + { + "epoch": 0.25323533390449565, + "flos": 50945826374400.0, + "grad_norm": 1.4195831608452603, + "learning_rate": 3.499700456310406e-06, + "loss": 0.7289, + "num_input_tokens_seen": 91776220, + "step": 2808 + }, + { + "epoch": 0.2533255174279659, + "flos": 22970286495360.0, + "grad_norm": 1.4343273977720825, + "learning_rate": 3.499313874134504e-06, + "loss": 0.8267, + "num_input_tokens_seen": 91803965, + "step": 2809 + }, + { + "epoch": 0.2534157009514362, + "flos": 22240938301920.0, + "grad_norm": 3.350024886537036, + "learning_rate": 3.498927164028875e-06, + "loss": 0.6488, + "num_input_tokens_seen": 91829950, + "step": 2810 + }, + { + "epoch": 0.25350588447490646, + "flos": 30146592145440.0, + "grad_norm": 1.3114447665958349, + "learning_rate": 3.498540326026515e-06, + "loss": 0.8162, + "num_input_tokens_seen": 91861515, + "step": 2811 + }, + { + "epoch": 0.2535960679983767, + "flos": 24607240549440.0, + "grad_norm": 1.5806741804418696, + "learning_rate": 3.4981533601604323e-06, + "loss": 0.8444, + "num_input_tokens_seen": 91891615, + "step": 2812 + }, + { + "epoch": 0.253686251521847, + "flos": 18051841702080.0, + "grad_norm": 1.902984761801011, + "learning_rate": 3.4977662664636443e-06, + "loss": 0.7886, + "num_input_tokens_seen": 91918580, + "step": 2813 + }, + { + "epoch": 0.2537764350453172, + "flos": 19946199309600.0, + "grad_norm": 1.6342771249206232, + "learning_rate": 3.497379044969179e-06, + "loss": 0.7351, + "num_input_tokens_seen": 91946445, + "step": 2814 + }, + { + "epoch": 0.2538666185687875, + "flos": 23589186131040.0, + "grad_norm": 2.2491068656842037, + "learning_rate": 3.4969916957100777e-06, + "loss": 0.7754, + "num_input_tokens_seen": 91975920, + "step": 2815 + }, + { + "epoch": 0.25395680209225774, + "flos": 28069330650720.0, + "grad_norm": 1.5256048191334504, + "learning_rate": 3.4966042187193905e-06, + "loss": 0.7703, + "num_input_tokens_seen": 92008880, + "step": 2816 + }, + { + "epoch": 0.254046985615728, + "flos": 22309416467040.0, + "grad_norm": 1.877134887913134, + "learning_rate": 3.496216614030179e-06, + "loss": 0.7742, + "num_input_tokens_seen": 92034175, + "step": 2817 + }, + { + "epoch": 0.25413716913919826, + "flos": 20673428826720.0, + "grad_norm": 2.0121800754799377, + "learning_rate": 3.495828881675516e-06, + "loss": 0.67, + "num_input_tokens_seen": 92061150, + "step": 2818 + }, + { + "epoch": 0.25422735266266855, + "flos": 22897273619520.0, + "grad_norm": 2.150853910095249, + "learning_rate": 3.4954410216884845e-06, + "loss": 0.8416, + "num_input_tokens_seen": 92082135, + "step": 2819 + }, + { + "epoch": 0.2543175361861388, + "flos": 21440026853280.0, + "grad_norm": 2.0024733746648056, + "learning_rate": 3.49505303410218e-06, + "loss": 0.7789, + "num_input_tokens_seen": 92110695, + "step": 2820 + }, + { + "epoch": 0.25440771970960907, + "flos": 24244703713920.0, + "grad_norm": 1.681699603350195, + "learning_rate": 3.4946649189497067e-06, + "loss": 0.8037, + "num_input_tokens_seen": 92143230, + "step": 2821 + }, + { + "epoch": 0.2544979032330793, + "flos": 18299729796960.0, + "grad_norm": 5.939547215939199, + "learning_rate": 3.4942766762641805e-06, + "loss": 0.7685, + "num_input_tokens_seen": 92163580, + "step": 2822 + }, + { + "epoch": 0.2545880867565496, + "flos": 21294372799200.0, + "grad_norm": 3.4547582017831706, + "learning_rate": 3.49388830607873e-06, + "loss": 0.7984, + "num_input_tokens_seen": 92191145, + "step": 2823 + }, + { + "epoch": 0.2546782702800198, + "flos": 26032255400160.0, + "grad_norm": 1.344081261190578, + "learning_rate": 3.493499808426491e-06, + "loss": 0.7409, + "num_input_tokens_seen": 92224590, + "step": 2824 + }, + { + "epoch": 0.2547684538034901, + "flos": 24354111518400.0, + "grad_norm": 1.8955246041687654, + "learning_rate": 3.493111183340614e-06, + "loss": 0.8254, + "num_input_tokens_seen": 92251900, + "step": 2825 + }, + { + "epoch": 0.25485863732696035, + "flos": 17247250447200.0, + "grad_norm": 2.2677116519014127, + "learning_rate": 3.4927224308542576e-06, + "loss": 0.6757, + "num_input_tokens_seen": 92277680, + "step": 2826 + }, + { + "epoch": 0.25494882085043064, + "flos": 23408252241120.0, + "grad_norm": 1.671766972839714, + "learning_rate": 3.4923335510005923e-06, + "loss": 0.8589, + "num_input_tokens_seen": 92309305, + "step": 2827 + }, + { + "epoch": 0.25503900437390087, + "flos": 25411422936960.0, + "grad_norm": 1.4527555426716863, + "learning_rate": 3.4919445438128e-06, + "loss": 0.7956, + "num_input_tokens_seen": 92339800, + "step": 2828 + }, + { + "epoch": 0.25512918789737116, + "flos": 23837706111840.0, + "grad_norm": 1.7593341937846607, + "learning_rate": 3.491555409324073e-06, + "loss": 0.8545, + "num_input_tokens_seen": 92368915, + "step": 2829 + }, + { + "epoch": 0.2552193714208414, + "flos": 70942879649280.0, + "grad_norm": 0.6627703942607692, + "learning_rate": 3.4911661475676136e-06, + "loss": 0.5596, + "num_input_tokens_seen": 92459625, + "step": 2830 + }, + { + "epoch": 0.2553095549443117, + "flos": 21181248018720.0, + "grad_norm": 1.5690529549769348, + "learning_rate": 3.490776758576637e-06, + "loss": 0.7556, + "num_input_tokens_seen": 92487875, + "step": 2831 + }, + { + "epoch": 0.2553997384677819, + "flos": 19761659952960.0, + "grad_norm": 1.7786494227569092, + "learning_rate": 3.4903872423843668e-06, + "loss": 0.8248, + "num_input_tokens_seen": 92515595, + "step": 2832 + }, + { + "epoch": 0.2554899219912522, + "flos": 25447557677280.0, + "grad_norm": 4.454537793296213, + "learning_rate": 3.4899975990240396e-06, + "loss": 0.7379, + "num_input_tokens_seen": 92545725, + "step": 2833 + }, + { + "epoch": 0.25558010551472243, + "flos": 19181311092000.0, + "grad_norm": 1.7550139126131097, + "learning_rate": 3.489607828528901e-06, + "loss": 0.7989, + "num_input_tokens_seen": 92571295, + "step": 2834 + }, + { + "epoch": 0.2556702890381927, + "flos": 57027565039200.0, + "grad_norm": 0.6564385960560779, + "learning_rate": 3.4892179309322093e-06, + "loss": 0.6688, + "num_input_tokens_seen": 92663850, + "step": 2835 + }, + { + "epoch": 0.255760472561663, + "flos": 19760135992800.0, + "grad_norm": 2.7043104273627177, + "learning_rate": 3.488827906267232e-06, + "loss": 0.6913, + "num_input_tokens_seen": 92691325, + "step": 2836 + }, + { + "epoch": 0.25585065608513324, + "flos": 25518340367520.0, + "grad_norm": 2.078224696012315, + "learning_rate": 3.4884377545672485e-06, + "loss": 0.8113, + "num_input_tokens_seen": 92719215, + "step": 2837 + }, + { + "epoch": 0.25594083960860353, + "flos": 24463482153120.0, + "grad_norm": 1.6342384646042005, + "learning_rate": 3.4880474758655485e-06, + "loss": 0.7473, + "num_input_tokens_seen": 92746530, + "step": 2838 + }, + { + "epoch": 0.25603102313207377, + "flos": 23509705716960.0, + "grad_norm": 1.8390490846622662, + "learning_rate": 3.487657070195433e-06, + "loss": 0.8398, + "num_input_tokens_seen": 92775195, + "step": 2839 + }, + { + "epoch": 0.25612120665554405, + "flos": 25446293905440.0, + "grad_norm": 1.7389702620812566, + "learning_rate": 3.487266537590213e-06, + "loss": 0.8513, + "num_input_tokens_seen": 92803685, + "step": 2840 + }, + { + "epoch": 0.2562113901790143, + "flos": 39143424792480.0, + "grad_norm": 1.653923127561721, + "learning_rate": 3.4868758780832116e-06, + "loss": 0.6929, + "num_input_tokens_seen": 92837435, + "step": 2841 + }, + { + "epoch": 0.2563015737024846, + "flos": 21439915344000.0, + "grad_norm": 2.080656531780526, + "learning_rate": 3.486485091707762e-06, + "loss": 0.7878, + "num_input_tokens_seen": 92863235, + "step": 2842 + }, + { + "epoch": 0.2563917572259548, + "flos": 24573782031840.0, + "grad_norm": 1.9017625112672387, + "learning_rate": 3.4860941784972077e-06, + "loss": 0.6562, + "num_input_tokens_seen": 92891850, + "step": 2843 + }, + { + "epoch": 0.2564819407494251, + "flos": 15063963596160.0, + "grad_norm": 3.0977566138769457, + "learning_rate": 3.485703138484904e-06, + "loss": 0.8702, + "num_input_tokens_seen": 92916360, + "step": 2844 + }, + { + "epoch": 0.25657212427289533, + "flos": 32588211793920.0, + "grad_norm": 1.796919847011678, + "learning_rate": 3.485311971704216e-06, + "loss": 0.8236, + "num_input_tokens_seen": 92945655, + "step": 2845 + }, + { + "epoch": 0.2566623077963656, + "flos": 23511564204960.0, + "grad_norm": 1.6308053561646878, + "learning_rate": 3.484920678188521e-06, + "loss": 0.8055, + "num_input_tokens_seen": 92974705, + "step": 2846 + }, + { + "epoch": 0.25675249131983585, + "flos": 19655485917600.0, + "grad_norm": 1.9387493412452037, + "learning_rate": 3.4845292579712063e-06, + "loss": 0.803, + "num_input_tokens_seen": 92999945, + "step": 2847 + }, + { + "epoch": 0.25684267484330614, + "flos": 20930126154720.0, + "grad_norm": 1.950416212412583, + "learning_rate": 3.484137711085669e-06, + "loss": 0.6581, + "num_input_tokens_seen": 93027805, + "step": 2848 + }, + { + "epoch": 0.2569328583667764, + "flos": 29602266515520.0, + "grad_norm": 1.533394848930252, + "learning_rate": 3.4837460375653198e-06, + "loss": 0.7594, + "num_input_tokens_seen": 93062005, + "step": 2849 + }, + { + "epoch": 0.25702304189024666, + "flos": 24682372101600.0, + "grad_norm": 2.020411050727997, + "learning_rate": 3.483354237443576e-06, + "loss": 0.808, + "num_input_tokens_seen": 93090135, + "step": 2850 + }, + { + "epoch": 0.2571132254137169, + "flos": 32221809303360.0, + "grad_norm": 1.5410802172527198, + "learning_rate": 3.48296231075387e-06, + "loss": 0.7187, + "num_input_tokens_seen": 93122275, + "step": 2851 + }, + { + "epoch": 0.2572034089371872, + "flos": 25482019778400.0, + "grad_norm": 1.8412264348093428, + "learning_rate": 3.4825702575296433e-06, + "loss": 0.7933, + "num_input_tokens_seen": 93151675, + "step": 2852 + }, + { + "epoch": 0.2572935924606574, + "flos": 67474433519040.0, + "grad_norm": 0.6967775789940804, + "learning_rate": 3.482178077804347e-06, + "loss": 0.613, + "num_input_tokens_seen": 93239075, + "step": 2853 + }, + { + "epoch": 0.2573837759841277, + "flos": 43117125401280.0, + "grad_norm": 1.6958095683214252, + "learning_rate": 3.4817857716114443e-06, + "loss": 0.739, + "num_input_tokens_seen": 93271960, + "step": 2854 + }, + { + "epoch": 0.25747395950759794, + "flos": 39510087471360.0, + "grad_norm": 2.0460207954908607, + "learning_rate": 3.4813933389844094e-06, + "loss": 0.7278, + "num_input_tokens_seen": 93303305, + "step": 2855 + }, + { + "epoch": 0.2575641430310682, + "flos": 36996495700320.0, + "grad_norm": 1.7739661684906685, + "learning_rate": 3.4810007799567264e-06, + "loss": 0.7797, + "num_input_tokens_seen": 93336355, + "step": 2856 + }, + { + "epoch": 0.25765432655453846, + "flos": 17716147166880.0, + "grad_norm": 2.2323319491265874, + "learning_rate": 3.480608094561891e-06, + "loss": 0.7913, + "num_input_tokens_seen": 93360125, + "step": 2857 + }, + { + "epoch": 0.25774451007800875, + "flos": 29670372983040.0, + "grad_norm": 1.7830946221824246, + "learning_rate": 3.4802152828334083e-06, + "loss": 0.7965, + "num_input_tokens_seen": 93390600, + "step": 2858 + }, + { + "epoch": 0.25783469360147904, + "flos": 27159494604480.0, + "grad_norm": 1.8181931105176339, + "learning_rate": 3.479822344804796e-06, + "loss": 0.7181, + "num_input_tokens_seen": 93420420, + "step": 2859 + }, + { + "epoch": 0.25792487712494927, + "flos": 19030750629600.0, + "grad_norm": 1.8000309305712454, + "learning_rate": 3.479429280509582e-06, + "loss": 0.8313, + "num_input_tokens_seen": 93446805, + "step": 2860 + }, + { + "epoch": 0.25801506064841956, + "flos": 21075780208800.0, + "grad_norm": 1.905263551718651, + "learning_rate": 3.4790360899813038e-06, + "loss": 0.8179, + "num_input_tokens_seen": 93471980, + "step": 2861 + }, + { + "epoch": 0.2581052441718898, + "flos": 16011012305760.0, + "grad_norm": 4.757857399213432, + "learning_rate": 3.4786427732535115e-06, + "loss": 0.7089, + "num_input_tokens_seen": 93499720, + "step": 2862 + }, + { + "epoch": 0.2581954276953601, + "flos": 20818153636800.0, + "grad_norm": 1.598742724201507, + "learning_rate": 3.478249330359764e-06, + "loss": 0.8199, + "num_input_tokens_seen": 93527250, + "step": 2863 + }, + { + "epoch": 0.2582856112188303, + "flos": 21622298854560.0, + "grad_norm": 2.151807105427837, + "learning_rate": 3.4778557613336333e-06, + "loss": 0.7061, + "num_input_tokens_seen": 93555270, + "step": 2864 + }, + { + "epoch": 0.2583757947423006, + "flos": 24464262718080.0, + "grad_norm": 1.7939636528560312, + "learning_rate": 3.4774620662087004e-06, + "loss": 0.7914, + "num_input_tokens_seen": 93586185, + "step": 2865 + }, + { + "epoch": 0.25846597826577083, + "flos": 20819528917920.0, + "grad_norm": 2.4091382218323028, + "learning_rate": 3.477068245018557e-06, + "loss": 0.7769, + "num_input_tokens_seen": 93613090, + "step": 2866 + }, + { + "epoch": 0.2585561617892411, + "flos": 21731446470720.0, + "grad_norm": 1.8024766599331825, + "learning_rate": 3.476674297796807e-06, + "loss": 0.781, + "num_input_tokens_seen": 93640235, + "step": 2867 + }, + { + "epoch": 0.25864634531271136, + "flos": 24682669459680.0, + "grad_norm": 1.812486188112447, + "learning_rate": 3.4762802245770627e-06, + "loss": 0.8146, + "num_input_tokens_seen": 93668900, + "step": 2868 + }, + { + "epoch": 0.25873652883618165, + "flos": 23698928463360.0, + "grad_norm": 2.0941357613087055, + "learning_rate": 3.4758860253929497e-06, + "loss": 0.7271, + "num_input_tokens_seen": 93698260, + "step": 2869 + }, + { + "epoch": 0.2588267123596519, + "flos": 29747214344160.0, + "grad_norm": 1.6802271486172093, + "learning_rate": 3.4754917002781038e-06, + "loss": 0.796, + "num_input_tokens_seen": 93728690, + "step": 2870 + }, + { + "epoch": 0.25891689588312217, + "flos": 19398788589600.0, + "grad_norm": 5.707736056004146, + "learning_rate": 3.475097249266169e-06, + "loss": 0.8147, + "num_input_tokens_seen": 93757685, + "step": 2871 + }, + { + "epoch": 0.2590070794065924, + "flos": 37761495427200.0, + "grad_norm": 1.7131705147877245, + "learning_rate": 3.4747026723908044e-06, + "loss": 0.7599, + "num_input_tokens_seen": 93788805, + "step": 2872 + }, + { + "epoch": 0.2590972629300627, + "flos": 26030248233120.0, + "grad_norm": 2.6244745182683324, + "learning_rate": 3.474307969685676e-06, + "loss": 0.7561, + "num_input_tokens_seen": 93818280, + "step": 2873 + }, + { + "epoch": 0.2591874464535329, + "flos": 28071151968960.0, + "grad_norm": 3.844514600178588, + "learning_rate": 3.473913141184462e-06, + "loss": 0.7217, + "num_input_tokens_seen": 93847915, + "step": 2874 + }, + { + "epoch": 0.2592776299770032, + "flos": 26868521024160.0, + "grad_norm": 1.7715888113509597, + "learning_rate": 3.4735181869208523e-06, + "loss": 0.6471, + "num_input_tokens_seen": 93878930, + "step": 2875 + }, + { + "epoch": 0.25936781350047344, + "flos": 24099086829600.0, + "grad_norm": 2.1400263708040645, + "learning_rate": 3.473123106928546e-06, + "loss": 0.6546, + "num_input_tokens_seen": 93905685, + "step": 2876 + }, + { + "epoch": 0.25945799702394373, + "flos": 29381889776640.0, + "grad_norm": 1.6252707475911403, + "learning_rate": 3.4727279012412533e-06, + "loss": 0.7315, + "num_input_tokens_seen": 93939020, + "step": 2877 + }, + { + "epoch": 0.25954818054741396, + "flos": 59401226899200.0, + "grad_norm": 0.5891084192728252, + "learning_rate": 3.4723325698926953e-06, + "loss": 0.575, + "num_input_tokens_seen": 94031035, + "step": 2878 + }, + { + "epoch": 0.25963836407088425, + "flos": 24493446713280.0, + "grad_norm": 1.850738986085477, + "learning_rate": 3.4719371129166045e-06, + "loss": 0.7816, + "num_input_tokens_seen": 94060220, + "step": 2879 + }, + { + "epoch": 0.2597285475943545, + "flos": 22496520537120.0, + "grad_norm": 1.8985026156086218, + "learning_rate": 3.471541530346723e-06, + "loss": 0.7008, + "num_input_tokens_seen": 94087880, + "step": 2880 + }, + { + "epoch": 0.2598187311178248, + "flos": 28980727826880.0, + "grad_norm": 2.0429262700232176, + "learning_rate": 3.4711458222168037e-06, + "loss": 0.8437, + "num_input_tokens_seen": 94115180, + "step": 2881 + }, + { + "epoch": 0.259908914641295, + "flos": 23262263659200.0, + "grad_norm": 1.5522411018764102, + "learning_rate": 3.4707499885606114e-06, + "loss": 0.8154, + "num_input_tokens_seen": 94141850, + "step": 2882 + }, + { + "epoch": 0.2599990981647653, + "flos": 21148755914880.0, + "grad_norm": 1.733126422869115, + "learning_rate": 3.4703540294119204e-06, + "loss": 0.8127, + "num_input_tokens_seen": 94169550, + "step": 2883 + }, + { + "epoch": 0.2600892816882356, + "flos": 18889556946720.0, + "grad_norm": 3.8859057625708133, + "learning_rate": 3.4699579448045163e-06, + "loss": 0.8241, + "num_input_tokens_seen": 94197945, + "step": 2884 + }, + { + "epoch": 0.2601794652117058, + "flos": 24134552514240.0, + "grad_norm": 1.6355512365224267, + "learning_rate": 3.4695617347721947e-06, + "loss": 0.6637, + "num_input_tokens_seen": 94229940, + "step": 2885 + }, + { + "epoch": 0.2602696487351761, + "flos": 20673057129120.0, + "grad_norm": 1.9562559133812008, + "learning_rate": 3.469165399348763e-06, + "loss": 0.7461, + "num_input_tokens_seen": 94257335, + "step": 2886 + }, + { + "epoch": 0.26035983225864634, + "flos": 21688732682880.0, + "grad_norm": 1.966320294519911, + "learning_rate": 3.4687689385680384e-06, + "loss": 0.7549, + "num_input_tokens_seen": 94284290, + "step": 2887 + }, + { + "epoch": 0.26045001578211663, + "flos": 27924308482560.0, + "grad_norm": 1.5173823599803031, + "learning_rate": 3.4683723524638494e-06, + "loss": 0.7557, + "num_input_tokens_seen": 94316830, + "step": 2888 + }, + { + "epoch": 0.26054019930558686, + "flos": 16667310453600.0, + "grad_norm": 2.5769915314351306, + "learning_rate": 3.4679756410700354e-06, + "loss": 0.8454, + "num_input_tokens_seen": 94344400, + "step": 2889 + }, + { + "epoch": 0.26063038282905715, + "flos": 12987445496640.0, + "grad_norm": 2.1179352175465325, + "learning_rate": 3.4675788044204445e-06, + "loss": 0.7763, + "num_input_tokens_seen": 94370140, + "step": 2890 + }, + { + "epoch": 0.2607205663525274, + "flos": 30146703654720.0, + "grad_norm": 3.0055612833936465, + "learning_rate": 3.467181842548938e-06, + "loss": 0.7067, + "num_input_tokens_seen": 94400035, + "step": 2891 + }, + { + "epoch": 0.26081074987599767, + "flos": 24934460379360.0, + "grad_norm": 1.5822161531123609, + "learning_rate": 3.466784755489387e-06, + "loss": 0.7726, + "num_input_tokens_seen": 94429995, + "step": 2892 + }, + { + "epoch": 0.2609009333994679, + "flos": 20346803712960.0, + "grad_norm": 1.8280631160858294, + "learning_rate": 3.4663875432756726e-06, + "loss": 0.8929, + "num_input_tokens_seen": 94457245, + "step": 2893 + }, + { + "epoch": 0.2609911169229382, + "flos": 19326779297280.0, + "grad_norm": 2.0612444668539998, + "learning_rate": 3.465990205941687e-06, + "loss": 0.9435, + "num_input_tokens_seen": 94486170, + "step": 2894 + }, + { + "epoch": 0.2610813004464084, + "flos": 16735379751360.0, + "grad_norm": 1.9699928962269617, + "learning_rate": 3.465592743521335e-06, + "loss": 0.7626, + "num_input_tokens_seen": 94512780, + "step": 2895 + }, + { + "epoch": 0.2611714839698787, + "flos": 26686249022880.0, + "grad_norm": 1.9794948372731402, + "learning_rate": 3.465195156048528e-06, + "loss": 0.8252, + "num_input_tokens_seen": 94542540, + "step": 2896 + }, + { + "epoch": 0.26126166749334895, + "flos": 69459422468640.0, + "grad_norm": 0.6293566258086623, + "learning_rate": 3.464797443557191e-06, + "loss": 0.5728, + "num_input_tokens_seen": 94638505, + "step": 2897 + }, + { + "epoch": 0.26135185101681924, + "flos": 34008208727040.0, + "grad_norm": 1.7202495622339122, + "learning_rate": 3.46439960608126e-06, + "loss": 0.7507, + "num_input_tokens_seen": 94671495, + "step": 2898 + }, + { + "epoch": 0.26144203454028947, + "flos": 37579409274720.0, + "grad_norm": 1.6235301632911947, + "learning_rate": 3.4640016436546797e-06, + "loss": 0.7121, + "num_input_tokens_seen": 94703685, + "step": 2899 + }, + { + "epoch": 0.26153221806375976, + "flos": 16229827914720.0, + "grad_norm": 1.6852878623504044, + "learning_rate": 3.4636035563114065e-06, + "loss": 0.7522, + "num_input_tokens_seen": 94729510, + "step": 2900 + }, + { + "epoch": 0.26162240158723, + "flos": 25484398643040.0, + "grad_norm": 1.5220062575242965, + "learning_rate": 3.4632053440854085e-06, + "loss": 0.7028, + "num_input_tokens_seen": 94760240, + "step": 2901 + }, + { + "epoch": 0.2617125851107003, + "flos": 18525235962720.0, + "grad_norm": 2.0616615209409885, + "learning_rate": 3.462807007010662e-06, + "loss": 0.8315, + "num_input_tokens_seen": 94787510, + "step": 2902 + }, + { + "epoch": 0.2618027686341705, + "flos": 18525867848640.0, + "grad_norm": 2.069804924697023, + "learning_rate": 3.462408545121155e-06, + "loss": 0.8021, + "num_input_tokens_seen": 94814630, + "step": 2903 + }, + { + "epoch": 0.2618929521576408, + "flos": 12259026547200.0, + "grad_norm": 2.0072608516474375, + "learning_rate": 3.4620099584508883e-06, + "loss": 0.8366, + "num_input_tokens_seen": 94840650, + "step": 2904 + }, + { + "epoch": 0.26198313568111103, + "flos": 17722094328480.0, + "grad_norm": 1.60050058282748, + "learning_rate": 3.46161124703387e-06, + "loss": 0.7341, + "num_input_tokens_seen": 94869410, + "step": 2905 + }, + { + "epoch": 0.2620733192045813, + "flos": 21731000433600.0, + "grad_norm": 2.0536725108224445, + "learning_rate": 3.461212410904122e-06, + "loss": 0.7694, + "num_input_tokens_seen": 94898550, + "step": 2906 + }, + { + "epoch": 0.2621635027280516, + "flos": 18051284155680.0, + "grad_norm": 2.6335635599665226, + "learning_rate": 3.4608134500956726e-06, + "loss": 0.8321, + "num_input_tokens_seen": 94923725, + "step": 2907 + }, + { + "epoch": 0.26225368625152184, + "flos": 29670670341120.0, + "grad_norm": 1.4333688835052796, + "learning_rate": 3.4604143646425655e-06, + "loss": 0.7639, + "num_input_tokens_seen": 94954790, + "step": 2908 + }, + { + "epoch": 0.26234386977499213, + "flos": 26139581698080.0, + "grad_norm": 1.4305197372808247, + "learning_rate": 3.460015154578852e-06, + "loss": 0.8465, + "num_input_tokens_seen": 94987670, + "step": 2909 + }, + { + "epoch": 0.26243405329846237, + "flos": 24353888499840.0, + "grad_norm": 1.9947573971771955, + "learning_rate": 3.459615819938595e-06, + "loss": 0.7508, + "num_input_tokens_seen": 95019200, + "step": 2910 + }, + { + "epoch": 0.26252423682193265, + "flos": 14582280479040.0, + "grad_norm": 1.9908378068126664, + "learning_rate": 3.4592163607558684e-06, + "loss": 0.7393, + "num_input_tokens_seen": 95045070, + "step": 2911 + }, + { + "epoch": 0.2626144203454029, + "flos": 22201160925120.0, + "grad_norm": 1.5067647347281234, + "learning_rate": 3.4588167770647553e-06, + "loss": 0.7778, + "num_input_tokens_seen": 95073140, + "step": 2912 + }, + { + "epoch": 0.2627046038688732, + "flos": 20638186160640.0, + "grad_norm": 2.278998112649709, + "learning_rate": 3.458417068899351e-06, + "loss": 0.8706, + "num_input_tokens_seen": 95098975, + "step": 2913 + }, + { + "epoch": 0.2627947873923434, + "flos": 24389019656640.0, + "grad_norm": 1.7122830190818086, + "learning_rate": 3.4580172362937612e-06, + "loss": 0.7471, + "num_input_tokens_seen": 95127680, + "step": 2914 + }, + { + "epoch": 0.2628849709158137, + "flos": 25336402894080.0, + "grad_norm": 1.551569663777481, + "learning_rate": 3.457617279282101e-06, + "loss": 0.743, + "num_input_tokens_seen": 95157850, + "step": 2915 + }, + { + "epoch": 0.26297515443928393, + "flos": 19799504502240.0, + "grad_norm": 2.5341369742908086, + "learning_rate": 3.4572171978984975e-06, + "loss": 0.7571, + "num_input_tokens_seen": 95183080, + "step": 2916 + }, + { + "epoch": 0.2630653379627542, + "flos": 20637479935200.0, + "grad_norm": 2.6607451158238766, + "learning_rate": 3.456816992177088e-06, + "loss": 0.7757, + "num_input_tokens_seen": 95212360, + "step": 2917 + }, + { + "epoch": 0.26315552148622445, + "flos": 22093797457440.0, + "grad_norm": 1.5424145439625645, + "learning_rate": 3.4564166621520193e-06, + "loss": 0.746, + "num_input_tokens_seen": 95243420, + "step": 2918 + }, + { + "epoch": 0.26324570500969474, + "flos": 21618507539040.0, + "grad_norm": 1.739032300222466, + "learning_rate": 3.4560162078574507e-06, + "loss": 0.7685, + "num_input_tokens_seen": 95271640, + "step": 2919 + }, + { + "epoch": 0.263335888533165, + "flos": 21547985037120.0, + "grad_norm": 1.7190952301099458, + "learning_rate": 3.455615629327551e-06, + "loss": 0.7778, + "num_input_tokens_seen": 95299195, + "step": 2920 + }, + { + "epoch": 0.26342607205663526, + "flos": 22788088833600.0, + "grad_norm": 1.8902609608643581, + "learning_rate": 3.4552149265964994e-06, + "loss": 0.7501, + "num_input_tokens_seen": 95327750, + "step": 2921 + }, + { + "epoch": 0.2635162555801055, + "flos": 30728613645600.0, + "grad_norm": 1.7368080325564235, + "learning_rate": 3.4548140996984866e-06, + "loss": 0.7274, + "num_input_tokens_seen": 95358900, + "step": 2922 + }, + { + "epoch": 0.2636064391035758, + "flos": 30256111459200.0, + "grad_norm": 1.5227332419145168, + "learning_rate": 3.4544131486677124e-06, + "loss": 0.7963, + "num_input_tokens_seen": 95389540, + "step": 2923 + }, + { + "epoch": 0.263696622627046, + "flos": 52987845203040.0, + "grad_norm": 1.6142248760724156, + "learning_rate": 3.454012073538389e-06, + "loss": 0.7167, + "num_input_tokens_seen": 95424305, + "step": 2924 + }, + { + "epoch": 0.2637868061505163, + "flos": 20383607508960.0, + "grad_norm": 1.5511986005138363, + "learning_rate": 3.453610874344738e-06, + "loss": 0.7834, + "num_input_tokens_seen": 95452405, + "step": 2925 + }, + { + "epoch": 0.26387698967398654, + "flos": 18598694875680.0, + "grad_norm": 1.684946034866214, + "learning_rate": 3.453209551120993e-06, + "loss": 0.8179, + "num_input_tokens_seen": 95480965, + "step": 2926 + }, + { + "epoch": 0.26396717319745683, + "flos": 20347175410560.0, + "grad_norm": 2.7021019383252955, + "learning_rate": 3.452808103901395e-06, + "loss": 0.7469, + "num_input_tokens_seen": 95510195, + "step": 2927 + }, + { + "epoch": 0.26405735672092706, + "flos": 26540929496640.0, + "grad_norm": 2.0082857635478195, + "learning_rate": 3.4524065327202e-06, + "loss": 0.8505, + "num_input_tokens_seen": 95537205, + "step": 2928 + }, + { + "epoch": 0.26414754024439735, + "flos": 66422697564480.0, + "grad_norm": 0.6388440148055818, + "learning_rate": 3.4520048376116702e-06, + "loss": 0.5252, + "num_input_tokens_seen": 95628895, + "step": 2929 + }, + { + "epoch": 0.26423772376786764, + "flos": 18051284155680.0, + "grad_norm": 1.460168401464494, + "learning_rate": 3.4516030186100817e-06, + "loss": 0.8123, + "num_input_tokens_seen": 95657130, + "step": 2930 + }, + { + "epoch": 0.26432790729133787, + "flos": 40749485042400.0, + "grad_norm": 1.49533810967306, + "learning_rate": 3.4512010757497197e-06, + "loss": 0.6921, + "num_input_tokens_seen": 95687670, + "step": 2931 + }, + { + "epoch": 0.26441809081480816, + "flos": 26576692539360.0, + "grad_norm": 1.7672435127643256, + "learning_rate": 3.4507990090648804e-06, + "loss": 0.694, + "num_input_tokens_seen": 95718025, + "step": 2932 + }, + { + "epoch": 0.2645082743382784, + "flos": 27377938515840.0, + "grad_norm": 1.7367534652857697, + "learning_rate": 3.4503968185898696e-06, + "loss": 0.7943, + "num_input_tokens_seen": 95744675, + "step": 2933 + }, + { + "epoch": 0.2645984578617487, + "flos": 18088013612160.0, + "grad_norm": 1.7039110662015338, + "learning_rate": 3.4499945043590047e-06, + "loss": 0.7807, + "num_input_tokens_seen": 95770025, + "step": 2934 + }, + { + "epoch": 0.2646886413852189, + "flos": 32188016257920.0, + "grad_norm": 1.6271770272852921, + "learning_rate": 3.4495920664066137e-06, + "loss": 0.7887, + "num_input_tokens_seen": 95798340, + "step": 2935 + }, + { + "epoch": 0.2647788249086892, + "flos": 26867963477760.0, + "grad_norm": 1.975443225331515, + "learning_rate": 3.449189504767035e-06, + "loss": 0.7613, + "num_input_tokens_seen": 95824005, + "step": 2936 + }, + { + "epoch": 0.26486900843215944, + "flos": 30511693694400.0, + "grad_norm": 2.01897170322762, + "learning_rate": 3.4487868194746163e-06, + "loss": 0.7998, + "num_input_tokens_seen": 95852670, + "step": 2937 + }, + { + "epoch": 0.2649591919556297, + "flos": 19217594511360.0, + "grad_norm": 1.8551207527280127, + "learning_rate": 3.4483840105637175e-06, + "loss": 0.8178, + "num_input_tokens_seen": 95880710, + "step": 2938 + }, + { + "epoch": 0.26504937547909996, + "flos": 22970063476800.0, + "grad_norm": 1.9124429507255016, + "learning_rate": 3.4479810780687097e-06, + "loss": 0.7605, + "num_input_tokens_seen": 95909205, + "step": 2939 + }, + { + "epoch": 0.26513955900257025, + "flos": 27961855673760.0, + "grad_norm": 1.580217088923614, + "learning_rate": 3.4475780220239714e-06, + "loss": 0.765, + "num_input_tokens_seen": 95942535, + "step": 2940 + }, + { + "epoch": 0.2652297425260405, + "flos": 22278262474560.0, + "grad_norm": 1.3595529637095005, + "learning_rate": 3.4471748424638948e-06, + "loss": 0.7663, + "num_input_tokens_seen": 95973150, + "step": 2941 + }, + { + "epoch": 0.26531992604951077, + "flos": 57769327932480.0, + "grad_norm": 0.5987592255632203, + "learning_rate": 3.4467715394228803e-06, + "loss": 0.5925, + "num_input_tokens_seen": 96069910, + "step": 2942 + }, + { + "epoch": 0.265410109572981, + "flos": 26287131409920.0, + "grad_norm": 1.5036876172755764, + "learning_rate": 3.4463681129353413e-06, + "loss": 0.7764, + "num_input_tokens_seen": 96100885, + "step": 2943 + }, + { + "epoch": 0.2655002930964513, + "flos": 21404300980320.0, + "grad_norm": 1.7791774135982694, + "learning_rate": 3.4459645630357e-06, + "loss": 0.7688, + "num_input_tokens_seen": 96131525, + "step": 2944 + }, + { + "epoch": 0.2655904766199215, + "flos": 16485224301120.0, + "grad_norm": 2.046985652349643, + "learning_rate": 3.4455608897583884e-06, + "loss": 0.8093, + "num_input_tokens_seen": 96156885, + "step": 2945 + }, + { + "epoch": 0.2656806601433918, + "flos": 19072349324640.0, + "grad_norm": 2.757037867178452, + "learning_rate": 3.4451570931378514e-06, + "loss": 0.775, + "num_input_tokens_seen": 96183875, + "step": 2946 + }, + { + "epoch": 0.26577084366686204, + "flos": 28798492995360.0, + "grad_norm": 1.3874900284829832, + "learning_rate": 3.444753173208543e-06, + "loss": 0.7836, + "num_input_tokens_seen": 96214905, + "step": 2947 + }, + { + "epoch": 0.26586102719033233, + "flos": 19691063111520.0, + "grad_norm": 1.8942581451832963, + "learning_rate": 3.444349130004927e-06, + "loss": 0.7694, + "num_input_tokens_seen": 96242550, + "step": 2948 + }, + { + "epoch": 0.26595121071380257, + "flos": 26206870430880.0, + "grad_norm": 1.8578381638482797, + "learning_rate": 3.4439449635614794e-06, + "loss": 0.6963, + "num_input_tokens_seen": 96269255, + "step": 2949 + }, + { + "epoch": 0.26604139423727285, + "flos": 21622930740480.0, + "grad_norm": 2.195305114950879, + "learning_rate": 3.4435406739126854e-06, + "loss": 0.8033, + "num_input_tokens_seen": 96297610, + "step": 2950 + }, + { + "epoch": 0.2661315777607431, + "flos": 14627261622240.0, + "grad_norm": 2.6427149881698386, + "learning_rate": 3.443136261093042e-06, + "loss": 0.5938, + "num_input_tokens_seen": 96323060, + "step": 2951 + }, + { + "epoch": 0.2662217612842134, + "flos": 27665715496800.0, + "grad_norm": 1.6804578567073782, + "learning_rate": 3.4427317251370553e-06, + "loss": 0.7195, + "num_input_tokens_seen": 96352680, + "step": 2952 + }, + { + "epoch": 0.2663119448076836, + "flos": 29083073376960.0, + "grad_norm": 1.8585010686650152, + "learning_rate": 3.4423270660792422e-06, + "loss": 0.6866, + "num_input_tokens_seen": 96382320, + "step": 2953 + }, + { + "epoch": 0.2664021283311539, + "flos": 24973828888800.0, + "grad_norm": 1.7215445255013566, + "learning_rate": 3.4419222839541314e-06, + "loss": 0.6634, + "num_input_tokens_seen": 96413010, + "step": 2954 + }, + { + "epoch": 0.2664923118546242, + "flos": 29455200010560.0, + "grad_norm": 1.5135579909275478, + "learning_rate": 3.4415173787962607e-06, + "loss": 0.7462, + "num_input_tokens_seen": 96444640, + "step": 2955 + }, + { + "epoch": 0.2665824953780944, + "flos": 32406162811200.0, + "grad_norm": 10.654772882641096, + "learning_rate": 3.4411123506401783e-06, + "loss": 0.6373, + "num_input_tokens_seen": 96473360, + "step": 2956 + }, + { + "epoch": 0.2666726789015647, + "flos": 39728271194400.0, + "grad_norm": 3.7124543312423466, + "learning_rate": 3.440707199520444e-06, + "loss": 0.6279, + "num_input_tokens_seen": 96504205, + "step": 2957 + }, + { + "epoch": 0.26676286242503494, + "flos": 23589260470560.0, + "grad_norm": 1.9713407757067862, + "learning_rate": 3.440301925471628e-06, + "loss": 0.7339, + "num_input_tokens_seen": 96533685, + "step": 2958 + }, + { + "epoch": 0.26685304594850523, + "flos": 29925174653280.0, + "grad_norm": 1.5561621491787876, + "learning_rate": 3.43989652852831e-06, + "loss": 0.8094, + "num_input_tokens_seen": 96564435, + "step": 2959 + }, + { + "epoch": 0.26694322947197546, + "flos": 31459337120160.0, + "grad_norm": 1.8036805353251286, + "learning_rate": 3.4394910087250804e-06, + "loss": 0.8341, + "num_input_tokens_seen": 96595105, + "step": 2960 + }, + { + "epoch": 0.26703341299544575, + "flos": 20782056066240.0, + "grad_norm": 1.5748125570262184, + "learning_rate": 3.4390853660965405e-06, + "loss": 0.7904, + "num_input_tokens_seen": 96624965, + "step": 2961 + }, + { + "epoch": 0.267123596518916, + "flos": 23844991384800.0, + "grad_norm": 2.537359705462027, + "learning_rate": 3.438679600677302e-06, + "loss": 0.8517, + "num_input_tokens_seen": 96653715, + "step": 2962 + }, + { + "epoch": 0.2672137800423863, + "flos": 28324057981440.0, + "grad_norm": 1.7344989904124186, + "learning_rate": 3.4382737125019874e-06, + "loss": 0.7286, + "num_input_tokens_seen": 96683550, + "step": 2963 + }, + { + "epoch": 0.2673039635658565, + "flos": 21947251329120.0, + "grad_norm": 1.7076962919215186, + "learning_rate": 3.4378677016052294e-06, + "loss": 0.7415, + "num_input_tokens_seen": 96713580, + "step": 2964 + }, + { + "epoch": 0.2673941470893268, + "flos": 19472693539680.0, + "grad_norm": 2.1545064760200656, + "learning_rate": 3.43746156802167e-06, + "loss": 0.7914, + "num_input_tokens_seen": 96740025, + "step": 2965 + }, + { + "epoch": 0.267484330612797, + "flos": 21069052482240.0, + "grad_norm": 1.97565219106188, + "learning_rate": 3.4370553117859643e-06, + "loss": 0.7064, + "num_input_tokens_seen": 96767105, + "step": 2966 + }, + { + "epoch": 0.2675745141362673, + "flos": 24317084703840.0, + "grad_norm": 1.7502081610990545, + "learning_rate": 3.4366489329327754e-06, + "loss": 0.801, + "num_input_tokens_seen": 96795835, + "step": 2967 + }, + { + "epoch": 0.26766469765973755, + "flos": 22634406111360.0, + "grad_norm": 2.8744531194261644, + "learning_rate": 3.4362424314967777e-06, + "loss": 0.7918, + "num_input_tokens_seen": 96823815, + "step": 2968 + }, + { + "epoch": 0.26775488118320784, + "flos": 18342963961440.0, + "grad_norm": 2.0878090909448432, + "learning_rate": 3.4358358075126567e-06, + "loss": 0.7998, + "num_input_tokens_seen": 96850360, + "step": 2969 + }, + { + "epoch": 0.26784506470667807, + "flos": 20930126154720.0, + "grad_norm": 1.8309439419634537, + "learning_rate": 3.4354290610151077e-06, + "loss": 0.7927, + "num_input_tokens_seen": 96880825, + "step": 2970 + }, + { + "epoch": 0.26793524823014836, + "flos": 34736925034560.0, + "grad_norm": 1.3898948295517402, + "learning_rate": 3.4350221920388354e-06, + "loss": 0.7809, + "num_input_tokens_seen": 96915200, + "step": 2971 + }, + { + "epoch": 0.2680254317536186, + "flos": 24421809118560.0, + "grad_norm": 1.5180229016361932, + "learning_rate": 3.4346152006185574e-06, + "loss": 0.8741, + "num_input_tokens_seen": 96943490, + "step": 2972 + }, + { + "epoch": 0.2681156152770889, + "flos": 17468890957920.0, + "grad_norm": 1.8850447215838715, + "learning_rate": 3.4342080867890006e-06, + "loss": 0.875, + "num_input_tokens_seen": 96968625, + "step": 2973 + }, + { + "epoch": 0.2682057988005591, + "flos": 14117026395840.0, + "grad_norm": 2.063480837556208, + "learning_rate": 3.4338008505849016e-06, + "loss": 0.6915, + "num_input_tokens_seen": 96993715, + "step": 2974 + }, + { + "epoch": 0.2682959823240294, + "flos": 27194365572960.0, + "grad_norm": 2.4035063124832288, + "learning_rate": 3.433393492041008e-06, + "loss": 0.7831, + "num_input_tokens_seen": 97024615, + "step": 2975 + }, + { + "epoch": 0.26838616584749964, + "flos": 23876702923680.0, + "grad_norm": 1.9134185330542552, + "learning_rate": 3.432986011192078e-06, + "loss": 0.8218, + "num_input_tokens_seen": 97053470, + "step": 2976 + }, + { + "epoch": 0.2684763493709699, + "flos": 19763852968800.0, + "grad_norm": 1.9456819237145762, + "learning_rate": 3.4325784080728796e-06, + "loss": 0.8418, + "num_input_tokens_seen": 97079800, + "step": 2977 + }, + { + "epoch": 0.2685665328944402, + "flos": 30399275139360.0, + "grad_norm": 1.595741150121638, + "learning_rate": 3.4321706827181926e-06, + "loss": 0.8043, + "num_input_tokens_seen": 97109655, + "step": 2978 + }, + { + "epoch": 0.26865671641791045, + "flos": 21439766664960.0, + "grad_norm": 1.9213013451887364, + "learning_rate": 3.4317628351628064e-06, + "loss": 0.726, + "num_input_tokens_seen": 97138940, + "step": 2979 + }, + { + "epoch": 0.26874689994138073, + "flos": 23909232197280.0, + "grad_norm": 1.427446286715805, + "learning_rate": 3.43135486544152e-06, + "loss": 0.7583, + "num_input_tokens_seen": 97165545, + "step": 2980 + }, + { + "epoch": 0.26883708346485097, + "flos": 25586409665280.0, + "grad_norm": 2.359656108758037, + "learning_rate": 3.4309467735891442e-06, + "loss": 0.8134, + "num_input_tokens_seen": 97194910, + "step": 2981 + }, + { + "epoch": 0.26892726698832126, + "flos": 20346878052480.0, + "grad_norm": 1.7662997011614763, + "learning_rate": 3.4305385596405e-06, + "loss": 0.8603, + "num_input_tokens_seen": 97222345, + "step": 2982 + }, + { + "epoch": 0.2690174505117915, + "flos": 22022791748640.0, + "grad_norm": 1.6430418831750924, + "learning_rate": 3.4301302236304174e-06, + "loss": 0.7393, + "num_input_tokens_seen": 97251825, + "step": 2983 + }, + { + "epoch": 0.2691076340352618, + "flos": 61363616634720.0, + "grad_norm": 0.5988410724352501, + "learning_rate": 3.429721765593739e-06, + "loss": 0.5472, + "num_input_tokens_seen": 97338520, + "step": 2984 + }, + { + "epoch": 0.269197817558732, + "flos": 21984315313440.0, + "grad_norm": 1.5535206465415763, + "learning_rate": 3.4293131855653155e-06, + "loss": 0.8009, + "num_input_tokens_seen": 97367480, + "step": 2985 + }, + { + "epoch": 0.2692880010822023, + "flos": 37543237364640.0, + "grad_norm": 1.7620297853071873, + "learning_rate": 3.4289044835800102e-06, + "loss": 0.7127, + "num_input_tokens_seen": 97398355, + "step": 2986 + }, + { + "epoch": 0.26937818460567253, + "flos": 18488989713120.0, + "grad_norm": 1.5732243049689758, + "learning_rate": 3.4284956596726953e-06, + "loss": 0.7777, + "num_input_tokens_seen": 97425665, + "step": 2987 + }, + { + "epoch": 0.2694683681291428, + "flos": 27777316317120.0, + "grad_norm": 1.9745151938778522, + "learning_rate": 3.4280867138782544e-06, + "loss": 0.768, + "num_input_tokens_seen": 97457125, + "step": 2988 + }, + { + "epoch": 0.26955855165261305, + "flos": 21650627945280.0, + "grad_norm": 2.489466697695377, + "learning_rate": 3.4276776462315803e-06, + "loss": 0.7777, + "num_input_tokens_seen": 97483685, + "step": 2989 + }, + { + "epoch": 0.26964873517608334, + "flos": 18809407476960.0, + "grad_norm": 1.6500268298193017, + "learning_rate": 3.427268456767578e-06, + "loss": 0.7751, + "num_input_tokens_seen": 97509660, + "step": 2990 + }, + { + "epoch": 0.2697389186995536, + "flos": 21069870216960.0, + "grad_norm": 2.219562972586298, + "learning_rate": 3.42685914552116e-06, + "loss": 0.8133, + "num_input_tokens_seen": 97536225, + "step": 2991 + }, + { + "epoch": 0.26982910222302386, + "flos": 22023237785760.0, + "grad_norm": 1.9870878284051816, + "learning_rate": 3.426449712527253e-06, + "loss": 0.749, + "num_input_tokens_seen": 97564175, + "step": 2992 + }, + { + "epoch": 0.2699192857464941, + "flos": 25848719627040.0, + "grad_norm": 5.58799797416328, + "learning_rate": 3.4260401578207904e-06, + "loss": 0.6846, + "num_input_tokens_seen": 97596650, + "step": 2993 + }, + { + "epoch": 0.2700094692699644, + "flos": 32404192813920.0, + "grad_norm": 2.1908371674320435, + "learning_rate": 3.4256304814367185e-06, + "loss": 0.7321, + "num_input_tokens_seen": 97624050, + "step": 2994 + }, + { + "epoch": 0.2700996527934346, + "flos": 24207937087680.0, + "grad_norm": 1.8061113561727018, + "learning_rate": 3.4252206834099936e-06, + "loss": 0.7297, + "num_input_tokens_seen": 97649350, + "step": 2995 + }, + { + "epoch": 0.2701898363169049, + "flos": 26099692812000.0, + "grad_norm": 5.406224245219542, + "learning_rate": 3.424810763775581e-06, + "loss": 0.7699, + "num_input_tokens_seen": 97680335, + "step": 2996 + }, + { + "epoch": 0.27028001984037514, + "flos": 34007911368960.0, + "grad_norm": 1.409405452965973, + "learning_rate": 3.4244007225684587e-06, + "loss": 0.7149, + "num_input_tokens_seen": 97712950, + "step": 2997 + }, + { + "epoch": 0.27037020336384543, + "flos": 22423470491520.0, + "grad_norm": 1.5507066069734967, + "learning_rate": 3.4239905598236115e-06, + "loss": 0.7802, + "num_input_tokens_seen": 97742325, + "step": 2998 + }, + { + "epoch": 0.27046038688731566, + "flos": 22677082729440.0, + "grad_norm": 1.6620425290428544, + "learning_rate": 3.4235802755760386e-06, + "loss": 0.7993, + "num_input_tokens_seen": 97770625, + "step": 2999 + }, + { + "epoch": 0.27055057041078595, + "flos": 64699009259520.0, + "grad_norm": 0.8484574499998787, + "learning_rate": 3.4231698698607464e-06, + "loss": 0.6941, + "num_input_tokens_seen": 97858755, + "step": 3000 + }, + { + "epoch": 0.2706407539342562, + "flos": 25334432896800.0, + "grad_norm": 1.7491510920795188, + "learning_rate": 3.4227593427127543e-06, + "loss": 0.8292, + "num_input_tokens_seen": 97887595, + "step": 3001 + }, + { + "epoch": 0.2707309374577265, + "flos": 35867063480160.0, + "grad_norm": 1.9537309240775453, + "learning_rate": 3.42234869416709e-06, + "loss": 0.745, + "num_input_tokens_seen": 97921260, + "step": 3002 + }, + { + "epoch": 0.27082112098119676, + "flos": 17504728340160.0, + "grad_norm": 1.979957830673605, + "learning_rate": 3.421937924258792e-06, + "loss": 0.7758, + "num_input_tokens_seen": 97946045, + "step": 3003 + }, + { + "epoch": 0.270911304504667, + "flos": 29125973013600.0, + "grad_norm": 1.7956419321304826, + "learning_rate": 3.4215270330229096e-06, + "loss": 0.8154, + "num_input_tokens_seen": 97975800, + "step": 3004 + }, + { + "epoch": 0.2710014880281373, + "flos": 25155915041280.0, + "grad_norm": 2.071651270597683, + "learning_rate": 3.421116020494503e-06, + "loss": 0.7394, + "num_input_tokens_seen": 98003325, + "step": 3005 + }, + { + "epoch": 0.2710916715516075, + "flos": 49272180033600.0, + "grad_norm": 1.3530339076014775, + "learning_rate": 3.420704886708642e-06, + "loss": 0.7207, + "num_input_tokens_seen": 98036540, + "step": 3006 + }, + { + "epoch": 0.2711818550750778, + "flos": 28543988683200.0, + "grad_norm": 1.6544291725988138, + "learning_rate": 3.4202936317004056e-06, + "loss": 0.725, + "num_input_tokens_seen": 98067405, + "step": 3007 + }, + { + "epoch": 0.27127203859854804, + "flos": 20018505960000.0, + "grad_norm": 2.3571329137253865, + "learning_rate": 3.4198822555048856e-06, + "loss": 0.7895, + "num_input_tokens_seen": 98095295, + "step": 3008 + }, + { + "epoch": 0.2713622221220183, + "flos": 13825123571520.0, + "grad_norm": 2.1015074768804705, + "learning_rate": 3.419470758157182e-06, + "loss": 0.8569, + "num_input_tokens_seen": 98119985, + "step": 3009 + }, + { + "epoch": 0.27145240564548856, + "flos": 25405661624160.0, + "grad_norm": 1.7959964096951015, + "learning_rate": 3.4190591396924068e-06, + "loss": 0.7697, + "num_input_tokens_seen": 98147260, + "step": 3010 + }, + { + "epoch": 0.27154258916895885, + "flos": 56116056353760.0, + "grad_norm": 0.6741098229743854, + "learning_rate": 3.418647400145681e-06, + "loss": 0.6155, + "num_input_tokens_seen": 98227750, + "step": 3011 + }, + { + "epoch": 0.2716327726924291, + "flos": 67372050799200.0, + "grad_norm": 0.6984053841189988, + "learning_rate": 3.4182355395521367e-06, + "loss": 0.575, + "num_input_tokens_seen": 98310765, + "step": 3012 + }, + { + "epoch": 0.27172295621589937, + "flos": 28613730620160.0, + "grad_norm": 2.5339872842121904, + "learning_rate": 3.417823557946916e-06, + "loss": 0.7708, + "num_input_tokens_seen": 98339790, + "step": 3013 + }, + { + "epoch": 0.2718131397393696, + "flos": 30509240490240.0, + "grad_norm": 1.6281853047805988, + "learning_rate": 3.417411455365172e-06, + "loss": 0.6726, + "num_input_tokens_seen": 98369350, + "step": 3014 + }, + { + "epoch": 0.2719033232628399, + "flos": 22096176322080.0, + "grad_norm": 2.102108682329008, + "learning_rate": 3.416999231842066e-06, + "loss": 0.6842, + "num_input_tokens_seen": 98398670, + "step": 3015 + }, + { + "epoch": 0.2719935067863101, + "flos": 19363620263040.0, + "grad_norm": 1.9916038705457597, + "learning_rate": 3.416586887412773e-06, + "loss": 0.8246, + "num_input_tokens_seen": 98426275, + "step": 3016 + }, + { + "epoch": 0.2720836903097804, + "flos": 25373764236480.0, + "grad_norm": 2.4434342163235607, + "learning_rate": 3.416174422112476e-06, + "loss": 0.7388, + "num_input_tokens_seen": 98455945, + "step": 3017 + }, + { + "epoch": 0.27217387383325065, + "flos": 67165166683200.0, + "grad_norm": 0.6037705509566017, + "learning_rate": 3.4157618359763687e-06, + "loss": 0.539, + "num_input_tokens_seen": 98553200, + "step": 3018 + }, + { + "epoch": 0.27226405735672093, + "flos": 22635186676320.0, + "grad_norm": 1.6759352535451115, + "learning_rate": 3.4153491290396542e-06, + "loss": 0.6927, + "num_input_tokens_seen": 98581920, + "step": 3019 + }, + { + "epoch": 0.27235424088019117, + "flos": 25590535508640.0, + "grad_norm": 2.248014739840095, + "learning_rate": 3.4149363013375485e-06, + "loss": 0.7919, + "num_input_tokens_seen": 98613155, + "step": 3020 + }, + { + "epoch": 0.27244442440366146, + "flos": 66782446668000.0, + "grad_norm": 0.6954117393395278, + "learning_rate": 3.414523352905276e-06, + "loss": 0.5551, + "num_input_tokens_seen": 98698085, + "step": 3021 + }, + { + "epoch": 0.2725346079271317, + "flos": 32477354368800.0, + "grad_norm": 2.1667545251883786, + "learning_rate": 3.414110283778071e-06, + "loss": 0.7415, + "num_input_tokens_seen": 98728555, + "step": 3022 + }, + { + "epoch": 0.272624791450602, + "flos": 21877174864320.0, + "grad_norm": 2.0826304076396847, + "learning_rate": 3.4136970939911797e-06, + "loss": 0.7183, + "num_input_tokens_seen": 98756010, + "step": 3023 + }, + { + "epoch": 0.2727149749740722, + "flos": 24242250509760.0, + "grad_norm": 4.813508559724498, + "learning_rate": 3.413283783579857e-06, + "loss": 0.7393, + "num_input_tokens_seen": 98783055, + "step": 3024 + }, + { + "epoch": 0.2728051584975425, + "flos": 15465534413280.0, + "grad_norm": 2.8155717817182024, + "learning_rate": 3.412870352579369e-06, + "loss": 0.7076, + "num_input_tokens_seen": 98808295, + "step": 3025 + }, + { + "epoch": 0.2728953420210128, + "flos": 66460727962560.0, + "grad_norm": 0.646395328531818, + "learning_rate": 3.4124568010249915e-06, + "loss": 0.5967, + "num_input_tokens_seen": 98900435, + "step": 3026 + }, + { + "epoch": 0.272985525544483, + "flos": 11566705168320.0, + "grad_norm": 1.7831828689678177, + "learning_rate": 3.4120431289520124e-06, + "loss": 0.7576, + "num_input_tokens_seen": 98927285, + "step": 3027 + }, + { + "epoch": 0.2730757090679533, + "flos": 27997841735040.0, + "grad_norm": 2.2727260459549803, + "learning_rate": 3.4116293363957276e-06, + "loss": 0.7775, + "num_input_tokens_seen": 98956210, + "step": 3028 + }, + { + "epoch": 0.27316589259142354, + "flos": 20928713703840.0, + "grad_norm": 1.9362694308720168, + "learning_rate": 3.4112154233914438e-06, + "loss": 0.7804, + "num_input_tokens_seen": 98982655, + "step": 3029 + }, + { + "epoch": 0.27325607611489383, + "flos": 31783286011200.0, + "grad_norm": 2.5419516690249426, + "learning_rate": 3.410801389974479e-06, + "loss": 0.6771, + "num_input_tokens_seen": 99012105, + "step": 3030 + }, + { + "epoch": 0.27334625963836406, + "flos": 24864198065760.0, + "grad_norm": 1.7952199196579017, + "learning_rate": 3.410387236180161e-06, + "loss": 0.8058, + "num_input_tokens_seen": 99040010, + "step": 3031 + }, + { + "epoch": 0.27343644316183435, + "flos": 33134544590880.0, + "grad_norm": 1.8123937772997212, + "learning_rate": 3.409972962043826e-06, + "loss": 0.6567, + "num_input_tokens_seen": 99071205, + "step": 3032 + }, + { + "epoch": 0.2735266266853046, + "flos": 20310074256480.0, + "grad_norm": 2.3605673095182254, + "learning_rate": 3.4095585676008234e-06, + "loss": 0.8278, + "num_input_tokens_seen": 99098795, + "step": 3033 + }, + { + "epoch": 0.2736168102087749, + "flos": 24056001344160.0, + "grad_norm": 1.9197074017493652, + "learning_rate": 3.4091440528865125e-06, + "loss": 0.7671, + "num_input_tokens_seen": 99127485, + "step": 3034 + }, + { + "epoch": 0.2737069937322451, + "flos": 14591052542400.0, + "grad_norm": 1.7885644211059453, + "learning_rate": 3.4087294179362606e-06, + "loss": 0.7693, + "num_input_tokens_seen": 99153260, + "step": 3035 + }, + { + "epoch": 0.2737971772557154, + "flos": 26026159559520.0, + "grad_norm": 1.9298987639761773, + "learning_rate": 3.4083146627854474e-06, + "loss": 0.8155, + "num_input_tokens_seen": 99181345, + "step": 3036 + }, + { + "epoch": 0.27388736077918563, + "flos": 22824372252960.0, + "grad_norm": 2.4886706180835785, + "learning_rate": 3.4078997874694614e-06, + "loss": 0.6987, + "num_input_tokens_seen": 99210700, + "step": 3037 + }, + { + "epoch": 0.2739775443026559, + "flos": 24828286344000.0, + "grad_norm": 1.9363160361350684, + "learning_rate": 3.407484792023703e-06, + "loss": 0.6624, + "num_input_tokens_seen": 99238555, + "step": 3038 + }, + { + "epoch": 0.27406772782612615, + "flos": 27451397428800.0, + "grad_norm": 2.0960344975158716, + "learning_rate": 3.407069676483581e-06, + "loss": 0.7375, + "num_input_tokens_seen": 99266050, + "step": 3039 + }, + { + "epoch": 0.27415791134959644, + "flos": 28467630528960.0, + "grad_norm": 1.718252630228699, + "learning_rate": 3.406654440884516e-06, + "loss": 0.8055, + "num_input_tokens_seen": 99294945, + "step": 3040 + }, + { + "epoch": 0.2742480948730667, + "flos": 23479852666080.0, + "grad_norm": 1.8517880201818033, + "learning_rate": 3.4062390852619372e-06, + "loss": 0.795, + "num_input_tokens_seen": 99322945, + "step": 3041 + }, + { + "epoch": 0.27433827839653696, + "flos": 21111766270080.0, + "grad_norm": 2.3894851061302336, + "learning_rate": 3.4058236096512867e-06, + "loss": 0.7659, + "num_input_tokens_seen": 99348125, + "step": 3042 + }, + { + "epoch": 0.2744284619200072, + "flos": 20856518562720.0, + "grad_norm": 3.052775923921506, + "learning_rate": 3.405408014088013e-06, + "loss": 0.6729, + "num_input_tokens_seen": 99376095, + "step": 3043 + }, + { + "epoch": 0.2745186454434775, + "flos": 31822803199680.0, + "grad_norm": 2.039311887025792, + "learning_rate": 3.404992298607579e-06, + "loss": 0.658, + "num_input_tokens_seen": 99406100, + "step": 3044 + }, + { + "epoch": 0.2746088289669477, + "flos": 23771532471840.0, + "grad_norm": 2.9283831731918557, + "learning_rate": 3.4045764632454547e-06, + "loss": 0.7482, + "num_input_tokens_seen": 99433470, + "step": 3045 + }, + { + "epoch": 0.274699012490418, + "flos": 27154216498560.0, + "grad_norm": 1.9269800381856368, + "learning_rate": 3.4041605080371223e-06, + "loss": 0.7318, + "num_input_tokens_seen": 99460850, + "step": 3046 + }, + { + "epoch": 0.27478919601388824, + "flos": 17904217650720.0, + "grad_norm": 1.7408671486313825, + "learning_rate": 3.4037444330180726e-06, + "loss": 0.7835, + "num_input_tokens_seen": 99488570, + "step": 3047 + }, + { + "epoch": 0.2748793795373585, + "flos": 15933539058720.0, + "grad_norm": 2.0748307354381934, + "learning_rate": 3.403328238223808e-06, + "loss": 0.7815, + "num_input_tokens_seen": 99515290, + "step": 3048 + }, + { + "epoch": 0.27496956306082876, + "flos": 31603355704800.0, + "grad_norm": 1.9248020863043687, + "learning_rate": 3.4029119236898395e-06, + "loss": 0.7983, + "num_input_tokens_seen": 99541820, + "step": 3049 + }, + { + "epoch": 0.27505974658429905, + "flos": 69029894258400.0, + "grad_norm": 1.035799970746818, + "learning_rate": 3.4024954894516906e-06, + "loss": 0.6079, + "num_input_tokens_seen": 99638890, + "step": 3050 + }, + { + "epoch": 0.27514993010776934, + "flos": 24243811639680.0, + "grad_norm": 5.6399063962920115, + "learning_rate": 3.4020789355448933e-06, + "loss": 0.7732, + "num_input_tokens_seen": 99666515, + "step": 3051 + }, + { + "epoch": 0.27524011363123957, + "flos": 23662161837120.0, + "grad_norm": 3.0007825466591185, + "learning_rate": 3.40166226200499e-06, + "loss": 0.8431, + "num_input_tokens_seen": 99694770, + "step": 3052 + }, + { + "epoch": 0.27533029715470986, + "flos": 25119408603360.0, + "grad_norm": 2.3423428189398767, + "learning_rate": 3.401245468867534e-06, + "loss": 0.7528, + "num_input_tokens_seen": 99723030, + "step": 3053 + }, + { + "epoch": 0.2754204806781801, + "flos": 64015391338080.0, + "grad_norm": 0.7046034422346691, + "learning_rate": 3.400828556168088e-06, + "loss": 0.6322, + "num_input_tokens_seen": 99806175, + "step": 3054 + }, + { + "epoch": 0.2755106642016504, + "flos": 21689513247840.0, + "grad_norm": 3.393341804249001, + "learning_rate": 3.4004115239422255e-06, + "loss": 0.7841, + "num_input_tokens_seen": 99833295, + "step": 3055 + }, + { + "epoch": 0.2756008477251206, + "flos": 20747556795360.0, + "grad_norm": 1.8287252109226824, + "learning_rate": 3.3999943722255305e-06, + "loss": 0.7564, + "num_input_tokens_seen": 99860490, + "step": 3056 + }, + { + "epoch": 0.2756910312485909, + "flos": 48036127740960.0, + "grad_norm": 1.6483439197463898, + "learning_rate": 3.3995771010535955e-06, + "loss": 0.6653, + "num_input_tokens_seen": 99894245, + "step": 3057 + }, + { + "epoch": 0.27578121477206113, + "flos": 63024104880480.0, + "grad_norm": 0.6945117668791859, + "learning_rate": 3.3991597104620253e-06, + "loss": 0.5396, + "num_input_tokens_seen": 99972020, + "step": 3058 + }, + { + "epoch": 0.2758713982955314, + "flos": 28471644863040.0, + "grad_norm": 1.747380953487103, + "learning_rate": 3.398742200486434e-06, + "loss": 0.7507, + "num_input_tokens_seen": 100001885, + "step": 3059 + }, + { + "epoch": 0.27596158181900166, + "flos": 26613682184160.0, + "grad_norm": 2.004578832593406, + "learning_rate": 3.3983245711624453e-06, + "loss": 0.7882, + "num_input_tokens_seen": 100031220, + "step": 3060 + }, + { + "epoch": 0.27605176534247194, + "flos": 70648449281280.0, + "grad_norm": 0.7564835053919244, + "learning_rate": 3.3979068225256946e-06, + "loss": 0.6549, + "num_input_tokens_seen": 100125055, + "step": 3061 + }, + { + "epoch": 0.2761419488659422, + "flos": 21257606172960.0, + "grad_norm": 1.7780734063089934, + "learning_rate": 3.3974889546118246e-06, + "loss": 0.7312, + "num_input_tokens_seen": 100151665, + "step": 3062 + }, + { + "epoch": 0.27623213238941247, + "flos": 16558014158400.0, + "grad_norm": 1.640214051140942, + "learning_rate": 3.3970709674564918e-06, + "loss": 0.7145, + "num_input_tokens_seen": 100179995, + "step": 3063 + }, + { + "epoch": 0.2763223159128827, + "flos": 23006569914720.0, + "grad_norm": 1.6730033418173968, + "learning_rate": 3.3966528610953607e-06, + "loss": 0.7648, + "num_input_tokens_seen": 100210050, + "step": 3064 + }, + { + "epoch": 0.276412499436353, + "flos": 23404052058240.0, + "grad_norm": 2.0986625772231964, + "learning_rate": 3.3962346355641067e-06, + "loss": 0.7774, + "num_input_tokens_seen": 100239565, + "step": 3065 + }, + { + "epoch": 0.2765026829598232, + "flos": 21330804897600.0, + "grad_norm": 1.8583055514096443, + "learning_rate": 3.3958162908984146e-06, + "loss": 0.8036, + "num_input_tokens_seen": 100268050, + "step": 3066 + }, + { + "epoch": 0.2765928664832935, + "flos": 23552977051200.0, + "grad_norm": 1.864266061339955, + "learning_rate": 3.39539782713398e-06, + "loss": 0.763, + "num_input_tokens_seen": 100296585, + "step": 3067 + }, + { + "epoch": 0.27668305000676374, + "flos": 26613124637760.0, + "grad_norm": 2.2826654928655965, + "learning_rate": 3.394979244306509e-06, + "loss": 0.7936, + "num_input_tokens_seen": 100325175, + "step": 3068 + }, + { + "epoch": 0.27677323353023403, + "flos": 25917457980480.0, + "grad_norm": 1.8116863660726907, + "learning_rate": 3.3945605424517166e-06, + "loss": 0.7789, + "num_input_tokens_seen": 100356025, + "step": 3069 + }, + { + "epoch": 0.27686341705370426, + "flos": 22715150297280.0, + "grad_norm": 3.2474460837548222, + "learning_rate": 3.3941417216053294e-06, + "loss": 0.7645, + "num_input_tokens_seen": 100383575, + "step": 3070 + }, + { + "epoch": 0.27695360057717455, + "flos": 20784286251840.0, + "grad_norm": 1.9906177394860294, + "learning_rate": 3.3937227818030835e-06, + "loss": 0.7329, + "num_input_tokens_seen": 100410890, + "step": 3071 + }, + { + "epoch": 0.2770437841006448, + "flos": 19836642826080.0, + "grad_norm": 1.911546571723514, + "learning_rate": 3.393303723080725e-06, + "loss": 0.7324, + "num_input_tokens_seen": 100438800, + "step": 3072 + }, + { + "epoch": 0.2771339676241151, + "flos": 25411460106720.0, + "grad_norm": 1.6669546434837412, + "learning_rate": 3.3928845454740097e-06, + "loss": 0.7775, + "num_input_tokens_seen": 100467490, + "step": 3073 + }, + { + "epoch": 0.27722415114758536, + "flos": 21513039729120.0, + "grad_norm": 1.6758962027152489, + "learning_rate": 3.392465249018705e-06, + "loss": 0.7538, + "num_input_tokens_seen": 100494565, + "step": 3074 + }, + { + "epoch": 0.2773143346710556, + "flos": 19796233563360.0, + "grad_norm": 1.701770558873739, + "learning_rate": 3.3920458337505872e-06, + "loss": 0.7649, + "num_input_tokens_seen": 100521835, + "step": 3075 + }, + { + "epoch": 0.2774045181945259, + "flos": 21657430011360.0, + "grad_norm": 4.607685154357821, + "learning_rate": 3.391626299705443e-06, + "loss": 0.6072, + "num_input_tokens_seen": 100546885, + "step": 3076 + }, + { + "epoch": 0.2774947017179961, + "flos": 22569905110560.0, + "grad_norm": 1.732146610036838, + "learning_rate": 3.39120664691907e-06, + "loss": 0.7774, + "num_input_tokens_seen": 100576105, + "step": 3077 + }, + { + "epoch": 0.2775848852414664, + "flos": 25303576262400.0, + "grad_norm": 2.599308263018742, + "learning_rate": 3.390786875427275e-06, + "loss": 0.7471, + "num_input_tokens_seen": 100606370, + "step": 3078 + }, + { + "epoch": 0.27767506876493664, + "flos": 29781341917440.0, + "grad_norm": 1.855166697762281, + "learning_rate": 3.390366985265875e-06, + "loss": 0.7648, + "num_input_tokens_seen": 100635850, + "step": 3079 + }, + { + "epoch": 0.2777652522884069, + "flos": 24135630437280.0, + "grad_norm": 1.6710883455692909, + "learning_rate": 3.389946976470697e-06, + "loss": 0.7223, + "num_input_tokens_seen": 100667260, + "step": 3080 + }, + { + "epoch": 0.27785543581187716, + "flos": 41441620572480.0, + "grad_norm": 2.024926657584766, + "learning_rate": 3.3895268490775787e-06, + "loss": 0.8615, + "num_input_tokens_seen": 100698050, + "step": 3081 + }, + { + "epoch": 0.27794561933534745, + "flos": 17249369123520.0, + "grad_norm": 4.416819563804266, + "learning_rate": 3.3891066031223685e-06, + "loss": 0.7229, + "num_input_tokens_seen": 100726775, + "step": 3082 + }, + { + "epoch": 0.2780358028588177, + "flos": 21440472890400.0, + "grad_norm": 1.8497458773135373, + "learning_rate": 3.3886862386409237e-06, + "loss": 0.8015, + "num_input_tokens_seen": 100753505, + "step": 3083 + }, + { + "epoch": 0.27812598638228797, + "flos": 15173445740160.0, + "grad_norm": 2.1069073972697914, + "learning_rate": 3.388265755669111e-06, + "loss": 0.7328, + "num_input_tokens_seen": 100781320, + "step": 3084 + }, + { + "epoch": 0.2782161699057582, + "flos": 22715521994880.0, + "grad_norm": 1.9347690285726487, + "learning_rate": 3.3878451542428093e-06, + "loss": 0.8353, + "num_input_tokens_seen": 100809690, + "step": 3085 + }, + { + "epoch": 0.2783063534292285, + "flos": 20638743707040.0, + "grad_norm": 1.741631731416078, + "learning_rate": 3.387424434397907e-06, + "loss": 0.7524, + "num_input_tokens_seen": 100839035, + "step": 3086 + }, + { + "epoch": 0.2783965369526987, + "flos": 20234310818400.0, + "grad_norm": 1.6268576494391562, + "learning_rate": 3.3870035961703013e-06, + "loss": 0.6772, + "num_input_tokens_seen": 100868165, + "step": 3087 + }, + { + "epoch": 0.278486720476169, + "flos": 20961912033120.0, + "grad_norm": 4.900044398203971, + "learning_rate": 3.3865826395959014e-06, + "loss": 0.8914, + "num_input_tokens_seen": 100894550, + "step": 3088 + }, + { + "epoch": 0.27857690399963925, + "flos": 23589446319360.0, + "grad_norm": 1.5651660299742958, + "learning_rate": 3.3861615647106253e-06, + "loss": 0.8524, + "num_input_tokens_seen": 100923605, + "step": 3089 + }, + { + "epoch": 0.27866708752310954, + "flos": 31347661960320.0, + "grad_norm": 2.639504640806743, + "learning_rate": 3.3857403715504012e-06, + "loss": 0.78, + "num_input_tokens_seen": 100950785, + "step": 3090 + }, + { + "epoch": 0.27875727104657977, + "flos": 19545743585280.0, + "grad_norm": 2.0372553157739386, + "learning_rate": 3.385319060151167e-06, + "loss": 0.731, + "num_input_tokens_seen": 100979195, + "step": 3091 + }, + { + "epoch": 0.27884745457005006, + "flos": 21512184824640.0, + "grad_norm": 1.9995646598989545, + "learning_rate": 3.3848976305488728e-06, + "loss": 0.7609, + "num_input_tokens_seen": 101007050, + "step": 3092 + }, + { + "epoch": 0.2789376380935203, + "flos": 19653664599360.0, + "grad_norm": 2.5229189313337104, + "learning_rate": 3.384476082779476e-06, + "loss": 0.8083, + "num_input_tokens_seen": 101031530, + "step": 3093 + }, + { + "epoch": 0.2790278216169906, + "flos": 16740286159680.0, + "grad_norm": 1.7274159527219441, + "learning_rate": 3.3840544168789463e-06, + "loss": 0.7428, + "num_input_tokens_seen": 101058720, + "step": 3094 + }, + { + "epoch": 0.2791180051404608, + "flos": 21658619443680.0, + "grad_norm": 1.8909807107023304, + "learning_rate": 3.3836326328832617e-06, + "loss": 0.8735, + "num_input_tokens_seen": 101087535, + "step": 3095 + }, + { + "epoch": 0.2792081886639311, + "flos": 19176218834880.0, + "grad_norm": 2.837170545634698, + "learning_rate": 3.383210730828412e-06, + "loss": 0.7184, + "num_input_tokens_seen": 101115035, + "step": 3096 + }, + { + "epoch": 0.2792983721874014, + "flos": 25884482669760.0, + "grad_norm": 1.567609900070042, + "learning_rate": 3.3827887107503953e-06, + "loss": 0.8341, + "num_input_tokens_seen": 101146815, + "step": 3097 + }, + { + "epoch": 0.2793885557108716, + "flos": 23225311184160.0, + "grad_norm": 2.4092495540512795, + "learning_rate": 3.3823665726852216e-06, + "loss": 0.7123, + "num_input_tokens_seen": 101176315, + "step": 3098 + }, + { + "epoch": 0.2794787392343419, + "flos": 29564607815040.0, + "grad_norm": 1.4907554425439766, + "learning_rate": 3.3819443166689095e-06, + "loss": 0.8891, + "num_input_tokens_seen": 101207320, + "step": 3099 + }, + { + "epoch": 0.27956892275781214, + "flos": 19070342157600.0, + "grad_norm": 3.098206210292753, + "learning_rate": 3.3815219427374886e-06, + "loss": 0.7645, + "num_input_tokens_seen": 101233825, + "step": 3100 + }, + { + "epoch": 0.27965910628128243, + "flos": 25302721357920.0, + "grad_norm": 1.5341627483051965, + "learning_rate": 3.3810994509269975e-06, + "loss": 0.7811, + "num_input_tokens_seen": 101262495, + "step": 3101 + }, + { + "epoch": 0.27974928980475267, + "flos": 24967770217920.0, + "grad_norm": 1.559734338245491, + "learning_rate": 3.3806768412734864e-06, + "loss": 0.803, + "num_input_tokens_seen": 101291015, + "step": 3102 + }, + { + "epoch": 0.27983947332822295, + "flos": 26104190352960.0, + "grad_norm": 1.9715325809372626, + "learning_rate": 3.380254113813014e-06, + "loss": 0.8466, + "num_input_tokens_seen": 101323415, + "step": 3103 + }, + { + "epoch": 0.2799296568516932, + "flos": 27810663325440.0, + "grad_norm": 1.8094861195757364, + "learning_rate": 3.3798312685816496e-06, + "loss": 0.8023, + "num_input_tokens_seen": 101351090, + "step": 3104 + }, + { + "epoch": 0.2800198403751635, + "flos": 25804110181440.0, + "grad_norm": 1.8250859096634267, + "learning_rate": 3.3794083056154738e-06, + "loss": 0.768, + "num_input_tokens_seen": 101380290, + "step": 3105 + }, + { + "epoch": 0.2801100238986337, + "flos": 27554783732160.0, + "grad_norm": 2.036102016591151, + "learning_rate": 3.3789852249505746e-06, + "loss": 0.7749, + "num_input_tokens_seen": 101408350, + "step": 3106 + }, + { + "epoch": 0.280200207422104, + "flos": 20784434930880.0, + "grad_norm": 1.6422170271609917, + "learning_rate": 3.378562026623053e-06, + "loss": 0.8209, + "num_input_tokens_seen": 101436860, + "step": 3107 + }, + { + "epoch": 0.28029039094557423, + "flos": 26868558193920.0, + "grad_norm": 2.265380635730705, + "learning_rate": 3.3781387106690175e-06, + "loss": 0.8389, + "num_input_tokens_seen": 101465980, + "step": 3108 + }, + { + "epoch": 0.2803805744690445, + "flos": 22861064539680.0, + "grad_norm": 1.6723586479178516, + "learning_rate": 3.3777152771245885e-06, + "loss": 0.7931, + "num_input_tokens_seen": 101495265, + "step": 3109 + }, + { + "epoch": 0.28047075799251475, + "flos": 25150451086560.0, + "grad_norm": 5.601234926570811, + "learning_rate": 3.377291726025895e-06, + "loss": 0.799, + "num_input_tokens_seen": 101521755, + "step": 3110 + }, + { + "epoch": 0.28056094151598504, + "flos": 23480150024160.0, + "grad_norm": 2.620564027746097, + "learning_rate": 3.3768680574090782e-06, + "loss": 0.8863, + "num_input_tokens_seen": 101546605, + "step": 3111 + }, + { + "epoch": 0.2806511250394553, + "flos": 35211657406560.0, + "grad_norm": 1.84047958499081, + "learning_rate": 3.3764442713102857e-06, + "loss": 0.7477, + "num_input_tokens_seen": 101574735, + "step": 3112 + }, + { + "epoch": 0.28074130856292556, + "flos": 27377529648480.0, + "grad_norm": 1.6891440400971733, + "learning_rate": 3.3760203677656786e-06, + "loss": 0.7639, + "num_input_tokens_seen": 101603305, + "step": 3113 + }, + { + "epoch": 0.2808314920863958, + "flos": 20564987436000.0, + "grad_norm": 8.09918478171834, + "learning_rate": 3.3755963468114262e-06, + "loss": 0.8231, + "num_input_tokens_seen": 101629495, + "step": 3114 + }, + { + "epoch": 0.2809216756098661, + "flos": 25302349660320.0, + "grad_norm": 1.733838591233991, + "learning_rate": 3.3751722084837095e-06, + "loss": 0.6708, + "num_input_tokens_seen": 101656185, + "step": 3115 + }, + { + "epoch": 0.2810118591333363, + "flos": 20711496394560.0, + "grad_norm": 1.500097192158642, + "learning_rate": 3.3747479528187166e-06, + "loss": 0.7448, + "num_input_tokens_seen": 101684010, + "step": 3116 + }, + { + "epoch": 0.2811020426568066, + "flos": 12805805381280.0, + "grad_norm": 2.0155951919125856, + "learning_rate": 3.3743235798526485e-06, + "loss": 0.8166, + "num_input_tokens_seen": 101710550, + "step": 3117 + }, + { + "epoch": 0.28119222618027684, + "flos": 23004488408160.0, + "grad_norm": 1.4689707488698187, + "learning_rate": 3.373899089621714e-06, + "loss": 0.7031, + "num_input_tokens_seen": 101739895, + "step": 3118 + }, + { + "epoch": 0.2812824097037471, + "flos": 27415039669920.0, + "grad_norm": 2.0895986823307107, + "learning_rate": 3.373474482162134e-06, + "loss": 0.7455, + "num_input_tokens_seen": 101769320, + "step": 3119 + }, + { + "epoch": 0.28137259322721736, + "flos": 18488357827200.0, + "grad_norm": 2.1674821737187764, + "learning_rate": 3.3730497575101376e-06, + "loss": 0.7721, + "num_input_tokens_seen": 101793590, + "step": 3120 + }, + { + "epoch": 0.28146277675068765, + "flos": 24900555824640.0, + "grad_norm": 1.9533422120427395, + "learning_rate": 3.3726249157019654e-06, + "loss": 0.7647, + "num_input_tokens_seen": 101826370, + "step": 3121 + }, + { + "epoch": 0.28155296027415794, + "flos": 24859180148160.0, + "grad_norm": 1.8127181480933212, + "learning_rate": 3.372199956773866e-06, + "loss": 0.8534, + "num_input_tokens_seen": 101854525, + "step": 3122 + }, + { + "epoch": 0.28164314379762817, + "flos": 30402546078240.0, + "grad_norm": 1.795565716952873, + "learning_rate": 3.371774880762101e-06, + "loss": 0.7185, + "num_input_tokens_seen": 101885245, + "step": 3123 + }, + { + "epoch": 0.28173332732109846, + "flos": 24350283033120.0, + "grad_norm": 2.6032321830014515, + "learning_rate": 3.3713496877029392e-06, + "loss": 0.7527, + "num_input_tokens_seen": 101911870, + "step": 3124 + }, + { + "epoch": 0.2818235108445687, + "flos": 20266505564160.0, + "grad_norm": 2.7564716483543146, + "learning_rate": 3.37092437763266e-06, + "loss": 0.8341, + "num_input_tokens_seen": 101938710, + "step": 3125 + }, + { + "epoch": 0.281913694368039, + "flos": 24026073953760.0, + "grad_norm": 1.9480081900516488, + "learning_rate": 3.3704989505875537e-06, + "loss": 0.804, + "num_input_tokens_seen": 101967375, + "step": 3126 + }, + { + "epoch": 0.2820038778915092, + "flos": 21476347442400.0, + "grad_norm": 2.222999586461381, + "learning_rate": 3.3700734066039205e-06, + "loss": 0.7395, + "num_input_tokens_seen": 101996720, + "step": 3127 + }, + { + "epoch": 0.2820940614149795, + "flos": 24391287012000.0, + "grad_norm": 1.941703420572234, + "learning_rate": 3.36964774571807e-06, + "loss": 0.7994, + "num_input_tokens_seen": 102023145, + "step": 3128 + }, + { + "epoch": 0.28218424493844974, + "flos": 20383719018240.0, + "grad_norm": 2.610523268178932, + "learning_rate": 3.3692219679663206e-06, + "loss": 0.7651, + "num_input_tokens_seen": 102051740, + "step": 3129 + }, + { + "epoch": 0.28227442846192, + "flos": 31895407208160.0, + "grad_norm": 2.0088734476912955, + "learning_rate": 3.3687960733850043e-06, + "loss": 0.7808, + "num_input_tokens_seen": 102078820, + "step": 3130 + }, + { + "epoch": 0.28236461198539026, + "flos": 16590729280800.0, + "grad_norm": 1.7515633275951434, + "learning_rate": 3.3683700620104586e-06, + "loss": 0.7059, + "num_input_tokens_seen": 102104000, + "step": 3131 + }, + { + "epoch": 0.28245479550886055, + "flos": 22496074500000.0, + "grad_norm": 1.7819904996702192, + "learning_rate": 3.3679439338790347e-06, + "loss": 0.7865, + "num_input_tokens_seen": 102132665, + "step": 3132 + }, + { + "epoch": 0.2825449790323308, + "flos": 22344584793600.0, + "grad_norm": 1.856875879883919, + "learning_rate": 3.3675176890270916e-06, + "loss": 0.8915, + "num_input_tokens_seen": 102159765, + "step": 3133 + }, + { + "epoch": 0.28263516255580107, + "flos": 21764979327840.0, + "grad_norm": 1.7177749642456972, + "learning_rate": 3.367091327490998e-06, + "loss": 0.7909, + "num_input_tokens_seen": 102187405, + "step": 3134 + }, + { + "epoch": 0.2827253460792713, + "flos": 17609787282720.0, + "grad_norm": 1.7364513580908343, + "learning_rate": 3.3666648493071347e-06, + "loss": 0.7542, + "num_input_tokens_seen": 102213995, + "step": 3135 + }, + { + "epoch": 0.2828155296027416, + "flos": 21513225577920.0, + "grad_norm": 1.8710241125603124, + "learning_rate": 3.3662382545118914e-06, + "loss": 0.7474, + "num_input_tokens_seen": 102240715, + "step": 3136 + }, + { + "epoch": 0.2829057131262118, + "flos": 24238161836160.0, + "grad_norm": 1.8774877574739925, + "learning_rate": 3.3658115431416663e-06, + "loss": 0.8511, + "num_input_tokens_seen": 102268860, + "step": 3137 + }, + { + "epoch": 0.2829958966496821, + "flos": 67085872117920.0, + "grad_norm": 0.7032140463498102, + "learning_rate": 3.36538471523287e-06, + "loss": 0.6082, + "num_input_tokens_seen": 102365235, + "step": 3138 + }, + { + "epoch": 0.28308608017315234, + "flos": 18489733108320.0, + "grad_norm": 1.8322179247786015, + "learning_rate": 3.3649577708219204e-06, + "loss": 0.8297, + "num_input_tokens_seen": 102391595, + "step": 3139 + }, + { + "epoch": 0.28317626369662263, + "flos": 25920766089120.0, + "grad_norm": 1.9807199172162513, + "learning_rate": 3.3645307099452477e-06, + "loss": 0.7145, + "num_input_tokens_seen": 102420385, + "step": 3140 + }, + { + "epoch": 0.28326644722009287, + "flos": 15866324665440.0, + "grad_norm": 1.9592609305389002, + "learning_rate": 3.3641035326392907e-06, + "loss": 0.7671, + "num_input_tokens_seen": 102446875, + "step": 3141 + }, + { + "epoch": 0.28335663074356315, + "flos": 20894140093440.0, + "grad_norm": 2.000678543380133, + "learning_rate": 3.363676238940499e-06, + "loss": 0.7827, + "num_input_tokens_seen": 102473760, + "step": 3142 + }, + { + "epoch": 0.2834468142670334, + "flos": 19218152057760.0, + "grad_norm": 1.8532331952936483, + "learning_rate": 3.363248828885331e-06, + "loss": 0.8047, + "num_input_tokens_seen": 102502315, + "step": 3143 + }, + { + "epoch": 0.2835369977905037, + "flos": 22055395361760.0, + "grad_norm": 1.5817716409050826, + "learning_rate": 3.3628213025102562e-06, + "loss": 0.7932, + "num_input_tokens_seen": 102531020, + "step": 3144 + }, + { + "epoch": 0.28362718131397396, + "flos": 15756396484320.0, + "grad_norm": 2.232253981957658, + "learning_rate": 3.3623936598517536e-06, + "loss": 0.8088, + "num_input_tokens_seen": 102557835, + "step": 3145 + }, + { + "epoch": 0.2837173648374442, + "flos": 23916889167840.0, + "grad_norm": 2.2907659552323567, + "learning_rate": 3.3619659009463117e-06, + "loss": 0.6164, + "num_input_tokens_seen": 102585205, + "step": 3146 + }, + { + "epoch": 0.2838075483609145, + "flos": 26390406204000.0, + "grad_norm": 1.6023443035122917, + "learning_rate": 3.3615380258304287e-06, + "loss": 0.8241, + "num_input_tokens_seen": 102614570, + "step": 3147 + }, + { + "epoch": 0.2838977318843847, + "flos": 26212148536800.0, + "grad_norm": 1.7709316076796924, + "learning_rate": 3.3611100345406146e-06, + "loss": 0.7609, + "num_input_tokens_seen": 102642280, + "step": 3148 + }, + { + "epoch": 0.283987915407855, + "flos": 19326704957760.0, + "grad_norm": 1.790524614059993, + "learning_rate": 3.3606819271133873e-06, + "loss": 0.7598, + "num_input_tokens_seen": 102667785, + "step": 3149 + }, + { + "epoch": 0.28407809893132524, + "flos": 66143246609760.0, + "grad_norm": 0.7188738425996206, + "learning_rate": 3.360253703585275e-06, + "loss": 0.5927, + "num_input_tokens_seen": 102760070, + "step": 3150 + }, + { + "epoch": 0.28416828245479553, + "flos": 15902162047680.0, + "grad_norm": 1.7519016151978644, + "learning_rate": 3.3598253639928164e-06, + "loss": 0.7065, + "num_input_tokens_seen": 102783510, + "step": 3151 + }, + { + "epoch": 0.28425846597826576, + "flos": 39437817990720.0, + "grad_norm": 3.5290340796511663, + "learning_rate": 3.3593969083725596e-06, + "loss": 0.7225, + "num_input_tokens_seen": 102814505, + "step": 3152 + }, + { + "epoch": 0.28434864950173605, + "flos": 22271794936320.0, + "grad_norm": 2.3018204782053484, + "learning_rate": 3.358968336761063e-06, + "loss": 0.7234, + "num_input_tokens_seen": 102841820, + "step": 3153 + }, + { + "epoch": 0.2844388330252063, + "flos": 18560701647360.0, + "grad_norm": 2.0677933247071896, + "learning_rate": 3.3585396491948945e-06, + "loss": 0.7383, + "num_input_tokens_seen": 102870620, + "step": 3154 + }, + { + "epoch": 0.28452901654867657, + "flos": 23585989531680.0, + "grad_norm": 1.4748097668836762, + "learning_rate": 3.358110845710633e-06, + "loss": 0.8258, + "num_input_tokens_seen": 102899665, + "step": 3155 + }, + { + "epoch": 0.2846192000721468, + "flos": 20673874863840.0, + "grad_norm": 1.909099700110701, + "learning_rate": 3.357681926344865e-06, + "loss": 0.7586, + "num_input_tokens_seen": 102928035, + "step": 3156 + }, + { + "epoch": 0.2847093835956171, + "flos": 22636822145760.0, + "grad_norm": 1.520140302733036, + "learning_rate": 3.357252891134189e-06, + "loss": 0.7678, + "num_input_tokens_seen": 102955255, + "step": 3157 + }, + { + "epoch": 0.2847995671190873, + "flos": 20675250144960.0, + "grad_norm": 2.0202734522280252, + "learning_rate": 3.356823740115212e-06, + "loss": 0.8152, + "num_input_tokens_seen": 102982805, + "step": 3158 + }, + { + "epoch": 0.2848897506425576, + "flos": 32841935541120.0, + "grad_norm": 1.4410580263546104, + "learning_rate": 3.3563944733245525e-06, + "loss": 0.7149, + "num_input_tokens_seen": 103014240, + "step": 3159 + }, + { + "epoch": 0.28497993416602785, + "flos": 27524075776800.0, + "grad_norm": 2.017910748045824, + "learning_rate": 3.3559650907988375e-06, + "loss": 0.7629, + "num_input_tokens_seen": 103044020, + "step": 3160 + }, + { + "epoch": 0.28507011768949814, + "flos": 31636702713120.0, + "grad_norm": 5.526013325418703, + "learning_rate": 3.3555355925747045e-06, + "loss": 0.619, + "num_input_tokens_seen": 103075465, + "step": 3161 + }, + { + "epoch": 0.28516030121296837, + "flos": 26321630680800.0, + "grad_norm": 1.7474171788313746, + "learning_rate": 3.3551059786888e-06, + "loss": 0.7546, + "num_input_tokens_seen": 103104815, + "step": 3162 + }, + { + "epoch": 0.28525048473643866, + "flos": 28762952971200.0, + "grad_norm": 1.51724934790725, + "learning_rate": 3.3546762491777807e-06, + "loss": 0.8203, + "num_input_tokens_seen": 103133030, + "step": 3163 + }, + { + "epoch": 0.2853406682599089, + "flos": 19725859740480.0, + "grad_norm": 2.225524231949549, + "learning_rate": 3.3542464040783156e-06, + "loss": 0.7561, + "num_input_tokens_seen": 103162265, + "step": 3164 + }, + { + "epoch": 0.2854308517833792, + "flos": 20528927035200.0, + "grad_norm": 1.9929999396539, + "learning_rate": 3.353816443427079e-06, + "loss": 0.7818, + "num_input_tokens_seen": 103188125, + "step": 3165 + }, + { + "epoch": 0.2855210353068494, + "flos": 26394606386880.0, + "grad_norm": 2.164304533826703, + "learning_rate": 3.3533863672607597e-06, + "loss": 0.7483, + "num_input_tokens_seen": 103219575, + "step": 3166 + }, + { + "epoch": 0.2856112188303197, + "flos": 21108681180000.0, + "grad_norm": 1.6324428158101767, + "learning_rate": 3.352956175616052e-06, + "loss": 0.7975, + "num_input_tokens_seen": 103246500, + "step": 3167 + }, + { + "epoch": 0.28570140235378993, + "flos": 23989716194880.0, + "grad_norm": 2.5373644975835483, + "learning_rate": 3.352525868529664e-06, + "loss": 0.7345, + "num_input_tokens_seen": 103273785, + "step": 3168 + }, + { + "epoch": 0.2857915858772602, + "flos": 31604396458080.0, + "grad_norm": 1.8737199228429053, + "learning_rate": 3.3520954460383103e-06, + "loss": 0.7061, + "num_input_tokens_seen": 103303685, + "step": 3169 + }, + { + "epoch": 0.2858817694007305, + "flos": 18598211668800.0, + "grad_norm": 2.819413144752724, + "learning_rate": 3.3516649081787182e-06, + "loss": 0.7088, + "num_input_tokens_seen": 103327970, + "step": 3170 + }, + { + "epoch": 0.28597195292420075, + "flos": 38963271467520.0, + "grad_norm": 1.5607848099421544, + "learning_rate": 3.3512342549876236e-06, + "loss": 0.8664, + "num_input_tokens_seen": 103356610, + "step": 3171 + }, + { + "epoch": 0.28606213644767103, + "flos": 14080705806720.0, + "grad_norm": 1.7090753324009123, + "learning_rate": 3.350803486501771e-06, + "loss": 0.7258, + "num_input_tokens_seen": 103383405, + "step": 3172 + }, + { + "epoch": 0.28615231997114127, + "flos": 23986370916480.0, + "grad_norm": 1.571599935501808, + "learning_rate": 3.3503726027579175e-06, + "loss": 0.7932, + "num_input_tokens_seen": 103411655, + "step": 3173 + }, + { + "epoch": 0.28624250349461156, + "flos": 23151480573600.0, + "grad_norm": 1.529782506780342, + "learning_rate": 3.349941603792827e-06, + "loss": 0.824, + "num_input_tokens_seen": 103441490, + "step": 3174 + }, + { + "epoch": 0.2863326870180818, + "flos": 18853719564480.0, + "grad_norm": 6.101700355056809, + "learning_rate": 3.3495104896432755e-06, + "loss": 0.7653, + "num_input_tokens_seen": 103469240, + "step": 3175 + }, + { + "epoch": 0.2864228705415521, + "flos": 26613384826080.0, + "grad_norm": 1.7683763250840925, + "learning_rate": 3.3490792603460477e-06, + "loss": 0.7289, + "num_input_tokens_seen": 103498520, + "step": 3176 + }, + { + "epoch": 0.2865130540650223, + "flos": 26685691476480.0, + "grad_norm": 2.0958680783899926, + "learning_rate": 3.3486479159379393e-06, + "loss": 0.6987, + "num_input_tokens_seen": 103529065, + "step": 3177 + }, + { + "epoch": 0.2866032375884926, + "flos": 60724342236960.0, + "grad_norm": 0.665915719423021, + "learning_rate": 3.3482164564557537e-06, + "loss": 0.5762, + "num_input_tokens_seen": 103626845, + "step": 3178 + }, + { + "epoch": 0.28669342111196283, + "flos": 22350829313280.0, + "grad_norm": 1.7470098562852368, + "learning_rate": 3.3477848819363065e-06, + "loss": 0.8254, + "num_input_tokens_seen": 103654520, + "step": 3179 + }, + { + "epoch": 0.2867836046354331, + "flos": 61486740080640.0, + "grad_norm": 0.6668678491739931, + "learning_rate": 3.3473531924164213e-06, + "loss": 0.6052, + "num_input_tokens_seen": 103737505, + "step": 3180 + }, + { + "epoch": 0.28687378815890335, + "flos": 17096838663840.0, + "grad_norm": 2.7667103162639006, + "learning_rate": 3.3469213879329325e-06, + "loss": 0.7246, + "num_input_tokens_seen": 103760695, + "step": 3181 + }, + { + "epoch": 0.28696397168237364, + "flos": 14039441639520.0, + "grad_norm": 2.066535700473274, + "learning_rate": 3.3464894685226837e-06, + "loss": 0.7329, + "num_input_tokens_seen": 103785810, + "step": 3182 + }, + { + "epoch": 0.2870541552058439, + "flos": 24500731986240.0, + "grad_norm": 1.9162015834397683, + "learning_rate": 3.34605743422253e-06, + "loss": 0.845, + "num_input_tokens_seen": 103813875, + "step": 3183 + }, + { + "epoch": 0.28714433872931416, + "flos": 24281842037760.0, + "grad_norm": 1.9790501677374197, + "learning_rate": 3.345625285069333e-06, + "loss": 0.7379, + "num_input_tokens_seen": 103842795, + "step": 3184 + }, + { + "epoch": 0.2872345222527844, + "flos": 28980504808320.0, + "grad_norm": 1.6389152700739869, + "learning_rate": 3.345193021099967e-06, + "loss": 0.7474, + "num_input_tokens_seen": 103873540, + "step": 3185 + }, + { + "epoch": 0.2873247057762547, + "flos": 22532841126240.0, + "grad_norm": 1.540855563787203, + "learning_rate": 3.3447606423513157e-06, + "loss": 0.7353, + "num_input_tokens_seen": 103903010, + "step": 3186 + }, + { + "epoch": 0.2874148892997249, + "flos": 26066940519840.0, + "grad_norm": 1.7495590744402016, + "learning_rate": 3.344328148860271e-06, + "loss": 0.7352, + "num_input_tokens_seen": 103931775, + "step": 3187 + }, + { + "epoch": 0.2875050728231952, + "flos": 20311077840000.0, + "grad_norm": 1.564485585308143, + "learning_rate": 3.3438955406637365e-06, + "loss": 0.7845, + "num_input_tokens_seen": 103958685, + "step": 3188 + }, + { + "epoch": 0.28759525634666544, + "flos": 20747779813920.0, + "grad_norm": 1.855977520707901, + "learning_rate": 3.343462817798624e-06, + "loss": 0.7918, + "num_input_tokens_seen": 103987185, + "step": 3189 + }, + { + "epoch": 0.28768543987013573, + "flos": 32115263570400.0, + "grad_norm": 4.048814699009939, + "learning_rate": 3.343029980301856e-06, + "loss": 0.7341, + "num_input_tokens_seen": 104017685, + "step": 3190 + }, + { + "epoch": 0.28777562339360596, + "flos": 22459902589920.0, + "grad_norm": 2.5113036196781824, + "learning_rate": 3.342597028210365e-06, + "loss": 0.7964, + "num_input_tokens_seen": 104044080, + "step": 3191 + }, + { + "epoch": 0.28786580691707625, + "flos": 38741073410400.0, + "grad_norm": 2.433549663211719, + "learning_rate": 3.342163961561092e-06, + "loss": 0.7172, + "num_input_tokens_seen": 104075840, + "step": 3192 + }, + { + "epoch": 0.28795599044054654, + "flos": 19653887617920.0, + "grad_norm": 1.5631407377662618, + "learning_rate": 3.34173078039099e-06, + "loss": 0.7387, + "num_input_tokens_seen": 104103785, + "step": 3193 + }, + { + "epoch": 0.28804617396401677, + "flos": 15502003681440.0, + "grad_norm": 2.0588923026066492, + "learning_rate": 3.3412974847370193e-06, + "loss": 0.9329, + "num_input_tokens_seen": 104128770, + "step": 3194 + }, + { + "epoch": 0.28813635748748706, + "flos": 28577633049600.0, + "grad_norm": 1.7025823866169836, + "learning_rate": 3.3408640746361514e-06, + "loss": 0.7893, + "num_input_tokens_seen": 104159695, + "step": 3195 + }, + { + "epoch": 0.2882265410109573, + "flos": 71127344666400.0, + "grad_norm": 0.7494959727071574, + "learning_rate": 3.3404305501253663e-06, + "loss": 0.599, + "num_input_tokens_seen": 104249740, + "step": 3196 + }, + { + "epoch": 0.2883167245344276, + "flos": 21549992204160.0, + "grad_norm": 1.514385463727255, + "learning_rate": 3.3399969112416565e-06, + "loss": 0.838, + "num_input_tokens_seen": 104277720, + "step": 3197 + }, + { + "epoch": 0.2884069080578978, + "flos": 31160186192640.0, + "grad_norm": 1.8788536524399624, + "learning_rate": 3.3395631580220213e-06, + "loss": 0.7012, + "num_input_tokens_seen": 104308040, + "step": 3198 + }, + { + "epoch": 0.2884970915813681, + "flos": 27121687224960.0, + "grad_norm": 2.036291351366834, + "learning_rate": 3.3391292905034714e-06, + "loss": 0.7445, + "num_input_tokens_seen": 104336000, + "step": 3199 + }, + { + "epoch": 0.28858727510483834, + "flos": 25986828219840.0, + "grad_norm": 1.804482914941777, + "learning_rate": 3.338695308723027e-06, + "loss": 0.8304, + "num_input_tokens_seen": 104361545, + "step": 3200 + }, + { + "epoch": 0.2886774586283086, + "flos": 16375296120000.0, + "grad_norm": 2.0675723962435986, + "learning_rate": 3.338261212717716e-06, + "loss": 0.8238, + "num_input_tokens_seen": 104386335, + "step": 3201 + }, + { + "epoch": 0.28876764215177886, + "flos": 23880680088000.0, + "grad_norm": 1.8206209895479062, + "learning_rate": 3.33782700252458e-06, + "loss": 0.8052, + "num_input_tokens_seen": 104416180, + "step": 3202 + }, + { + "epoch": 0.28885782567524915, + "flos": 25483803926880.0, + "grad_norm": 1.7718113909678914, + "learning_rate": 3.337392678180668e-06, + "loss": 0.8067, + "num_input_tokens_seen": 104444780, + "step": 3203 + }, + { + "epoch": 0.2889480091987194, + "flos": 23772424546080.0, + "grad_norm": 3.668239400707555, + "learning_rate": 3.3369582397230377e-06, + "loss": 0.7722, + "num_input_tokens_seen": 104471355, + "step": 3204 + }, + { + "epoch": 0.28903819272218967, + "flos": 19581469458240.0, + "grad_norm": 2.0829314670798733, + "learning_rate": 3.336523687188759e-06, + "loss": 0.8113, + "num_input_tokens_seen": 104499050, + "step": 3205 + }, + { + "epoch": 0.2891283762456599, + "flos": 65295086662560.0, + "grad_norm": 0.7112362412125868, + "learning_rate": 3.336089020614909e-06, + "loss": 0.6119, + "num_input_tokens_seen": 104578735, + "step": 3206 + }, + { + "epoch": 0.2892185597691302, + "flos": 18161361015840.0, + "grad_norm": 1.6025346736162098, + "learning_rate": 3.3356542400385774e-06, + "loss": 0.6433, + "num_input_tokens_seen": 104607760, + "step": 3207 + }, + { + "epoch": 0.2893087432926004, + "flos": 25769610910560.0, + "grad_norm": 1.8288605343312605, + "learning_rate": 3.3352193454968607e-06, + "loss": 0.7062, + "num_input_tokens_seen": 104636600, + "step": 3208 + }, + { + "epoch": 0.2893989268160707, + "flos": 22127739181920.0, + "grad_norm": 1.7132501866368228, + "learning_rate": 3.3347843370268675e-06, + "loss": 0.8327, + "num_input_tokens_seen": 104662065, + "step": 3209 + }, + { + "epoch": 0.28948911033954094, + "flos": 22569496243200.0, + "grad_norm": 1.4195097649249495, + "learning_rate": 3.334349214665715e-06, + "loss": 0.7874, + "num_input_tokens_seen": 104689945, + "step": 3210 + }, + { + "epoch": 0.28957929386301123, + "flos": 24136559681280.0, + "grad_norm": 1.8524045035681267, + "learning_rate": 3.3339139784505293e-06, + "loss": 0.8514, + "num_input_tokens_seen": 104720240, + "step": 3211 + }, + { + "epoch": 0.28966947738648147, + "flos": 26249584218720.0, + "grad_norm": 1.5917029604676785, + "learning_rate": 3.333478628418448e-06, + "loss": 0.8355, + "num_input_tokens_seen": 104750480, + "step": 3212 + }, + { + "epoch": 0.28975966090995176, + "flos": 35210988350880.0, + "grad_norm": 1.9550690398257626, + "learning_rate": 3.333043164606618e-06, + "loss": 0.658, + "num_input_tokens_seen": 104781960, + "step": 3213 + }, + { + "epoch": 0.289849844433422, + "flos": 21836839941120.0, + "grad_norm": 1.8513361152643593, + "learning_rate": 3.3326075870521948e-06, + "loss": 0.8481, + "num_input_tokens_seen": 104810550, + "step": 3214 + }, + { + "epoch": 0.2899400279568923, + "flos": 21798363505920.0, + "grad_norm": 2.345346644468776, + "learning_rate": 3.3321718957923437e-06, + "loss": 0.7742, + "num_input_tokens_seen": 104838670, + "step": 3215 + }, + { + "epoch": 0.29003021148036257, + "flos": 27341580756960.0, + "grad_norm": 1.7851239291213865, + "learning_rate": 3.3317360908642413e-06, + "loss": 0.7701, + "num_input_tokens_seen": 104867430, + "step": 3216 + }, + { + "epoch": 0.2901203950038328, + "flos": 19836531316800.0, + "grad_norm": 2.1567521547161044, + "learning_rate": 3.331300172305072e-06, + "loss": 0.7232, + "num_input_tokens_seen": 104896230, + "step": 3217 + }, + { + "epoch": 0.2902105785273031, + "flos": 20748002832480.0, + "grad_norm": 1.8557661922364581, + "learning_rate": 3.330864140152032e-06, + "loss": 0.759, + "num_input_tokens_seen": 104926240, + "step": 3218 + }, + { + "epoch": 0.2903007620507733, + "flos": 38126225278560.0, + "grad_norm": 1.9945486998102897, + "learning_rate": 3.330427994442325e-06, + "loss": 0.7901, + "num_input_tokens_seen": 104956415, + "step": 3219 + }, + { + "epoch": 0.2903909455742436, + "flos": 24788285948640.0, + "grad_norm": 1.5313786043930075, + "learning_rate": 3.3299917352131657e-06, + "loss": 0.804, + "num_input_tokens_seen": 104987080, + "step": 3220 + }, + { + "epoch": 0.29048112909771384, + "flos": 11747527548960.0, + "grad_norm": 2.387330521564946, + "learning_rate": 3.329555362501778e-06, + "loss": 0.7353, + "num_input_tokens_seen": 105013735, + "step": 3221 + }, + { + "epoch": 0.29057131262118413, + "flos": 18452223086880.0, + "grad_norm": 1.778253075356615, + "learning_rate": 3.3291188763453954e-06, + "loss": 0.7358, + "num_input_tokens_seen": 105040295, + "step": 3222 + }, + { + "epoch": 0.29066149614465436, + "flos": 24864755612160.0, + "grad_norm": 2.077407598357124, + "learning_rate": 3.3286822767812618e-06, + "loss": 0.76, + "num_input_tokens_seen": 105069350, + "step": 3223 + }, + { + "epoch": 0.29075167966812465, + "flos": 22930248930240.0, + "grad_norm": 1.8635042644888984, + "learning_rate": 3.32824556384663e-06, + "loss": 0.7507, + "num_input_tokens_seen": 105099815, + "step": 3224 + }, + { + "epoch": 0.2908418631915949, + "flos": 20711124696960.0, + "grad_norm": 1.8110564830488007, + "learning_rate": 3.3278087375787628e-06, + "loss": 0.774, + "num_input_tokens_seen": 105127210, + "step": 3225 + }, + { + "epoch": 0.2909320467150652, + "flos": 26686360532160.0, + "grad_norm": 1.8429806284018366, + "learning_rate": 3.327371798014933e-06, + "loss": 0.7691, + "num_input_tokens_seen": 105152590, + "step": 3226 + }, + { + "epoch": 0.2910222302385354, + "flos": 25915822511040.0, + "grad_norm": 1.8823128824845083, + "learning_rate": 3.3269347451924218e-06, + "loss": 0.7713, + "num_input_tokens_seen": 105178635, + "step": 3227 + }, + { + "epoch": 0.2911124137620057, + "flos": 28471905051360.0, + "grad_norm": 1.447829163832333, + "learning_rate": 3.326497579148522e-06, + "loss": 0.8328, + "num_input_tokens_seen": 105209855, + "step": 3228 + }, + { + "epoch": 0.29120259728547593, + "flos": 21002730163200.0, + "grad_norm": 1.9358900826399206, + "learning_rate": 3.3260602999205345e-06, + "loss": 0.7672, + "num_input_tokens_seen": 105236610, + "step": 3229 + }, + { + "epoch": 0.2912927808089462, + "flos": 36668755493760.0, + "grad_norm": 3.71338298328384, + "learning_rate": 3.32562290754577e-06, + "loss": 0.7824, + "num_input_tokens_seen": 105266075, + "step": 3230 + }, + { + "epoch": 0.29138296433241645, + "flos": 16813113186720.0, + "grad_norm": 1.763942017677079, + "learning_rate": 3.3251854020615494e-06, + "loss": 0.7517, + "num_input_tokens_seen": 105293265, + "step": 3231 + }, + { + "epoch": 0.29147314785588674, + "flos": 25993964813760.0, + "grad_norm": 1.8260447085569744, + "learning_rate": 3.324747783505204e-06, + "loss": 0.7746, + "num_input_tokens_seen": 105322485, + "step": 3232 + }, + { + "epoch": 0.29156333137935697, + "flos": 32221772133600.0, + "grad_norm": 2.2156977546056242, + "learning_rate": 3.324310051914073e-06, + "loss": 0.6865, + "num_input_tokens_seen": 105353675, + "step": 3233 + }, + { + "epoch": 0.29165351490282726, + "flos": 17213680420320.0, + "grad_norm": 2.3533403640008066, + "learning_rate": 3.3238722073255056e-06, + "loss": 0.7376, + "num_input_tokens_seen": 105380335, + "step": 3234 + }, + { + "epoch": 0.2917436984262975, + "flos": 36232165029120.0, + "grad_norm": 1.7045492813760976, + "learning_rate": 3.323434249776863e-06, + "loss": 0.7309, + "num_input_tokens_seen": 105411135, + "step": 3235 + }, + { + "epoch": 0.2918338819497678, + "flos": 69076330755840.0, + "grad_norm": 0.7211417747367554, + "learning_rate": 3.3229961793055117e-06, + "loss": 0.5896, + "num_input_tokens_seen": 105498965, + "step": 3236 + }, + { + "epoch": 0.291924065473238, + "flos": 35431327920000.0, + "grad_norm": 1.724776297728613, + "learning_rate": 3.3225579959488314e-06, + "loss": 0.6963, + "num_input_tokens_seen": 105528675, + "step": 3237 + }, + { + "epoch": 0.2920142489967083, + "flos": 20747259437280.0, + "grad_norm": 2.151352448138059, + "learning_rate": 3.322119699744211e-06, + "loss": 0.7952, + "num_input_tokens_seen": 105555155, + "step": 3238 + }, + { + "epoch": 0.29210443252017854, + "flos": 19356892536480.0, + "grad_norm": 8.092534861426204, + "learning_rate": 3.3216812907290476e-06, + "loss": 0.782, + "num_input_tokens_seen": 105580980, + "step": 3239 + }, + { + "epoch": 0.2921946160436488, + "flos": 23952466361760.0, + "grad_norm": 1.6975263213194267, + "learning_rate": 3.3212427689407484e-06, + "loss": 0.7903, + "num_input_tokens_seen": 105611965, + "step": 3240 + }, + { + "epoch": 0.2922847995671191, + "flos": 61263018063360.0, + "grad_norm": 0.6526744940620888, + "learning_rate": 3.3208041344167317e-06, + "loss": 0.5667, + "num_input_tokens_seen": 105701205, + "step": 3241 + }, + { + "epoch": 0.29237498309058935, + "flos": 23843616103680.0, + "grad_norm": 1.6342930993881244, + "learning_rate": 3.3203653871944224e-06, + "loss": 0.6757, + "num_input_tokens_seen": 105728275, + "step": 3242 + }, + { + "epoch": 0.29246516661405964, + "flos": 22456594481280.0, + "grad_norm": 1.664840015136883, + "learning_rate": 3.3199265273112587e-06, + "loss": 0.7302, + "num_input_tokens_seen": 105756790, + "step": 3243 + }, + { + "epoch": 0.29255535013752987, + "flos": 17032226153760.0, + "grad_norm": 1.613980859720974, + "learning_rate": 3.3194875548046852e-06, + "loss": 0.7107, + "num_input_tokens_seen": 105785195, + "step": 3244 + }, + { + "epoch": 0.29264553366100016, + "flos": 23079917318400.0, + "grad_norm": 1.7688904669422882, + "learning_rate": 3.319048469712158e-06, + "loss": 0.7485, + "num_input_tokens_seen": 105814700, + "step": 3245 + }, + { + "epoch": 0.2927357171844704, + "flos": 24682409271360.0, + "grad_norm": 2.361003894490864, + "learning_rate": 3.3186092720711423e-06, + "loss": 0.8644, + "num_input_tokens_seen": 105842865, + "step": 3246 + }, + { + "epoch": 0.2928259007079407, + "flos": 23261408754720.0, + "grad_norm": 2.209853914360617, + "learning_rate": 3.3181699619191125e-06, + "loss": 0.6958, + "num_input_tokens_seen": 105871500, + "step": 3247 + }, + { + "epoch": 0.2929160842314109, + "flos": 19363397244480.0, + "grad_norm": 1.5835571465498681, + "learning_rate": 3.3177305392935536e-06, + "loss": 0.8104, + "num_input_tokens_seen": 105899850, + "step": 3248 + }, + { + "epoch": 0.2930062677548812, + "flos": 61796081256000.0, + "grad_norm": 0.738307407460725, + "learning_rate": 3.3172910042319595e-06, + "loss": 0.6552, + "num_input_tokens_seen": 105988695, + "step": 3249 + }, + { + "epoch": 0.29309645127835143, + "flos": 18524975774400.0, + "grad_norm": 1.6853960838365625, + "learning_rate": 3.316851356771833e-06, + "loss": 0.8033, + "num_input_tokens_seen": 106018300, + "step": 3250 + }, + { + "epoch": 0.2931866348018217, + "flos": 21984092294880.0, + "grad_norm": 1.8024258664741615, + "learning_rate": 3.3164115969506876e-06, + "loss": 0.8028, + "num_input_tokens_seen": 106047285, + "step": 3251 + }, + { + "epoch": 0.29327681832529195, + "flos": 16265776806240.0, + "grad_norm": 2.4443843164659014, + "learning_rate": 3.315971724806046e-06, + "loss": 0.8547, + "num_input_tokens_seen": 106068770, + "step": 3252 + }, + { + "epoch": 0.29336700184876224, + "flos": 32110877538720.0, + "grad_norm": 1.6144517988545168, + "learning_rate": 3.315531740375441e-06, + "loss": 0.7087, + "num_input_tokens_seen": 106100045, + "step": 3253 + }, + { + "epoch": 0.2934571853722325, + "flos": 28215876779040.0, + "grad_norm": 1.747336404152716, + "learning_rate": 3.315091643696414e-06, + "loss": 0.736, + "num_input_tokens_seen": 106132315, + "step": 3254 + }, + { + "epoch": 0.29354736889570276, + "flos": 24646274531040.0, + "grad_norm": 1.5041236621207505, + "learning_rate": 3.3146514348065164e-06, + "loss": 0.6936, + "num_input_tokens_seen": 106161790, + "step": 3255 + }, + { + "epoch": 0.293637552419173, + "flos": 20274088195200.0, + "grad_norm": 4.9624161566658715, + "learning_rate": 3.31421111374331e-06, + "loss": 0.7726, + "num_input_tokens_seen": 106190435, + "step": 3256 + }, + { + "epoch": 0.2937277359426433, + "flos": 17177768698560.0, + "grad_norm": 1.8480629987382189, + "learning_rate": 3.3137706805443647e-06, + "loss": 0.7777, + "num_input_tokens_seen": 106216620, + "step": 3257 + }, + { + "epoch": 0.2938179194661135, + "flos": 29852161777440.0, + "grad_norm": 1.6192773409230372, + "learning_rate": 3.313330135247261e-06, + "loss": 0.8187, + "num_input_tokens_seen": 106243715, + "step": 3258 + }, + { + "epoch": 0.2939081029895838, + "flos": 34702165575360.0, + "grad_norm": 2.3304093182281562, + "learning_rate": 3.312889477889588e-06, + "loss": 0.7421, + "num_input_tokens_seen": 106277560, + "step": 3259 + }, + { + "epoch": 0.29399828651305404, + "flos": 19837943767680.0, + "grad_norm": 4.416933999278175, + "learning_rate": 3.3124487085089464e-06, + "loss": 0.7806, + "num_input_tokens_seen": 106304140, + "step": 3260 + }, + { + "epoch": 0.29408847003652433, + "flos": 23116386586560.0, + "grad_norm": 1.6007777082293282, + "learning_rate": 3.312007827142943e-06, + "loss": 0.8027, + "num_input_tokens_seen": 106332325, + "step": 3261 + }, + { + "epoch": 0.29417865355999456, + "flos": 26173857950400.0, + "grad_norm": 1.434813269185799, + "learning_rate": 3.3115668338291983e-06, + "loss": 0.6889, + "num_input_tokens_seen": 106361120, + "step": 3262 + }, + { + "epoch": 0.29426883708346485, + "flos": 23329292203680.0, + "grad_norm": 1.77675221419126, + "learning_rate": 3.3111257286053394e-06, + "loss": 0.8517, + "num_input_tokens_seen": 106388615, + "step": 3263 + }, + { + "epoch": 0.29435902060693514, + "flos": 22568790017760.0, + "grad_norm": 16.62880852464237, + "learning_rate": 3.3106845115090043e-06, + "loss": 0.7654, + "num_input_tokens_seen": 106418790, + "step": 3264 + }, + { + "epoch": 0.2944492041304054, + "flos": 19399866512640.0, + "grad_norm": 1.8190501669980248, + "learning_rate": 3.310243182577839e-06, + "loss": 0.7169, + "num_input_tokens_seen": 106446660, + "step": 3265 + }, + { + "epoch": 0.29453938765387566, + "flos": 23513905899840.0, + "grad_norm": 1.7274341713139592, + "learning_rate": 3.3098017418495007e-06, + "loss": 0.6614, + "num_input_tokens_seen": 106476690, + "step": 3266 + }, + { + "epoch": 0.2946295711773459, + "flos": 29090767517280.0, + "grad_norm": 1.8463319706266412, + "learning_rate": 3.309360189361656e-06, + "loss": 0.7354, + "num_input_tokens_seen": 106507875, + "step": 3267 + }, + { + "epoch": 0.2947197547008162, + "flos": 24312847351200.0, + "grad_norm": 1.740166693847313, + "learning_rate": 3.3089185251519797e-06, + "loss": 0.8453, + "num_input_tokens_seen": 106536360, + "step": 3268 + }, + { + "epoch": 0.2948099382242864, + "flos": 28723212764160.0, + "grad_norm": 1.9355883858518526, + "learning_rate": 3.3084767492581574e-06, + "loss": 0.7475, + "num_input_tokens_seen": 106565510, + "step": 3269 + }, + { + "epoch": 0.2949001217477567, + "flos": 24063472465920.0, + "grad_norm": 1.9413886270932104, + "learning_rate": 3.3080348617178846e-06, + "loss": 0.7867, + "num_input_tokens_seen": 106592295, + "step": 3270 + }, + { + "epoch": 0.29499030527122694, + "flos": 19837163202720.0, + "grad_norm": 1.7071598798880168, + "learning_rate": 3.307592862568865e-06, + "loss": 0.7438, + "num_input_tokens_seen": 106620175, + "step": 3271 + }, + { + "epoch": 0.2950804887946972, + "flos": 29415794331360.0, + "grad_norm": 1.5936311415687783, + "learning_rate": 3.307150751848812e-06, + "loss": 0.833, + "num_input_tokens_seen": 106650240, + "step": 3272 + }, + { + "epoch": 0.29517067231816746, + "flos": 33898131866880.0, + "grad_norm": 4.9707975187500955, + "learning_rate": 3.3067085295954497e-06, + "loss": 0.6742, + "num_input_tokens_seen": 106683595, + "step": 3273 + }, + { + "epoch": 0.29526085584163775, + "flos": 28030965724800.0, + "grad_norm": 1.693911698846962, + "learning_rate": 3.3062661958465098e-06, + "loss": 0.6326, + "num_input_tokens_seen": 106713125, + "step": 3274 + }, + { + "epoch": 0.295351039365108, + "flos": 25046432897280.0, + "grad_norm": 2.0077486139367986, + "learning_rate": 3.305823750639736e-06, + "loss": 0.6578, + "num_input_tokens_seen": 106743170, + "step": 3275 + }, + { + "epoch": 0.29544122288857827, + "flos": 17214089287680.0, + "grad_norm": 1.6845270742865834, + "learning_rate": 3.3053811940128795e-06, + "loss": 0.7591, + "num_input_tokens_seen": 106769880, + "step": 3276 + }, + { + "epoch": 0.2955314064120485, + "flos": 21836988620160.0, + "grad_norm": 1.724009840580941, + "learning_rate": 3.3049385260037016e-06, + "loss": 0.7432, + "num_input_tokens_seen": 106797970, + "step": 3277 + }, + { + "epoch": 0.2956215899355188, + "flos": 36631654339680.0, + "grad_norm": 1.6766154740382255, + "learning_rate": 3.3044957466499736e-06, + "loss": 0.6597, + "num_input_tokens_seen": 106830270, + "step": 3278 + }, + { + "epoch": 0.295711773458989, + "flos": 23261891961600.0, + "grad_norm": 2.7767503321736657, + "learning_rate": 3.304052855989475e-06, + "loss": 0.684, + "num_input_tokens_seen": 106858190, + "step": 3279 + }, + { + "epoch": 0.2958019569824593, + "flos": 21694717014240.0, + "grad_norm": 1.7339468075162445, + "learning_rate": 3.3036098540599966e-06, + "loss": 0.8261, + "num_input_tokens_seen": 106888280, + "step": 3280 + }, + { + "epoch": 0.29589214050592955, + "flos": 20600936327520.0, + "grad_norm": 2.1610098009571272, + "learning_rate": 3.3031667408993373e-06, + "loss": 0.7227, + "num_input_tokens_seen": 106915895, + "step": 3281 + }, + { + "epoch": 0.29598232402939983, + "flos": 24572889957600.0, + "grad_norm": 1.874463068059242, + "learning_rate": 3.302723516545306e-06, + "loss": 0.7623, + "num_input_tokens_seen": 106945755, + "step": 3282 + }, + { + "epoch": 0.29607250755287007, + "flos": 21474674803200.0, + "grad_norm": 1.593789717235465, + "learning_rate": 3.302280181035722e-06, + "loss": 0.7912, + "num_input_tokens_seen": 106975495, + "step": 3283 + }, + { + "epoch": 0.29616269107634036, + "flos": 66651028632000.0, + "grad_norm": 0.5992928828371404, + "learning_rate": 3.3018367344084117e-06, + "loss": 0.6175, + "num_input_tokens_seen": 107073795, + "step": 3284 + }, + { + "epoch": 0.2962528745998106, + "flos": 29125378297440.0, + "grad_norm": 1.4460026098678906, + "learning_rate": 3.3013931767012125e-06, + "loss": 0.7534, + "num_input_tokens_seen": 107106080, + "step": 3285 + }, + { + "epoch": 0.2963430581232809, + "flos": 23334644649120.0, + "grad_norm": 1.685340912966846, + "learning_rate": 3.300949507951972e-06, + "loss": 0.7561, + "num_input_tokens_seen": 107134325, + "step": 3286 + }, + { + "epoch": 0.2964332416467511, + "flos": 21111506081760.0, + "grad_norm": 2.286831129497245, + "learning_rate": 3.300505728198546e-06, + "loss": 0.8086, + "num_input_tokens_seen": 107161400, + "step": 3287 + }, + { + "epoch": 0.2965234251702214, + "flos": 22569198885120.0, + "grad_norm": 1.9126304407453105, + "learning_rate": 3.3000618374788e-06, + "loss": 0.7477, + "num_input_tokens_seen": 107190065, + "step": 3288 + }, + { + "epoch": 0.2966136086936917, + "flos": 21512891050080.0, + "grad_norm": 1.627505571557981, + "learning_rate": 3.2996178358306104e-06, + "loss": 0.801, + "num_input_tokens_seen": 107217865, + "step": 3289 + }, + { + "epoch": 0.2967037922171619, + "flos": 34701124822080.0, + "grad_norm": 2.5874249654753916, + "learning_rate": 3.2991737232918606e-06, + "loss": 0.729, + "num_input_tokens_seen": 107249215, + "step": 3290 + }, + { + "epoch": 0.2967939757406322, + "flos": 18890374681440.0, + "grad_norm": 1.7391507492226568, + "learning_rate": 3.298729499900445e-06, + "loss": 0.7146, + "num_input_tokens_seen": 107276180, + "step": 3291 + }, + { + "epoch": 0.29688415926410244, + "flos": 14407219411200.0, + "grad_norm": 1.7996065689969238, + "learning_rate": 3.2982851656942677e-06, + "loss": 0.7043, + "num_input_tokens_seen": 107301095, + "step": 3292 + }, + { + "epoch": 0.29697434278757273, + "flos": 26905696517760.0, + "grad_norm": 1.5817407411943736, + "learning_rate": 3.2978407207112416e-06, + "loss": 0.7878, + "num_input_tokens_seen": 107331675, + "step": 3293 + }, + { + "epoch": 0.29706452631104296, + "flos": 19072274985120.0, + "grad_norm": 3.4078708257148334, + "learning_rate": 3.2973961649892888e-06, + "loss": 0.7685, + "num_input_tokens_seen": 107359170, + "step": 3294 + }, + { + "epoch": 0.29715470983451325, + "flos": 23480893419360.0, + "grad_norm": 2.711045496655584, + "learning_rate": 3.296951498566341e-06, + "loss": 0.8249, + "num_input_tokens_seen": 107388030, + "step": 3295 + }, + { + "epoch": 0.2972448933579835, + "flos": 17906708024640.0, + "grad_norm": 2.230325938127251, + "learning_rate": 3.2965067214803404e-06, + "loss": 0.7562, + "num_input_tokens_seen": 107415825, + "step": 3296 + }, + { + "epoch": 0.2973350768814538, + "flos": 27593222997600.0, + "grad_norm": 1.296742903634653, + "learning_rate": 3.2960618337692372e-06, + "loss": 0.7998, + "num_input_tokens_seen": 107447435, + "step": 3297 + }, + { + "epoch": 0.297425260404924, + "flos": 37432194090720.0, + "grad_norm": 1.6512713187355412, + "learning_rate": 3.2956168354709927e-06, + "loss": 0.7234, + "num_input_tokens_seen": 107480895, + "step": 3298 + }, + { + "epoch": 0.2975154439283943, + "flos": 20784695119200.0, + "grad_norm": 2.1766712826986083, + "learning_rate": 3.2951717266235754e-06, + "loss": 0.814, + "num_input_tokens_seen": 107506635, + "step": 3299 + }, + { + "epoch": 0.29760562745186453, + "flos": 26832200435040.0, + "grad_norm": 1.5127123344749187, + "learning_rate": 3.294726507264964e-06, + "loss": 0.6353, + "num_input_tokens_seen": 107537435, + "step": 3300 + }, + { + "epoch": 0.2976958109753348, + "flos": 25156323908640.0, + "grad_norm": 1.804660417097582, + "learning_rate": 3.2942811774331487e-06, + "loss": 0.7357, + "num_input_tokens_seen": 107564145, + "step": 3301 + }, + { + "epoch": 0.29778599449880505, + "flos": 21002432805120.0, + "grad_norm": 1.8545179641482998, + "learning_rate": 3.293835737166127e-06, + "loss": 0.7574, + "num_input_tokens_seen": 107590845, + "step": 3302 + }, + { + "epoch": 0.29787617802227534, + "flos": 21257531833440.0, + "grad_norm": 2.1512133612164264, + "learning_rate": 3.293390186501906e-06, + "loss": 0.7102, + "num_input_tokens_seen": 107622135, + "step": 3303 + }, + { + "epoch": 0.2979663615457456, + "flos": 14845817042880.0, + "grad_norm": 1.8684148864347425, + "learning_rate": 3.2929445254785024e-06, + "loss": 0.7716, + "num_input_tokens_seen": 107649785, + "step": 3304 + }, + { + "epoch": 0.29805654506921586, + "flos": 16448160316800.0, + "grad_norm": 2.203876015646091, + "learning_rate": 3.2924987541339423e-06, + "loss": 0.7188, + "num_input_tokens_seen": 107673780, + "step": 3305 + }, + { + "epoch": 0.2981467285926861, + "flos": 26103335448480.0, + "grad_norm": 1.9676591182091987, + "learning_rate": 3.292052872506262e-06, + "loss": 0.7838, + "num_input_tokens_seen": 107701105, + "step": 3306 + }, + { + "epoch": 0.2982369121161564, + "flos": 19947165723360.0, + "grad_norm": 1.5062428739387077, + "learning_rate": 3.291606880633506e-06, + "loss": 0.8026, + "num_input_tokens_seen": 107729665, + "step": 3307 + }, + { + "epoch": 0.2983270956396266, + "flos": 19472396181600.0, + "grad_norm": 2.445957379117176, + "learning_rate": 3.2911607785537297e-06, + "loss": 0.8182, + "num_input_tokens_seen": 107756625, + "step": 3308 + }, + { + "epoch": 0.2984172791630969, + "flos": 22456594481280.0, + "grad_norm": 1.8844928761522435, + "learning_rate": 3.290714566304997e-06, + "loss": 0.8522, + "num_input_tokens_seen": 107780680, + "step": 3309 + }, + { + "epoch": 0.29850746268656714, + "flos": 22532692447200.0, + "grad_norm": 3.6574518559612397, + "learning_rate": 3.2902682439253794e-06, + "loss": 0.8353, + "num_input_tokens_seen": 107805350, + "step": 3310 + }, + { + "epoch": 0.2985976462100374, + "flos": 18707656643040.0, + "grad_norm": 2.165283409580587, + "learning_rate": 3.289821811452961e-06, + "loss": 0.7774, + "num_input_tokens_seen": 107830140, + "step": 3311 + }, + { + "epoch": 0.2986878297335077, + "flos": 22022828918400.0, + "grad_norm": 2.604645514650415, + "learning_rate": 3.289375268925834e-06, + "loss": 0.8167, + "num_input_tokens_seen": 107858335, + "step": 3312 + }, + { + "epoch": 0.29877801325697795, + "flos": 26831345530560.0, + "grad_norm": 2.1323925904462433, + "learning_rate": 3.288928616382099e-06, + "loss": 0.7528, + "num_input_tokens_seen": 107888415, + "step": 3313 + }, + { + "epoch": 0.29886819678044824, + "flos": 21693713430720.0, + "grad_norm": 1.8184470904343708, + "learning_rate": 3.288481853859868e-06, + "loss": 0.7979, + "num_input_tokens_seen": 107917720, + "step": 3314 + }, + { + "epoch": 0.29895838030391847, + "flos": 22242016224960.0, + "grad_norm": 2.9219448468840787, + "learning_rate": 3.2880349813972604e-06, + "loss": 0.7391, + "num_input_tokens_seen": 107947360, + "step": 3315 + }, + { + "epoch": 0.29904856382738876, + "flos": 22382726700960.0, + "grad_norm": 2.4325694566084572, + "learning_rate": 3.2875879990324052e-06, + "loss": 0.7651, + "num_input_tokens_seen": 107977110, + "step": 3316 + }, + { + "epoch": 0.299138747350859, + "flos": 24354594725280.0, + "grad_norm": 1.8001572195904252, + "learning_rate": 3.287140906803443e-06, + "loss": 0.8137, + "num_input_tokens_seen": 108005420, + "step": 3317 + }, + { + "epoch": 0.2992289308743293, + "flos": 19982408389440.0, + "grad_norm": 2.045685050961724, + "learning_rate": 3.2866937047485216e-06, + "loss": 0.6817, + "num_input_tokens_seen": 108032825, + "step": 3318 + }, + { + "epoch": 0.2993191143977995, + "flos": 27450951391680.0, + "grad_norm": 2.560140872894009, + "learning_rate": 3.2862463929057985e-06, + "loss": 0.8242, + "num_input_tokens_seen": 108061500, + "step": 3319 + }, + { + "epoch": 0.2994092979212698, + "flos": 39396590993280.0, + "grad_norm": 5.821909517686522, + "learning_rate": 3.285798971313441e-06, + "loss": 0.6109, + "num_input_tokens_seen": 108097020, + "step": 3320 + }, + { + "epoch": 0.29949948144474003, + "flos": 23625358041120.0, + "grad_norm": 1.6546209709185862, + "learning_rate": 3.2853514400096248e-06, + "loss": 0.7847, + "num_input_tokens_seen": 108126895, + "step": 3321 + }, + { + "epoch": 0.2995896649682103, + "flos": 22823294329920.0, + "grad_norm": 2.150001095616261, + "learning_rate": 3.2849037990325367e-06, + "loss": 0.7885, + "num_input_tokens_seen": 108156520, + "step": 3322 + }, + { + "epoch": 0.29967984849168056, + "flos": 19218226397280.0, + "grad_norm": 1.477161911496048, + "learning_rate": 3.2844560484203717e-06, + "loss": 0.7214, + "num_input_tokens_seen": 108185260, + "step": 3323 + }, + { + "epoch": 0.29977003201515084, + "flos": 28322794209600.0, + "grad_norm": 1.5164816097222322, + "learning_rate": 3.2840081882113333e-06, + "loss": 0.8457, + "num_input_tokens_seen": 108217230, + "step": 3324 + }, + { + "epoch": 0.2998602155386211, + "flos": 28176842797440.0, + "grad_norm": 2.2713835046860993, + "learning_rate": 3.283560218443638e-06, + "loss": 0.737, + "num_input_tokens_seen": 108244505, + "step": 3325 + }, + { + "epoch": 0.29995039906209137, + "flos": 16522436964480.0, + "grad_norm": 1.644851109801777, + "learning_rate": 3.2831121391555064e-06, + "loss": 0.7777, + "num_input_tokens_seen": 108270880, + "step": 3326 + }, + { + "epoch": 0.3000405825855616, + "flos": 38744939065440.0, + "grad_norm": 1.8876115096544641, + "learning_rate": 3.2826639503851724e-06, + "loss": 0.7151, + "num_input_tokens_seen": 108307305, + "step": 3327 + }, + { + "epoch": 0.3001307661090319, + "flos": 26684204686080.0, + "grad_norm": 1.4207633567127174, + "learning_rate": 3.282215652170877e-06, + "loss": 0.8053, + "num_input_tokens_seen": 108336430, + "step": 3328 + }, + { + "epoch": 0.3002209496325021, + "flos": 23335462383840.0, + "grad_norm": 3.2534299504155215, + "learning_rate": 3.281767244550873e-06, + "loss": 0.7964, + "num_input_tokens_seen": 108366935, + "step": 3329 + }, + { + "epoch": 0.3003111331559724, + "flos": 65310623622240.0, + "grad_norm": 0.6195893932637068, + "learning_rate": 3.2813187275634193e-06, + "loss": 0.5525, + "num_input_tokens_seen": 108462005, + "step": 3330 + }, + { + "epoch": 0.30040131667944264, + "flos": 23368772222400.0, + "grad_norm": 1.4843698601884698, + "learning_rate": 3.280870101246787e-06, + "loss": 0.7876, + "num_input_tokens_seen": 108490305, + "step": 3331 + }, + { + "epoch": 0.30049150020291293, + "flos": 21253368820320.0, + "grad_norm": 1.8099231818531953, + "learning_rate": 3.280421365639255e-06, + "loss": 0.686, + "num_input_tokens_seen": 108515785, + "step": 3332 + }, + { + "epoch": 0.30058168372638316, + "flos": 18890188832640.0, + "grad_norm": 1.5803694674524706, + "learning_rate": 3.279972520779112e-06, + "loss": 0.7454, + "num_input_tokens_seen": 108543170, + "step": 3333 + }, + { + "epoch": 0.30067186724985345, + "flos": 18598992233760.0, + "grad_norm": 1.4934453007042738, + "learning_rate": 3.279523566704656e-06, + "loss": 0.8318, + "num_input_tokens_seen": 108569525, + "step": 3334 + }, + { + "epoch": 0.30076205077332374, + "flos": 20777707204320.0, + "grad_norm": 1.8450112635571057, + "learning_rate": 3.2790745034541935e-06, + "loss": 0.8158, + "num_input_tokens_seen": 108593425, + "step": 3335 + }, + { + "epoch": 0.300852234296794, + "flos": 17432161501440.0, + "grad_norm": 1.7131716977156382, + "learning_rate": 3.278625331066042e-06, + "loss": 0.8088, + "num_input_tokens_seen": 108619660, + "step": 3336 + }, + { + "epoch": 0.30094241782026426, + "flos": 21731074773120.0, + "grad_norm": 1.932695276379997, + "learning_rate": 3.278176049578527e-06, + "loss": 0.8115, + "num_input_tokens_seen": 108647705, + "step": 3337 + }, + { + "epoch": 0.3010326013437345, + "flos": 36013089231840.0, + "grad_norm": 1.7692957841448111, + "learning_rate": 3.2777266590299835e-06, + "loss": 0.7239, + "num_input_tokens_seen": 108677930, + "step": 3338 + }, + { + "epoch": 0.3011227848672048, + "flos": 34810569796320.0, + "grad_norm": 1.7251801631663548, + "learning_rate": 3.2772771594587562e-06, + "loss": 0.7001, + "num_input_tokens_seen": 108710980, + "step": 3339 + }, + { + "epoch": 0.301212968390675, + "flos": 22715001618240.0, + "grad_norm": 1.9375119362740312, + "learning_rate": 3.2768275509031988e-06, + "loss": 0.8694, + "num_input_tokens_seen": 108738285, + "step": 3340 + }, + { + "epoch": 0.3013031519141453, + "flos": 25191975442080.0, + "grad_norm": 2.690219922204852, + "learning_rate": 3.276377833401675e-06, + "loss": 0.7443, + "num_input_tokens_seen": 108765630, + "step": 3341 + }, + { + "epoch": 0.30139333543761554, + "flos": 28180262415360.0, + "grad_norm": 1.6887626868843695, + "learning_rate": 3.2759280069925557e-06, + "loss": 0.8278, + "num_input_tokens_seen": 108795045, + "step": 3342 + }, + { + "epoch": 0.30148351896108583, + "flos": 24385488529440.0, + "grad_norm": 1.4934954278682457, + "learning_rate": 3.2754780717142233e-06, + "loss": 0.8048, + "num_input_tokens_seen": 108823055, + "step": 3343 + }, + { + "epoch": 0.30157370248455606, + "flos": 19690654244160.0, + "grad_norm": 1.9770852102653296, + "learning_rate": 3.27502802760507e-06, + "loss": 0.741, + "num_input_tokens_seen": 108850920, + "step": 3344 + }, + { + "epoch": 0.30166388600802635, + "flos": 28252234537920.0, + "grad_norm": 1.7459813024168622, + "learning_rate": 3.2745778747034943e-06, + "loss": 0.6796, + "num_input_tokens_seen": 108882075, + "step": 3345 + }, + { + "epoch": 0.3017540695314966, + "flos": 24391138332960.0, + "grad_norm": 2.062078490679051, + "learning_rate": 3.274127613047906e-06, + "loss": 0.7463, + "num_input_tokens_seen": 108910880, + "step": 3346 + }, + { + "epoch": 0.30184425305496687, + "flos": 22642248930720.0, + "grad_norm": 1.9830133767727631, + "learning_rate": 3.273677242676725e-06, + "loss": 0.7834, + "num_input_tokens_seen": 108939930, + "step": 3347 + }, + { + "epoch": 0.3019344365784371, + "flos": 23808187588800.0, + "grad_norm": 2.40844606533881, + "learning_rate": 3.2732267636283782e-06, + "loss": 0.7653, + "num_input_tokens_seen": 108969380, + "step": 3348 + }, + { + "epoch": 0.3020246201019074, + "flos": 16735045223520.0, + "grad_norm": 2.6552316845455315, + "learning_rate": 3.2727761759413034e-06, + "loss": 0.7297, + "num_input_tokens_seen": 108995500, + "step": 3349 + }, + { + "epoch": 0.3021148036253776, + "flos": 20087467332000.0, + "grad_norm": 1.729027003274315, + "learning_rate": 3.2723254796539477e-06, + "loss": 0.7787, + "num_input_tokens_seen": 109023150, + "step": 3350 + }, + { + "epoch": 0.3022049871488479, + "flos": 36632174716320.0, + "grad_norm": 8.990259197432058, + "learning_rate": 3.271874674804766e-06, + "loss": 0.7552, + "num_input_tokens_seen": 109053995, + "step": 3351 + }, + { + "epoch": 0.30229517067231815, + "flos": 19581989834880.0, + "grad_norm": 3.0923228970068806, + "learning_rate": 3.2714237614322242e-06, + "loss": 0.7912, + "num_input_tokens_seen": 109082595, + "step": 3352 + }, + { + "epoch": 0.30238535419578844, + "flos": 22165843919520.0, + "grad_norm": 1.7769385242679923, + "learning_rate": 3.2709727395747974e-06, + "loss": 0.6039, + "num_input_tokens_seen": 109114065, + "step": 3353 + }, + { + "epoch": 0.30247553771925867, + "flos": 68515756207200.0, + "grad_norm": 0.6812241601855566, + "learning_rate": 3.2705216092709673e-06, + "loss": 0.5717, + "num_input_tokens_seen": 109205010, + "step": 3354 + }, + { + "epoch": 0.30256572124272896, + "flos": 17904291990240.0, + "grad_norm": 2.0635619464040156, + "learning_rate": 3.2700703705592282e-06, + "loss": 0.8533, + "num_input_tokens_seen": 109228635, + "step": 3355 + }, + { + "epoch": 0.3026559047661992, + "flos": 22496148839520.0, + "grad_norm": 1.8010923567861088, + "learning_rate": 3.269619023478082e-06, + "loss": 0.6683, + "num_input_tokens_seen": 109256625, + "step": 3356 + }, + { + "epoch": 0.3027460882896695, + "flos": 24312698672160.0, + "grad_norm": 2.0610990679315657, + "learning_rate": 3.26916756806604e-06, + "loss": 0.8651, + "num_input_tokens_seen": 109284230, + "step": 3357 + }, + { + "epoch": 0.3028362718131397, + "flos": 12987965873280.0, + "grad_norm": 1.6471617041941553, + "learning_rate": 3.268716004361623e-06, + "loss": 0.7834, + "num_input_tokens_seen": 109311410, + "step": 3358 + }, + { + "epoch": 0.30292645533661, + "flos": 65904650954880.0, + "grad_norm": 0.6562092859263414, + "learning_rate": 3.268264332403361e-06, + "loss": 0.5869, + "num_input_tokens_seen": 109412370, + "step": 3359 + }, + { + "epoch": 0.3030166388600803, + "flos": 25483618078080.0, + "grad_norm": 3.2746557913867815, + "learning_rate": 3.2678125522297933e-06, + "loss": 0.7984, + "num_input_tokens_seen": 109443160, + "step": 3360 + }, + { + "epoch": 0.3031068223835505, + "flos": 26285384431200.0, + "grad_norm": 2.281011595091559, + "learning_rate": 3.267360663879468e-06, + "loss": 0.707, + "num_input_tokens_seen": 109470535, + "step": 3361 + }, + { + "epoch": 0.3031970059070208, + "flos": 30111795516480.0, + "grad_norm": 2.5902352475881223, + "learning_rate": 3.266908667390942e-06, + "loss": 0.7564, + "num_input_tokens_seen": 109501925, + "step": 3362 + }, + { + "epoch": 0.30328718943049104, + "flos": 25847641704000.0, + "grad_norm": 1.7780329112256301, + "learning_rate": 3.2664565628027833e-06, + "loss": 0.8212, + "num_input_tokens_seen": 109527155, + "step": 3363 + }, + { + "epoch": 0.30337737295396133, + "flos": 22386703865280.0, + "grad_norm": 2.7052833892394657, + "learning_rate": 3.2660043501535675e-06, + "loss": 0.7654, + "num_input_tokens_seen": 109553090, + "step": 3364 + }, + { + "epoch": 0.30346755647743157, + "flos": 21257494663680.0, + "grad_norm": 2.3004579171151924, + "learning_rate": 3.2655520294818797e-06, + "loss": 0.8376, + "num_input_tokens_seen": 109577210, + "step": 3365 + }, + { + "epoch": 0.30355774000090185, + "flos": 22092013308960.0, + "grad_norm": 1.5333850533981552, + "learning_rate": 3.2650996008263146e-06, + "loss": 0.8367, + "num_input_tokens_seen": 109608025, + "step": 3366 + }, + { + "epoch": 0.3036479235243721, + "flos": 22930434779040.0, + "grad_norm": 2.193521223610329, + "learning_rate": 3.2646470642254756e-06, + "loss": 0.7266, + "num_input_tokens_seen": 109637105, + "step": 3367 + }, + { + "epoch": 0.3037381070478424, + "flos": 21804236328000.0, + "grad_norm": 2.026764127161431, + "learning_rate": 3.2641944197179767e-06, + "loss": 0.7883, + "num_input_tokens_seen": 109664650, + "step": 3368 + }, + { + "epoch": 0.3038282905713126, + "flos": 55394365130880.0, + "grad_norm": 0.6624349958090117, + "learning_rate": 3.2637416673424383e-06, + "loss": 0.572, + "num_input_tokens_seen": 109739905, + "step": 3369 + }, + { + "epoch": 0.3039184740947829, + "flos": 24132322328640.0, + "grad_norm": 2.7691304894777216, + "learning_rate": 3.2632888071374937e-06, + "loss": 0.7791, + "num_input_tokens_seen": 109769190, + "step": 3370 + }, + { + "epoch": 0.30400865761825313, + "flos": 15574979387520.0, + "grad_norm": 1.8303742225195676, + "learning_rate": 3.2628358391417815e-06, + "loss": 0.7978, + "num_input_tokens_seen": 109797120, + "step": 3371 + }, + { + "epoch": 0.3040988411417234, + "flos": 20784026063520.0, + "grad_norm": 1.7697877003350435, + "learning_rate": 3.2623827633939526e-06, + "loss": 0.7591, + "num_input_tokens_seen": 109825965, + "step": 3372 + }, + { + "epoch": 0.30418902466519365, + "flos": 22638308936160.0, + "grad_norm": 1.8886453021131877, + "learning_rate": 3.2619295799326657e-06, + "loss": 0.7597, + "num_input_tokens_seen": 109854420, + "step": 3373 + }, + { + "epoch": 0.30427920818866394, + "flos": 32076341098080.0, + "grad_norm": 2.3655693138351177, + "learning_rate": 3.2614762887965883e-06, + "loss": 0.6667, + "num_input_tokens_seen": 109880025, + "step": 3374 + }, + { + "epoch": 0.3043693917121342, + "flos": 13351654971360.0, + "grad_norm": 1.8603208043717145, + "learning_rate": 3.2610228900243984e-06, + "loss": 0.7555, + "num_input_tokens_seen": 109905225, + "step": 3375 + }, + { + "epoch": 0.30445957523560446, + "flos": 70044306040320.0, + "grad_norm": 0.6074888676930273, + "learning_rate": 3.260569383654783e-06, + "loss": 0.6234, + "num_input_tokens_seen": 110000745, + "step": 3376 + }, + { + "epoch": 0.3045497587590747, + "flos": 70303345063200.0, + "grad_norm": 0.6907053050651961, + "learning_rate": 3.2601157697264365e-06, + "loss": 0.5939, + "num_input_tokens_seen": 110091210, + "step": 3377 + }, + { + "epoch": 0.304639942282545, + "flos": 20347101071040.0, + "grad_norm": 1.85227611326556, + "learning_rate": 3.2596620482780647e-06, + "loss": 0.7447, + "num_input_tokens_seen": 110118245, + "step": 3378 + }, + { + "epoch": 0.3047301258060152, + "flos": 62647103274720.0, + "grad_norm": 0.5820937155159712, + "learning_rate": 3.2592082193483803e-06, + "loss": 0.5924, + "num_input_tokens_seen": 110216100, + "step": 3379 + }, + { + "epoch": 0.3048203093294855, + "flos": 24464225548320.0, + "grad_norm": 1.7550192107537588, + "learning_rate": 3.258754282976109e-06, + "loss": 0.7537, + "num_input_tokens_seen": 110246625, + "step": 3380 + }, + { + "epoch": 0.30491049285295574, + "flos": 17724621872160.0, + "grad_norm": 2.1377282137928906, + "learning_rate": 3.25830023919998e-06, + "loss": 0.8398, + "num_input_tokens_seen": 110274335, + "step": 3381 + }, + { + "epoch": 0.305000676376426, + "flos": 25695297093120.0, + "grad_norm": 1.7976296871764266, + "learning_rate": 3.2578460880587374e-06, + "loss": 0.7224, + "num_input_tokens_seen": 110303405, + "step": 3382 + }, + { + "epoch": 0.3050908598998963, + "flos": 19946273649120.0, + "grad_norm": 2.592862356496686, + "learning_rate": 3.2573918295911306e-06, + "loss": 0.821, + "num_input_tokens_seen": 110330455, + "step": 3383 + }, + { + "epoch": 0.30518104342336655, + "flos": 19581989834880.0, + "grad_norm": 2.127299803336988, + "learning_rate": 3.2569374638359196e-06, + "loss": 0.7077, + "num_input_tokens_seen": 110358145, + "step": 3384 + }, + { + "epoch": 0.30527122694683684, + "flos": 24207230862240.0, + "grad_norm": 1.3972082009485167, + "learning_rate": 3.2564829908318736e-06, + "loss": 0.7803, + "num_input_tokens_seen": 110388945, + "step": 3385 + }, + { + "epoch": 0.30536141047030707, + "flos": 18995545133280.0, + "grad_norm": 1.3359359384271072, + "learning_rate": 3.2560284106177705e-06, + "loss": 0.8259, + "num_input_tokens_seen": 110417590, + "step": 3386 + }, + { + "epoch": 0.30545159399377736, + "flos": 17688115434240.0, + "grad_norm": 1.7139304820488324, + "learning_rate": 3.2555737232323978e-06, + "loss": 0.7386, + "num_input_tokens_seen": 110443600, + "step": 3387 + }, + { + "epoch": 0.3055417775172476, + "flos": 26827256856960.0, + "grad_norm": 1.9155066200008524, + "learning_rate": 3.255118928714552e-06, + "loss": 0.7977, + "num_input_tokens_seen": 110474855, + "step": 3388 + }, + { + "epoch": 0.3056319610407179, + "flos": 19909358343840.0, + "grad_norm": 2.727575775128864, + "learning_rate": 3.2546640271030386e-06, + "loss": 0.6965, + "num_input_tokens_seen": 110501140, + "step": 3389 + }, + { + "epoch": 0.3057221445641881, + "flos": 70476770661600.0, + "grad_norm": 0.6226510970262608, + "learning_rate": 3.2542090184366717e-06, + "loss": 0.6244, + "num_input_tokens_seen": 110592975, + "step": 3390 + }, + { + "epoch": 0.3058123280876584, + "flos": 21652189075200.0, + "grad_norm": 2.3856542432880055, + "learning_rate": 3.253753902754276e-06, + "loss": 0.8504, + "num_input_tokens_seen": 110618905, + "step": 3391 + }, + { + "epoch": 0.30590251161112864, + "flos": 46471629016320.0, + "grad_norm": 1.3426099865007113, + "learning_rate": 3.253298680094685e-06, + "loss": 0.6995, + "num_input_tokens_seen": 110658605, + "step": 3392 + }, + { + "epoch": 0.3059926951345989, + "flos": 21257717682240.0, + "grad_norm": 4.979532693899034, + "learning_rate": 3.2528433504967394e-06, + "loss": 0.7132, + "num_input_tokens_seen": 110687475, + "step": 3393 + }, + { + "epoch": 0.30608287865806916, + "flos": 23225348353920.0, + "grad_norm": 2.711841091868022, + "learning_rate": 3.252387913999291e-06, + "loss": 0.7812, + "num_input_tokens_seen": 110715865, + "step": 3394 + }, + { + "epoch": 0.30617306218153945, + "flos": 26066568822240.0, + "grad_norm": 1.7243334159524129, + "learning_rate": 3.2519323706411998e-06, + "loss": 0.7702, + "num_input_tokens_seen": 110745160, + "step": 3395 + }, + { + "epoch": 0.3062632457050097, + "flos": 30504148233120.0, + "grad_norm": 1.5433775541315766, + "learning_rate": 3.251476720461336e-06, + "loss": 0.6568, + "num_input_tokens_seen": 110774795, + "step": 3396 + }, + { + "epoch": 0.30635342922847997, + "flos": 14114164324320.0, + "grad_norm": 1.8439274072156568, + "learning_rate": 3.251020963498578e-06, + "loss": 0.6967, + "num_input_tokens_seen": 110801090, + "step": 3397 + }, + { + "epoch": 0.3064436127519502, + "flos": 23114007721920.0, + "grad_norm": 1.524900702073444, + "learning_rate": 3.250565099791813e-06, + "loss": 0.812, + "num_input_tokens_seen": 110828475, + "step": 3398 + }, + { + "epoch": 0.3065337962754205, + "flos": 24607240549440.0, + "grad_norm": 1.5852231075049426, + "learning_rate": 3.2501091293799387e-06, + "loss": 0.8227, + "num_input_tokens_seen": 110858965, + "step": 3399 + }, + { + "epoch": 0.3066239797988907, + "flos": 21694754184000.0, + "grad_norm": 1.7768715702846298, + "learning_rate": 3.24965305230186e-06, + "loss": 0.7812, + "num_input_tokens_seen": 110886510, + "step": 3400 + }, + { + "epoch": 0.306714163322361, + "flos": 11676261651840.0, + "grad_norm": 2.1125440739655637, + "learning_rate": 3.249196868596492e-06, + "loss": 0.8399, + "num_input_tokens_seen": 110909435, + "step": 3401 + }, + { + "epoch": 0.30680434684583124, + "flos": 31130704839360.0, + "grad_norm": 3.453606343272492, + "learning_rate": 3.24874057830276e-06, + "loss": 0.6958, + "num_input_tokens_seen": 110938050, + "step": 3402 + }, + { + "epoch": 0.30689453036930153, + "flos": 33458196123840.0, + "grad_norm": 1.7799029024836994, + "learning_rate": 3.2482841814595954e-06, + "loss": 0.6883, + "num_input_tokens_seen": 110967790, + "step": 3403 + }, + { + "epoch": 0.30698471389277177, + "flos": 27159234416160.0, + "grad_norm": 2.425508984506827, + "learning_rate": 3.247827678105943e-06, + "loss": 0.7107, + "num_input_tokens_seen": 110997015, + "step": 3404 + }, + { + "epoch": 0.30707489741624205, + "flos": 24063769824000.0, + "grad_norm": 1.6517135103564993, + "learning_rate": 3.247371068280751e-06, + "loss": 0.7661, + "num_input_tokens_seen": 111027495, + "step": 3405 + }, + { + "epoch": 0.3071650809397123, + "flos": 23512047411840.0, + "grad_norm": 1.5372957264490368, + "learning_rate": 3.2469143520229823e-06, + "loss": 0.6262, + "num_input_tokens_seen": 111057735, + "step": 3406 + }, + { + "epoch": 0.3072552644631826, + "flos": 38709361871520.0, + "grad_norm": 2.214058657840041, + "learning_rate": 3.2464575293716054e-06, + "loss": 0.717, + "num_input_tokens_seen": 111088950, + "step": 3407 + }, + { + "epoch": 0.30734544798665286, + "flos": 22241718866880.0, + "grad_norm": 1.9732915176437404, + "learning_rate": 3.2460006003655997e-06, + "loss": 0.7076, + "num_input_tokens_seen": 111116475, + "step": 3408 + }, + { + "epoch": 0.3074356315101231, + "flos": 21875167697280.0, + "grad_norm": 2.1164853666661654, + "learning_rate": 3.245543565043952e-06, + "loss": 0.8207, + "num_input_tokens_seen": 111140025, + "step": 3409 + }, + { + "epoch": 0.3075258150335934, + "flos": 18523191625920.0, + "grad_norm": 2.159706882968848, + "learning_rate": 3.2450864234456592e-06, + "loss": 0.761, + "num_input_tokens_seen": 111166145, + "step": 3410 + }, + { + "epoch": 0.3076159985570636, + "flos": 37718298432480.0, + "grad_norm": 1.4577007647478644, + "learning_rate": 3.244629175609728e-06, + "loss": 0.7588, + "num_input_tokens_seen": 111195965, + "step": 3411 + }, + { + "epoch": 0.3077061820805339, + "flos": 22197927156000.0, + "grad_norm": 2.1990328179964127, + "learning_rate": 3.2441718215751726e-06, + "loss": 0.7509, + "num_input_tokens_seen": 111223740, + "step": 3412 + }, + { + "epoch": 0.30779636560400414, + "flos": 18889408267680.0, + "grad_norm": 1.7959682664680232, + "learning_rate": 3.2437143613810173e-06, + "loss": 0.7272, + "num_input_tokens_seen": 111249990, + "step": 3413 + }, + { + "epoch": 0.30788654912747443, + "flos": 28325396092800.0, + "grad_norm": 1.6319973749906336, + "learning_rate": 3.2432567950662947e-06, + "loss": 0.8092, + "num_input_tokens_seen": 111281660, + "step": 3414 + }, + { + "epoch": 0.30797673265094466, + "flos": 24756425730720.0, + "grad_norm": 2.1601636448435246, + "learning_rate": 3.2427991226700468e-06, + "loss": 0.7528, + "num_input_tokens_seen": 111312705, + "step": 3415 + }, + { + "epoch": 0.30806691617441495, + "flos": 20747668304640.0, + "grad_norm": 1.890979260759084, + "learning_rate": 3.2423413442313246e-06, + "loss": 0.6892, + "num_input_tokens_seen": 111340865, + "step": 3416 + }, + { + "epoch": 0.3081570996978852, + "flos": 25700426520000.0, + "grad_norm": 1.8955859207969918, + "learning_rate": 3.2418834597891904e-06, + "loss": 0.6815, + "num_input_tokens_seen": 111369030, + "step": 3417 + }, + { + "epoch": 0.3082472832213555, + "flos": 14626964264160.0, + "grad_norm": 2.3513734050775943, + "learning_rate": 3.2414254693827098e-06, + "loss": 0.854, + "num_input_tokens_seen": 111394090, + "step": 3418 + }, + { + "epoch": 0.3083374667448257, + "flos": 21948106233600.0, + "grad_norm": 1.5425285387413163, + "learning_rate": 3.2409673730509644e-06, + "loss": 0.7409, + "num_input_tokens_seen": 111423825, + "step": 3419 + }, + { + "epoch": 0.308427650268296, + "flos": 28545215285280.0, + "grad_norm": 2.074376791700122, + "learning_rate": 3.2405091708330393e-06, + "loss": 0.7402, + "num_input_tokens_seen": 111450475, + "step": 3420 + }, + { + "epoch": 0.3085178337917662, + "flos": 29268541977600.0, + "grad_norm": 1.9319170839828415, + "learning_rate": 3.2400508627680323e-06, + "loss": 0.754, + "num_input_tokens_seen": 111481685, + "step": 3421 + }, + { + "epoch": 0.3086080173152365, + "flos": 15246644464800.0, + "grad_norm": 1.8984616752937638, + "learning_rate": 3.2395924488950474e-06, + "loss": 0.858, + "num_input_tokens_seen": 111505800, + "step": 3422 + }, + { + "epoch": 0.30869820083870675, + "flos": 22124839940640.0, + "grad_norm": 2.0253680236202793, + "learning_rate": 3.2391339292532004e-06, + "loss": 0.8932, + "num_input_tokens_seen": 111532870, + "step": 3423 + }, + { + "epoch": 0.30878838436217704, + "flos": 34373421785280.0, + "grad_norm": 1.940372798337242, + "learning_rate": 3.238675303881614e-06, + "loss": 0.7171, + "num_input_tokens_seen": 111563965, + "step": 3424 + }, + { + "epoch": 0.30887856788564727, + "flos": 20710827338880.0, + "grad_norm": 1.6101011101148557, + "learning_rate": 3.2382165728194203e-06, + "loss": 0.7323, + "num_input_tokens_seen": 111591700, + "step": 3425 + }, + { + "epoch": 0.30896875140911756, + "flos": 35282774624640.0, + "grad_norm": 1.475952587611911, + "learning_rate": 3.237757736105761e-06, + "loss": 0.7449, + "num_input_tokens_seen": 111623395, + "step": 3426 + }, + { + "epoch": 0.3090589349325878, + "flos": 25848682457280.0, + "grad_norm": 1.5285269990532444, + "learning_rate": 3.2372987937797867e-06, + "loss": 0.8455, + "num_input_tokens_seen": 111654700, + "step": 3427 + }, + { + "epoch": 0.3091491184560581, + "flos": 19253134535520.0, + "grad_norm": 1.5680400290488712, + "learning_rate": 3.2368397458806573e-06, + "loss": 0.6719, + "num_input_tokens_seen": 111683215, + "step": 3428 + }, + { + "epoch": 0.3092393019795283, + "flos": 29928222573600.0, + "grad_norm": 2.08194753968824, + "learning_rate": 3.2363805924475412e-06, + "loss": 0.7973, + "num_input_tokens_seen": 111709290, + "step": 3429 + }, + { + "epoch": 0.3093294855029986, + "flos": 25659050843520.0, + "grad_norm": 2.3144834106009005, + "learning_rate": 3.2359213335196153e-06, + "loss": 0.7027, + "num_input_tokens_seen": 111736635, + "step": 3430 + }, + { + "epoch": 0.3094196690264689, + "flos": 21402888529440.0, + "grad_norm": 1.845125687305929, + "learning_rate": 3.2354619691360663e-06, + "loss": 0.7193, + "num_input_tokens_seen": 111766020, + "step": 3431 + }, + { + "epoch": 0.3095098525499391, + "flos": 26540260440960.0, + "grad_norm": 1.778165790646437, + "learning_rate": 3.2350024993360898e-06, + "loss": 0.8012, + "num_input_tokens_seen": 111796230, + "step": 3432 + }, + { + "epoch": 0.3096000360734094, + "flos": 21512891050080.0, + "grad_norm": 1.9159343052832618, + "learning_rate": 3.2345429241588902e-06, + "loss": 0.7643, + "num_input_tokens_seen": 111826335, + "step": 3433 + }, + { + "epoch": 0.30969021959687965, + "flos": 23844173650080.0, + "grad_norm": 2.017782904396229, + "learning_rate": 3.234083243643681e-06, + "loss": 0.7883, + "num_input_tokens_seen": 111853605, + "step": 3434 + }, + { + "epoch": 0.30978040312034993, + "flos": 18597914310720.0, + "grad_norm": 1.6056119015026857, + "learning_rate": 3.233623457829686e-06, + "loss": 0.8337, + "num_input_tokens_seen": 111879150, + "step": 3435 + }, + { + "epoch": 0.30987058664382017, + "flos": 22788274682400.0, + "grad_norm": 2.011067405363937, + "learning_rate": 3.2331635667561344e-06, + "loss": 0.7315, + "num_input_tokens_seen": 111907480, + "step": 3436 + }, + { + "epoch": 0.30996077016729046, + "flos": 25447483337760.0, + "grad_norm": 1.6853300200431784, + "learning_rate": 3.2327035704622674e-06, + "loss": 0.7953, + "num_input_tokens_seen": 111934770, + "step": 3437 + }, + { + "epoch": 0.3100509536907607, + "flos": 33682029650400.0, + "grad_norm": 1.4365576267387976, + "learning_rate": 3.2322434689873353e-06, + "loss": 0.6613, + "num_input_tokens_seen": 111968580, + "step": 3438 + }, + { + "epoch": 0.310141137214231, + "flos": 20048544859680.0, + "grad_norm": 2.0011322869415924, + "learning_rate": 3.2317832623705957e-06, + "loss": 0.8438, + "num_input_tokens_seen": 111995000, + "step": 3439 + }, + { + "epoch": 0.3102313207377012, + "flos": 50403136213920.0, + "grad_norm": 1.9818605126146445, + "learning_rate": 3.231322950651316e-06, + "loss": 0.7092, + "num_input_tokens_seen": 112027145, + "step": 3440 + }, + { + "epoch": 0.3103215042611715, + "flos": 24096782304480.0, + "grad_norm": 1.4815088810417956, + "learning_rate": 3.2308625338687735e-06, + "loss": 0.7905, + "num_input_tokens_seen": 112057565, + "step": 3441 + }, + { + "epoch": 0.31041168778464173, + "flos": 12250700520960.0, + "grad_norm": 2.498908758420394, + "learning_rate": 3.230402012062252e-06, + "loss": 0.8644, + "num_input_tokens_seen": 112078890, + "step": 3442 + }, + { + "epoch": 0.310501871308112, + "flos": 66401170539840.0, + "grad_norm": 0.6117467208839896, + "learning_rate": 3.2299413852710466e-06, + "loss": 0.6131, + "num_input_tokens_seen": 112174140, + "step": 3443 + }, + { + "epoch": 0.31059205483158225, + "flos": 18671336053920.0, + "grad_norm": 2.2028677499341516, + "learning_rate": 3.2294806535344606e-06, + "loss": 0.7198, + "num_input_tokens_seen": 112201685, + "step": 3444 + }, + { + "epoch": 0.31068223835505254, + "flos": 23662199006880.0, + "grad_norm": 1.9100868508929383, + "learning_rate": 3.2290198168918056e-06, + "loss": 0.8034, + "num_input_tokens_seen": 112230215, + "step": 3445 + }, + { + "epoch": 0.3107724218785228, + "flos": 22204952240640.0, + "grad_norm": 1.7696477249582272, + "learning_rate": 3.2285588753824035e-06, + "loss": 0.8008, + "num_input_tokens_seen": 112258395, + "step": 3446 + }, + { + "epoch": 0.31086260540199306, + "flos": 17283608206080.0, + "grad_norm": 1.8127426774247042, + "learning_rate": 3.228097829045584e-06, + "loss": 0.7767, + "num_input_tokens_seen": 112284925, + "step": 3447 + }, + { + "epoch": 0.3109527889254633, + "flos": 25482837513120.0, + "grad_norm": 1.8047176870488357, + "learning_rate": 3.227636677920685e-06, + "loss": 0.8046, + "num_input_tokens_seen": 112316215, + "step": 3448 + }, + { + "epoch": 0.3110429724489336, + "flos": 23844954215040.0, + "grad_norm": 1.7235476032408292, + "learning_rate": 3.2271754220470567e-06, + "loss": 0.8101, + "num_input_tokens_seen": 112344380, + "step": 3449 + }, + { + "epoch": 0.3111331559724038, + "flos": 16003764202560.0, + "grad_norm": 3.2856989173055866, + "learning_rate": 3.2267140614640547e-06, + "loss": 0.8452, + "num_input_tokens_seen": 112368610, + "step": 3450 + }, + { + "epoch": 0.3112233394958741, + "flos": 22532729616960.0, + "grad_norm": 2.145821383051018, + "learning_rate": 3.2262525962110445e-06, + "loss": 0.8081, + "num_input_tokens_seen": 112397450, + "step": 3451 + }, + { + "epoch": 0.31131352301934434, + "flos": 23880977446080.0, + "grad_norm": 1.9770643479213146, + "learning_rate": 3.2257910263274015e-06, + "loss": 0.691, + "num_input_tokens_seen": 112425880, + "step": 3452 + }, + { + "epoch": 0.31140370654281463, + "flos": 24094663628160.0, + "grad_norm": 2.183829798078389, + "learning_rate": 3.225329351852509e-06, + "loss": 0.7117, + "num_input_tokens_seen": 112454495, + "step": 3453 + }, + { + "epoch": 0.3114938900662849, + "flos": 23042964843360.0, + "grad_norm": 1.8460592905154782, + "learning_rate": 3.2248675728257596e-06, + "loss": 0.7684, + "num_input_tokens_seen": 112483970, + "step": 3454 + }, + { + "epoch": 0.31158407358975515, + "flos": 25914633078720.0, + "grad_norm": 1.649480552981916, + "learning_rate": 3.2244056892865557e-06, + "loss": 0.8386, + "num_input_tokens_seen": 112512685, + "step": 3455 + }, + { + "epoch": 0.31167425711322544, + "flos": 29196235327200.0, + "grad_norm": 1.5590927357818394, + "learning_rate": 3.2239437012743063e-06, + "loss": 0.7551, + "num_input_tokens_seen": 112540850, + "step": 3456 + }, + { + "epoch": 0.31176444063669567, + "flos": 17869792719360.0, + "grad_norm": 1.5646281436170264, + "learning_rate": 3.223481608828432e-06, + "loss": 0.7805, + "num_input_tokens_seen": 112568755, + "step": 3457 + }, + { + "epoch": 0.31185462416016596, + "flos": 17832431376960.0, + "grad_norm": 2.139948382958354, + "learning_rate": 3.223019411988361e-06, + "loss": 0.7799, + "num_input_tokens_seen": 112596620, + "step": 3458 + }, + { + "epoch": 0.3119448076836362, + "flos": 40056383098560.0, + "grad_norm": 1.8119741532038094, + "learning_rate": 3.22255711079353e-06, + "loss": 0.7642, + "num_input_tokens_seen": 112629630, + "step": 3459 + }, + { + "epoch": 0.3120349912071065, + "flos": 20853582151680.0, + "grad_norm": 1.7379650042602484, + "learning_rate": 3.222094705283385e-06, + "loss": 0.7269, + "num_input_tokens_seen": 112657420, + "step": 3460 + }, + { + "epoch": 0.3121251747305767, + "flos": 28281901740000.0, + "grad_norm": 1.8283731015081361, + "learning_rate": 3.2216321954973805e-06, + "loss": 0.7599, + "num_input_tokens_seen": 112682700, + "step": 3461 + }, + { + "epoch": 0.312215358254047, + "flos": 23297766513600.0, + "grad_norm": 3.3974111125173025, + "learning_rate": 3.2211695814749816e-06, + "loss": 0.7514, + "num_input_tokens_seen": 112713950, + "step": 3462 + }, + { + "epoch": 0.31230554177751724, + "flos": 15676470033120.0, + "grad_norm": 4.518251493400649, + "learning_rate": 3.220706863255661e-06, + "loss": 0.8695, + "num_input_tokens_seen": 112736145, + "step": 3463 + }, + { + "epoch": 0.3123957253009875, + "flos": 21877286373600.0, + "grad_norm": 3.0202386380324424, + "learning_rate": 3.2202440408788994e-06, + "loss": 0.7638, + "num_input_tokens_seen": 112764045, + "step": 3464 + }, + { + "epoch": 0.31248590882445776, + "flos": 18160803469440.0, + "grad_norm": 1.7570833855709544, + "learning_rate": 3.2197811143841883e-06, + "loss": 0.8359, + "num_input_tokens_seen": 112790135, + "step": 3465 + }, + { + "epoch": 0.31257609234792805, + "flos": 21689810605920.0, + "grad_norm": 1.6530747115737623, + "learning_rate": 3.2193180838110267e-06, + "loss": 0.762, + "num_input_tokens_seen": 112818570, + "step": 3466 + }, + { + "epoch": 0.3126662758713983, + "flos": 21691929282240.0, + "grad_norm": 2.30224800959292, + "learning_rate": 3.2188549491989225e-06, + "loss": 0.6602, + "num_input_tokens_seen": 112845340, + "step": 3467 + }, + { + "epoch": 0.31275645939486857, + "flos": 22605333625440.0, + "grad_norm": 2.2578513594190537, + "learning_rate": 3.2183917105873934e-06, + "loss": 0.7905, + "num_input_tokens_seen": 112873590, + "step": 3468 + }, + { + "epoch": 0.3128466429183388, + "flos": 22134689927040.0, + "grad_norm": 1.9558155296413637, + "learning_rate": 3.217928368015966e-06, + "loss": 0.8297, + "num_input_tokens_seen": 112906045, + "step": 3469 + }, + { + "epoch": 0.3129368264418091, + "flos": 23188730406720.0, + "grad_norm": 1.4310917029650938, + "learning_rate": 3.217464921524174e-06, + "loss": 0.794, + "num_input_tokens_seen": 112936380, + "step": 3470 + }, + { + "epoch": 0.3130270099652793, + "flos": 22456148444160.0, + "grad_norm": 4.853814086333796, + "learning_rate": 3.2170013711515616e-06, + "loss": 0.7793, + "num_input_tokens_seen": 112962270, + "step": 3471 + }, + { + "epoch": 0.3131171934887496, + "flos": 28361976870240.0, + "grad_norm": 2.983601734583205, + "learning_rate": 3.216537716937682e-06, + "loss": 0.6886, + "num_input_tokens_seen": 112993380, + "step": 3472 + }, + { + "epoch": 0.31320737701221985, + "flos": 23626882001280.0, + "grad_norm": 2.2503221069351245, + "learning_rate": 3.2160739589220968e-06, + "loss": 0.7411, + "num_input_tokens_seen": 113021675, + "step": 3473 + }, + { + "epoch": 0.31329756053569013, + "flos": 21987363233760.0, + "grad_norm": 2.036606753144938, + "learning_rate": 3.215610097144376e-06, + "loss": 0.824, + "num_input_tokens_seen": 113051305, + "step": 3474 + }, + { + "epoch": 0.31338774405916037, + "flos": 23698668275040.0, + "grad_norm": 2.006970893424522, + "learning_rate": 3.215146131644099e-06, + "loss": 0.7181, + "num_input_tokens_seen": 113079695, + "step": 3475 + }, + { + "epoch": 0.31347792758263066, + "flos": 19867610969760.0, + "grad_norm": 1.5385178120114784, + "learning_rate": 3.214682062460854e-06, + "loss": 0.8354, + "num_input_tokens_seen": 113107580, + "step": 3476 + }, + { + "epoch": 0.3135681111061009, + "flos": 23479964175360.0, + "grad_norm": 2.1858365955199974, + "learning_rate": 3.2142178896342367e-06, + "loss": 0.6818, + "num_input_tokens_seen": 113136300, + "step": 3477 + }, + { + "epoch": 0.3136582946295712, + "flos": 36337595669280.0, + "grad_norm": 1.60422395215817, + "learning_rate": 3.2137536132038552e-06, + "loss": 0.7528, + "num_input_tokens_seen": 113168390, + "step": 3478 + }, + { + "epoch": 0.31374847815304147, + "flos": 19866718895520.0, + "grad_norm": 1.7014734502352566, + "learning_rate": 3.2132892332093226e-06, + "loss": 0.7725, + "num_input_tokens_seen": 113194320, + "step": 3479 + }, + { + "epoch": 0.3138386616765117, + "flos": 26828371949760.0, + "grad_norm": 1.615783007806822, + "learning_rate": 3.2128247496902623e-06, + "loss": 0.8753, + "num_input_tokens_seen": 113224780, + "step": 3480 + }, + { + "epoch": 0.313928845199982, + "flos": 24572258071680.0, + "grad_norm": 1.4056854386352062, + "learning_rate": 3.2123601626863064e-06, + "loss": 0.7767, + "num_input_tokens_seen": 113257860, + "step": 3481 + }, + { + "epoch": 0.3140190287234522, + "flos": 64946525656800.0, + "grad_norm": 0.6359824621628541, + "learning_rate": 3.2118954722370974e-06, + "loss": 0.5705, + "num_input_tokens_seen": 113352535, + "step": 3482 + }, + { + "epoch": 0.3141092122469225, + "flos": 22241756036640.0, + "grad_norm": 1.903961872075697, + "learning_rate": 3.2114306783822837e-06, + "loss": 0.7102, + "num_input_tokens_seen": 113380060, + "step": 3483 + }, + { + "epoch": 0.31419939577039274, + "flos": 26829301193760.0, + "grad_norm": 1.7294725430976032, + "learning_rate": 3.210965781161525e-06, + "loss": 0.7442, + "num_input_tokens_seen": 113410905, + "step": 3484 + }, + { + "epoch": 0.31428957929386303, + "flos": 25739051634240.0, + "grad_norm": 1.8747784801598322, + "learning_rate": 3.2105007806144892e-06, + "loss": 0.8806, + "num_input_tokens_seen": 113438755, + "step": 3485 + }, + { + "epoch": 0.31437976281733326, + "flos": 23079694299840.0, + "grad_norm": 2.720742208250779, + "learning_rate": 3.2100356767808513e-06, + "loss": 0.7677, + "num_input_tokens_seen": 113468120, + "step": 3486 + }, + { + "epoch": 0.31446994634080355, + "flos": 20383681848480.0, + "grad_norm": 2.0195450625509874, + "learning_rate": 3.2095704697002977e-06, + "loss": 0.82, + "num_input_tokens_seen": 113493670, + "step": 3487 + }, + { + "epoch": 0.3145601298642738, + "flos": 21076151906400.0, + "grad_norm": 1.857481625595704, + "learning_rate": 3.209105159412522e-06, + "loss": 0.7651, + "num_input_tokens_seen": 113523385, + "step": 3488 + }, + { + "epoch": 0.3146503133877441, + "flos": 20960871279840.0, + "grad_norm": 2.2663939124911954, + "learning_rate": 3.208639745957228e-06, + "loss": 0.7273, + "num_input_tokens_seen": 113549440, + "step": 3489 + }, + { + "epoch": 0.3147404969112143, + "flos": 25593360410400.0, + "grad_norm": 1.4348324613429309, + "learning_rate": 3.2081742293741256e-06, + "loss": 0.7934, + "num_input_tokens_seen": 113581475, + "step": 3490 + }, + { + "epoch": 0.3148306804346846, + "flos": 22387484430240.0, + "grad_norm": 1.4434625281612572, + "learning_rate": 3.2077086097029366e-06, + "loss": 0.8296, + "num_input_tokens_seen": 113610565, + "step": 3491 + }, + { + "epoch": 0.31492086395815483, + "flos": 21148495726560.0, + "grad_norm": 1.8453711155586519, + "learning_rate": 3.2072428869833895e-06, + "loss": 0.7653, + "num_input_tokens_seen": 113639060, + "step": 3492 + }, + { + "epoch": 0.3150110474816251, + "flos": 23772238697280.0, + "grad_norm": 1.5237566491373895, + "learning_rate": 3.206777061255223e-06, + "loss": 0.7783, + "num_input_tokens_seen": 113668850, + "step": 3493 + }, + { + "epoch": 0.31510123100509535, + "flos": 24317567910720.0, + "grad_norm": 3.189177576555926, + "learning_rate": 3.206311132558183e-06, + "loss": 0.8097, + "num_input_tokens_seen": 113696790, + "step": 3494 + }, + { + "epoch": 0.31519141452856564, + "flos": 17899720109760.0, + "grad_norm": 1.8540100145779346, + "learning_rate": 3.205845100932026e-06, + "loss": 0.7814, + "num_input_tokens_seen": 113722725, + "step": 3495 + }, + { + "epoch": 0.31528159805203587, + "flos": 24901410729120.0, + "grad_norm": 1.6509510420547664, + "learning_rate": 3.205378966416516e-06, + "loss": 0.7914, + "num_input_tokens_seen": 113752895, + "step": 3496 + }, + { + "epoch": 0.31537178157550616, + "flos": 22969877628000.0, + "grad_norm": 1.6957355446079156, + "learning_rate": 3.204912729051426e-06, + "loss": 0.7637, + "num_input_tokens_seen": 113780845, + "step": 3497 + }, + { + "epoch": 0.3154619650989764, + "flos": 19800768274080.0, + "grad_norm": 1.95439460688443, + "learning_rate": 3.2044463888765384e-06, + "loss": 0.741, + "num_input_tokens_seen": 113809370, + "step": 3498 + }, + { + "epoch": 0.3155521486224467, + "flos": 38124775657920.0, + "grad_norm": 1.8021776986818872, + "learning_rate": 3.2039799459316436e-06, + "loss": 0.6622, + "num_input_tokens_seen": 113839370, + "step": 3499 + }, + { + "epoch": 0.3156423321459169, + "flos": 25445327491680.0, + "grad_norm": 5.666902574602731, + "learning_rate": 3.2035134002565407e-06, + "loss": 0.8018, + "num_input_tokens_seen": 113870200, + "step": 3500 + }, + { + "epoch": 0.3157325156693872, + "flos": 14954927489280.0, + "grad_norm": 1.8597235473154539, + "learning_rate": 3.203046751891039e-06, + "loss": 0.7765, + "num_input_tokens_seen": 113897030, + "step": 3501 + }, + { + "epoch": 0.3158226991928575, + "flos": 22604813248800.0, + "grad_norm": 1.7299592449922359, + "learning_rate": 3.2025800008749545e-06, + "loss": 0.7798, + "num_input_tokens_seen": 113921555, + "step": 3502 + }, + { + "epoch": 0.3159128827163277, + "flos": 23549185735680.0, + "grad_norm": 1.58385477347228, + "learning_rate": 3.202113147248114e-06, + "loss": 0.7408, + "num_input_tokens_seen": 113950745, + "step": 3503 + }, + { + "epoch": 0.316003066239798, + "flos": 24132508177440.0, + "grad_norm": 1.7445525124087369, + "learning_rate": 3.20164619105035e-06, + "loss": 0.7411, + "num_input_tokens_seen": 113979095, + "step": 3504 + }, + { + "epoch": 0.31609324976326825, + "flos": 23516842310880.0, + "grad_norm": 1.4769226119494623, + "learning_rate": 3.201179132321508e-06, + "loss": 0.621, + "num_input_tokens_seen": 114009790, + "step": 3505 + }, + { + "epoch": 0.31618343328673854, + "flos": 16521582060000.0, + "grad_norm": 2.021824138302321, + "learning_rate": 3.200711971101439e-06, + "loss": 0.7109, + "num_input_tokens_seen": 114036615, + "step": 3506 + }, + { + "epoch": 0.31627361681020877, + "flos": 18227051448960.0, + "grad_norm": 2.0168235822849123, + "learning_rate": 3.2002447074300047e-06, + "loss": 0.7674, + "num_input_tokens_seen": 114064050, + "step": 3507 + }, + { + "epoch": 0.31636380033367906, + "flos": 22023683822880.0, + "grad_norm": 2.2531357213372365, + "learning_rate": 3.1997773413470736e-06, + "loss": 0.7728, + "num_input_tokens_seen": 114093815, + "step": 3508 + }, + { + "epoch": 0.3164539838571493, + "flos": 23735211882720.0, + "grad_norm": 2.7015505047471735, + "learning_rate": 3.199309872892524e-06, + "loss": 0.7594, + "num_input_tokens_seen": 114123275, + "step": 3509 + }, + { + "epoch": 0.3165441673806196, + "flos": 24822041824320.0, + "grad_norm": 1.7930917469131078, + "learning_rate": 3.198842302106243e-06, + "loss": 0.8195, + "num_input_tokens_seen": 114150985, + "step": 3510 + }, + { + "epoch": 0.3166343509040898, + "flos": 19436038422720.0, + "grad_norm": 1.8761126078216022, + "learning_rate": 3.1983746290281265e-06, + "loss": 0.752, + "num_input_tokens_seen": 114178680, + "step": 3511 + }, + { + "epoch": 0.3167245344275601, + "flos": 33676342677120.0, + "grad_norm": 1.4064899663692723, + "learning_rate": 3.197906853698079e-06, + "loss": 0.7748, + "num_input_tokens_seen": 114207315, + "step": 3512 + }, + { + "epoch": 0.31681471795103033, + "flos": 39872512797600.0, + "grad_norm": 1.7722210850540918, + "learning_rate": 3.1974389761560137e-06, + "loss": 0.7166, + "num_input_tokens_seen": 114237840, + "step": 3513 + }, + { + "epoch": 0.3169049014745006, + "flos": 21731297791680.0, + "grad_norm": 1.5995112984089213, + "learning_rate": 3.1969709964418525e-06, + "loss": 0.7163, + "num_input_tokens_seen": 114264090, + "step": 3514 + }, + { + "epoch": 0.31699508499797086, + "flos": 24609024697920.0, + "grad_norm": 1.7492204701056122, + "learning_rate": 3.196502914595525e-06, + "loss": 0.7903, + "num_input_tokens_seen": 114294380, + "step": 3515 + }, + { + "epoch": 0.31708526852144114, + "flos": 17650939940640.0, + "grad_norm": 1.8016870169441601, + "learning_rate": 3.1960347306569723e-06, + "loss": 0.788, + "num_input_tokens_seen": 114320270, + "step": 3516 + }, + { + "epoch": 0.3171754520449114, + "flos": 15282742035360.0, + "grad_norm": 1.8956921353700558, + "learning_rate": 3.195566444666141e-06, + "loss": 0.7202, + "num_input_tokens_seen": 114346135, + "step": 3517 + }, + { + "epoch": 0.31726563556838167, + "flos": 23114044891680.0, + "grad_norm": 1.6850659512027872, + "learning_rate": 3.1950980566629886e-06, + "loss": 0.8752, + "num_input_tokens_seen": 114374320, + "step": 3518 + }, + { + "epoch": 0.3173558190918519, + "flos": 24098789471520.0, + "grad_norm": 1.5549645105218781, + "learning_rate": 3.1946295666874797e-06, + "loss": 0.7823, + "num_input_tokens_seen": 114405100, + "step": 3519 + }, + { + "epoch": 0.3174460026153222, + "flos": 22385254244640.0, + "grad_norm": 1.8467796133230263, + "learning_rate": 3.19416097477959e-06, + "loss": 0.7875, + "num_input_tokens_seen": 114432410, + "step": 3520 + }, + { + "epoch": 0.3175361861387924, + "flos": 24278236571040.0, + "grad_norm": 1.4159859068810388, + "learning_rate": 3.1936922809793005e-06, + "loss": 0.8251, + "num_input_tokens_seen": 114462085, + "step": 3521 + }, + { + "epoch": 0.3176263696622627, + "flos": 25661801405760.0, + "grad_norm": 1.5119300714297452, + "learning_rate": 3.193223485326604e-06, + "loss": 0.8874, + "num_input_tokens_seen": 114491130, + "step": 3522 + }, + { + "epoch": 0.31771655318573294, + "flos": 17868343098720.0, + "grad_norm": 2.0327629910186356, + "learning_rate": 3.1927545878615005e-06, + "loss": 0.8586, + "num_input_tokens_seen": 114516050, + "step": 3523 + }, + { + "epoch": 0.31780673670920323, + "flos": 21581889591840.0, + "grad_norm": 2.179291938793926, + "learning_rate": 3.192285588623999e-06, + "loss": 0.7196, + "num_input_tokens_seen": 114542430, + "step": 3524 + }, + { + "epoch": 0.31789692023267346, + "flos": 31348107997440.0, + "grad_norm": 1.440563172443138, + "learning_rate": 3.191816487654117e-06, + "loss": 0.6989, + "num_input_tokens_seen": 114574125, + "step": 3525 + }, + { + "epoch": 0.31798710375614375, + "flos": 15537841063680.0, + "grad_norm": 2.5778102798258127, + "learning_rate": 3.19134728499188e-06, + "loss": 0.732, + "num_input_tokens_seen": 114601445, + "step": 3526 + }, + { + "epoch": 0.31807728727961404, + "flos": 25921063447200.0, + "grad_norm": 1.7356110713126478, + "learning_rate": 3.1908779806773235e-06, + "loss": 0.6615, + "num_input_tokens_seen": 114628645, + "step": 3527 + }, + { + "epoch": 0.3181674708030843, + "flos": 28360267061280.0, + "grad_norm": 1.6977739176587021, + "learning_rate": 3.190408574750492e-06, + "loss": 0.8247, + "num_input_tokens_seen": 114657565, + "step": 3528 + }, + { + "epoch": 0.31825765432655456, + "flos": 28761168822720.0, + "grad_norm": 1.5495255236276082, + "learning_rate": 3.1899390672514367e-06, + "loss": 0.8388, + "num_input_tokens_seen": 114688160, + "step": 3529 + }, + { + "epoch": 0.3183478378500248, + "flos": 43003443074400.0, + "grad_norm": 2.260370725759625, + "learning_rate": 3.189469458220219e-06, + "loss": 0.6927, + "num_input_tokens_seen": 114718190, + "step": 3530 + }, + { + "epoch": 0.3184380213734951, + "flos": 24279574682400.0, + "grad_norm": 1.410701403291734, + "learning_rate": 3.1889997476969086e-06, + "loss": 0.7696, + "num_input_tokens_seen": 114749055, + "step": 3531 + }, + { + "epoch": 0.3185282048969653, + "flos": 23730677172000.0, + "grad_norm": 1.5632306488708612, + "learning_rate": 3.188529935721583e-06, + "loss": 0.7588, + "num_input_tokens_seen": 114777510, + "step": 3532 + }, + { + "epoch": 0.3186183884204356, + "flos": 25666113097920.0, + "grad_norm": 1.6230625193148336, + "learning_rate": 3.18806002233433e-06, + "loss": 0.8308, + "num_input_tokens_seen": 114807230, + "step": 3533 + }, + { + "epoch": 0.31870857194390584, + "flos": 17141819807040.0, + "grad_norm": 1.6854397241298564, + "learning_rate": 3.187590007575245e-06, + "loss": 0.8332, + "num_input_tokens_seen": 114834145, + "step": 3534 + }, + { + "epoch": 0.3187987554673761, + "flos": 34185834508320.0, + "grad_norm": 7.9373712003575605, + "learning_rate": 3.1871198914844327e-06, + "loss": 0.7072, + "num_input_tokens_seen": 114862765, + "step": 3535 + }, + { + "epoch": 0.31888893899084636, + "flos": 20746367363040.0, + "grad_norm": 1.946482540147358, + "learning_rate": 3.1866496741020057e-06, + "loss": 0.7044, + "num_input_tokens_seen": 114891220, + "step": 3536 + }, + { + "epoch": 0.31897912251431665, + "flos": 35613599921280.0, + "grad_norm": 1.3175812862300638, + "learning_rate": 3.186179355468085e-06, + "loss": 0.7382, + "num_input_tokens_seen": 114927910, + "step": 3537 + }, + { + "epoch": 0.3190693060377869, + "flos": 24420768365280.0, + "grad_norm": 3.381104494020749, + "learning_rate": 3.1857089356228015e-06, + "loss": 0.7594, + "num_input_tokens_seen": 114954465, + "step": 3538 + }, + { + "epoch": 0.31915948956125717, + "flos": 19545632076000.0, + "grad_norm": 1.5263147033922932, + "learning_rate": 3.1852384146062933e-06, + "loss": 0.7155, + "num_input_tokens_seen": 114983110, + "step": 3539 + }, + { + "epoch": 0.3192496730847274, + "flos": 70885812600480.0, + "grad_norm": 0.6279908383317712, + "learning_rate": 3.184767792458708e-06, + "loss": 0.5579, + "num_input_tokens_seen": 115083385, + "step": 3540 + }, + { + "epoch": 0.3193398566081977, + "flos": 22236923967840.0, + "grad_norm": 1.8129960400790077, + "learning_rate": 3.1842970692202023e-06, + "loss": 0.897, + "num_input_tokens_seen": 115111715, + "step": 3541 + }, + { + "epoch": 0.3194300401316679, + "flos": 23844247989600.0, + "grad_norm": 1.7127927710878144, + "learning_rate": 3.1838262449309403e-06, + "loss": 0.7491, + "num_input_tokens_seen": 115139330, + "step": 3542 + }, + { + "epoch": 0.3195202236551382, + "flos": 23115680361120.0, + "grad_norm": 2.0099410711731385, + "learning_rate": 3.1833553196310956e-06, + "loss": 0.6877, + "num_input_tokens_seen": 115167220, + "step": 3543 + }, + { + "epoch": 0.31961040717860845, + "flos": 23770380209280.0, + "grad_norm": 1.6160179737526743, + "learning_rate": 3.18288429336085e-06, + "loss": 0.6286, + "num_input_tokens_seen": 115195235, + "step": 3544 + }, + { + "epoch": 0.31970059070207874, + "flos": 23145942279360.0, + "grad_norm": 1.9511889207799908, + "learning_rate": 3.182413166160394e-06, + "loss": 0.874, + "num_input_tokens_seen": 115222175, + "step": 3545 + }, + { + "epoch": 0.31979077422554897, + "flos": 21002581484160.0, + "grad_norm": 1.8650742706769057, + "learning_rate": 3.1819419380699275e-06, + "loss": 0.8059, + "num_input_tokens_seen": 115250055, + "step": 3546 + }, + { + "epoch": 0.31988095774901926, + "flos": 26283525943200.0, + "grad_norm": 2.010492814224558, + "learning_rate": 3.181470609129658e-06, + "loss": 0.7768, + "num_input_tokens_seen": 115278925, + "step": 3547 + }, + { + "epoch": 0.3199711412724895, + "flos": 18051395664960.0, + "grad_norm": 2.6014900082864263, + "learning_rate": 3.1809991793798e-06, + "loss": 0.8143, + "num_input_tokens_seen": 115303655, + "step": 3548 + }, + { + "epoch": 0.3200613247959598, + "flos": 24859254487680.0, + "grad_norm": 1.937685127027742, + "learning_rate": 3.1805276488605806e-06, + "loss": 0.7252, + "num_input_tokens_seen": 115327930, + "step": 3549 + }, + { + "epoch": 0.32015150831943007, + "flos": 18885877140480.0, + "grad_norm": 1.5632615331715132, + "learning_rate": 3.1800560176122336e-06, + "loss": 0.7607, + "num_input_tokens_seen": 115354715, + "step": 3550 + }, + { + "epoch": 0.3202416918429003, + "flos": 21440324211360.0, + "grad_norm": 2.9828242721844864, + "learning_rate": 3.179584285675e-06, + "loss": 0.7728, + "num_input_tokens_seen": 115382660, + "step": 3551 + }, + { + "epoch": 0.3203318753663706, + "flos": 25551166999200.0, + "grad_norm": 2.5845074805330226, + "learning_rate": 3.1791124530891315e-06, + "loss": 0.6868, + "num_input_tokens_seen": 115410670, + "step": 3552 + }, + { + "epoch": 0.3204220588898408, + "flos": 19836494147040.0, + "grad_norm": 1.7353684612904692, + "learning_rate": 3.178640519894886e-06, + "loss": 0.7697, + "num_input_tokens_seen": 115438665, + "step": 3553 + }, + { + "epoch": 0.3205122424133111, + "flos": 33754336300800.0, + "grad_norm": 1.7286450008962324, + "learning_rate": 3.1781684861325324e-06, + "loss": 0.7269, + "num_input_tokens_seen": 115469595, + "step": 3554 + }, + { + "epoch": 0.32060242593678134, + "flos": 37761681276000.0, + "grad_norm": 1.8884703914309917, + "learning_rate": 3.177696351842348e-06, + "loss": 0.7766, + "num_input_tokens_seen": 115500870, + "step": 3555 + }, + { + "epoch": 0.32069260946025163, + "flos": 70969827725280.0, + "grad_norm": 0.7024766263772655, + "learning_rate": 3.1772241170646167e-06, + "loss": 0.5601, + "num_input_tokens_seen": 115598935, + "step": 3556 + }, + { + "epoch": 0.32078279298372187, + "flos": 14772209450880.0, + "grad_norm": 15.912133228662924, + "learning_rate": 3.1767517818396334e-06, + "loss": 0.8731, + "num_input_tokens_seen": 115624265, + "step": 3557 + }, + { + "epoch": 0.32087297650719215, + "flos": 20814176472480.0, + "grad_norm": 1.5223641551381881, + "learning_rate": 3.1762793462076986e-06, + "loss": 0.925, + "num_input_tokens_seen": 115652080, + "step": 3558 + }, + { + "epoch": 0.3209631600306624, + "flos": 17359780511520.0, + "grad_norm": 3.5072536235695164, + "learning_rate": 3.1758068102091236e-06, + "loss": 0.6516, + "num_input_tokens_seen": 115679820, + "step": 3559 + }, + { + "epoch": 0.3210533435541327, + "flos": 27014658285120.0, + "grad_norm": 1.9212276950353007, + "learning_rate": 3.175334173884229e-06, + "loss": 0.7372, + "num_input_tokens_seen": 115708980, + "step": 3560 + }, + { + "epoch": 0.3211435270776029, + "flos": 24172211214720.0, + "grad_norm": 1.483954761648461, + "learning_rate": 3.174861437273342e-06, + "loss": 0.769, + "num_input_tokens_seen": 115737920, + "step": 3561 + }, + { + "epoch": 0.3212337106010732, + "flos": 30620766971040.0, + "grad_norm": 2.029683609533656, + "learning_rate": 3.174388600416799e-06, + "loss": 0.7283, + "num_input_tokens_seen": 115768735, + "step": 3562 + }, + { + "epoch": 0.32132389412454343, + "flos": 69258714266400.0, + "grad_norm": 0.6573855248939917, + "learning_rate": 3.1739156633549445e-06, + "loss": 0.5699, + "num_input_tokens_seen": 115858805, + "step": 3563 + }, + { + "epoch": 0.3214140776480137, + "flos": 21216267666240.0, + "grad_norm": 1.8276287611702264, + "learning_rate": 3.173442626128133e-06, + "loss": 0.8319, + "num_input_tokens_seen": 115886790, + "step": 3564 + }, + { + "epoch": 0.32150426117148395, + "flos": 14954369942880.0, + "grad_norm": 3.7057182149218146, + "learning_rate": 3.1729694887767265e-06, + "loss": 0.8781, + "num_input_tokens_seen": 115907785, + "step": 3565 + }, + { + "epoch": 0.32159444469495424, + "flos": 19283991169920.0, + "grad_norm": 1.7403750721304632, + "learning_rate": 3.172496251341096e-06, + "loss": 0.7697, + "num_input_tokens_seen": 115933310, + "step": 3566 + }, + { + "epoch": 0.3216846282184245, + "flos": 31532461505280.0, + "grad_norm": 2.009913907081444, + "learning_rate": 3.172022913861619e-06, + "loss": 0.6233, + "num_input_tokens_seen": 115965495, + "step": 3567 + }, + { + "epoch": 0.32177481174189476, + "flos": 23844508177920.0, + "grad_norm": 1.83918319474632, + "learning_rate": 3.171549476378686e-06, + "loss": 0.7219, + "num_input_tokens_seen": 115994960, + "step": 3568 + }, + { + "epoch": 0.321864995265365, + "flos": 21112435325760.0, + "grad_norm": 1.6955001656867723, + "learning_rate": 3.1710759389326906e-06, + "loss": 0.7346, + "num_input_tokens_seen": 116025195, + "step": 3569 + }, + { + "epoch": 0.3219551787888353, + "flos": 22642434779520.0, + "grad_norm": 3.630917290630417, + "learning_rate": 3.1706023015640396e-06, + "loss": 0.7612, + "num_input_tokens_seen": 116052020, + "step": 3570 + }, + { + "epoch": 0.3220453623123055, + "flos": 27778988956320.0, + "grad_norm": 1.977478357961023, + "learning_rate": 3.1701285643131453e-06, + "loss": 0.7882, + "num_input_tokens_seen": 116080155, + "step": 3571 + }, + { + "epoch": 0.3221355458357758, + "flos": 20164903409280.0, + "grad_norm": 2.118448255089316, + "learning_rate": 3.16965472722043e-06, + "loss": 0.7767, + "num_input_tokens_seen": 116107160, + "step": 3572 + }, + { + "epoch": 0.32222572935924604, + "flos": 19071308571360.0, + "grad_norm": 2.2200996699598115, + "learning_rate": 3.169180790326324e-06, + "loss": 0.6897, + "num_input_tokens_seen": 116133435, + "step": 3573 + }, + { + "epoch": 0.3223159128827163, + "flos": 23371225426560.0, + "grad_norm": 1.457929482626734, + "learning_rate": 3.168706753671266e-06, + "loss": 0.7691, + "num_input_tokens_seen": 116161325, + "step": 3574 + }, + { + "epoch": 0.3224060964061866, + "flos": 41659209579360.0, + "grad_norm": 1.6555340497044557, + "learning_rate": 3.168232617295704e-06, + "loss": 0.6909, + "num_input_tokens_seen": 116193920, + "step": 3575 + }, + { + "epoch": 0.32249627992965685, + "flos": 18161175167040.0, + "grad_norm": 1.9565577743270923, + "learning_rate": 3.167758381240093e-06, + "loss": 0.7633, + "num_input_tokens_seen": 116220530, + "step": 3576 + }, + { + "epoch": 0.32258646345312714, + "flos": 25739311822560.0, + "grad_norm": 1.9115246822750303, + "learning_rate": 3.1672840455448978e-06, + "loss": 0.8099, + "num_input_tokens_seen": 116251690, + "step": 3577 + }, + { + "epoch": 0.32267664697659737, + "flos": 22019000433120.0, + "grad_norm": 1.593234633350835, + "learning_rate": 3.166809610250592e-06, + "loss": 0.7097, + "num_input_tokens_seen": 116281385, + "step": 3578 + }, + { + "epoch": 0.32276683050006766, + "flos": 24315635083200.0, + "grad_norm": 1.7026941996991736, + "learning_rate": 3.166335075397656e-06, + "loss": 0.6938, + "num_input_tokens_seen": 116310935, + "step": 3579 + }, + { + "epoch": 0.3228570140235379, + "flos": 23256353667360.0, + "grad_norm": 1.6967139011295855, + "learning_rate": 3.1658604410265808e-06, + "loss": 0.7315, + "num_input_tokens_seen": 116338890, + "step": 3580 + }, + { + "epoch": 0.3229471975470082, + "flos": 72399754717920.0, + "grad_norm": 0.9130831305312804, + "learning_rate": 3.1653857071778644e-06, + "loss": 0.5996, + "num_input_tokens_seen": 116430970, + "step": 3581 + }, + { + "epoch": 0.3230373810704784, + "flos": 19946050630560.0, + "grad_norm": 2.3673169191945003, + "learning_rate": 3.1649108738920133e-06, + "loss": 0.7612, + "num_input_tokens_seen": 116459910, + "step": 3582 + }, + { + "epoch": 0.3231275645939487, + "flos": 33790285192320.0, + "grad_norm": 1.886081311448201, + "learning_rate": 3.1644359412095432e-06, + "loss": 0.7671, + "num_input_tokens_seen": 116490685, + "step": 3583 + }, + { + "epoch": 0.32321774811741893, + "flos": 67084942873920.0, + "grad_norm": 0.7499716276524269, + "learning_rate": 3.163960909170978e-06, + "loss": 0.6093, + "num_input_tokens_seen": 116580970, + "step": 3584 + }, + { + "epoch": 0.3233079316408892, + "flos": 23626101436320.0, + "grad_norm": 1.7788115424677042, + "learning_rate": 3.1634857778168496e-06, + "loss": 0.6649, + "num_input_tokens_seen": 116609810, + "step": 3585 + }, + { + "epoch": 0.32339811516435946, + "flos": 27159048567360.0, + "grad_norm": 1.7197906811451606, + "learning_rate": 3.1630105471877002e-06, + "loss": 0.8025, + "num_input_tokens_seen": 116639670, + "step": 3586 + }, + { + "epoch": 0.32348829868782975, + "flos": 25811804321760.0, + "grad_norm": 1.9999959468800483, + "learning_rate": 3.162535217324077e-06, + "loss": 0.7456, + "num_input_tokens_seen": 116668730, + "step": 3587 + }, + { + "epoch": 0.3235784822113, + "flos": 21359691534720.0, + "grad_norm": 2.505011780153394, + "learning_rate": 3.1620597882665393e-06, + "loss": 0.7961, + "num_input_tokens_seen": 116690065, + "step": 3588 + }, + { + "epoch": 0.32366866573477027, + "flos": 20710976017920.0, + "grad_norm": 1.8516076963678831, + "learning_rate": 3.1615842600556535e-06, + "loss": 0.8257, + "num_input_tokens_seen": 116718785, + "step": 3589 + }, + { + "epoch": 0.3237588492582405, + "flos": 26431484522400.0, + "grad_norm": 1.692059186049063, + "learning_rate": 3.1611086327319932e-06, + "loss": 0.766, + "num_input_tokens_seen": 116747180, + "step": 3590 + }, + { + "epoch": 0.3238490327817108, + "flos": 24387123998880.0, + "grad_norm": 1.6066560462083572, + "learning_rate": 3.160632906336142e-06, + "loss": 0.8355, + "num_input_tokens_seen": 116775025, + "step": 3591 + }, + { + "epoch": 0.323939216305181, + "flos": 24243588621120.0, + "grad_norm": 2.9183401665497724, + "learning_rate": 3.160157080908692e-06, + "loss": 0.7892, + "num_input_tokens_seen": 116804390, + "step": 3592 + }, + { + "epoch": 0.3240293998286513, + "flos": 22894894754880.0, + "grad_norm": 1.6769558007063223, + "learning_rate": 3.1596811564902426e-06, + "loss": 0.8204, + "num_input_tokens_seen": 116835605, + "step": 3593 + }, + { + "epoch": 0.32411958335212154, + "flos": 19582064174400.0, + "grad_norm": 2.6191139532746637, + "learning_rate": 3.1592051331214023e-06, + "loss": 0.7056, + "num_input_tokens_seen": 116862100, + "step": 3594 + }, + { + "epoch": 0.32420976687559183, + "flos": 25590498338880.0, + "grad_norm": 1.9592863023263452, + "learning_rate": 3.158729010842789e-06, + "loss": 0.7874, + "num_input_tokens_seen": 116890930, + "step": 3595 + }, + { + "epoch": 0.32429995039906206, + "flos": 22129932197760.0, + "grad_norm": 2.261561887371238, + "learning_rate": 3.1582527896950266e-06, + "loss": 0.8197, + "num_input_tokens_seen": 116919500, + "step": 3596 + }, + { + "epoch": 0.32439013392253235, + "flos": 24062766240480.0, + "grad_norm": 1.8924755151291017, + "learning_rate": 3.157776469718749e-06, + "loss": 0.7839, + "num_input_tokens_seen": 116945705, + "step": 3597 + }, + { + "epoch": 0.32448031744600264, + "flos": 20019360864480.0, + "grad_norm": 1.80453893537154, + "learning_rate": 3.1573000509546004e-06, + "loss": 0.7568, + "num_input_tokens_seen": 116973570, + "step": 3598 + }, + { + "epoch": 0.3245705009694729, + "flos": 24973717379520.0, + "grad_norm": 1.595788736276333, + "learning_rate": 3.1568235334432296e-06, + "loss": 0.8188, + "num_input_tokens_seen": 116999995, + "step": 3599 + }, + { + "epoch": 0.32466068449294316, + "flos": 23438625668640.0, + "grad_norm": 2.167656183478586, + "learning_rate": 3.1563469172252964e-06, + "loss": 0.835, + "num_input_tokens_seen": 117027010, + "step": 3600 + }, + { + "epoch": 0.3247508680164134, + "flos": 21803976139680.0, + "grad_norm": 2.0467194942196643, + "learning_rate": 3.155870202341468e-06, + "loss": 0.8415, + "num_input_tokens_seen": 117055210, + "step": 3601 + }, + { + "epoch": 0.3248410515398837, + "flos": 25479455064960.0, + "grad_norm": 1.8837256532144933, + "learning_rate": 3.155393388832421e-06, + "loss": 0.763, + "num_input_tokens_seen": 117083135, + "step": 3602 + }, + { + "epoch": 0.3249312350633539, + "flos": 20779231164480.0, + "grad_norm": 3.991533259403621, + "learning_rate": 3.1549164767388386e-06, + "loss": 0.7948, + "num_input_tokens_seen": 117108240, + "step": 3603 + }, + { + "epoch": 0.3250214185868242, + "flos": 22272798519840.0, + "grad_norm": 2.0257363240489945, + "learning_rate": 3.1544394661014145e-06, + "loss": 0.7665, + "num_input_tokens_seen": 117135570, + "step": 3604 + }, + { + "epoch": 0.32511160211029444, + "flos": 33497378784480.0, + "grad_norm": 1.5950038689832489, + "learning_rate": 3.15396235696085e-06, + "loss": 0.6837, + "num_input_tokens_seen": 117168860, + "step": 3605 + }, + { + "epoch": 0.32520178563376473, + "flos": 20959533168480.0, + "grad_norm": 1.8335424506248577, + "learning_rate": 3.153485149357854e-06, + "loss": 0.758, + "num_input_tokens_seen": 117194585, + "step": 3606 + }, + { + "epoch": 0.32529196915723496, + "flos": 23881126125120.0, + "grad_norm": 1.6553633133786025, + "learning_rate": 3.153007843333145e-06, + "loss": 0.7526, + "num_input_tokens_seen": 117224770, + "step": 3607 + }, + { + "epoch": 0.32538215268070525, + "flos": 21986545499040.0, + "grad_norm": 1.9659130303869101, + "learning_rate": 3.152530438927449e-06, + "loss": 0.7983, + "num_input_tokens_seen": 117249605, + "step": 3608 + }, + { + "epoch": 0.3254723362041755, + "flos": 20748969246240.0, + "grad_norm": 1.9236682496228226, + "learning_rate": 3.1520529361815008e-06, + "loss": 0.8125, + "num_input_tokens_seen": 117276350, + "step": 3609 + }, + { + "epoch": 0.32556251972764577, + "flos": 28612541187840.0, + "grad_norm": 1.9691092753700563, + "learning_rate": 3.151575335136044e-06, + "loss": 0.8084, + "num_input_tokens_seen": 117305410, + "step": 3610 + }, + { + "epoch": 0.325652703251116, + "flos": 21294484308480.0, + "grad_norm": 1.9515167328696017, + "learning_rate": 3.1510976358318298e-06, + "loss": 0.7531, + "num_input_tokens_seen": 117331745, + "step": 3611 + }, + { + "epoch": 0.3257428867745863, + "flos": 26573495940000.0, + "grad_norm": 1.8370256161925804, + "learning_rate": 3.1506198383096186e-06, + "loss": 0.7152, + "num_input_tokens_seen": 117362490, + "step": 3612 + }, + { + "epoch": 0.3258330702980565, + "flos": 20674692598560.0, + "grad_norm": 1.9927058125340547, + "learning_rate": 3.150141942610178e-06, + "loss": 0.7822, + "num_input_tokens_seen": 117391070, + "step": 3613 + }, + { + "epoch": 0.3259232538215268, + "flos": 63590509347840.0, + "grad_norm": 1.0784114512390213, + "learning_rate": 3.1496639487742853e-06, + "loss": 0.5858, + "num_input_tokens_seen": 117476625, + "step": 3614 + }, + { + "epoch": 0.32601343734499705, + "flos": 26358731834880.0, + "grad_norm": 1.8763863842195512, + "learning_rate": 3.1491858568427247e-06, + "loss": 0.8185, + "num_input_tokens_seen": 117505595, + "step": 3615 + }, + { + "epoch": 0.32610362086846734, + "flos": 15137199490560.0, + "grad_norm": 1.7465987531305682, + "learning_rate": 3.1487076668562903e-06, + "loss": 0.8211, + "num_input_tokens_seen": 117533125, + "step": 3616 + }, + { + "epoch": 0.32619380439193757, + "flos": 21182028583680.0, + "grad_norm": 2.0040266510186844, + "learning_rate": 3.1482293788557847e-06, + "loss": 0.8148, + "num_input_tokens_seen": 117560820, + "step": 3617 + }, + { + "epoch": 0.32628398791540786, + "flos": 19363248565440.0, + "grad_norm": 1.6836467831926059, + "learning_rate": 3.1477509928820165e-06, + "loss": 0.7341, + "num_input_tokens_seen": 117588360, + "step": 3618 + }, + { + "epoch": 0.3263741714388781, + "flos": 38122025095680.0, + "grad_norm": 1.7627580034514947, + "learning_rate": 3.147272508975805e-06, + "loss": 0.7687, + "num_input_tokens_seen": 117618785, + "step": 3619 + }, + { + "epoch": 0.3264643549623484, + "flos": 54679141446240.0, + "grad_norm": 0.6744503299827993, + "learning_rate": 3.1467939271779775e-06, + "loss": 0.6059, + "num_input_tokens_seen": 117707135, + "step": 3620 + }, + { + "epoch": 0.32655453848581867, + "flos": 42750908759520.0, + "grad_norm": 1.471106971523291, + "learning_rate": 3.146315247529368e-06, + "loss": 0.7318, + "num_input_tokens_seen": 117741085, + "step": 3621 + }, + { + "epoch": 0.3266447220092889, + "flos": 25957755733920.0, + "grad_norm": 2.089599138721797, + "learning_rate": 3.1458364700708212e-06, + "loss": 0.7274, + "num_input_tokens_seen": 117770230, + "step": 3622 + }, + { + "epoch": 0.3267349055327592, + "flos": 26213449478400.0, + "grad_norm": 1.501211074845374, + "learning_rate": 3.1453575948431892e-06, + "loss": 0.7916, + "num_input_tokens_seen": 117801395, + "step": 3623 + }, + { + "epoch": 0.3268250890562294, + "flos": 27156595363200.0, + "grad_norm": 1.8763884172204466, + "learning_rate": 3.144878621887331e-06, + "loss": 0.7766, + "num_input_tokens_seen": 117829925, + "step": 3624 + }, + { + "epoch": 0.3269152725796997, + "flos": 22168966179360.0, + "grad_norm": 2.736755119951065, + "learning_rate": 3.1443995512441167e-06, + "loss": 0.8116, + "num_input_tokens_seen": 117858275, + "step": 3625 + }, + { + "epoch": 0.32700545610316994, + "flos": 19725748231200.0, + "grad_norm": 1.7887412369977949, + "learning_rate": 3.1439203829544224e-06, + "loss": 0.8366, + "num_input_tokens_seen": 117888865, + "step": 3626 + }, + { + "epoch": 0.32709563962664023, + "flos": 22058852149440.0, + "grad_norm": 3.5162245853724747, + "learning_rate": 3.143441117059133e-06, + "loss": 0.8505, + "num_input_tokens_seen": 117919220, + "step": 3627 + }, + { + "epoch": 0.32718582315011047, + "flos": 35466198888480.0, + "grad_norm": 1.9755352502148142, + "learning_rate": 3.142961753599143e-06, + "loss": 0.7167, + "num_input_tokens_seen": 117948915, + "step": 3628 + }, + { + "epoch": 0.32727600667358075, + "flos": 22569087375840.0, + "grad_norm": 4.2972030237009164, + "learning_rate": 3.1424822926153543e-06, + "loss": 0.7365, + "num_input_tokens_seen": 117977185, + "step": 3629 + }, + { + "epoch": 0.327366190197051, + "flos": 33932036421600.0, + "grad_norm": 1.901342610267453, + "learning_rate": 3.142002734148676e-06, + "loss": 0.813, + "num_input_tokens_seen": 118004730, + "step": 3630 + }, + { + "epoch": 0.3274563737205213, + "flos": 26358806174400.0, + "grad_norm": 1.9225664213138134, + "learning_rate": 3.141523078240028e-06, + "loss": 0.798, + "num_input_tokens_seen": 118032810, + "step": 3631 + }, + { + "epoch": 0.3275465572439915, + "flos": 23624428797120.0, + "grad_norm": 1.962494494193548, + "learning_rate": 3.1410433249303366e-06, + "loss": 0.8272, + "num_input_tokens_seen": 118056970, + "step": 3632 + }, + { + "epoch": 0.3276367407674618, + "flos": 25368374621280.0, + "grad_norm": 1.8790978474783981, + "learning_rate": 3.1405634742605366e-06, + "loss": 0.8595, + "num_input_tokens_seen": 118084040, + "step": 3633 + }, + { + "epoch": 0.32772692429093203, + "flos": 16521284701920.0, + "grad_norm": 2.1912365199978447, + "learning_rate": 3.1400835262715727e-06, + "loss": 0.8081, + "num_input_tokens_seen": 118108870, + "step": 3634 + }, + { + "epoch": 0.3278171078144023, + "flos": 22488083001600.0, + "grad_norm": 1.7193506751771663, + "learning_rate": 3.139603481004396e-06, + "loss": 0.759, + "num_input_tokens_seen": 118134090, + "step": 3635 + }, + { + "epoch": 0.32790729133787255, + "flos": 24573075806400.0, + "grad_norm": 1.6508506720587321, + "learning_rate": 3.139123338499966e-06, + "loss": 0.7875, + "num_input_tokens_seen": 118165210, + "step": 3636 + }, + { + "epoch": 0.32799747486134284, + "flos": 20019249355200.0, + "grad_norm": 2.0562827296101616, + "learning_rate": 3.1386430987992524e-06, + "loss": 0.8111, + "num_input_tokens_seen": 118193210, + "step": 3637 + }, + { + "epoch": 0.3280876583848131, + "flos": 40274678330880.0, + "grad_norm": 1.8664037104861202, + "learning_rate": 3.1381627619432307e-06, + "loss": 0.7155, + "num_input_tokens_seen": 118224650, + "step": 3638 + }, + { + "epoch": 0.32817784190828336, + "flos": 25627785341760.0, + "grad_norm": 2.191401462923693, + "learning_rate": 3.1376823279728864e-06, + "loss": 0.7814, + "num_input_tokens_seen": 118249620, + "step": 3639 + }, + { + "epoch": 0.3282680254317536, + "flos": 30396301558560.0, + "grad_norm": 1.9137701940187402, + "learning_rate": 3.1372017969292125e-06, + "loss": 0.8538, + "num_input_tokens_seen": 118278655, + "step": 3640 + }, + { + "epoch": 0.3283582089552239, + "flos": 22970398004640.0, + "grad_norm": 1.719377021857695, + "learning_rate": 3.136721168853211e-06, + "loss": 0.7547, + "num_input_tokens_seen": 118306195, + "step": 3641 + }, + { + "epoch": 0.3284483924786941, + "flos": 26722458102720.0, + "grad_norm": 2.0335789167381253, + "learning_rate": 3.1362404437858924e-06, + "loss": 0.7364, + "num_input_tokens_seen": 118333670, + "step": 3642 + }, + { + "epoch": 0.3285385760021644, + "flos": 25337741005440.0, + "grad_norm": 1.9466286615141182, + "learning_rate": 3.135759621768273e-06, + "loss": 0.8271, + "num_input_tokens_seen": 118362115, + "step": 3643 + }, + { + "epoch": 0.32862875952563464, + "flos": 16995236508960.0, + "grad_norm": 1.7721207221265383, + "learning_rate": 3.13527870284138e-06, + "loss": 0.8257, + "num_input_tokens_seen": 118389270, + "step": 3644 + }, + { + "epoch": 0.32871894304910493, + "flos": 20638148990880.0, + "grad_norm": 6.704985958258186, + "learning_rate": 3.134797687046249e-06, + "loss": 0.8275, + "num_input_tokens_seen": 118415880, + "step": 3645 + }, + { + "epoch": 0.3288091265725752, + "flos": 16849545285120.0, + "grad_norm": 3.531341281487516, + "learning_rate": 3.1343165744239218e-06, + "loss": 0.7074, + "num_input_tokens_seen": 118443785, + "step": 3646 + }, + { + "epoch": 0.32889931009604545, + "flos": 25301792113920.0, + "grad_norm": 1.7783299654803908, + "learning_rate": 3.13383536501545e-06, + "loss": 0.7951, + "num_input_tokens_seen": 118471805, + "step": 3647 + }, + { + "epoch": 0.32898949361951574, + "flos": 18996251358720.0, + "grad_norm": 1.624530944618724, + "learning_rate": 3.133354058861893e-06, + "loss": 0.7937, + "num_input_tokens_seen": 118499990, + "step": 3648 + }, + { + "epoch": 0.32907967714298597, + "flos": 23473087769760.0, + "grad_norm": 1.6704023060506856, + "learning_rate": 3.132872656004318e-06, + "loss": 0.7433, + "num_input_tokens_seen": 118527995, + "step": 3649 + }, + { + "epoch": 0.32916986066645626, + "flos": 31058249509920.0, + "grad_norm": 1.6358194361423675, + "learning_rate": 3.132391156483802e-06, + "loss": 0.6692, + "num_input_tokens_seen": 118557765, + "step": 3650 + }, + { + "epoch": 0.3292600441899265, + "flos": 24500285949120.0, + "grad_norm": 1.8482588255196972, + "learning_rate": 3.131909560341428e-06, + "loss": 0.7308, + "num_input_tokens_seen": 118585295, + "step": 3651 + }, + { + "epoch": 0.3293502277133968, + "flos": 23444461320960.0, + "grad_norm": 1.7496785141062972, + "learning_rate": 3.1314278676182893e-06, + "loss": 0.724, + "num_input_tokens_seen": 118614380, + "step": 3652 + }, + { + "epoch": 0.329440411236867, + "flos": 23844247989600.0, + "grad_norm": 1.517944841496008, + "learning_rate": 3.130946078355486e-06, + "loss": 0.8416, + "num_input_tokens_seen": 118642825, + "step": 3653 + }, + { + "epoch": 0.3295305947603373, + "flos": 22715410485600.0, + "grad_norm": 1.5587714913728998, + "learning_rate": 3.130464192594128e-06, + "loss": 0.8065, + "num_input_tokens_seen": 118672050, + "step": 3654 + }, + { + "epoch": 0.32962077828380754, + "flos": 25374581971200.0, + "grad_norm": 1.596680284785821, + "learning_rate": 3.1299822103753315e-06, + "loss": 0.7003, + "num_input_tokens_seen": 118702410, + "step": 3655 + }, + { + "epoch": 0.3297109618072778, + "flos": 18197421416640.0, + "grad_norm": 1.8188552590025666, + "learning_rate": 3.1295001317402217e-06, + "loss": 0.7757, + "num_input_tokens_seen": 118729730, + "step": 3656 + }, + { + "epoch": 0.32980114533074806, + "flos": 17212119290400.0, + "grad_norm": 2.2421782490077558, + "learning_rate": 3.1290179567299335e-06, + "loss": 0.72, + "num_input_tokens_seen": 118753075, + "step": 3657 + }, + { + "epoch": 0.32989132885421835, + "flos": 26427767546400.0, + "grad_norm": 1.6441034408428457, + "learning_rate": 3.128535685385607e-06, + "loss": 0.8032, + "num_input_tokens_seen": 118783500, + "step": 3658 + }, + { + "epoch": 0.3299815123776886, + "flos": 19144470126240.0, + "grad_norm": 2.05428504628535, + "learning_rate": 3.1280533177483935e-06, + "loss": 0.8538, + "num_input_tokens_seen": 118812100, + "step": 3659 + }, + { + "epoch": 0.33007169590115887, + "flos": 27232247292000.0, + "grad_norm": 1.5237817619837484, + "learning_rate": 3.127570853859451e-06, + "loss": 0.786, + "num_input_tokens_seen": 118841965, + "step": 3660 + }, + { + "epoch": 0.3301618794246291, + "flos": 25297666270560.0, + "grad_norm": 1.5775716585789883, + "learning_rate": 3.1270882937599456e-06, + "loss": 0.6599, + "num_input_tokens_seen": 118871560, + "step": 3661 + }, + { + "epoch": 0.3302520629480994, + "flos": 20054752209600.0, + "grad_norm": 1.97976756427005, + "learning_rate": 3.1266056374910532e-06, + "loss": 0.7951, + "num_input_tokens_seen": 118897770, + "step": 3662 + }, + { + "epoch": 0.3303422464715696, + "flos": 27771778022880.0, + "grad_norm": 1.958488728735996, + "learning_rate": 3.126122885093955e-06, + "loss": 0.67, + "num_input_tokens_seen": 118926875, + "step": 3663 + }, + { + "epoch": 0.3304324299950399, + "flos": 23516619292320.0, + "grad_norm": 2.027805048034868, + "learning_rate": 3.1256400366098427e-06, + "loss": 0.7657, + "num_input_tokens_seen": 118955005, + "step": 3664 + }, + { + "epoch": 0.33052261351851014, + "flos": 22818833958720.0, + "grad_norm": 2.094713018057901, + "learning_rate": 3.125157092079916e-06, + "loss": 0.7312, + "num_input_tokens_seen": 118981925, + "step": 3665 + }, + { + "epoch": 0.33061279704198043, + "flos": 22313876838240.0, + "grad_norm": 2.351636612950414, + "learning_rate": 3.1246740515453824e-06, + "loss": 0.74, + "num_input_tokens_seen": 119009470, + "step": 3666 + }, + { + "epoch": 0.33070298056545067, + "flos": 33532101073920.0, + "grad_norm": 1.5707283038872664, + "learning_rate": 3.124190915047457e-06, + "loss": 0.7116, + "num_input_tokens_seen": 119042645, + "step": 3667 + }, + { + "epoch": 0.33079316408892095, + "flos": 13716198973920.0, + "grad_norm": 2.0983291200328376, + "learning_rate": 3.123707682627364e-06, + "loss": 0.7434, + "num_input_tokens_seen": 119068845, + "step": 3668 + }, + { + "epoch": 0.33088334761239124, + "flos": 31127917107360.0, + "grad_norm": 1.6738649451035204, + "learning_rate": 3.1232243543263356e-06, + "loss": 0.7563, + "num_input_tokens_seen": 119100440, + "step": 3669 + }, + { + "epoch": 0.3309735311358615, + "flos": 28616778540480.0, + "grad_norm": 3.787662963382847, + "learning_rate": 3.1227409301856122e-06, + "loss": 0.6747, + "num_input_tokens_seen": 119129435, + "step": 3670 + }, + { + "epoch": 0.33106371465933176, + "flos": 24570659772000.0, + "grad_norm": 1.399020068334868, + "learning_rate": 3.1222574102464413e-06, + "loss": 0.7544, + "num_input_tokens_seen": 119159715, + "step": 3671 + }, + { + "epoch": 0.331153898182802, + "flos": 26940902014080.0, + "grad_norm": 1.8273487236147712, + "learning_rate": 3.12177379455008e-06, + "loss": 0.8005, + "num_input_tokens_seen": 119187395, + "step": 3672 + }, + { + "epoch": 0.3312440817062723, + "flos": 22532878296000.0, + "grad_norm": 2.1623050976720886, + "learning_rate": 3.121290083137794e-06, + "loss": 0.7628, + "num_input_tokens_seen": 119217485, + "step": 3673 + }, + { + "epoch": 0.3313342652297425, + "flos": 23552939881440.0, + "grad_norm": 1.73076150109937, + "learning_rate": 3.1208062760508547e-06, + "loss": 0.7403, + "num_input_tokens_seen": 119246500, + "step": 3674 + }, + { + "epoch": 0.3314244487532128, + "flos": 28726520872800.0, + "grad_norm": 1.4849667624404859, + "learning_rate": 3.1203223733305438e-06, + "loss": 0.7089, + "num_input_tokens_seen": 119277235, + "step": 3675 + }, + { + "epoch": 0.33151463227668304, + "flos": 64250004095040.0, + "grad_norm": 0.6583091129805092, + "learning_rate": 3.1198383750181512e-06, + "loss": 0.5635, + "num_input_tokens_seen": 119374265, + "step": 3676 + }, + { + "epoch": 0.33160481580015333, + "flos": 24245558618400.0, + "grad_norm": 1.8939921762451868, + "learning_rate": 3.1193542811549734e-06, + "loss": 0.6958, + "num_input_tokens_seen": 119404695, + "step": 3677 + }, + { + "epoch": 0.33169499932362356, + "flos": 57579887901120.0, + "grad_norm": 1.700099252159592, + "learning_rate": 3.1188700917823166e-06, + "loss": 0.6912, + "num_input_tokens_seen": 119442925, + "step": 3678 + }, + { + "epoch": 0.33178518284709385, + "flos": 66575005005600.0, + "grad_norm": 0.7054364041742132, + "learning_rate": 3.1183858069414937e-06, + "loss": 0.6263, + "num_input_tokens_seen": 119532115, + "step": 3679 + }, + { + "epoch": 0.3318753663705641, + "flos": 23771643981120.0, + "grad_norm": 2.7947914418322206, + "learning_rate": 3.117901426673827e-06, + "loss": 0.8198, + "num_input_tokens_seen": 119560220, + "step": 3680 + }, + { + "epoch": 0.3319655498940344, + "flos": 24500397458400.0, + "grad_norm": 8.764220099648236, + "learning_rate": 3.1174169510206466e-06, + "loss": 0.7272, + "num_input_tokens_seen": 119588755, + "step": 3681 + }, + { + "epoch": 0.3320557334175046, + "flos": 22386964053600.0, + "grad_norm": 1.8434813998851227, + "learning_rate": 3.1169323800232908e-06, + "loss": 0.772, + "num_input_tokens_seen": 119619005, + "step": 3682 + }, + { + "epoch": 0.3321459169409749, + "flos": 23297729343840.0, + "grad_norm": 2.21788760336104, + "learning_rate": 3.1164477137231054e-06, + "loss": 0.7972, + "num_input_tokens_seen": 119646640, + "step": 3683 + }, + { + "epoch": 0.3322361004644451, + "flos": 16448420505120.0, + "grad_norm": 2.253984208634714, + "learning_rate": 3.115962952161445e-06, + "loss": 0.7747, + "num_input_tokens_seen": 119670775, + "step": 3684 + }, + { + "epoch": 0.3323262839879154, + "flos": 24245744467200.0, + "grad_norm": 2.1766838790228586, + "learning_rate": 3.1154780953796727e-06, + "loss": 0.8068, + "num_input_tokens_seen": 119697590, + "step": 3685 + }, + { + "epoch": 0.33241646751138565, + "flos": 17360300888160.0, + "grad_norm": 2.0064411155488204, + "learning_rate": 3.114993143419158e-06, + "loss": 0.6721, + "num_input_tokens_seen": 119724480, + "step": 3686 + }, + { + "epoch": 0.33250665103485594, + "flos": 38417050179840.0, + "grad_norm": 1.5540597693095024, + "learning_rate": 3.1145080963212806e-06, + "loss": 0.6808, + "num_input_tokens_seen": 119759315, + "step": 3687 + }, + { + "epoch": 0.33259683455832617, + "flos": 23735174712960.0, + "grad_norm": 1.7235956031305903, + "learning_rate": 3.114022954127427e-06, + "loss": 0.7695, + "num_input_tokens_seen": 119786895, + "step": 3688 + }, + { + "epoch": 0.33268701808179646, + "flos": 23328362959680.0, + "grad_norm": 1.704243563881207, + "learning_rate": 3.1135377168789923e-06, + "loss": 0.9752, + "num_input_tokens_seen": 119814860, + "step": 3689 + }, + { + "epoch": 0.3327772016052667, + "flos": 18197309907360.0, + "grad_norm": 2.706504341137511, + "learning_rate": 3.1130523846173803e-06, + "loss": 0.8338, + "num_input_tokens_seen": 119842625, + "step": 3690 + }, + { + "epoch": 0.332867385128737, + "flos": 32004889352160.0, + "grad_norm": 2.1897671531321596, + "learning_rate": 3.1125669573840006e-06, + "loss": 0.7073, + "num_input_tokens_seen": 119874575, + "step": 3691 + }, + { + "epoch": 0.3329575686522072, + "flos": 16408531619040.0, + "grad_norm": 7.687562523564341, + "learning_rate": 3.112081435220274e-06, + "loss": 0.7143, + "num_input_tokens_seen": 119901155, + "step": 3692 + }, + { + "epoch": 0.3330477521756775, + "flos": 23188879085760.0, + "grad_norm": 1.672908071620171, + "learning_rate": 3.111595818167627e-06, + "loss": 0.7555, + "num_input_tokens_seen": 119929480, + "step": 3693 + }, + { + "epoch": 0.3331379356991478, + "flos": 21111468912000.0, + "grad_norm": 3.032528951558586, + "learning_rate": 3.1111101062674953e-06, + "loss": 0.8263, + "num_input_tokens_seen": 119956740, + "step": 3694 + }, + { + "epoch": 0.333228119222618, + "flos": 23043076352640.0, + "grad_norm": 1.570753348799127, + "learning_rate": 3.1106242995613233e-06, + "loss": 0.7573, + "num_input_tokens_seen": 119984730, + "step": 3695 + }, + { + "epoch": 0.3333183027460883, + "flos": 23953618624320.0, + "grad_norm": 1.8100769684164806, + "learning_rate": 3.1101383980905616e-06, + "loss": 0.766, + "num_input_tokens_seen": 120014380, + "step": 3696 + }, + { + "epoch": 0.33340848626955855, + "flos": 67487703123360.0, + "grad_norm": 0.6731323630692645, + "learning_rate": 3.109652401896671e-06, + "loss": 0.5513, + "num_input_tokens_seen": 120107950, + "step": 3697 + }, + { + "epoch": 0.33349866979302883, + "flos": 29742307935840.0, + "grad_norm": 1.6329893171588215, + "learning_rate": 3.109166311021119e-06, + "loss": 0.8135, + "num_input_tokens_seen": 120138735, + "step": 3698 + }, + { + "epoch": 0.33358885331649907, + "flos": 16193544495360.0, + "grad_norm": 1.9881170838338427, + "learning_rate": 3.1086801255053807e-06, + "loss": 0.8074, + "num_input_tokens_seen": 120162715, + "step": 3699 + }, + { + "epoch": 0.33367903683996936, + "flos": 25338558740160.0, + "grad_norm": 1.5463518741633928, + "learning_rate": 3.108193845390942e-06, + "loss": 0.8068, + "num_input_tokens_seen": 120192965, + "step": 3700 + }, + { + "epoch": 0.3337692203634396, + "flos": 28725480119520.0, + "grad_norm": 1.3119114509887055, + "learning_rate": 3.1077074707192933e-06, + "loss": 0.7456, + "num_input_tokens_seen": 120224480, + "step": 3701 + }, + { + "epoch": 0.3338594038869099, + "flos": 23551415921280.0, + "grad_norm": 1.479316444156731, + "learning_rate": 3.1072210015319353e-06, + "loss": 0.7881, + "num_input_tokens_seen": 120254050, + "step": 3702 + }, + { + "epoch": 0.3339495874103801, + "flos": 22787642796480.0, + "grad_norm": 2.226110472721755, + "learning_rate": 3.106734437870376e-06, + "loss": 0.7221, + "num_input_tokens_seen": 120284205, + "step": 3703 + }, + { + "epoch": 0.3340397709338504, + "flos": 24573596183040.0, + "grad_norm": 1.8459422768963223, + "learning_rate": 3.1062477797761327e-06, + "loss": 0.7137, + "num_input_tokens_seen": 120314615, + "step": 3704 + }, + { + "epoch": 0.33412995445732063, + "flos": 21112063628160.0, + "grad_norm": 1.6699083829635881, + "learning_rate": 3.105761027290729e-06, + "loss": 0.8115, + "num_input_tokens_seen": 120340610, + "step": 3705 + }, + { + "epoch": 0.3342201379807909, + "flos": 28725554459040.0, + "grad_norm": 1.553382671020162, + "learning_rate": 3.105274180455697e-06, + "loss": 0.7345, + "num_input_tokens_seen": 120371540, + "step": 3706 + }, + { + "epoch": 0.33431032150426115, + "flos": 22751991263040.0, + "grad_norm": 1.6742289006597035, + "learning_rate": 3.1047872393125775e-06, + "loss": 0.7881, + "num_input_tokens_seen": 120399225, + "step": 3707 + }, + { + "epoch": 0.33440050502773144, + "flos": 61435105550400.0, + "grad_norm": 0.6310927960132888, + "learning_rate": 3.1043002039029186e-06, + "loss": 0.5791, + "num_input_tokens_seen": 120492180, + "step": 3708 + }, + { + "epoch": 0.3344906885512017, + "flos": 23113896212640.0, + "grad_norm": 1.8220045877895108, + "learning_rate": 3.1038130742682782e-06, + "loss": 0.7596, + "num_input_tokens_seen": 120522870, + "step": 3709 + }, + { + "epoch": 0.33458087207467196, + "flos": 28762321085280.0, + "grad_norm": 1.7201724321960943, + "learning_rate": 3.103325850450219e-06, + "loss": 0.7436, + "num_input_tokens_seen": 120554725, + "step": 3710 + }, + { + "epoch": 0.3346710555981422, + "flos": 25443989380320.0, + "grad_norm": 1.8209287067055027, + "learning_rate": 3.1028385324903154e-06, + "loss": 0.8057, + "num_input_tokens_seen": 120582855, + "step": 3711 + }, + { + "epoch": 0.3347612391216125, + "flos": 25957681394400.0, + "grad_norm": 1.6636733912048811, + "learning_rate": 3.1023511204301465e-06, + "loss": 0.7734, + "num_input_tokens_seen": 120612560, + "step": 3712 + }, + { + "epoch": 0.3348514226450827, + "flos": 29996849417760.0, + "grad_norm": 1.5592493193898467, + "learning_rate": 3.1018636143113022e-06, + "loss": 0.7278, + "num_input_tokens_seen": 120641515, + "step": 3713 + }, + { + "epoch": 0.334941606168553, + "flos": 20420039607360.0, + "grad_norm": 1.5209260241685136, + "learning_rate": 3.1013760141753787e-06, + "loss": 0.6715, + "num_input_tokens_seen": 120669460, + "step": 3714 + }, + { + "epoch": 0.33503178969202324, + "flos": 23771420962560.0, + "grad_norm": 1.7570500055230556, + "learning_rate": 3.100888320063981e-06, + "loss": 0.7365, + "num_input_tokens_seen": 120696735, + "step": 3715 + }, + { + "epoch": 0.33512197321549353, + "flos": 22751842584000.0, + "grad_norm": 1.4635681294755976, + "learning_rate": 3.100400532018721e-06, + "loss": 0.7714, + "num_input_tokens_seen": 120727235, + "step": 3716 + }, + { + "epoch": 0.3352121567389638, + "flos": 16477344312000.0, + "grad_norm": 2.9618264828998893, + "learning_rate": 3.0999126500812204e-06, + "loss": 0.8353, + "num_input_tokens_seen": 120752110, + "step": 3717 + }, + { + "epoch": 0.33530234026243405, + "flos": 58904855993280.0, + "grad_norm": 0.7320224741080937, + "learning_rate": 3.0994246742931076e-06, + "loss": 0.6329, + "num_input_tokens_seen": 120842135, + "step": 3718 + }, + { + "epoch": 0.33539252378590434, + "flos": 33644928496320.0, + "grad_norm": 1.6565171512138428, + "learning_rate": 3.098936604696019e-06, + "loss": 0.7121, + "num_input_tokens_seen": 120875465, + "step": 3719 + }, + { + "epoch": 0.3354827073093746, + "flos": 26540000252640.0, + "grad_norm": 2.9116739501454427, + "learning_rate": 3.0984484413316e-06, + "loss": 0.7986, + "num_input_tokens_seen": 120902600, + "step": 3720 + }, + { + "epoch": 0.33557289083284486, + "flos": 17906299157280.0, + "grad_norm": 4.814722117865747, + "learning_rate": 3.0979601842415033e-06, + "loss": 0.7426, + "num_input_tokens_seen": 120929670, + "step": 3721 + }, + { + "epoch": 0.3356630743563151, + "flos": 28943477993760.0, + "grad_norm": 1.60976353371884, + "learning_rate": 3.0974718334673896e-06, + "loss": 0.6871, + "num_input_tokens_seen": 120961495, + "step": 3722 + }, + { + "epoch": 0.3357532578797854, + "flos": 20637591444480.0, + "grad_norm": 1.6113091854947534, + "learning_rate": 3.0969833890509282e-06, + "loss": 0.7585, + "num_input_tokens_seen": 120990470, + "step": 3723 + }, + { + "epoch": 0.3358434414032556, + "flos": 15755838937920.0, + "grad_norm": 2.0074599138297993, + "learning_rate": 3.096494851033795e-06, + "loss": 0.7569, + "num_input_tokens_seen": 121016780, + "step": 3724 + }, + { + "epoch": 0.3359336249267259, + "flos": 25338744588960.0, + "grad_norm": 1.8632147115346898, + "learning_rate": 3.0960062194576747e-06, + "loss": 0.6914, + "num_input_tokens_seen": 121045315, + "step": 3725 + }, + { + "epoch": 0.33602380845019614, + "flos": 16047778932000.0, + "grad_norm": 1.769259467737411, + "learning_rate": 3.0955174943642606e-06, + "loss": 0.7204, + "num_input_tokens_seen": 121073150, + "step": 3726 + }, + { + "epoch": 0.3361139919736664, + "flos": 32801674957440.0, + "grad_norm": 2.014255738735842, + "learning_rate": 3.0950286757952534e-06, + "loss": 0.6884, + "num_input_tokens_seen": 121102155, + "step": 3727 + }, + { + "epoch": 0.33620417549713666, + "flos": 20929605778080.0, + "grad_norm": 2.113265229621, + "learning_rate": 3.0945397637923617e-06, + "loss": 0.8279, + "num_input_tokens_seen": 121131430, + "step": 3728 + }, + { + "epoch": 0.33629435902060695, + "flos": 22859875107360.0, + "grad_norm": 9.435098708889292, + "learning_rate": 3.0940507583973025e-06, + "loss": 0.7726, + "num_input_tokens_seen": 121158680, + "step": 3729 + }, + { + "epoch": 0.3363845425440772, + "flos": 27633706599840.0, + "grad_norm": 1.8374598258680588, + "learning_rate": 3.093561659651799e-06, + "loss": 0.7371, + "num_input_tokens_seen": 121189625, + "step": 3730 + }, + { + "epoch": 0.33647472606754747, + "flos": 25811692812480.0, + "grad_norm": 1.7328652348934253, + "learning_rate": 3.093072467597586e-06, + "loss": 0.6995, + "num_input_tokens_seen": 121218500, + "step": 3731 + }, + { + "epoch": 0.3365649095910177, + "flos": 71423893710720.0, + "grad_norm": 0.6384171144028872, + "learning_rate": 3.092583182276402e-06, + "loss": 0.5839, + "num_input_tokens_seen": 121315835, + "step": 3732 + }, + { + "epoch": 0.336655093114488, + "flos": 54790370568960.0, + "grad_norm": 0.6381109965447718, + "learning_rate": 3.092093803729997e-06, + "loss": 0.5894, + "num_input_tokens_seen": 121412965, + "step": 3733 + }, + { + "epoch": 0.3367452766379582, + "flos": 28724030498880.0, + "grad_norm": 1.6084089620462434, + "learning_rate": 3.0916043320001264e-06, + "loss": 0.778, + "num_input_tokens_seen": 121442675, + "step": 3734 + }, + { + "epoch": 0.3368354601614285, + "flos": 22965194238240.0, + "grad_norm": 3.5375259128875065, + "learning_rate": 3.0911147671285557e-06, + "loss": 0.7755, + "num_input_tokens_seen": 121470980, + "step": 3735 + }, + { + "epoch": 0.33692564368489875, + "flos": 21986917196640.0, + "grad_norm": 1.8899568370149142, + "learning_rate": 3.0906251091570565e-06, + "loss": 0.7931, + "num_input_tokens_seen": 121499475, + "step": 3736 + }, + { + "epoch": 0.33701582720836903, + "flos": 27414444953760.0, + "grad_norm": 1.6005081055426096, + "learning_rate": 3.0901353581274094e-06, + "loss": 0.7766, + "num_input_tokens_seen": 121530425, + "step": 3737 + }, + { + "epoch": 0.33710601073183927, + "flos": 32915580302880.0, + "grad_norm": 2.702467204414566, + "learning_rate": 3.089645514081402e-06, + "loss": 0.6479, + "num_input_tokens_seen": 121564150, + "step": 3738 + }, + { + "epoch": 0.33719619425530956, + "flos": 16959027429120.0, + "grad_norm": 3.062208356399899, + "learning_rate": 3.0891555770608323e-06, + "loss": 0.6849, + "num_input_tokens_seen": 121589145, + "step": 3739 + }, + { + "epoch": 0.33728637777877984, + "flos": 24717949295520.0, + "grad_norm": 1.6811246889062894, + "learning_rate": 3.088665547107503e-06, + "loss": 0.7863, + "num_input_tokens_seen": 121619725, + "step": 3740 + }, + { + "epoch": 0.3373765613022501, + "flos": 24425340245760.0, + "grad_norm": 1.552802548621325, + "learning_rate": 3.0881754242632254e-06, + "loss": 0.841, + "num_input_tokens_seen": 121648905, + "step": 3741 + }, + { + "epoch": 0.33746674482572037, + "flos": 22240789622880.0, + "grad_norm": 1.39662419399483, + "learning_rate": 3.0876852085698213e-06, + "loss": 0.8102, + "num_input_tokens_seen": 121678505, + "step": 3742 + }, + { + "epoch": 0.3375569283491906, + "flos": 25191938272320.0, + "grad_norm": 1.848229414125687, + "learning_rate": 3.087194900069117e-06, + "loss": 0.7905, + "num_input_tokens_seen": 121707720, + "step": 3743 + }, + { + "epoch": 0.3376471118726609, + "flos": 24275894876160.0, + "grad_norm": 2.3900244463384595, + "learning_rate": 3.08670449880295e-06, + "loss": 0.8212, + "num_input_tokens_seen": 121734965, + "step": 3744 + }, + { + "epoch": 0.3377372953961311, + "flos": 27958844923200.0, + "grad_norm": 2.7320203833817973, + "learning_rate": 3.086214004813163e-06, + "loss": 0.7831, + "num_input_tokens_seen": 121763865, + "step": 3745 + }, + { + "epoch": 0.3378274789196014, + "flos": 22639349689440.0, + "grad_norm": 1.8568286931322011, + "learning_rate": 3.0857234181416074e-06, + "loss": 0.829, + "num_input_tokens_seen": 121792785, + "step": 3746 + }, + { + "epoch": 0.33791766244307164, + "flos": 26282373680640.0, + "grad_norm": 1.841556213772109, + "learning_rate": 3.085232738830143e-06, + "loss": 0.8153, + "num_input_tokens_seen": 121819220, + "step": 3747 + }, + { + "epoch": 0.33800784596654193, + "flos": 21440361381120.0, + "grad_norm": 2.098797535633453, + "learning_rate": 3.084741966920638e-06, + "loss": 0.8283, + "num_input_tokens_seen": 121846915, + "step": 3748 + }, + { + "epoch": 0.33809802949001216, + "flos": 22569459073440.0, + "grad_norm": 1.7872437139996329, + "learning_rate": 3.084251102454966e-06, + "loss": 0.7256, + "num_input_tokens_seen": 121876560, + "step": 3749 + }, + { + "epoch": 0.33818821301348245, + "flos": 70092972723360.0, + "grad_norm": 0.7310852876007538, + "learning_rate": 3.083760145475013e-06, + "loss": 0.6136, + "num_input_tokens_seen": 121956960, + "step": 3750 + }, + { + "epoch": 0.3382783965369527, + "flos": 19180865054880.0, + "grad_norm": 3.2470802983869005, + "learning_rate": 3.0832690960226678e-06, + "loss": 0.7094, + "num_input_tokens_seen": 121984395, + "step": 3751 + }, + { + "epoch": 0.338368580060423, + "flos": 46902197979840.0, + "grad_norm": 2.0957119986851978, + "learning_rate": 3.08277795413983e-06, + "loss": 0.7343, + "num_input_tokens_seen": 122016580, + "step": 3752 + }, + { + "epoch": 0.3384587635838932, + "flos": 25191009028320.0, + "grad_norm": 1.7523207943178472, + "learning_rate": 3.0822867198684073e-06, + "loss": 0.8166, + "num_input_tokens_seen": 122046925, + "step": 3753 + }, + { + "epoch": 0.3385489471073635, + "flos": 22060115921280.0, + "grad_norm": 1.755337341578324, + "learning_rate": 3.081795393250314e-06, + "loss": 0.7788, + "num_input_tokens_seen": 122077790, + "step": 3754 + }, + { + "epoch": 0.33863913063083373, + "flos": 20746070004960.0, + "grad_norm": 1.9975286951021123, + "learning_rate": 3.081303974327473e-06, + "loss": 0.7351, + "num_input_tokens_seen": 122104185, + "step": 3755 + }, + { + "epoch": 0.338729314154304, + "flos": 19582064174400.0, + "grad_norm": 1.7797560868682278, + "learning_rate": 3.080812463141814e-06, + "loss": 0.7636, + "num_input_tokens_seen": 122133305, + "step": 3756 + }, + { + "epoch": 0.33881949767777425, + "flos": 32624569552800.0, + "grad_norm": 1.5055177451981652, + "learning_rate": 3.080320859735276e-06, + "loss": 0.6397, + "num_input_tokens_seen": 122165075, + "step": 3757 + }, + { + "epoch": 0.33890968120124454, + "flos": 26901682183680.0, + "grad_norm": 2.1356649363029323, + "learning_rate": 3.079829164149806e-06, + "loss": 0.8219, + "num_input_tokens_seen": 122195565, + "step": 3758 + }, + { + "epoch": 0.3389998647247148, + "flos": 23989604685600.0, + "grad_norm": 1.6274118131779125, + "learning_rate": 3.0793373764273573e-06, + "loss": 0.7595, + "num_input_tokens_seen": 122225415, + "step": 3759 + }, + { + "epoch": 0.33909004824818506, + "flos": 23400037724160.0, + "grad_norm": 3.0904745622670204, + "learning_rate": 3.078845496609892e-06, + "loss": 0.7921, + "num_input_tokens_seen": 122251535, + "step": 3760 + }, + { + "epoch": 0.3391802317716553, + "flos": 24894274135200.0, + "grad_norm": 1.6906666889797373, + "learning_rate": 3.078353524739381e-06, + "loss": 0.8549, + "num_input_tokens_seen": 122278635, + "step": 3761 + }, + { + "epoch": 0.3392704152951256, + "flos": 20419630740000.0, + "grad_norm": 1.697868564915084, + "learning_rate": 3.077861460857801e-06, + "loss": 0.8524, + "num_input_tokens_seen": 122306240, + "step": 3762 + }, + { + "epoch": 0.3393605988185958, + "flos": 21768398945760.0, + "grad_norm": 1.737189680970647, + "learning_rate": 3.077369305007138e-06, + "loss": 0.7872, + "num_input_tokens_seen": 122336515, + "step": 3763 + }, + { + "epoch": 0.3394507823420661, + "flos": 17899311242400.0, + "grad_norm": 1.7982785921129363, + "learning_rate": 3.0768770572293852e-06, + "loss": 0.805, + "num_input_tokens_seen": 122364420, + "step": 3764 + }, + { + "epoch": 0.3395409658655364, + "flos": 25848236420160.0, + "grad_norm": 2.4838564349533625, + "learning_rate": 3.0763847175665437e-06, + "loss": 0.8483, + "num_input_tokens_seen": 122391620, + "step": 3765 + }, + { + "epoch": 0.3396311493890066, + "flos": 22058108754240.0, + "grad_norm": 1.940319655225957, + "learning_rate": 3.0758922860606237e-06, + "loss": 0.7097, + "num_input_tokens_seen": 122422595, + "step": 3766 + }, + { + "epoch": 0.3397213329124769, + "flos": 22168631651520.0, + "grad_norm": 3.7537017353726507, + "learning_rate": 3.0753997627536404e-06, + "loss": 0.7134, + "num_input_tokens_seen": 122451250, + "step": 3767 + }, + { + "epoch": 0.33981151643594715, + "flos": 24203811244320.0, + "grad_norm": 2.072838965513994, + "learning_rate": 3.0749071476876203e-06, + "loss": 0.7223, + "num_input_tokens_seen": 122477420, + "step": 3768 + }, + { + "epoch": 0.33990169995941744, + "flos": 29779297580640.0, + "grad_norm": 1.733059495911675, + "learning_rate": 3.0744144409045952e-06, + "loss": 0.7372, + "num_input_tokens_seen": 122507355, + "step": 3769 + }, + { + "epoch": 0.33999188348288767, + "flos": 17796631164480.0, + "grad_norm": 3.9235822753446787, + "learning_rate": 3.0739216424466056e-06, + "loss": 0.7948, + "num_input_tokens_seen": 122534135, + "step": 3770 + }, + { + "epoch": 0.34008206700635796, + "flos": 25374544801440.0, + "grad_norm": 2.4043698582545914, + "learning_rate": 3.0734287523557002e-06, + "loss": 0.7511, + "num_input_tokens_seen": 122564515, + "step": 3771 + }, + { + "epoch": 0.3401722505298282, + "flos": 20159662473120.0, + "grad_norm": 1.776451675058387, + "learning_rate": 3.0729357706739348e-06, + "loss": 0.7213, + "num_input_tokens_seen": 122590485, + "step": 3772 + }, + { + "epoch": 0.3402624340532985, + "flos": 29270177447040.0, + "grad_norm": 1.788397053060064, + "learning_rate": 3.0724426974433737e-06, + "loss": 0.7505, + "num_input_tokens_seen": 122617985, + "step": 3773 + }, + { + "epoch": 0.3403526175767687, + "flos": 26577027067200.0, + "grad_norm": 1.5359346779751781, + "learning_rate": 3.0719495327060874e-06, + "loss": 0.7464, + "num_input_tokens_seen": 122648280, + "step": 3774 + }, + { + "epoch": 0.340442801100239, + "flos": 20966075046240.0, + "grad_norm": 1.752246028439014, + "learning_rate": 3.071456276504157e-06, + "loss": 0.7757, + "num_input_tokens_seen": 122676925, + "step": 3775 + }, + { + "epoch": 0.34053298462370923, + "flos": 25371980088000.0, + "grad_norm": 1.5516361215263996, + "learning_rate": 3.070962928879669e-06, + "loss": 0.7661, + "num_input_tokens_seen": 122706940, + "step": 3776 + }, + { + "epoch": 0.3406231681471795, + "flos": 24384745134240.0, + "grad_norm": 1.6519388915806552, + "learning_rate": 3.0704694898747185e-06, + "loss": 0.7823, + "num_input_tokens_seen": 122734965, + "step": 3777 + }, + { + "epoch": 0.34071335167064976, + "flos": 17937341640480.0, + "grad_norm": 1.8022795627310872, + "learning_rate": 3.069975959531408e-06, + "loss": 0.8896, + "num_input_tokens_seen": 122760480, + "step": 3778 + }, + { + "epoch": 0.34080353519412004, + "flos": 70210186177440.0, + "grad_norm": 0.6663208650758605, + "learning_rate": 3.06948233789185e-06, + "loss": 0.6474, + "num_input_tokens_seen": 122847270, + "step": 3779 + }, + { + "epoch": 0.3408937187175903, + "flos": 19764299005920.0, + "grad_norm": 1.8900762975520955, + "learning_rate": 3.0689886249981614e-06, + "loss": 0.7858, + "num_input_tokens_seen": 122875800, + "step": 3780 + }, + { + "epoch": 0.34098390224106057, + "flos": 18635387162400.0, + "grad_norm": 2.056872812117529, + "learning_rate": 3.0684948208924693e-06, + "loss": 0.7977, + "num_input_tokens_seen": 122903060, + "step": 3781 + }, + { + "epoch": 0.3410740857645308, + "flos": 20783840214720.0, + "grad_norm": 1.9774177118438196, + "learning_rate": 3.068000925616907e-06, + "loss": 0.7871, + "num_input_tokens_seen": 122930280, + "step": 3782 + }, + { + "epoch": 0.3411642692880011, + "flos": 19800396576480.0, + "grad_norm": 1.977593315228025, + "learning_rate": 3.067506939213617e-06, + "loss": 0.7625, + "num_input_tokens_seen": 122955905, + "step": 3783 + }, + { + "epoch": 0.3412544528114713, + "flos": 26175419080320.0, + "grad_norm": 1.3780177985122917, + "learning_rate": 3.0670128617247493e-06, + "loss": 0.7735, + "num_input_tokens_seen": 122986455, + "step": 3784 + }, + { + "epoch": 0.3413446363349416, + "flos": 24172991779680.0, + "grad_norm": 2.1921654132537904, + "learning_rate": 3.06651869319246e-06, + "loss": 0.6801, + "num_input_tokens_seen": 123017375, + "step": 3785 + }, + { + "epoch": 0.34143481985841184, + "flos": 13096704622080.0, + "grad_norm": 2.2544934649919495, + "learning_rate": 3.0660244336589154e-06, + "loss": 0.7687, + "num_input_tokens_seen": 123042060, + "step": 3786 + }, + { + "epoch": 0.34152500338188213, + "flos": 36705113252640.0, + "grad_norm": 1.9348769120685576, + "learning_rate": 3.065530083166288e-06, + "loss": 0.7836, + "num_input_tokens_seen": 123073160, + "step": 3787 + }, + { + "epoch": 0.3416151869053524, + "flos": 24425451755040.0, + "grad_norm": 2.6148691413770364, + "learning_rate": 3.0650356417567586e-06, + "loss": 0.7315, + "num_input_tokens_seen": 123103515, + "step": 3788 + }, + { + "epoch": 0.34170537042882265, + "flos": 16084248200160.0, + "grad_norm": 2.148196231126085, + "learning_rate": 3.0645411094725156e-06, + "loss": 0.8192, + "num_input_tokens_seen": 123128875, + "step": 3789 + }, + { + "epoch": 0.34179555395229294, + "flos": 24718581181440.0, + "grad_norm": 1.7533637144384084, + "learning_rate": 3.0640464863557556e-06, + "loss": 0.7672, + "num_input_tokens_seen": 123156620, + "step": 3790 + }, + { + "epoch": 0.3418857374757632, + "flos": 31422793512480.0, + "grad_norm": 4.591148418296682, + "learning_rate": 3.063551772448682e-06, + "loss": 0.8064, + "num_input_tokens_seen": 123192075, + "step": 3791 + }, + { + "epoch": 0.34197592099923346, + "flos": 26532640640160.0, + "grad_norm": 1.702609202822004, + "learning_rate": 3.0630569677935075e-06, + "loss": 0.7013, + "num_input_tokens_seen": 123218915, + "step": 3792 + }, + { + "epoch": 0.3420661045227037, + "flos": 26351297882880.0, + "grad_norm": 2.3468554459124213, + "learning_rate": 3.06256207243245e-06, + "loss": 0.8594, + "num_input_tokens_seen": 123246505, + "step": 3793 + }, + { + "epoch": 0.342156288046174, + "flos": 59407917456000.0, + "grad_norm": 0.6033682288261526, + "learning_rate": 3.0620670864077385e-06, + "loss": 0.5678, + "num_input_tokens_seen": 123339065, + "step": 3794 + }, + { + "epoch": 0.3422464715696442, + "flos": 64373759426880.0, + "grad_norm": 0.6217938681734795, + "learning_rate": 3.0615720097616063e-06, + "loss": 0.571, + "num_input_tokens_seen": 123434345, + "step": 3795 + }, + { + "epoch": 0.3423366550931145, + "flos": 23619856916640.0, + "grad_norm": 1.6341519497149217, + "learning_rate": 3.0610768425362967e-06, + "loss": 0.7209, + "num_input_tokens_seen": 123462070, + "step": 3796 + }, + { + "epoch": 0.34242683861658474, + "flos": 19181496940800.0, + "grad_norm": 2.4752338106323135, + "learning_rate": 3.0605815847740603e-06, + "loss": 0.7466, + "num_input_tokens_seen": 123489340, + "step": 3797 + }, + { + "epoch": 0.342517022140055, + "flos": 20601828401760.0, + "grad_norm": 1.6888554392326935, + "learning_rate": 3.0600862365171553e-06, + "loss": 0.7353, + "num_input_tokens_seen": 123515670, + "step": 3798 + }, + { + "epoch": 0.34260720566352526, + "flos": 19035731377440.0, + "grad_norm": 1.445436214616509, + "learning_rate": 3.0595907978078474e-06, + "loss": 0.7453, + "num_input_tokens_seen": 123544545, + "step": 3799 + }, + { + "epoch": 0.34269738918699555, + "flos": 23844099310560.0, + "grad_norm": 1.553671806988782, + "learning_rate": 3.05909526868841e-06, + "loss": 0.7357, + "num_input_tokens_seen": 123572120, + "step": 3800 + }, + { + "epoch": 0.3427875727104658, + "flos": 24536383519680.0, + "grad_norm": 1.8339170335912667, + "learning_rate": 3.0585996492011243e-06, + "loss": 0.776, + "num_input_tokens_seen": 123601470, + "step": 3801 + }, + { + "epoch": 0.34287775623393607, + "flos": 20233827611520.0, + "grad_norm": 1.7078936251317367, + "learning_rate": 3.05810393938828e-06, + "loss": 0.8157, + "num_input_tokens_seen": 123626950, + "step": 3802 + }, + { + "epoch": 0.3429679397574063, + "flos": 31456809576480.0, + "grad_norm": 2.5021920131062845, + "learning_rate": 3.0576081392921723e-06, + "loss": 0.6791, + "num_input_tokens_seen": 123656765, + "step": 3803 + }, + { + "epoch": 0.3430581232808766, + "flos": 24166970278560.0, + "grad_norm": 1.7940212642766071, + "learning_rate": 3.057112248955107e-06, + "loss": 0.8161, + "num_input_tokens_seen": 123684945, + "step": 3804 + }, + { + "epoch": 0.3431483068043468, + "flos": 15714314582400.0, + "grad_norm": 2.053940089524868, + "learning_rate": 3.0566162684193963e-06, + "loss": 0.7852, + "num_input_tokens_seen": 123711275, + "step": 3805 + }, + { + "epoch": 0.3432384903278171, + "flos": 28363240642080.0, + "grad_norm": 1.6873163900445332, + "learning_rate": 3.056120197727359e-06, + "loss": 0.7174, + "num_input_tokens_seen": 123746375, + "step": 3806 + }, + { + "epoch": 0.34332867385128735, + "flos": 26431484522400.0, + "grad_norm": 1.5840950020612463, + "learning_rate": 3.0556240369213236e-06, + "loss": 0.7594, + "num_input_tokens_seen": 123776090, + "step": 3807 + }, + { + "epoch": 0.34341885737475764, + "flos": 24022208298720.0, + "grad_norm": 2.141648987376745, + "learning_rate": 3.055127786043624e-06, + "loss": 0.6814, + "num_input_tokens_seen": 123806375, + "step": 3808 + }, + { + "epoch": 0.34350904089822787, + "flos": 23443680756000.0, + "grad_norm": 2.4213223718984644, + "learning_rate": 3.054631445136604e-06, + "loss": 0.8521, + "num_input_tokens_seen": 123833500, + "step": 3809 + }, + { + "epoch": 0.34359922442169816, + "flos": 23844062140800.0, + "grad_norm": 2.529908660777828, + "learning_rate": 3.0541350142426147e-06, + "loss": 0.779, + "num_input_tokens_seen": 123858925, + "step": 3810 + }, + { + "epoch": 0.3436894079451684, + "flos": 24536829556800.0, + "grad_norm": 2.5194457047008467, + "learning_rate": 3.053638493404012e-06, + "loss": 0.7261, + "num_input_tokens_seen": 123889635, + "step": 3811 + }, + { + "epoch": 0.3437795914686387, + "flos": 24427273073280.0, + "grad_norm": 2.04225067297805, + "learning_rate": 3.0531418826631643e-06, + "loss": 0.7025, + "num_input_tokens_seen": 123918700, + "step": 3812 + }, + { + "epoch": 0.34386977499210897, + "flos": 26245904412480.0, + "grad_norm": 1.854510571879319, + "learning_rate": 3.052645182062444e-06, + "loss": 0.8471, + "num_input_tokens_seen": 123946480, + "step": 3813 + }, + { + "epoch": 0.3439599585155792, + "flos": 29449104169920.0, + "grad_norm": 2.2572844524705418, + "learning_rate": 3.0521483916442324e-06, + "loss": 0.8069, + "num_input_tokens_seen": 123975005, + "step": 3814 + }, + { + "epoch": 0.3440501420390495, + "flos": 24245632957920.0, + "grad_norm": 2.9115896087331086, + "learning_rate": 3.0516515114509183e-06, + "loss": 0.7643, + "num_input_tokens_seen": 124004050, + "step": 3815 + }, + { + "epoch": 0.3441403255625197, + "flos": 23225125335360.0, + "grad_norm": 1.5714403343998962, + "learning_rate": 3.0511545415249e-06, + "loss": 0.7176, + "num_input_tokens_seen": 124034995, + "step": 3816 + }, + { + "epoch": 0.34423050908599, + "flos": 20784397761120.0, + "grad_norm": 1.9635709428164658, + "learning_rate": 3.050657481908579e-06, + "loss": 0.8, + "num_input_tokens_seen": 124063010, + "step": 3817 + }, + { + "epoch": 0.34432069260946024, + "flos": 21694865693280.0, + "grad_norm": 1.7624953898917433, + "learning_rate": 3.0501603326443677e-06, + "loss": 0.8751, + "num_input_tokens_seen": 124090635, + "step": 3818 + }, + { + "epoch": 0.34441087613293053, + "flos": 22602694572480.0, + "grad_norm": 1.680352722447084, + "learning_rate": 3.049663093774687e-06, + "loss": 0.7532, + "num_input_tokens_seen": 124118430, + "step": 3819 + }, + { + "epoch": 0.34450105965640077, + "flos": 33608161870080.0, + "grad_norm": 3.173706578201432, + "learning_rate": 3.0491657653419643e-06, + "loss": 0.6636, + "num_input_tokens_seen": 124147410, + "step": 3820 + }, + { + "epoch": 0.34459124317987105, + "flos": 23334681818880.0, + "grad_norm": 4.037583454793095, + "learning_rate": 3.0486683473886325e-06, + "loss": 0.8136, + "num_input_tokens_seen": 124175875, + "step": 3821 + }, + { + "epoch": 0.3446814267033413, + "flos": 18999224939520.0, + "grad_norm": 2.0834393919333527, + "learning_rate": 3.0481708399571355e-06, + "loss": 0.8277, + "num_input_tokens_seen": 124203615, + "step": 3822 + }, + { + "epoch": 0.3447716102268116, + "flos": 23620823330400.0, + "grad_norm": 2.2210960342551127, + "learning_rate": 3.047673243089922e-06, + "loss": 0.7974, + "num_input_tokens_seen": 124233670, + "step": 3823 + }, + { + "epoch": 0.3448617937502818, + "flos": 23949269762400.0, + "grad_norm": 1.9198976594750217, + "learning_rate": 3.047175556829451e-06, + "loss": 0.8003, + "num_input_tokens_seen": 124262620, + "step": 3824 + }, + { + "epoch": 0.3449519772737521, + "flos": 21476310272640.0, + "grad_norm": 1.9752432299180362, + "learning_rate": 3.046677781218188e-06, + "loss": 0.8181, + "num_input_tokens_seen": 124289930, + "step": 3825 + }, + { + "epoch": 0.34504216079722233, + "flos": 20566325547360.0, + "grad_norm": 2.5922245917445905, + "learning_rate": 3.0461799162986043e-06, + "loss": 0.744, + "num_input_tokens_seen": 124319085, + "step": 3826 + }, + { + "epoch": 0.3451323443206926, + "flos": 20383421660160.0, + "grad_norm": 1.9777185730639086, + "learning_rate": 3.045681962113183e-06, + "loss": 0.6945, + "num_input_tokens_seen": 124344280, + "step": 3827 + }, + { + "epoch": 0.34522252784416285, + "flos": 23552642523360.0, + "grad_norm": 1.7709007100769856, + "learning_rate": 3.0451839187044095e-06, + "loss": 0.7171, + "num_input_tokens_seen": 124370400, + "step": 3828 + }, + { + "epoch": 0.34531271136763314, + "flos": 33098967396960.0, + "grad_norm": 1.95718791281318, + "learning_rate": 3.0446857861147816e-06, + "loss": 0.7878, + "num_input_tokens_seen": 124399735, + "step": 3829 + }, + { + "epoch": 0.3454028948911034, + "flos": 23407397336640.0, + "grad_norm": 2.0570070351502134, + "learning_rate": 3.044187564386802e-06, + "loss": 0.7447, + "num_input_tokens_seen": 124427560, + "step": 3830 + }, + { + "epoch": 0.34549307841457366, + "flos": 21658136236800.0, + "grad_norm": 1.8177424940453484, + "learning_rate": 3.0436892535629818e-06, + "loss": 0.83, + "num_input_tokens_seen": 124453915, + "step": 3831 + }, + { + "epoch": 0.3455832619380439, + "flos": 23552791202400.0, + "grad_norm": 2.948645529713584, + "learning_rate": 3.0431908536858393e-06, + "loss": 0.8502, + "num_input_tokens_seen": 124482725, + "step": 3832 + }, + { + "epoch": 0.3456734454615142, + "flos": 32297015195040.0, + "grad_norm": 2.0242075982738092, + "learning_rate": 3.0426923647979016e-06, + "loss": 0.7581, + "num_input_tokens_seen": 124510245, + "step": 3833 + }, + { + "epoch": 0.3457636289849844, + "flos": 23188581727680.0, + "grad_norm": 2.093168946673643, + "learning_rate": 3.0421937869417016e-06, + "loss": 0.7056, + "num_input_tokens_seen": 124536295, + "step": 3834 + }, + { + "epoch": 0.3458538125084547, + "flos": 26719893389280.0, + "grad_norm": 2.068353611521125, + "learning_rate": 3.041695120159782e-06, + "loss": 0.8319, + "num_input_tokens_seen": 124563055, + "step": 3835 + }, + { + "epoch": 0.345943996031925, + "flos": 22678234992000.0, + "grad_norm": 1.7817584198942384, + "learning_rate": 3.04119636449469e-06, + "loss": 0.7035, + "num_input_tokens_seen": 124591455, + "step": 3836 + }, + { + "epoch": 0.3460341795553952, + "flos": 30291168276480.0, + "grad_norm": 1.5100522972044241, + "learning_rate": 3.040697519988983e-06, + "loss": 0.7991, + "num_input_tokens_seen": 124625315, + "step": 3837 + }, + { + "epoch": 0.3461243630788655, + "flos": 59801236586400.0, + "grad_norm": 0.6431486617351904, + "learning_rate": 3.040198586685226e-06, + "loss": 0.591, + "num_input_tokens_seen": 124717160, + "step": 3838 + }, + { + "epoch": 0.34621454660233575, + "flos": 26716548110880.0, + "grad_norm": 2.28088072831967, + "learning_rate": 3.039699564625989e-06, + "loss": 0.7186, + "num_input_tokens_seen": 124744225, + "step": 3839 + }, + { + "epoch": 0.34630473012580604, + "flos": 70719938196960.0, + "grad_norm": 0.6720595439035967, + "learning_rate": 3.039200453853853e-06, + "loss": 0.5847, + "num_input_tokens_seen": 124833445, + "step": 3840 + }, + { + "epoch": 0.34639491364927627, + "flos": 23219772889920.0, + "grad_norm": 2.0278537233782887, + "learning_rate": 3.038701254411404e-06, + "loss": 0.8199, + "num_input_tokens_seen": 124859295, + "step": 3841 + }, + { + "epoch": 0.34648509717274656, + "flos": 28214798856000.0, + "grad_norm": 1.775954623806848, + "learning_rate": 3.0382019663412367e-06, + "loss": 0.7879, + "num_input_tokens_seen": 124885400, + "step": 3842 + }, + { + "epoch": 0.3465752806962168, + "flos": 16485075622080.0, + "grad_norm": 2.1634802740867096, + "learning_rate": 3.0377025896859532e-06, + "loss": 0.8069, + "num_input_tokens_seen": 124913590, + "step": 3843 + }, + { + "epoch": 0.3466654642196871, + "flos": 25155617683200.0, + "grad_norm": 1.5702594349563572, + "learning_rate": 3.0372031244881627e-06, + "loss": 0.8757, + "num_input_tokens_seen": 124940825, + "step": 3844 + }, + { + "epoch": 0.3467556477431573, + "flos": 22023312125280.0, + "grad_norm": 3.9435779951161676, + "learning_rate": 3.0367035707904826e-06, + "loss": 0.7282, + "num_input_tokens_seen": 124968620, + "step": 3845 + }, + { + "epoch": 0.3468458312666276, + "flos": 22017848170560.0, + "grad_norm": 2.100740179679713, + "learning_rate": 3.036203928635537e-06, + "loss": 0.6841, + "num_input_tokens_seen": 124994675, + "step": 3846 + }, + { + "epoch": 0.34693601479009784, + "flos": 24208085766720.0, + "grad_norm": 2.0406451751506136, + "learning_rate": 3.035704198065959e-06, + "loss": 0.8346, + "num_input_tokens_seen": 125020095, + "step": 3847 + }, + { + "epoch": 0.3470261983135681, + "flos": 25556853972480.0, + "grad_norm": 1.7884895706661847, + "learning_rate": 3.0352043791243886e-06, + "loss": 0.7563, + "num_input_tokens_seen": 125047675, + "step": 3848 + }, + { + "epoch": 0.34711638183703836, + "flos": 60858064798080.0, + "grad_norm": 0.6596098449987291, + "learning_rate": 3.034704471853472e-06, + "loss": 0.5883, + "num_input_tokens_seen": 125139500, + "step": 3849 + }, + { + "epoch": 0.34720656536050865, + "flos": 21581294875680.0, + "grad_norm": 1.5040994891578523, + "learning_rate": 3.0342044762958646e-06, + "loss": 0.8171, + "num_input_tokens_seen": 125167535, + "step": 3850 + }, + { + "epoch": 0.3472967488839789, + "flos": 21440101192800.0, + "grad_norm": 1.828479765027191, + "learning_rate": 3.0337043924942286e-06, + "loss": 0.8081, + "num_input_tokens_seen": 125194955, + "step": 3851 + }, + { + "epoch": 0.34738693240744917, + "flos": 22423544831040.0, + "grad_norm": 2.2394039939815453, + "learning_rate": 3.0332042204912343e-06, + "loss": 0.757, + "num_input_tokens_seen": 125225570, + "step": 3852 + }, + { + "epoch": 0.3474771159309194, + "flos": 67685214726240.0, + "grad_norm": 0.6646351308776928, + "learning_rate": 3.0327039603295587e-06, + "loss": 0.5923, + "num_input_tokens_seen": 125320550, + "step": 3853 + }, + { + "epoch": 0.3475672994543897, + "flos": 20492346257760.0, + "grad_norm": 2.2153323003051137, + "learning_rate": 3.032203612051887e-06, + "loss": 0.7291, + "num_input_tokens_seen": 125346180, + "step": 3854 + }, + { + "epoch": 0.3476574829778599, + "flos": 15501260286240.0, + "grad_norm": 1.9151224121132038, + "learning_rate": 3.0317031757009116e-06, + "loss": 0.7517, + "num_input_tokens_seen": 125371525, + "step": 3855 + }, + { + "epoch": 0.3477476665013302, + "flos": 28318445347680.0, + "grad_norm": 2.1371984971878106, + "learning_rate": 3.0312026513193326e-06, + "loss": 0.731, + "num_input_tokens_seen": 125399795, + "step": 3856 + }, + { + "epoch": 0.34783785002480044, + "flos": 33689575111680.0, + "grad_norm": 2.148705705027477, + "learning_rate": 3.0307020389498573e-06, + "loss": 0.748, + "num_input_tokens_seen": 125435335, + "step": 3857 + }, + { + "epoch": 0.34792803354827073, + "flos": 41105480000160.0, + "grad_norm": 1.4851162314604422, + "learning_rate": 3.0302013386352004e-06, + "loss": 0.6976, + "num_input_tokens_seen": 125467335, + "step": 3858 + }, + { + "epoch": 0.348018217071741, + "flos": 22168445802720.0, + "grad_norm": 1.5225408018060016, + "learning_rate": 3.0297005504180854e-06, + "loss": 0.7601, + "num_input_tokens_seen": 125498955, + "step": 3859 + }, + { + "epoch": 0.34810840059521125, + "flos": 70179664070880.0, + "grad_norm": 0.6028648002959177, + "learning_rate": 3.0291996743412417e-06, + "loss": 0.5119, + "num_input_tokens_seen": 125593670, + "step": 3860 + }, + { + "epoch": 0.34819858411868154, + "flos": 23002815768960.0, + "grad_norm": 1.6295594293994626, + "learning_rate": 3.0286987104474063e-06, + "loss": 0.858, + "num_input_tokens_seen": 125620255, + "step": 3861 + }, + { + "epoch": 0.3482887676421518, + "flos": 35537427615840.0, + "grad_norm": 1.3133193138184343, + "learning_rate": 3.028197658779325e-06, + "loss": 0.7413, + "num_input_tokens_seen": 125653185, + "step": 3862 + }, + { + "epoch": 0.34837895116562206, + "flos": 27155591779680.0, + "grad_norm": 1.9056553773094655, + "learning_rate": 3.0276965193797503e-06, + "loss": 0.7869, + "num_input_tokens_seen": 125683275, + "step": 3863 + }, + { + "epoch": 0.3484691346890923, + "flos": 45120333266880.0, + "grad_norm": 1.4140171022662362, + "learning_rate": 3.0271952922914423e-06, + "loss": 0.7348, + "num_input_tokens_seen": 125715095, + "step": 3864 + }, + { + "epoch": 0.3485593182125626, + "flos": 17573912730720.0, + "grad_norm": 2.495054694361718, + "learning_rate": 3.0266939775571675e-06, + "loss": 0.7468, + "num_input_tokens_seen": 125741125, + "step": 3865 + }, + { + "epoch": 0.3486495017360328, + "flos": 32547282154560.0, + "grad_norm": 2.0544652782411026, + "learning_rate": 3.026192575219701e-06, + "loss": 0.759, + "num_input_tokens_seen": 125771710, + "step": 3866 + }, + { + "epoch": 0.3487396852595031, + "flos": 27560507875200.0, + "grad_norm": 1.9901156553848476, + "learning_rate": 3.025691085321826e-06, + "loss": 0.659, + "num_input_tokens_seen": 125803625, + "step": 3867 + }, + { + "epoch": 0.34882986878297334, + "flos": 25520124516000.0, + "grad_norm": 1.8220261133250943, + "learning_rate": 3.025189507906332e-06, + "loss": 0.7531, + "num_input_tokens_seen": 125831890, + "step": 3868 + }, + { + "epoch": 0.34892005230644363, + "flos": 26321965208640.0, + "grad_norm": 1.7505873648082153, + "learning_rate": 3.0246878430160166e-06, + "loss": 0.7893, + "num_input_tokens_seen": 125862635, + "step": 3869 + }, + { + "epoch": 0.34901023582991386, + "flos": 25734814281600.0, + "grad_norm": 1.955114099436446, + "learning_rate": 3.024186090693684e-06, + "loss": 0.7115, + "num_input_tokens_seen": 125892390, + "step": 3870 + }, + { + "epoch": 0.34910041935338415, + "flos": 21112286646720.0, + "grad_norm": 1.6749935121552946, + "learning_rate": 3.023684250982147e-06, + "loss": 0.794, + "num_input_tokens_seen": 125918715, + "step": 3871 + }, + { + "epoch": 0.3491906028768544, + "flos": 19071457250400.0, + "grad_norm": 1.8397794740630176, + "learning_rate": 3.0231823239242252e-06, + "loss": 0.6884, + "num_input_tokens_seen": 125947145, + "step": 3872 + }, + { + "epoch": 0.34928078640032467, + "flos": 25228593389280.0, + "grad_norm": 1.8181013176176533, + "learning_rate": 3.0226803095627457e-06, + "loss": 0.7245, + "num_input_tokens_seen": 125976690, + "step": 3873 + }, + { + "epoch": 0.3493709699237949, + "flos": 25512727733760.0, + "grad_norm": 2.3103416267869563, + "learning_rate": 3.022178207940543e-06, + "loss": 0.7495, + "num_input_tokens_seen": 126004610, + "step": 3874 + }, + { + "epoch": 0.3494611534472652, + "flos": 25301160228000.0, + "grad_norm": 1.5506664781354382, + "learning_rate": 3.02167601910046e-06, + "loss": 0.7815, + "num_input_tokens_seen": 126032145, + "step": 3875 + }, + { + "epoch": 0.3495513369707354, + "flos": 23039359376640.0, + "grad_norm": 1.8998712219967724, + "learning_rate": 3.021173743085345e-06, + "loss": 0.7891, + "num_input_tokens_seen": 126060375, + "step": 3876 + }, + { + "epoch": 0.3496415204942057, + "flos": 70547516182080.0, + "grad_norm": 0.7987819369144395, + "learning_rate": 3.0206713799380557e-06, + "loss": 0.5786, + "num_input_tokens_seen": 126149780, + "step": 3877 + }, + { + "epoch": 0.34973170401767595, + "flos": 27633595090560.0, + "grad_norm": 1.7098407452583144, + "learning_rate": 3.0201689297014565e-06, + "loss": 0.7759, + "num_input_tokens_seen": 126177655, + "step": 3878 + }, + { + "epoch": 0.34982188754114624, + "flos": 25774665997920.0, + "grad_norm": 1.5634152592798942, + "learning_rate": 3.0196663924184187e-06, + "loss": 0.8114, + "num_input_tokens_seen": 126208145, + "step": 3879 + }, + { + "epoch": 0.34991207106461647, + "flos": 23115420172800.0, + "grad_norm": 2.6980848124968246, + "learning_rate": 3.019163768131822e-06, + "loss": 0.8029, + "num_input_tokens_seen": 126236945, + "step": 3880 + }, + { + "epoch": 0.35000225458808676, + "flos": 22131827855520.0, + "grad_norm": 8.453562726940655, + "learning_rate": 3.0186610568845533e-06, + "loss": 0.6891, + "num_input_tokens_seen": 126264630, + "step": 3881 + }, + { + "epoch": 0.350092438111557, + "flos": 24390915314400.0, + "grad_norm": 2.016552970245369, + "learning_rate": 3.018158258719507e-06, + "loss": 0.702, + "num_input_tokens_seen": 126292650, + "step": 3882 + }, + { + "epoch": 0.3501826216350273, + "flos": 24135407418720.0, + "grad_norm": 1.634862097077852, + "learning_rate": 3.0176553736795827e-06, + "loss": 0.794, + "num_input_tokens_seen": 126322775, + "step": 3883 + }, + { + "epoch": 0.35027280515849757, + "flos": 64955818096800.0, + "grad_norm": 0.5349514980256324, + "learning_rate": 3.017152401807691e-06, + "loss": 0.5168, + "num_input_tokens_seen": 126418555, + "step": 3884 + }, + { + "epoch": 0.3503629886819678, + "flos": 25048365724800.0, + "grad_norm": 1.6570424037703642, + "learning_rate": 3.0166493431467476e-06, + "loss": 0.7511, + "num_input_tokens_seen": 126452465, + "step": 3885 + }, + { + "epoch": 0.3504531722054381, + "flos": 22278411153600.0, + "grad_norm": 1.5461901294956202, + "learning_rate": 3.016146197739677e-06, + "loss": 0.7391, + "num_input_tokens_seen": 126482475, + "step": 3886 + }, + { + "epoch": 0.3505433557289083, + "flos": 23698779784320.0, + "grad_norm": 1.6978839410853823, + "learning_rate": 3.0156429656294097e-06, + "loss": 0.7008, + "num_input_tokens_seen": 126510105, + "step": 3887 + }, + { + "epoch": 0.3506335392523786, + "flos": 34445914284480.0, + "grad_norm": 1.6684626042488138, + "learning_rate": 3.0151396468588844e-06, + "loss": 0.7011, + "num_input_tokens_seen": 126541030, + "step": 3888 + }, + { + "epoch": 0.35072372277584885, + "flos": 28208405657280.0, + "grad_norm": 2.1540386636605846, + "learning_rate": 3.014636241471047e-06, + "loss": 0.717, + "num_input_tokens_seen": 126569030, + "step": 3889 + }, + { + "epoch": 0.35081390629931913, + "flos": 19070528006400.0, + "grad_norm": 1.6896674753381435, + "learning_rate": 3.0141327495088514e-06, + "loss": 0.7558, + "num_input_tokens_seen": 126596295, + "step": 3890 + }, + { + "epoch": 0.35090408982278937, + "flos": 28395138029760.0, + "grad_norm": 1.6233253653224355, + "learning_rate": 3.0136291710152566e-06, + "loss": 0.7202, + "num_input_tokens_seen": 126626770, + "step": 3891 + }, + { + "epoch": 0.35099427334625966, + "flos": 32624532383040.0, + "grad_norm": 1.8100010317594, + "learning_rate": 3.0131255060332325e-06, + "loss": 0.8028, + "num_input_tokens_seen": 126656605, + "step": 3892 + }, + { + "epoch": 0.3510844568697299, + "flos": 25191975442080.0, + "grad_norm": 1.571689210988257, + "learning_rate": 3.012621754605754e-06, + "loss": 0.783, + "num_input_tokens_seen": 126686320, + "step": 3893 + }, + { + "epoch": 0.3511746403932002, + "flos": 22674964053120.0, + "grad_norm": 1.7240318981813112, + "learning_rate": 3.0121179167758035e-06, + "loss": 0.7899, + "num_input_tokens_seen": 126712550, + "step": 3894 + }, + { + "epoch": 0.3512648239166704, + "flos": 22053202345920.0, + "grad_norm": 3.117724379576687, + "learning_rate": 3.0116139925863717e-06, + "loss": 0.8448, + "num_input_tokens_seen": 126735745, + "step": 3895 + }, + { + "epoch": 0.3513550074401407, + "flos": 22824892629600.0, + "grad_norm": 1.927885998422261, + "learning_rate": 3.011109982080456e-06, + "loss": 0.8028, + "num_input_tokens_seen": 126764010, + "step": 3896 + }, + { + "epoch": 0.35144519096361093, + "flos": 17612946712320.0, + "grad_norm": 2.164587955512753, + "learning_rate": 3.0106058853010614e-06, + "loss": 0.7084, + "num_input_tokens_seen": 126789715, + "step": 3897 + }, + { + "epoch": 0.3515353744870812, + "flos": 38815907604480.0, + "grad_norm": 1.6480350816751836, + "learning_rate": 3.010101702291201e-06, + "loss": 0.7474, + "num_input_tokens_seen": 126820355, + "step": 3898 + }, + { + "epoch": 0.35162555801055145, + "flos": 18306940730400.0, + "grad_norm": 1.8321641820900405, + "learning_rate": 3.009597433093893e-06, + "loss": 0.7615, + "num_input_tokens_seen": 126846970, + "step": 3899 + }, + { + "epoch": 0.35171574153402174, + "flos": 22751619565440.0, + "grad_norm": 1.7588426750183481, + "learning_rate": 3.009093077752165e-06, + "loss": 0.7713, + "num_input_tokens_seen": 126875835, + "step": 3900 + }, + { + "epoch": 0.351805925057492, + "flos": 22896790412640.0, + "grad_norm": 5.703196903977375, + "learning_rate": 3.008588636309052e-06, + "loss": 0.7926, + "num_input_tokens_seen": 126902670, + "step": 3901 + }, + { + "epoch": 0.35189610858096226, + "flos": 23662422025440.0, + "grad_norm": 3.923193721269621, + "learning_rate": 3.0080841088075947e-06, + "loss": 0.7297, + "num_input_tokens_seen": 126932645, + "step": 3902 + }, + { + "epoch": 0.3519862921044325, + "flos": 20601828401760.0, + "grad_norm": 5.123348249034394, + "learning_rate": 3.0075794952908436e-06, + "loss": 0.8066, + "num_input_tokens_seen": 126959100, + "step": 3903 + }, + { + "epoch": 0.3520764756279028, + "flos": 20201670035520.0, + "grad_norm": 1.551239484064329, + "learning_rate": 3.0070747958018528e-06, + "loss": 0.7795, + "num_input_tokens_seen": 126988940, + "step": 3904 + }, + { + "epoch": 0.352166659151373, + "flos": 23990310911040.0, + "grad_norm": 1.4588097566134526, + "learning_rate": 3.0065700103836894e-06, + "loss": 0.7817, + "num_input_tokens_seen": 127018515, + "step": 3905 + }, + { + "epoch": 0.3522568426748433, + "flos": 23735918108160.0, + "grad_norm": 2.2951608087054343, + "learning_rate": 3.0060651390794214e-06, + "loss": 0.7829, + "num_input_tokens_seen": 127044925, + "step": 3906 + }, + { + "epoch": 0.3523470261983136, + "flos": 35647987682880.0, + "grad_norm": 1.4865362395107575, + "learning_rate": 3.005560181932128e-06, + "loss": 0.7462, + "num_input_tokens_seen": 127079435, + "step": 3907 + }, + { + "epoch": 0.35243720972178383, + "flos": 38745199253760.0, + "grad_norm": 1.7935346653578943, + "learning_rate": 3.005055138984896e-06, + "loss": 0.7223, + "num_input_tokens_seen": 127109940, + "step": 3908 + }, + { + "epoch": 0.3525273932452541, + "flos": 29416351877760.0, + "grad_norm": 2.0304141158723246, + "learning_rate": 3.0045500102808174e-06, + "loss": 0.7363, + "num_input_tokens_seen": 127141215, + "step": 3909 + }, + { + "epoch": 0.35261757676872435, + "flos": 21469731225120.0, + "grad_norm": 1.9078191333752283, + "learning_rate": 3.0040447958629927e-06, + "loss": 0.8476, + "num_input_tokens_seen": 127165640, + "step": 3910 + }, + { + "epoch": 0.35270776029219464, + "flos": 23334495970080.0, + "grad_norm": 1.675688769938717, + "learning_rate": 3.00353949577453e-06, + "loss": 0.7836, + "num_input_tokens_seen": 127191265, + "step": 3911 + }, + { + "epoch": 0.35279794381566487, + "flos": 19508679600960.0, + "grad_norm": 3.956160755567089, + "learning_rate": 3.003034110058544e-06, + "loss": 0.603, + "num_input_tokens_seen": 127214800, + "step": 3912 + }, + { + "epoch": 0.35288812733913516, + "flos": 21177456703200.0, + "grad_norm": 1.5605370971220585, + "learning_rate": 3.002528638758157e-06, + "loss": 0.7424, + "num_input_tokens_seen": 127242300, + "step": 3913 + }, + { + "epoch": 0.3529783108626054, + "flos": 18780669518880.0, + "grad_norm": 2.607522729644751, + "learning_rate": 3.0020230819164985e-06, + "loss": 0.7561, + "num_input_tokens_seen": 127269710, + "step": 3914 + }, + { + "epoch": 0.3530684943860757, + "flos": 24206190108960.0, + "grad_norm": 1.475041974003873, + "learning_rate": 3.0015174395767064e-06, + "loss": 0.7683, + "num_input_tokens_seen": 127299175, + "step": 3915 + }, + { + "epoch": 0.3531586779095459, + "flos": 19545594906240.0, + "grad_norm": 1.8104289488673933, + "learning_rate": 3.001011711781923e-06, + "loss": 0.7826, + "num_input_tokens_seen": 127327555, + "step": 3916 + }, + { + "epoch": 0.3532488614330162, + "flos": 66427126499520.0, + "grad_norm": 1.7407279978880175, + "learning_rate": 3.0005058985753017e-06, + "loss": 0.7167, + "num_input_tokens_seen": 127358885, + "step": 3917 + }, + { + "epoch": 0.35333904495648644, + "flos": 19977910848480.0, + "grad_norm": 1.68277205309701, + "learning_rate": 3e-06, + "loss": 0.7509, + "num_input_tokens_seen": 127386350, + "step": 3918 + }, + { + "epoch": 0.3534292284799567, + "flos": 70587590916960.0, + "grad_norm": 0.6876481720168859, + "learning_rate": 2.9994940160991843e-06, + "loss": 0.6051, + "num_input_tokens_seen": 127474185, + "step": 3919 + }, + { + "epoch": 0.35351941200342696, + "flos": 63941226199680.0, + "grad_norm": 0.6412903656626504, + "learning_rate": 2.9989879469160285e-06, + "loss": 0.5628, + "num_input_tokens_seen": 127564355, + "step": 3920 + }, + { + "epoch": 0.35360959552689725, + "flos": 23516098915680.0, + "grad_norm": 2.187282769452478, + "learning_rate": 2.9984817924937124e-06, + "loss": 0.7797, + "num_input_tokens_seen": 127592055, + "step": 3921 + }, + { + "epoch": 0.3536997790503675, + "flos": 17360300888160.0, + "grad_norm": 1.6727066110853985, + "learning_rate": 2.997975552875424e-06, + "loss": 0.7018, + "num_input_tokens_seen": 127618725, + "step": 3922 + }, + { + "epoch": 0.35378996257383777, + "flos": 26576358011520.0, + "grad_norm": 1.7666309158138693, + "learning_rate": 2.997469228104358e-06, + "loss": 0.7501, + "num_input_tokens_seen": 127650575, + "step": 3923 + }, + { + "epoch": 0.353880146097308, + "flos": 35320247476320.0, + "grad_norm": 1.5118988481827185, + "learning_rate": 2.996962818223718e-06, + "loss": 0.6595, + "num_input_tokens_seen": 127679930, + "step": 3924 + }, + { + "epoch": 0.3539703296207783, + "flos": 17796817013280.0, + "grad_norm": 2.241789566509367, + "learning_rate": 2.9964563232767135e-06, + "loss": 0.7099, + "num_input_tokens_seen": 127707775, + "step": 3925 + }, + { + "epoch": 0.3540605131442485, + "flos": 20747891323200.0, + "grad_norm": 3.505795993072607, + "learning_rate": 2.9959497433065617e-06, + "loss": 0.7603, + "num_input_tokens_seen": 127733740, + "step": 3926 + }, + { + "epoch": 0.3541506966677188, + "flos": 45809569555680.0, + "grad_norm": 1.7615429437379768, + "learning_rate": 2.9954430783564848e-06, + "loss": 0.6824, + "num_input_tokens_seen": 127767340, + "step": 3927 + }, + { + "epoch": 0.35424088019118904, + "flos": 18887215251840.0, + "grad_norm": 1.7380604807357638, + "learning_rate": 2.994936328469716e-06, + "loss": 0.7944, + "num_input_tokens_seen": 127795085, + "step": 3928 + }, + { + "epoch": 0.35433106371465933, + "flos": 22708385400960.0, + "grad_norm": 1.562661887365547, + "learning_rate": 2.994429493689494e-06, + "loss": 0.7642, + "num_input_tokens_seen": 127822870, + "step": 3929 + }, + { + "epoch": 0.35442124723812957, + "flos": 22309602315840.0, + "grad_norm": 1.3734990510829652, + "learning_rate": 2.9939225740590642e-06, + "loss": 0.7857, + "num_input_tokens_seen": 127853155, + "step": 3930 + }, + { + "epoch": 0.35451143076159986, + "flos": 32807250421440.0, + "grad_norm": 1.4883706373870185, + "learning_rate": 2.99341556962168e-06, + "loss": 0.7166, + "num_input_tokens_seen": 127883655, + "step": 3931 + }, + { + "epoch": 0.35460161428507014, + "flos": 30250907692800.0, + "grad_norm": 1.8964197558346159, + "learning_rate": 2.992908480420602e-06, + "loss": 0.8037, + "num_input_tokens_seen": 127915360, + "step": 3932 + }, + { + "epoch": 0.3546917978085404, + "flos": 25625220628320.0, + "grad_norm": 1.7452323546829733, + "learning_rate": 2.9924013064990974e-06, + "loss": 0.7485, + "num_input_tokens_seen": 127945360, + "step": 3933 + }, + { + "epoch": 0.35478198133201067, + "flos": 71134890127680.0, + "grad_norm": 0.6568192329290331, + "learning_rate": 2.991894047900441e-06, + "loss": 0.6102, + "num_input_tokens_seen": 128047630, + "step": 3934 + }, + { + "epoch": 0.3548721648554809, + "flos": 21182586130080.0, + "grad_norm": 1.6823704074190675, + "learning_rate": 2.991386704667916e-06, + "loss": 0.7762, + "num_input_tokens_seen": 128078550, + "step": 3935 + }, + { + "epoch": 0.3549623483789512, + "flos": 27774491415360.0, + "grad_norm": 1.803337682282756, + "learning_rate": 2.9908792768448097e-06, + "loss": 0.8251, + "num_input_tokens_seen": 128110530, + "step": 3936 + }, + { + "epoch": 0.3550525319024214, + "flos": 26831865907200.0, + "grad_norm": 1.9605851179741667, + "learning_rate": 2.990371764474421e-06, + "loss": 0.758, + "num_input_tokens_seen": 128140985, + "step": 3937 + }, + { + "epoch": 0.3551427154258917, + "flos": 21112063628160.0, + "grad_norm": 1.6673587871964117, + "learning_rate": 2.9898641676000518e-06, + "loss": 0.7275, + "num_input_tokens_seen": 128169535, + "step": 3938 + }, + { + "epoch": 0.35523289894936194, + "flos": 23078988074400.0, + "grad_norm": 1.9079493468096722, + "learning_rate": 2.9893564862650138e-06, + "loss": 0.7613, + "num_input_tokens_seen": 128197990, + "step": 3939 + }, + { + "epoch": 0.35532308247283223, + "flos": 21913049416320.0, + "grad_norm": 2.4315941495935136, + "learning_rate": 2.9888487205126254e-06, + "loss": 0.8149, + "num_input_tokens_seen": 128224140, + "step": 3940 + }, + { + "epoch": 0.35541326599630246, + "flos": 27888359591040.0, + "grad_norm": 1.6777445356529856, + "learning_rate": 2.9883408703862115e-06, + "loss": 0.7568, + "num_input_tokens_seen": 128253040, + "step": 3941 + }, + { + "epoch": 0.35550344951977275, + "flos": 13023803255520.0, + "grad_norm": 2.106347065395421, + "learning_rate": 2.987832935929105e-06, + "loss": 0.8404, + "num_input_tokens_seen": 128279995, + "step": 3942 + }, + { + "epoch": 0.355593633043243, + "flos": 17177359831200.0, + "grad_norm": 1.6614434491709584, + "learning_rate": 2.9873249171846454e-06, + "loss": 0.8266, + "num_input_tokens_seen": 128304675, + "step": 3943 + }, + { + "epoch": 0.3556838165667133, + "flos": 20456285856960.0, + "grad_norm": 1.939622393364005, + "learning_rate": 2.98681681419618e-06, + "loss": 0.7536, + "num_input_tokens_seen": 128330915, + "step": 3944 + }, + { + "epoch": 0.3557740000901835, + "flos": 21836393904000.0, + "grad_norm": 4.583548084198652, + "learning_rate": 2.9863086270070627e-06, + "loss": 0.8173, + "num_input_tokens_seen": 128360735, + "step": 3945 + }, + { + "epoch": 0.3558641836136538, + "flos": 21331065085920.0, + "grad_norm": 1.6582210975634526, + "learning_rate": 2.985800355660655e-06, + "loss": 0.7664, + "num_input_tokens_seen": 128389740, + "step": 3946 + }, + { + "epoch": 0.35595436713712403, + "flos": 64245989760960.0, + "grad_norm": 0.6724743276964195, + "learning_rate": 2.9852920002003252e-06, + "loss": 0.6083, + "num_input_tokens_seen": 128484480, + "step": 3947 + }, + { + "epoch": 0.3560445506605943, + "flos": 20164271523360.0, + "grad_norm": 1.9835531981065262, + "learning_rate": 2.9847835606694494e-06, + "loss": 0.8007, + "num_input_tokens_seen": 128512895, + "step": 3948 + }, + { + "epoch": 0.35613473418406455, + "flos": 29527432321440.0, + "grad_norm": 2.001432144482367, + "learning_rate": 2.98427503711141e-06, + "loss": 0.6893, + "num_input_tokens_seen": 128543005, + "step": 3949 + }, + { + "epoch": 0.35622491770753484, + "flos": 23553274409280.0, + "grad_norm": 1.8229205394885386, + "learning_rate": 2.9837664295695973e-06, + "loss": 0.7923, + "num_input_tokens_seen": 128572230, + "step": 3950 + }, + { + "epoch": 0.35631510123100507, + "flos": 33316259045760.0, + "grad_norm": 1.5306748068927563, + "learning_rate": 2.983257738087408e-06, + "loss": 0.7573, + "num_input_tokens_seen": 128604545, + "step": 3951 + }, + { + "epoch": 0.35640528475447536, + "flos": 25556445105120.0, + "grad_norm": 1.857194086300967, + "learning_rate": 2.982748962708247e-06, + "loss": 0.7902, + "num_input_tokens_seen": 128630890, + "step": 3952 + }, + { + "epoch": 0.3564954682779456, + "flos": 23990459590080.0, + "grad_norm": 1.7906337984145206, + "learning_rate": 2.982240103475526e-06, + "loss": 0.8153, + "num_input_tokens_seen": 128660615, + "step": 3953 + }, + { + "epoch": 0.3565856518014159, + "flos": 21694791353760.0, + "grad_norm": 2.2607769891354814, + "learning_rate": 2.981731160432663e-06, + "loss": 0.761, + "num_input_tokens_seen": 128688095, + "step": 3954 + }, + { + "epoch": 0.35667583532488617, + "flos": 29018051999520.0, + "grad_norm": 1.7176828712934291, + "learning_rate": 2.981222133623084e-06, + "loss": 0.6594, + "num_input_tokens_seen": 128718660, + "step": 3955 + }, + { + "epoch": 0.3567660188483564, + "flos": 20748337360320.0, + "grad_norm": 2.5204021515956185, + "learning_rate": 2.980713023090222e-06, + "loss": 0.7698, + "num_input_tokens_seen": 128746630, + "step": 3956 + }, + { + "epoch": 0.3568562023718267, + "flos": 23443271888640.0, + "grad_norm": 1.6819347164307785, + "learning_rate": 2.980203828877518e-06, + "loss": 0.7376, + "num_input_tokens_seen": 128775705, + "step": 3957 + }, + { + "epoch": 0.3569463858952969, + "flos": 17395989591360.0, + "grad_norm": 9.403628164503848, + "learning_rate": 2.9796945510284182e-06, + "loss": 0.8015, + "num_input_tokens_seen": 128798960, + "step": 3958 + }, + { + "epoch": 0.3570365694187672, + "flos": 23079694299840.0, + "grad_norm": 1.7493096761169942, + "learning_rate": 2.9791851895863774e-06, + "loss": 0.8193, + "num_input_tokens_seen": 128828730, + "step": 3959 + }, + { + "epoch": 0.35712675294223745, + "flos": 19433325030240.0, + "grad_norm": 2.0702300433166787, + "learning_rate": 2.978675744594857e-06, + "loss": 0.7944, + "num_input_tokens_seen": 128854955, + "step": 3960 + }, + { + "epoch": 0.35721693646570774, + "flos": 33061829073120.0, + "grad_norm": 2.3294186678130777, + "learning_rate": 2.978166216097326e-06, + "loss": 0.7219, + "num_input_tokens_seen": 128889610, + "step": 3961 + }, + { + "epoch": 0.35730711998917797, + "flos": 27597274501440.0, + "grad_norm": 1.9103110783086648, + "learning_rate": 2.9776566041372596e-06, + "loss": 0.7525, + "num_input_tokens_seen": 128920580, + "step": 3962 + }, + { + "epoch": 0.35739730351264826, + "flos": 22387744618560.0, + "grad_norm": 3.6945753592207233, + "learning_rate": 2.977146908758141e-06, + "loss": 0.8153, + "num_input_tokens_seen": 128949840, + "step": 3963 + }, + { + "epoch": 0.3574874870361185, + "flos": 26176459833600.0, + "grad_norm": 1.863969655694495, + "learning_rate": 2.9766371300034604e-06, + "loss": 0.7776, + "num_input_tokens_seen": 128978170, + "step": 3964 + }, + { + "epoch": 0.3575776705595888, + "flos": 20456731894080.0, + "grad_norm": 1.552458425508828, + "learning_rate": 2.9761272679167142e-06, + "loss": 0.8132, + "num_input_tokens_seen": 129005340, + "step": 3965 + }, + { + "epoch": 0.357667854083059, + "flos": 25447557677280.0, + "grad_norm": 1.6618894617571185, + "learning_rate": 2.9756173225414072e-06, + "loss": 0.611, + "num_input_tokens_seen": 129035075, + "step": 3966 + }, + { + "epoch": 0.3577580376065293, + "flos": 26905250480640.0, + "grad_norm": 2.3466295990219894, + "learning_rate": 2.975107293921051e-06, + "loss": 0.939, + "num_input_tokens_seen": 129063905, + "step": 3967 + }, + { + "epoch": 0.35784822112999953, + "flos": 24572332411200.0, + "grad_norm": 4.8244939192777965, + "learning_rate": 2.9745971820991643e-06, + "loss": 0.8029, + "num_input_tokens_seen": 129087995, + "step": 3968 + }, + { + "epoch": 0.3579384046534698, + "flos": 34516771314240.0, + "grad_norm": 2.3353300293486963, + "learning_rate": 2.9740869871192715e-06, + "loss": 0.8054, + "num_input_tokens_seen": 129120180, + "step": 3969 + }, + { + "epoch": 0.35802858817694005, + "flos": 18379470399360.0, + "grad_norm": 1.577174816560548, + "learning_rate": 2.9735767090249065e-06, + "loss": 0.827, + "num_input_tokens_seen": 129147140, + "step": 3970 + }, + { + "epoch": 0.35811877170041034, + "flos": 32072289594240.0, + "grad_norm": 1.9087913957365126, + "learning_rate": 2.973066347859608e-06, + "loss": 0.6715, + "num_input_tokens_seen": 129176960, + "step": 3971 + }, + { + "epoch": 0.3582089552238806, + "flos": 23480038514880.0, + "grad_norm": 1.8744123173656173, + "learning_rate": 2.972555903666923e-06, + "loss": 0.7757, + "num_input_tokens_seen": 129205110, + "step": 3972 + }, + { + "epoch": 0.35829913874735086, + "flos": 24209015010720.0, + "grad_norm": 2.988278696894334, + "learning_rate": 2.972045376490406e-06, + "loss": 0.6815, + "num_input_tokens_seen": 129234140, + "step": 3973 + }, + { + "epoch": 0.3583893222708211, + "flos": 19180567696800.0, + "grad_norm": 1.5599985317076595, + "learning_rate": 2.9715347663736177e-06, + "loss": 0.6907, + "num_input_tokens_seen": 129263290, + "step": 3974 + }, + { + "epoch": 0.3584795057942914, + "flos": 18523972190880.0, + "grad_norm": 1.710986968418349, + "learning_rate": 2.9710240733601266e-06, + "loss": 0.7598, + "num_input_tokens_seen": 129291815, + "step": 3975 + }, + { + "epoch": 0.3585696893177616, + "flos": 23552977051200.0, + "grad_norm": 2.2474669927394637, + "learning_rate": 2.970513297493507e-06, + "loss": 0.8439, + "num_input_tokens_seen": 129317595, + "step": 3976 + }, + { + "epoch": 0.3586598728412319, + "flos": 25336663082400.0, + "grad_norm": 1.531129014821441, + "learning_rate": 2.9700024388173416e-06, + "loss": 0.8118, + "num_input_tokens_seen": 129346955, + "step": 3977 + }, + { + "epoch": 0.35875005636470214, + "flos": 19764670703520.0, + "grad_norm": 1.849023357555355, + "learning_rate": 2.969491497375219e-06, + "loss": 0.8162, + "num_input_tokens_seen": 129375540, + "step": 3978 + }, + { + "epoch": 0.35884023988817243, + "flos": 19909953060000.0, + "grad_norm": 3.396134642044831, + "learning_rate": 2.9689804732107364e-06, + "loss": 0.7864, + "num_input_tokens_seen": 129404185, + "step": 3979 + }, + { + "epoch": 0.3589304234116427, + "flos": 14627484640800.0, + "grad_norm": 2.0135030296422247, + "learning_rate": 2.9684693663674968e-06, + "loss": 0.7217, + "num_input_tokens_seen": 129431225, + "step": 3980 + }, + { + "epoch": 0.35902060693511295, + "flos": 24932601891360.0, + "grad_norm": 1.6647023148519269, + "learning_rate": 2.9679581768891115e-06, + "loss": 0.8465, + "num_input_tokens_seen": 129460350, + "step": 3981 + }, + { + "epoch": 0.35911079045858324, + "flos": 22127107296000.0, + "grad_norm": 1.6054500383832033, + "learning_rate": 2.967446904819197e-06, + "loss": 0.6864, + "num_input_tokens_seen": 129490505, + "step": 3982 + }, + { + "epoch": 0.3592009739820535, + "flos": 20492011729920.0, + "grad_norm": 1.6013324107062366, + "learning_rate": 2.966935550201378e-06, + "loss": 0.7901, + "num_input_tokens_seen": 129518930, + "step": 3983 + }, + { + "epoch": 0.35929115750552376, + "flos": 19978617073920.0, + "grad_norm": 2.0425567503274187, + "learning_rate": 2.966424113079286e-06, + "loss": 0.7605, + "num_input_tokens_seen": 129547125, + "step": 3984 + }, + { + "epoch": 0.359381341028994, + "flos": 27305631865440.0, + "grad_norm": 1.8517012403312296, + "learning_rate": 2.9659125934965596e-06, + "loss": 0.8173, + "num_input_tokens_seen": 129577055, + "step": 3985 + }, + { + "epoch": 0.3594715245524643, + "flos": 23219252513280.0, + "grad_norm": 1.7199446947609969, + "learning_rate": 2.9654009914968457e-06, + "loss": 0.8238, + "num_input_tokens_seen": 129607280, + "step": 3986 + }, + { + "epoch": 0.3595617080759345, + "flos": 24427496091840.0, + "grad_norm": 2.216015810052819, + "learning_rate": 2.9648893071237956e-06, + "loss": 0.6523, + "num_input_tokens_seen": 129635545, + "step": 3987 + }, + { + "epoch": 0.3596518915994048, + "flos": 23035828249440.0, + "grad_norm": 1.6405118630772955, + "learning_rate": 2.964377540421069e-06, + "loss": 0.7427, + "num_input_tokens_seen": 129662895, + "step": 3988 + }, + { + "epoch": 0.35974207512287504, + "flos": 16735008053760.0, + "grad_norm": 2.3610550917418753, + "learning_rate": 2.963865691432334e-06, + "loss": 0.7896, + "num_input_tokens_seen": 129687700, + "step": 3989 + }, + { + "epoch": 0.3598322586463453, + "flos": 27446342341440.0, + "grad_norm": 1.6706288765156958, + "learning_rate": 2.963353760201263e-06, + "loss": 0.679, + "num_input_tokens_seen": 129716695, + "step": 3990 + }, + { + "epoch": 0.35992244216981556, + "flos": 15278132985120.0, + "grad_norm": 1.8748780528784488, + "learning_rate": 2.962841746771537e-06, + "loss": 0.7477, + "num_input_tokens_seen": 129742110, + "step": 3991 + }, + { + "epoch": 0.36001262569328585, + "flos": 27996540793440.0, + "grad_norm": 1.8255317096322716, + "learning_rate": 2.9623296511868445e-06, + "loss": 0.7698, + "num_input_tokens_seen": 129771220, + "step": 3992 + }, + { + "epoch": 0.3601028092167561, + "flos": 22748460135840.0, + "grad_norm": 1.7432540571131914, + "learning_rate": 2.96181747349088e-06, + "loss": 0.6926, + "num_input_tokens_seen": 129800880, + "step": 3993 + }, + { + "epoch": 0.36019299274022637, + "flos": 18307126579200.0, + "grad_norm": 2.2750125423546774, + "learning_rate": 2.961305213727345e-06, + "loss": 0.7845, + "num_input_tokens_seen": 129828460, + "step": 3994 + }, + { + "epoch": 0.3602831762636966, + "flos": 70471083688320.0, + "grad_norm": 0.6454200165523957, + "learning_rate": 2.960792871939949e-06, + "loss": 0.6027, + "num_input_tokens_seen": 129921115, + "step": 3995 + }, + { + "epoch": 0.3603733597871669, + "flos": 28726112005440.0, + "grad_norm": 2.4411544792055286, + "learning_rate": 2.9602804481724064e-06, + "loss": 0.8562, + "num_input_tokens_seen": 129950105, + "step": 3996 + }, + { + "epoch": 0.3604635433106371, + "flos": 22715038788000.0, + "grad_norm": 2.4569787533826934, + "learning_rate": 2.9597679424684427e-06, + "loss": 0.7609, + "num_input_tokens_seen": 129979710, + "step": 3997 + }, + { + "epoch": 0.3605537268341074, + "flos": 23625655399200.0, + "grad_norm": 2.2262990695315557, + "learning_rate": 2.9592553548717848e-06, + "loss": 0.7816, + "num_input_tokens_seen": 130007185, + "step": 3998 + }, + { + "epoch": 0.36064391035757765, + "flos": 13127895784320.0, + "grad_norm": 2.104857403779786, + "learning_rate": 2.958742685426171e-06, + "loss": 0.8485, + "num_input_tokens_seen": 130031660, + "step": 3999 + }, + { + "epoch": 0.36073409388104793, + "flos": 22350940822560.0, + "grad_norm": 1.666833495691459, + "learning_rate": 2.9582299341753446e-06, + "loss": 0.8351, + "num_input_tokens_seen": 130060220, + "step": 4000 + }, + { + "epoch": 0.36082427740451817, + "flos": 21840148049760.0, + "grad_norm": 4.5342134453500496, + "learning_rate": 2.957717101163057e-06, + "loss": 0.708, + "num_input_tokens_seen": 130084100, + "step": 4001 + }, + { + "epoch": 0.36091446092798846, + "flos": 22350383276160.0, + "grad_norm": 1.6313149917926673, + "learning_rate": 2.9572041864330655e-06, + "loss": 0.7671, + "num_input_tokens_seen": 130111450, + "step": 4002 + }, + { + "epoch": 0.36100464445145874, + "flos": 28690720660320.0, + "grad_norm": 2.1591136547879994, + "learning_rate": 2.9566911900291346e-06, + "loss": 0.7836, + "num_input_tokens_seen": 130144570, + "step": 4003 + }, + { + "epoch": 0.361094827974929, + "flos": 24609730923360.0, + "grad_norm": 2.5425875095247576, + "learning_rate": 2.9561781119950368e-06, + "loss": 0.6818, + "num_input_tokens_seen": 130173965, + "step": 4004 + }, + { + "epoch": 0.36118501149839927, + "flos": 19794337905600.0, + "grad_norm": 1.4823128279432969, + "learning_rate": 2.9556649523745493e-06, + "loss": 0.7111, + "num_input_tokens_seen": 130201285, + "step": 4005 + }, + { + "epoch": 0.3612751950218695, + "flos": 21731557980000.0, + "grad_norm": 1.609237665038743, + "learning_rate": 2.955151711211459e-06, + "loss": 0.8262, + "num_input_tokens_seen": 130228520, + "step": 4006 + }, + { + "epoch": 0.3613653785453398, + "flos": 23951202589920.0, + "grad_norm": 1.7148336006159646, + "learning_rate": 2.9546383885495583e-06, + "loss": 0.7344, + "num_input_tokens_seen": 130258790, + "step": 4007 + }, + { + "epoch": 0.36145556206881, + "flos": 34045978936800.0, + "grad_norm": 1.9312977732456178, + "learning_rate": 2.9541249844326464e-06, + "loss": 0.7189, + "num_input_tokens_seen": 130289450, + "step": 4008 + }, + { + "epoch": 0.3615457455922803, + "flos": 24099830224800.0, + "grad_norm": 1.6554510240744456, + "learning_rate": 2.9536114989045295e-06, + "loss": 0.747, + "num_input_tokens_seen": 130318360, + "step": 4009 + }, + { + "epoch": 0.36163592911575054, + "flos": 18926397912480.0, + "grad_norm": 1.5990451288015894, + "learning_rate": 2.9530979320090216e-06, + "loss": 0.8256, + "num_input_tokens_seen": 130346010, + "step": 4010 + }, + { + "epoch": 0.36172611263922083, + "flos": 23042890503840.0, + "grad_norm": 2.2047789297198865, + "learning_rate": 2.9525842837899422e-06, + "loss": 0.8429, + "num_input_tokens_seen": 130375545, + "step": 4011 + }, + { + "epoch": 0.36181629616269106, + "flos": 20638557858240.0, + "grad_norm": 2.1814741790033745, + "learning_rate": 2.95207055429112e-06, + "loss": 0.8468, + "num_input_tokens_seen": 130401910, + "step": 4012 + }, + { + "epoch": 0.36190647968616135, + "flos": 22930137420960.0, + "grad_norm": 1.6337075586504985, + "learning_rate": 2.951556743556388e-06, + "loss": 0.7607, + "num_input_tokens_seen": 130429600, + "step": 4013 + }, + { + "epoch": 0.3619966632096316, + "flos": 17760087556800.0, + "grad_norm": 2.6946165209047215, + "learning_rate": 2.951042851629588e-06, + "loss": 0.8147, + "num_input_tokens_seen": 130455635, + "step": 4014 + }, + { + "epoch": 0.3620868467331019, + "flos": 23622718988160.0, + "grad_norm": 1.763879030889296, + "learning_rate": 2.950528878554568e-06, + "loss": 0.7217, + "num_input_tokens_seen": 130483285, + "step": 4015 + }, + { + "epoch": 0.3621770302565721, + "flos": 25812250358880.0, + "grad_norm": 1.4416320127758018, + "learning_rate": 2.950014824375183e-06, + "loss": 0.8044, + "num_input_tokens_seen": 130512845, + "step": 4016 + }, + { + "epoch": 0.3622672137800424, + "flos": 67389074549280.0, + "grad_norm": 0.632575284978129, + "learning_rate": 2.949500689135295e-06, + "loss": 0.5272, + "num_input_tokens_seen": 130610710, + "step": 4017 + }, + { + "epoch": 0.36235739730351263, + "flos": 23990013552960.0, + "grad_norm": 1.5036679562125022, + "learning_rate": 2.9489864728787722e-06, + "loss": 0.8469, + "num_input_tokens_seen": 130639545, + "step": 4018 + }, + { + "epoch": 0.3624475808269829, + "flos": 22962369336480.0, + "grad_norm": 1.6213668409375608, + "learning_rate": 2.9484721756494915e-06, + "loss": 0.8266, + "num_input_tokens_seen": 130665255, + "step": 4019 + }, + { + "epoch": 0.36253776435045315, + "flos": 24937879997280.0, + "grad_norm": 2.0778642074891236, + "learning_rate": 2.9479577974913343e-06, + "loss": 0.7718, + "num_input_tokens_seen": 130696585, + "step": 4020 + }, + { + "epoch": 0.36262794787392344, + "flos": 23841608936640.0, + "grad_norm": 1.6552592499161525, + "learning_rate": 2.9474433384481908e-06, + "loss": 0.7632, + "num_input_tokens_seen": 130727400, + "step": 4021 + }, + { + "epoch": 0.3627181313973937, + "flos": 27232804838400.0, + "grad_norm": 1.7131225011418352, + "learning_rate": 2.9469287985639577e-06, + "loss": 0.7637, + "num_input_tokens_seen": 130755235, + "step": 4022 + }, + { + "epoch": 0.36280831492086396, + "flos": 23771420962560.0, + "grad_norm": 1.7081414634449965, + "learning_rate": 2.9464141778825384e-06, + "loss": 0.8136, + "num_input_tokens_seen": 130783805, + "step": 4023 + }, + { + "epoch": 0.3628984984443342, + "flos": 27117895909440.0, + "grad_norm": 2.004054251814024, + "learning_rate": 2.9458994764478427e-06, + "loss": 0.8082, + "num_input_tokens_seen": 130813875, + "step": 4024 + }, + { + "epoch": 0.3629886819678045, + "flos": 20595695391360.0, + "grad_norm": 1.8127331419331736, + "learning_rate": 2.9453846943037883e-06, + "loss": 0.7987, + "num_input_tokens_seen": 130839350, + "step": 4025 + }, + { + "epoch": 0.36307886549127477, + "flos": 23007053121600.0, + "grad_norm": 2.6692409805241217, + "learning_rate": 2.9448698314942987e-06, + "loss": 0.7135, + "num_input_tokens_seen": 130870615, + "step": 4026 + }, + { + "epoch": 0.363169049014745, + "flos": 71164631669280.0, + "grad_norm": 0.7570621275534384, + "learning_rate": 2.944354888063305e-06, + "loss": 0.6231, + "num_input_tokens_seen": 130967870, + "step": 4027 + }, + { + "epoch": 0.3632592325382153, + "flos": 21695051542080.0, + "grad_norm": 1.957150453755629, + "learning_rate": 2.9438398640547453e-06, + "loss": 0.7439, + "num_input_tokens_seen": 130995690, + "step": 4028 + }, + { + "epoch": 0.3633494160616855, + "flos": 23007127461120.0, + "grad_norm": 1.6993266455058664, + "learning_rate": 2.943324759512564e-06, + "loss": 0.7047, + "num_input_tokens_seen": 131024365, + "step": 4029 + }, + { + "epoch": 0.3634395995851558, + "flos": 24207788408640.0, + "grad_norm": 1.7926767813790994, + "learning_rate": 2.9428095744807134e-06, + "loss": 0.7959, + "num_input_tokens_seen": 131055870, + "step": 4030 + }, + { + "epoch": 0.36352978310862605, + "flos": 35683825065120.0, + "grad_norm": 4.206564478008342, + "learning_rate": 2.942294309003151e-06, + "loss": 0.72, + "num_input_tokens_seen": 131088590, + "step": 4031 + }, + { + "epoch": 0.36361996663209634, + "flos": 28070743101600.0, + "grad_norm": 1.8001128505723, + "learning_rate": 2.941778963123843e-06, + "loss": 0.785, + "num_input_tokens_seen": 131116780, + "step": 4032 + }, + { + "epoch": 0.36371015015556657, + "flos": 31240521511200.0, + "grad_norm": 1.6825761662739402, + "learning_rate": 2.94126353688676e-06, + "loss": 0.767, + "num_input_tokens_seen": 131149320, + "step": 4033 + }, + { + "epoch": 0.36380033367903686, + "flos": 24537126914880.0, + "grad_norm": 2.578568668914997, + "learning_rate": 2.9407480303358825e-06, + "loss": 0.7173, + "num_input_tokens_seen": 131176780, + "step": 4034 + }, + { + "epoch": 0.3638905172025071, + "flos": 19757980146720.0, + "grad_norm": 1.6617387477171834, + "learning_rate": 2.940232443515195e-06, + "loss": 0.7735, + "num_input_tokens_seen": 131207505, + "step": 4035 + }, + { + "epoch": 0.3639807007259774, + "flos": 22532172070560.0, + "grad_norm": 3.2098820569008244, + "learning_rate": 2.9397167764686916e-06, + "loss": 0.789, + "num_input_tokens_seen": 131238310, + "step": 4036 + }, + { + "epoch": 0.3640708842494476, + "flos": 24828435023040.0, + "grad_norm": 2.042118749066892, + "learning_rate": 2.9392010292403714e-06, + "loss": 0.7427, + "num_input_tokens_seen": 131266710, + "step": 4037 + }, + { + "epoch": 0.3641610677729179, + "flos": 21841114463520.0, + "grad_norm": 2.310644384730538, + "learning_rate": 2.9386852018742404e-06, + "loss": 0.7607, + "num_input_tokens_seen": 131293270, + "step": 4038 + }, + { + "epoch": 0.36425125129638813, + "flos": 23734765845600.0, + "grad_norm": 15.114255004653822, + "learning_rate": 2.938169294414312e-06, + "loss": 0.74, + "num_input_tokens_seen": 131322815, + "step": 4039 + }, + { + "epoch": 0.3643414348198584, + "flos": 22090600858080.0, + "grad_norm": 2.3701832765623525, + "learning_rate": 2.9376533069046067e-06, + "loss": 0.7968, + "num_input_tokens_seen": 131351850, + "step": 4040 + }, + { + "epoch": 0.36443161834332866, + "flos": 66606419186400.0, + "grad_norm": 0.7153369057134475, + "learning_rate": 2.9371372393891514e-06, + "loss": 0.6344, + "num_input_tokens_seen": 131440865, + "step": 4041 + }, + { + "epoch": 0.36452180186679894, + "flos": 25338298551840.0, + "grad_norm": 1.2664055948414599, + "learning_rate": 2.936621091911979e-06, + "loss": 0.8352, + "num_input_tokens_seen": 131472395, + "step": 4042 + }, + { + "epoch": 0.3646119853902692, + "flos": 24135964965120.0, + "grad_norm": 2.560489610254252, + "learning_rate": 2.936104864517131e-06, + "loss": 0.6992, + "num_input_tokens_seen": 131500155, + "step": 4043 + }, + { + "epoch": 0.36470216891373947, + "flos": 38848808575680.0, + "grad_norm": 1.9391616341012334, + "learning_rate": 2.9355885572486535e-06, + "loss": 0.7233, + "num_input_tokens_seen": 131528150, + "step": 4044 + }, + { + "epoch": 0.3647923524372097, + "flos": 22129820688480.0, + "grad_norm": 1.6650527371826658, + "learning_rate": 2.9350721701506026e-06, + "loss": 0.798, + "num_input_tokens_seen": 131558840, + "step": 4045 + }, + { + "epoch": 0.36488253596068, + "flos": 29564607815040.0, + "grad_norm": 1.8201456934913969, + "learning_rate": 2.9345557032670375e-06, + "loss": 0.7577, + "num_input_tokens_seen": 131588920, + "step": 4046 + }, + { + "epoch": 0.3649727194841502, + "flos": 19035471189120.0, + "grad_norm": 1.9876779416638959, + "learning_rate": 2.934039156642027e-06, + "loss": 0.8101, + "num_input_tokens_seen": 131614840, + "step": 4047 + }, + { + "epoch": 0.3650629030076205, + "flos": 37317768368640.0, + "grad_norm": 1.87121103831, + "learning_rate": 2.9335225303196454e-06, + "loss": 0.6283, + "num_input_tokens_seen": 131644590, + "step": 4048 + }, + { + "epoch": 0.36515308653109074, + "flos": 22969989137280.0, + "grad_norm": 1.7926558344097483, + "learning_rate": 2.933005824343974e-06, + "loss": 0.7347, + "num_input_tokens_seen": 131673005, + "step": 4049 + }, + { + "epoch": 0.36524327005456103, + "flos": 18450067240800.0, + "grad_norm": 2.404849251604581, + "learning_rate": 2.932489038759101e-06, + "loss": 0.7394, + "num_input_tokens_seen": 131700565, + "step": 4050 + }, + { + "epoch": 0.3653334535780313, + "flos": 53060820909120.0, + "grad_norm": 1.4647706687499, + "learning_rate": 2.9319721736091215e-06, + "loss": 0.6574, + "num_input_tokens_seen": 131734240, + "step": 4051 + }, + { + "epoch": 0.36542363710150155, + "flos": 25005986464800.0, + "grad_norm": 1.5211273675543262, + "learning_rate": 2.9314552289381377e-06, + "loss": 0.6729, + "num_input_tokens_seen": 131765620, + "step": 4052 + }, + { + "epoch": 0.36551382062497184, + "flos": 18670852847040.0, + "grad_norm": 1.845525371731605, + "learning_rate": 2.9309382047902574e-06, + "loss": 0.7363, + "num_input_tokens_seen": 131790330, + "step": 4053 + }, + { + "epoch": 0.3656040041484421, + "flos": 24100722299040.0, + "grad_norm": 1.5423581230909058, + "learning_rate": 2.9304211012095963e-06, + "loss": 0.8106, + "num_input_tokens_seen": 131821195, + "step": 4054 + }, + { + "epoch": 0.36569418767191236, + "flos": 18483191230560.0, + "grad_norm": 1.823423878178387, + "learning_rate": 2.929903918240277e-06, + "loss": 0.823, + "num_input_tokens_seen": 131847825, + "step": 4055 + }, + { + "epoch": 0.3657843711953826, + "flos": 66725453958720.0, + "grad_norm": 0.6319203386727867, + "learning_rate": 2.9293866559264273e-06, + "loss": 0.5778, + "num_input_tokens_seen": 131938830, + "step": 4056 + }, + { + "epoch": 0.3658745547188529, + "flos": 21986508329280.0, + "grad_norm": 2.241713736322382, + "learning_rate": 2.928869314312184e-06, + "loss": 0.7652, + "num_input_tokens_seen": 131967710, + "step": 4057 + }, + { + "epoch": 0.3659647382423231, + "flos": 21695237390880.0, + "grad_norm": 1.7027788422902554, + "learning_rate": 2.9283518934416892e-06, + "loss": 0.7639, + "num_input_tokens_seen": 131994970, + "step": 4058 + }, + { + "epoch": 0.3660549217657934, + "flos": 20711050357440.0, + "grad_norm": 1.541282872377727, + "learning_rate": 2.927834393359092e-06, + "loss": 0.774, + "num_input_tokens_seen": 132022160, + "step": 4059 + }, + { + "epoch": 0.36614510528926364, + "flos": 20674543919520.0, + "grad_norm": 2.059091120648512, + "learning_rate": 2.927316814108548e-06, + "loss": 0.8348, + "num_input_tokens_seen": 132050115, + "step": 4060 + }, + { + "epoch": 0.36623528881273393, + "flos": 24792077264160.0, + "grad_norm": 2.4811346647039842, + "learning_rate": 2.92679915573422e-06, + "loss": 0.6703, + "num_input_tokens_seen": 132080380, + "step": 4061 + }, + { + "epoch": 0.36632547233620416, + "flos": 25445364661440.0, + "grad_norm": 2.1914252894045347, + "learning_rate": 2.926281418280278e-06, + "loss": 0.7055, + "num_input_tokens_seen": 132110935, + "step": 4062 + }, + { + "epoch": 0.36641565585967445, + "flos": 24391435691040.0, + "grad_norm": 2.554858933975782, + "learning_rate": 2.925763601790899e-06, + "loss": 0.7655, + "num_input_tokens_seen": 132139755, + "step": 4063 + }, + { + "epoch": 0.3665058393831447, + "flos": 21841002954240.0, + "grad_norm": 1.9903241587444278, + "learning_rate": 2.9252457063102635e-06, + "loss": 0.7098, + "num_input_tokens_seen": 132169130, + "step": 4064 + }, + { + "epoch": 0.36659602290661497, + "flos": 53608380308160.0, + "grad_norm": 1.9879461279579393, + "learning_rate": 2.9247277318825626e-06, + "loss": 0.7264, + "num_input_tokens_seen": 132204265, + "step": 4065 + }, + { + "epoch": 0.3666862064300852, + "flos": 22820803956000.0, + "grad_norm": 2.0169806837557465, + "learning_rate": 2.924209678551993e-06, + "loss": 0.8558, + "num_input_tokens_seen": 132232430, + "step": 4066 + }, + { + "epoch": 0.3667763899535555, + "flos": 20602162929600.0, + "grad_norm": 1.7259887156156308, + "learning_rate": 2.923691546362757e-06, + "loss": 0.7763, + "num_input_tokens_seen": 132258860, + "step": 4067 + }, + { + "epoch": 0.3668665734770257, + "flos": 22132199553120.0, + "grad_norm": 1.8228472305879095, + "learning_rate": 2.9231733353590663e-06, + "loss": 0.7274, + "num_input_tokens_seen": 132286030, + "step": 4068 + }, + { + "epoch": 0.366956757000496, + "flos": 23293863688800.0, + "grad_norm": 9.818585300824672, + "learning_rate": 2.922655045585136e-06, + "loss": 0.8348, + "num_input_tokens_seen": 132312705, + "step": 4069 + }, + { + "epoch": 0.36704694052396625, + "flos": 71095521618240.0, + "grad_norm": 0.6915737962037217, + "learning_rate": 2.92213667708519e-06, + "loss": 0.6199, + "num_input_tokens_seen": 132407125, + "step": 4070 + }, + { + "epoch": 0.36713712404743654, + "flos": 24387235508160.0, + "grad_norm": 2.5494334470385587, + "learning_rate": 2.921618229903457e-06, + "loss": 0.7742, + "num_input_tokens_seen": 132434325, + "step": 4071 + }, + { + "epoch": 0.36722730757090677, + "flos": 34664767063200.0, + "grad_norm": 2.001601174283922, + "learning_rate": 2.9210997040841752e-06, + "loss": 0.6849, + "num_input_tokens_seen": 132467885, + "step": 4072 + }, + { + "epoch": 0.36731749109437706, + "flos": 17465285491200.0, + "grad_norm": 1.9145038329489819, + "learning_rate": 2.9205810996715885e-06, + "loss": 0.7301, + "num_input_tokens_seen": 132494665, + "step": 4073 + }, + { + "epoch": 0.36740767461784735, + "flos": 23262523847520.0, + "grad_norm": 1.7431969562066278, + "learning_rate": 2.9200624167099456e-06, + "loss": 0.7274, + "num_input_tokens_seen": 132522040, + "step": 4074 + }, + { + "epoch": 0.3674978581413176, + "flos": 24828249174240.0, + "grad_norm": 1.4865700002356383, + "learning_rate": 2.919543655243505e-06, + "loss": 0.6805, + "num_input_tokens_seen": 132552585, + "step": 4075 + }, + { + "epoch": 0.36758804166478787, + "flos": 22531465845120.0, + "grad_norm": 2.4334633487647794, + "learning_rate": 2.919024815316529e-06, + "loss": 0.7929, + "num_input_tokens_seen": 132580640, + "step": 4076 + }, + { + "epoch": 0.3676782251882581, + "flos": 22350643464480.0, + "grad_norm": 1.8286335555660111, + "learning_rate": 2.9185058969732877e-06, + "loss": 0.8836, + "num_input_tokens_seen": 132607120, + "step": 4077 + }, + { + "epoch": 0.3677684087117284, + "flos": 24828174834720.0, + "grad_norm": 1.7232132268601672, + "learning_rate": 2.917986900258059e-06, + "loss": 0.7989, + "num_input_tokens_seen": 132635590, + "step": 4078 + }, + { + "epoch": 0.3678585922351986, + "flos": 21585941095680.0, + "grad_norm": 7.411084106462201, + "learning_rate": 2.917467825215126e-06, + "loss": 0.7566, + "num_input_tokens_seen": 132666590, + "step": 4079 + }, + { + "epoch": 0.3679487757586689, + "flos": 28253684158560.0, + "grad_norm": 1.5410596408528936, + "learning_rate": 2.9169486718887803e-06, + "loss": 0.7118, + "num_input_tokens_seen": 132698280, + "step": 4080 + }, + { + "epoch": 0.36803895928213914, + "flos": 24972193419360.0, + "grad_norm": 4.284580244160721, + "learning_rate": 2.9164294403233173e-06, + "loss": 0.814, + "num_input_tokens_seen": 132729535, + "step": 4081 + }, + { + "epoch": 0.36812914280560943, + "flos": 15246086918400.0, + "grad_norm": 2.396986730724237, + "learning_rate": 2.915910130563041e-06, + "loss": 0.7007, + "num_input_tokens_seen": 132756925, + "step": 4082 + }, + { + "epoch": 0.36821932632907967, + "flos": 22095953303520.0, + "grad_norm": 2.0245964825628366, + "learning_rate": 2.915390742652262e-06, + "loss": 0.7106, + "num_input_tokens_seen": 132785165, + "step": 4083 + }, + { + "epoch": 0.36830950985254995, + "flos": 18306606202560.0, + "grad_norm": 2.2736344399637676, + "learning_rate": 2.914871276635298e-06, + "loss": 0.6949, + "num_input_tokens_seen": 132811890, + "step": 4084 + }, + { + "epoch": 0.3683996933760202, + "flos": 28362348567840.0, + "grad_norm": 1.9959331529963267, + "learning_rate": 2.914351732556472e-06, + "loss": 0.7424, + "num_input_tokens_seen": 132839970, + "step": 4085 + }, + { + "epoch": 0.3684898768994905, + "flos": 20165237937120.0, + "grad_norm": 1.5999455174468342, + "learning_rate": 2.9138321104601144e-06, + "loss": 0.6844, + "num_input_tokens_seen": 132868160, + "step": 4086 + }, + { + "epoch": 0.3685800604229607, + "flos": 29746136421120.0, + "grad_norm": 2.3272691215953203, + "learning_rate": 2.9133124103905623e-06, + "loss": 0.6963, + "num_input_tokens_seen": 132900785, + "step": 4087 + }, + { + "epoch": 0.368670243946431, + "flos": 23042816164320.0, + "grad_norm": 2.3710987776022687, + "learning_rate": 2.9127926323921596e-06, + "loss": 0.7531, + "num_input_tokens_seen": 132928485, + "step": 4088 + }, + { + "epoch": 0.36876042746990123, + "flos": 24865275988800.0, + "grad_norm": 1.846788392077529, + "learning_rate": 2.912272776509256e-06, + "loss": 0.7813, + "num_input_tokens_seen": 132957840, + "step": 4089 + }, + { + "epoch": 0.3688506109933715, + "flos": 27480692933280.0, + "grad_norm": 1.857709186527093, + "learning_rate": 2.911752842786209e-06, + "loss": 0.7292, + "num_input_tokens_seen": 132986320, + "step": 4090 + }, + { + "epoch": 0.36894079451684175, + "flos": 21221322753600.0, + "grad_norm": 1.9710015032301405, + "learning_rate": 2.911232831267383e-06, + "loss": 0.7011, + "num_input_tokens_seen": 133015580, + "step": 4091 + }, + { + "epoch": 0.36903097804031204, + "flos": 27957060774720.0, + "grad_norm": 1.8036558827505698, + "learning_rate": 2.910712741997146e-06, + "loss": 0.7823, + "num_input_tokens_seen": 133045415, + "step": 4092 + }, + { + "epoch": 0.3691211615637823, + "flos": 27560470705440.0, + "grad_norm": 1.6358500430859655, + "learning_rate": 2.910192575019877e-06, + "loss": 0.7487, + "num_input_tokens_seen": 133074295, + "step": 4093 + }, + { + "epoch": 0.36921134508725256, + "flos": 23844433838400.0, + "grad_norm": 1.8748956651269142, + "learning_rate": 2.9096723303799583e-06, + "loss": 0.7938, + "num_input_tokens_seen": 133105310, + "step": 4094 + }, + { + "epoch": 0.3693015286107228, + "flos": 26171033048640.0, + "grad_norm": 2.385931956066793, + "learning_rate": 2.9091520081217805e-06, + "loss": 0.7699, + "num_input_tokens_seen": 133135345, + "step": 4095 + }, + { + "epoch": 0.3693917121341931, + "flos": 22678904047680.0, + "grad_norm": 1.6186984462703007, + "learning_rate": 2.908631608289741e-06, + "loss": 0.7111, + "num_input_tokens_seen": 133165010, + "step": 4096 + }, + { + "epoch": 0.3694818956576633, + "flos": 31239555097440.0, + "grad_norm": 1.6848018409733436, + "learning_rate": 2.9081111309282423e-06, + "loss": 0.7517, + "num_input_tokens_seen": 133194210, + "step": 4097 + }, + { + "epoch": 0.3695720791811336, + "flos": 68024371782720.0, + "grad_norm": 0.6420544980840045, + "learning_rate": 2.9075905760816942e-06, + "loss": 0.631, + "num_input_tokens_seen": 133290575, + "step": 4098 + }, + { + "epoch": 0.3696622627046039, + "flos": 19690505565120.0, + "grad_norm": 2.0232329394065354, + "learning_rate": 2.907069943794514e-06, + "loss": 0.7995, + "num_input_tokens_seen": 133317080, + "step": 4099 + }, + { + "epoch": 0.3697524462280741, + "flos": 39725743650720.0, + "grad_norm": 1.585783701988833, + "learning_rate": 2.906549234111125e-06, + "loss": 0.7293, + "num_input_tokens_seen": 133350605, + "step": 4100 + }, + { + "epoch": 0.3698426297515444, + "flos": 20964179388480.0, + "grad_norm": 1.483580929026314, + "learning_rate": 2.906028447075956e-06, + "loss": 0.8178, + "num_input_tokens_seen": 133379210, + "step": 4101 + }, + { + "epoch": 0.36993281327501465, + "flos": 25915934020320.0, + "grad_norm": 1.9684450428029847, + "learning_rate": 2.905507582733445e-06, + "loss": 0.845, + "num_input_tokens_seen": 133408165, + "step": 4102 + }, + { + "epoch": 0.37002299679848494, + "flos": 21476496121440.0, + "grad_norm": 1.6835043190837549, + "learning_rate": 2.904986641128033e-06, + "loss": 0.8782, + "num_input_tokens_seen": 133432500, + "step": 4103 + }, + { + "epoch": 0.37011318032195517, + "flos": 27268679390400.0, + "grad_norm": 1.424469979748807, + "learning_rate": 2.9044656223041716e-06, + "loss": 0.7512, + "num_input_tokens_seen": 133464040, + "step": 4104 + }, + { + "epoch": 0.37020336384542546, + "flos": 23516136085440.0, + "grad_norm": 1.5686702966390023, + "learning_rate": 2.9039445263063157e-06, + "loss": 0.7204, + "num_input_tokens_seen": 133492220, + "step": 4105 + }, + { + "epoch": 0.3702935473688957, + "flos": 25520161685760.0, + "grad_norm": 2.678470346269211, + "learning_rate": 2.903423353178929e-06, + "loss": 0.6578, + "num_input_tokens_seen": 133520190, + "step": 4106 + }, + { + "epoch": 0.370383730892366, + "flos": 28105316712000.0, + "grad_norm": 1.841953307352131, + "learning_rate": 2.9029021029664802e-06, + "loss": 0.7581, + "num_input_tokens_seen": 133551510, + "step": 4107 + }, + { + "epoch": 0.3704739144158362, + "flos": 24172880270400.0, + "grad_norm": 1.4620256094718262, + "learning_rate": 2.9023807757134455e-06, + "loss": 0.8154, + "num_input_tokens_seen": 133582195, + "step": 4108 + }, + { + "epoch": 0.3705640979393065, + "flos": 20015012002560.0, + "grad_norm": 1.6492622084247432, + "learning_rate": 2.901859371464307e-06, + "loss": 0.8277, + "num_input_tokens_seen": 133609775, + "step": 4109 + }, + { + "epoch": 0.37065428146277674, + "flos": 23225013826080.0, + "grad_norm": 1.76529454202376, + "learning_rate": 2.9013378902635535e-06, + "loss": 0.7997, + "num_input_tokens_seen": 133638265, + "step": 4110 + }, + { + "epoch": 0.370744464986247, + "flos": 22569087375840.0, + "grad_norm": 1.8675059282446222, + "learning_rate": 2.9008163321556823e-06, + "loss": 0.731, + "num_input_tokens_seen": 133667005, + "step": 4111 + }, + { + "epoch": 0.37083464850971726, + "flos": 21324300189600.0, + "grad_norm": 4.1049541873210815, + "learning_rate": 2.900294697185194e-06, + "loss": 0.7767, + "num_input_tokens_seen": 133691095, + "step": 4112 + }, + { + "epoch": 0.37092483203318755, + "flos": 27014695454880.0, + "grad_norm": 1.6865046355983655, + "learning_rate": 2.899772985396599e-06, + "loss": 0.7806, + "num_input_tokens_seen": 133722205, + "step": 4113 + }, + { + "epoch": 0.3710150155566578, + "flos": 26833055339520.0, + "grad_norm": 1.6484484920767992, + "learning_rate": 2.8992511968344104e-06, + "loss": 0.7512, + "num_input_tokens_seen": 133750050, + "step": 4114 + }, + { + "epoch": 0.37110519908012807, + "flos": 24712113643200.0, + "grad_norm": 1.702188987051964, + "learning_rate": 2.8987293315431523e-06, + "loss": 0.7049, + "num_input_tokens_seen": 133778895, + "step": 4115 + }, + { + "epoch": 0.3711953826035983, + "flos": 27851741643840.0, + "grad_norm": 3.287281705041535, + "learning_rate": 2.898207389567351e-06, + "loss": 0.693, + "num_input_tokens_seen": 133809640, + "step": 4116 + }, + { + "epoch": 0.3712855661270686, + "flos": 22239079813920.0, + "grad_norm": 1.8462916813344248, + "learning_rate": 2.897685370951543e-06, + "loss": 0.7529, + "num_input_tokens_seen": 133838200, + "step": 4117 + }, + { + "epoch": 0.3713757496505388, + "flos": 15829781057760.0, + "grad_norm": 3.2716674389079565, + "learning_rate": 2.89716327574027e-06, + "loss": 0.6522, + "num_input_tokens_seen": 133864560, + "step": 4118 + }, + { + "epoch": 0.3714659331740091, + "flos": 30539688257280.0, + "grad_norm": 2.023780116233556, + "learning_rate": 2.8966411039780787e-06, + "loss": 0.8481, + "num_input_tokens_seen": 133891405, + "step": 4119 + }, + { + "epoch": 0.37155611669747934, + "flos": 23618778993600.0, + "grad_norm": 1.8043242910597723, + "learning_rate": 2.8961188557095248e-06, + "loss": 0.7682, + "num_input_tokens_seen": 133918560, + "step": 4120 + }, + { + "epoch": 0.37164630022094963, + "flos": 17431975652640.0, + "grad_norm": 1.8623386774026833, + "learning_rate": 2.895596530979168e-06, + "loss": 0.705, + "num_input_tokens_seen": 133945555, + "step": 4121 + }, + { + "epoch": 0.3717364837444199, + "flos": 21398985704640.0, + "grad_norm": 1.9772033858621985, + "learning_rate": 2.895074129831578e-06, + "loss": 0.7762, + "num_input_tokens_seen": 133973450, + "step": 4122 + }, + { + "epoch": 0.37182666726789015, + "flos": 32879296883520.0, + "grad_norm": 1.7044468221160163, + "learning_rate": 2.8945516523113275e-06, + "loss": 0.678, + "num_input_tokens_seen": 134002515, + "step": 4123 + }, + { + "epoch": 0.37191685079136044, + "flos": 69369757540320.0, + "grad_norm": 0.6796944058823804, + "learning_rate": 2.894029098462998e-06, + "loss": 0.5509, + "num_input_tokens_seen": 134088330, + "step": 4124 + }, + { + "epoch": 0.3720070343148307, + "flos": 22715038788000.0, + "grad_norm": 1.7630842714902952, + "learning_rate": 2.8935064683311756e-06, + "loss": 0.7839, + "num_input_tokens_seen": 134116410, + "step": 4125 + }, + { + "epoch": 0.37209721783830096, + "flos": 65160434857440.0, + "grad_norm": 0.677547088162044, + "learning_rate": 2.8929837619604544e-06, + "loss": 0.6552, + "num_input_tokens_seen": 134207430, + "step": 4126 + }, + { + "epoch": 0.3721874013617712, + "flos": 24346603226880.0, + "grad_norm": 1.9047382801725254, + "learning_rate": 2.8924609793954346e-06, + "loss": 0.8052, + "num_input_tokens_seen": 134235290, + "step": 4127 + }, + { + "epoch": 0.3722775848852415, + "flos": 22675224241440.0, + "grad_norm": 1.8415174383518023, + "learning_rate": 2.891938120680724e-06, + "loss": 0.7149, + "num_input_tokens_seen": 134264065, + "step": 4128 + }, + { + "epoch": 0.3723677684087117, + "flos": 22930546288320.0, + "grad_norm": 1.5503302622236828, + "learning_rate": 2.8914151858609343e-06, + "loss": 0.7517, + "num_input_tokens_seen": 134294705, + "step": 4129 + }, + { + "epoch": 0.372457951932182, + "flos": 28762841461920.0, + "grad_norm": 1.8944885878331206, + "learning_rate": 2.8908921749806858e-06, + "loss": 0.7814, + "num_input_tokens_seen": 134324600, + "step": 4130 + }, + { + "epoch": 0.37254813545565224, + "flos": 21179724058560.0, + "grad_norm": 1.7606307431886687, + "learning_rate": 2.890369088084605e-06, + "loss": 0.7748, + "num_input_tokens_seen": 134351565, + "step": 4131 + }, + { + "epoch": 0.37263831897912253, + "flos": 27743263083360.0, + "grad_norm": 1.6234395531062196, + "learning_rate": 2.889845925217323e-06, + "loss": 0.7843, + "num_input_tokens_seen": 134380570, + "step": 4132 + }, + { + "epoch": 0.37272850250259276, + "flos": 28252420386720.0, + "grad_norm": 2.0494789857068563, + "learning_rate": 2.8893226864234813e-06, + "loss": 0.777, + "num_input_tokens_seen": 134411955, + "step": 4133 + }, + { + "epoch": 0.37281868602606305, + "flos": 21112286646720.0, + "grad_norm": 1.9235551514578606, + "learning_rate": 2.8887993717477236e-06, + "loss": 0.7719, + "num_input_tokens_seen": 134439255, + "step": 4134 + }, + { + "epoch": 0.3729088695495333, + "flos": 14117212244640.0, + "grad_norm": 1.8272157021582716, + "learning_rate": 2.8882759812347035e-06, + "loss": 0.651, + "num_input_tokens_seen": 134466390, + "step": 4135 + }, + { + "epoch": 0.3729990530730036, + "flos": 22569421903680.0, + "grad_norm": 1.7392516438601304, + "learning_rate": 2.887752514929078e-06, + "loss": 0.7651, + "num_input_tokens_seen": 134497080, + "step": 4136 + }, + { + "epoch": 0.3730892365964738, + "flos": 38016891813600.0, + "grad_norm": 2.002088290978154, + "learning_rate": 2.887228972875513e-06, + "loss": 0.7159, + "num_input_tokens_seen": 134526215, + "step": 4137 + }, + { + "epoch": 0.3731794201199441, + "flos": 23223341186880.0, + "grad_norm": 2.0658855972152113, + "learning_rate": 2.88670535511868e-06, + "loss": 0.7633, + "num_input_tokens_seen": 134551065, + "step": 4138 + }, + { + "epoch": 0.3732696036434143, + "flos": 22091901799680.0, + "grad_norm": 2.7078359489504202, + "learning_rate": 2.886181661703257e-06, + "loss": 0.7079, + "num_input_tokens_seen": 134578355, + "step": 4139 + }, + { + "epoch": 0.3733597871668846, + "flos": 25484287133760.0, + "grad_norm": 1.8125607052864494, + "learning_rate": 2.8856578926739285e-06, + "loss": 0.7259, + "num_input_tokens_seen": 134608015, + "step": 4140 + }, + { + "epoch": 0.37344997069035485, + "flos": 27485748020640.0, + "grad_norm": 1.9151740136030384, + "learning_rate": 2.8851340480753846e-06, + "loss": 0.719, + "num_input_tokens_seen": 134639310, + "step": 4141 + }, + { + "epoch": 0.37354015421382514, + "flos": 29054075230560.0, + "grad_norm": 2.664784442708009, + "learning_rate": 2.8846101279523232e-06, + "loss": 0.7635, + "num_input_tokens_seen": 134668435, + "step": 4142 + }, + { + "epoch": 0.37363033773729537, + "flos": 26429328676320.0, + "grad_norm": 8.31185428721501, + "learning_rate": 2.8840861323494487e-06, + "loss": 0.7378, + "num_input_tokens_seen": 134700100, + "step": 4143 + }, + { + "epoch": 0.37372052126076566, + "flos": 22310866087680.0, + "grad_norm": 1.7277081861242531, + "learning_rate": 2.88356206131147e-06, + "loss": 0.8191, + "num_input_tokens_seen": 134727695, + "step": 4144 + }, + { + "epoch": 0.37381070478423595, + "flos": 27342770189280.0, + "grad_norm": 1.8925330882033364, + "learning_rate": 2.883037914883104e-06, + "loss": 0.7491, + "num_input_tokens_seen": 134757175, + "step": 4145 + }, + { + "epoch": 0.3739008883077062, + "flos": 19691806506720.0, + "grad_norm": 1.7033895584551422, + "learning_rate": 2.882513693109075e-06, + "loss": 0.7296, + "num_input_tokens_seen": 134784945, + "step": 4146 + }, + { + "epoch": 0.37399107183117647, + "flos": 25663808572800.0, + "grad_norm": 1.3804497657597443, + "learning_rate": 2.8819893960341106e-06, + "loss": 0.7383, + "num_input_tokens_seen": 134815260, + "step": 4147 + }, + { + "epoch": 0.3740812553546467, + "flos": 19756530526080.0, + "grad_norm": 2.442525427070012, + "learning_rate": 2.881465023702948e-06, + "loss": 0.8212, + "num_input_tokens_seen": 134838675, + "step": 4148 + }, + { + "epoch": 0.374171438878117, + "flos": 24643672647840.0, + "grad_norm": 2.4354777628400908, + "learning_rate": 2.8809405761603294e-06, + "loss": 0.8441, + "num_input_tokens_seen": 134868015, + "step": 4149 + }, + { + "epoch": 0.3742616224015872, + "flos": 16448197486560.0, + "grad_norm": 2.0176277560751785, + "learning_rate": 2.880416053451003e-06, + "loss": 0.6816, + "num_input_tokens_seen": 134893110, + "step": 4150 + }, + { + "epoch": 0.3743518059250575, + "flos": 20747631134880.0, + "grad_norm": 1.7309997295068698, + "learning_rate": 2.879891455619725e-06, + "loss": 0.852, + "num_input_tokens_seen": 134921205, + "step": 4151 + }, + { + "epoch": 0.37444198944852775, + "flos": 36484550664960.0, + "grad_norm": 1.883186208272747, + "learning_rate": 2.879366782711256e-06, + "loss": 0.7683, + "num_input_tokens_seen": 134953875, + "step": 4152 + }, + { + "epoch": 0.37453217297199803, + "flos": 24792188773440.0, + "grad_norm": 2.1080114407889416, + "learning_rate": 2.8788420347703643e-06, + "loss": 0.7854, + "num_input_tokens_seen": 134983000, + "step": 4153 + }, + { + "epoch": 0.37462235649546827, + "flos": 62273415511200.0, + "grad_norm": 0.6240268999699152, + "learning_rate": 2.8783172118418244e-06, + "loss": 0.5764, + "num_input_tokens_seen": 135072025, + "step": 4154 + }, + { + "epoch": 0.37471254001893856, + "flos": 30323771889600.0, + "grad_norm": 1.6161666817958995, + "learning_rate": 2.877792313970417e-06, + "loss": 0.7694, + "num_input_tokens_seen": 135101040, + "step": 4155 + }, + { + "epoch": 0.3748027235424088, + "flos": 26794764753120.0, + "grad_norm": 1.7428738050535377, + "learning_rate": 2.8772673412009293e-06, + "loss": 0.8191, + "num_input_tokens_seen": 135131545, + "step": 4156 + }, + { + "epoch": 0.3748929070658791, + "flos": 14728715098080.0, + "grad_norm": 1.7694943994722883, + "learning_rate": 2.8767422935781545e-06, + "loss": 0.8122, + "num_input_tokens_seen": 135158430, + "step": 4157 + }, + { + "epoch": 0.3749830905893493, + "flos": 18197718774720.0, + "grad_norm": 1.4425945397772266, + "learning_rate": 2.8762171711468935e-06, + "loss": 0.799, + "num_input_tokens_seen": 135186975, + "step": 4158 + }, + { + "epoch": 0.3750732741128196, + "flos": 18706244192160.0, + "grad_norm": 1.594060082283273, + "learning_rate": 2.875691973951952e-06, + "loss": 0.7662, + "num_input_tokens_seen": 135214700, + "step": 4159 + }, + { + "epoch": 0.37516345763628983, + "flos": 67167434038560.0, + "grad_norm": 0.6373703553635959, + "learning_rate": 2.8751667020381425e-06, + "loss": 0.6476, + "num_input_tokens_seen": 135303965, + "step": 4160 + }, + { + "epoch": 0.3752536411597601, + "flos": 24244108997760.0, + "grad_norm": 2.924420069400012, + "learning_rate": 2.8746413554502837e-06, + "loss": 0.8306, + "num_input_tokens_seen": 135333655, + "step": 4161 + }, + { + "epoch": 0.37534382468323035, + "flos": 24500843495520.0, + "grad_norm": 1.7292314846651888, + "learning_rate": 2.8741159342332027e-06, + "loss": 0.6965, + "num_input_tokens_seen": 135362070, + "step": 4162 + }, + { + "epoch": 0.37543400820670064, + "flos": 24682186252800.0, + "grad_norm": 1.8123923302602036, + "learning_rate": 2.87359043843173e-06, + "loss": 0.8208, + "num_input_tokens_seen": 135392310, + "step": 4163 + }, + { + "epoch": 0.3755241917301709, + "flos": 25266437938560.0, + "grad_norm": 1.5719997773594978, + "learning_rate": 2.873064868090704e-06, + "loss": 0.8018, + "num_input_tokens_seen": 135422945, + "step": 4164 + }, + { + "epoch": 0.37561437525364116, + "flos": 17031259740000.0, + "grad_norm": 2.339190511315823, + "learning_rate": 2.8725392232549697e-06, + "loss": 0.8549, + "num_input_tokens_seen": 135450170, + "step": 4165 + }, + { + "epoch": 0.3757045587771114, + "flos": 22313690989440.0, + "grad_norm": 2.0900962035382884, + "learning_rate": 2.872013503969378e-06, + "loss": 0.8353, + "num_input_tokens_seen": 135473635, + "step": 4166 + }, + { + "epoch": 0.3757947423005817, + "flos": 26029727856480.0, + "grad_norm": 1.728976396257247, + "learning_rate": 2.8714877102787853e-06, + "loss": 0.7764, + "num_input_tokens_seen": 135504710, + "step": 4167 + }, + { + "epoch": 0.3758849258240519, + "flos": 24278794117440.0, + "grad_norm": 2.2382891323623206, + "learning_rate": 2.8709618422280564e-06, + "loss": 0.7744, + "num_input_tokens_seen": 135533280, + "step": 4168 + }, + { + "epoch": 0.3759751093475222, + "flos": 27082021357440.0, + "grad_norm": 1.6897330168034126, + "learning_rate": 2.8704358998620605e-06, + "loss": 0.7416, + "num_input_tokens_seen": 135560770, + "step": 4169 + }, + { + "epoch": 0.3760652928709925, + "flos": 21257792021760.0, + "grad_norm": 1.6142776035727202, + "learning_rate": 2.8699098832256735e-06, + "loss": 0.7945, + "num_input_tokens_seen": 135588325, + "step": 4170 + }, + { + "epoch": 0.37615547639446273, + "flos": 16230125272800.0, + "grad_norm": 2.1619897273452158, + "learning_rate": 2.86938379236378e-06, + "loss": 0.6788, + "num_input_tokens_seen": 135615785, + "step": 4171 + }, + { + "epoch": 0.376245659917933, + "flos": 43408098981600.0, + "grad_norm": 1.6877186245053764, + "learning_rate": 2.868857627321266e-06, + "loss": 0.6913, + "num_input_tokens_seen": 135645845, + "step": 4172 + }, + { + "epoch": 0.37633584344140325, + "flos": 25447669186560.0, + "grad_norm": 2.1070301727756098, + "learning_rate": 2.8683313881430296e-06, + "loss": 0.7704, + "num_input_tokens_seen": 135675095, + "step": 4173 + }, + { + "epoch": 0.37642602696487354, + "flos": 16262543037120.0, + "grad_norm": 1.934661878472539, + "learning_rate": 2.8678050748739706e-06, + "loss": 0.7605, + "num_input_tokens_seen": 135699865, + "step": 4174 + }, + { + "epoch": 0.37651621048834377, + "flos": 22601951177280.0, + "grad_norm": 1.591108564157833, + "learning_rate": 2.8672786875589976e-06, + "loss": 0.7386, + "num_input_tokens_seen": 135728690, + "step": 4175 + }, + { + "epoch": 0.37660639401181406, + "flos": 23298138211200.0, + "grad_norm": 1.9461993911229456, + "learning_rate": 2.866752226243025e-06, + "loss": 0.7203, + "num_input_tokens_seen": 135756100, + "step": 4176 + }, + { + "epoch": 0.3766965775352843, + "flos": 24864644102880.0, + "grad_norm": 5.722499264591648, + "learning_rate": 2.8662256909709733e-06, + "loss": 0.8075, + "num_input_tokens_seen": 135783835, + "step": 4177 + }, + { + "epoch": 0.3767867610587546, + "flos": 19028966481120.0, + "grad_norm": 1.9118262481708923, + "learning_rate": 2.865699081787769e-06, + "loss": 0.8047, + "num_input_tokens_seen": 135812010, + "step": 4178 + }, + { + "epoch": 0.3768769445822248, + "flos": 39510570678240.0, + "grad_norm": 1.6679278688088752, + "learning_rate": 2.8651723987383465e-06, + "loss": 0.7485, + "num_input_tokens_seen": 135842690, + "step": 4179 + }, + { + "epoch": 0.3769671281056951, + "flos": 26103409788000.0, + "grad_norm": 1.9081525844725502, + "learning_rate": 2.8646456418676437e-06, + "loss": 0.744, + "num_input_tokens_seen": 135870725, + "step": 4180 + }, + { + "epoch": 0.37705731162916534, + "flos": 59044377036960.0, + "grad_norm": 0.7253477545063101, + "learning_rate": 2.8641188112206067e-06, + "loss": 0.5961, + "num_input_tokens_seen": 135962540, + "step": 4181 + }, + { + "epoch": 0.3771474951526356, + "flos": 16449052391040.0, + "grad_norm": 2.2928266075046535, + "learning_rate": 2.863591906842189e-06, + "loss": 0.7933, + "num_input_tokens_seen": 135990415, + "step": 4182 + }, + { + "epoch": 0.37723767867610586, + "flos": 28363389321120.0, + "grad_norm": 1.885275215477051, + "learning_rate": 2.863064928777347e-06, + "loss": 0.7986, + "num_input_tokens_seen": 136022955, + "step": 4183 + }, + { + "epoch": 0.37732786219957615, + "flos": 25409675958240.0, + "grad_norm": 1.946553703725491, + "learning_rate": 2.862537877071047e-06, + "loss": 0.7424, + "num_input_tokens_seen": 136051355, + "step": 4184 + }, + { + "epoch": 0.3774180457230464, + "flos": 69780732306720.0, + "grad_norm": 0.6448858268362797, + "learning_rate": 2.8620107517682597e-06, + "loss": 0.5602, + "num_input_tokens_seen": 136139255, + "step": 4185 + }, + { + "epoch": 0.37750822924651667, + "flos": 37288584373440.0, + "grad_norm": 1.5102195698842849, + "learning_rate": 2.8614835529139618e-06, + "loss": 0.6993, + "num_input_tokens_seen": 136174330, + "step": 4186 + }, + { + "epoch": 0.3775984127699869, + "flos": 18634383578880.0, + "grad_norm": 2.4204985614182126, + "learning_rate": 2.8609562805531367e-06, + "loss": 0.7248, + "num_input_tokens_seen": 136199530, + "step": 4187 + }, + { + "epoch": 0.3776885962934572, + "flos": 19321017984480.0, + "grad_norm": 1.9504828319913678, + "learning_rate": 2.8604289347307746e-06, + "loss": 0.7977, + "num_input_tokens_seen": 136227085, + "step": 4188 + }, + { + "epoch": 0.3777787798169274, + "flos": 20420039607360.0, + "grad_norm": 2.148688839169036, + "learning_rate": 2.859901515491871e-06, + "loss": 0.7603, + "num_input_tokens_seen": 136253815, + "step": 4189 + }, + { + "epoch": 0.3778689633403977, + "flos": 26284901224320.0, + "grad_norm": 1.549025764699267, + "learning_rate": 2.8593740228814298e-06, + "loss": 0.6515, + "num_input_tokens_seen": 136284455, + "step": 4190 + }, + { + "epoch": 0.37795914686386795, + "flos": 15282370337760.0, + "grad_norm": 2.2001182654410854, + "learning_rate": 2.8588464569444574e-06, + "loss": 0.7511, + "num_input_tokens_seen": 136309115, + "step": 4191 + }, + { + "epoch": 0.37804933038733823, + "flos": 22275028705440.0, + "grad_norm": 1.7555183189737906, + "learning_rate": 2.8583188177259697e-06, + "loss": 0.7566, + "num_input_tokens_seen": 136337555, + "step": 4192 + }, + { + "epoch": 0.3781395139108085, + "flos": 21876542978400.0, + "grad_norm": 1.81659886149054, + "learning_rate": 2.857791105270988e-06, + "loss": 0.7486, + "num_input_tokens_seen": 136364520, + "step": 4193 + }, + { + "epoch": 0.37822969743427876, + "flos": 22315029100800.0, + "grad_norm": 1.649322416877444, + "learning_rate": 2.857263319624539e-06, + "loss": 0.7966, + "num_input_tokens_seen": 136393380, + "step": 4194 + }, + { + "epoch": 0.37831988095774904, + "flos": 22023237785760.0, + "grad_norm": 1.6173854835689965, + "learning_rate": 2.856735460831657e-06, + "loss": 0.7484, + "num_input_tokens_seen": 136425695, + "step": 4195 + }, + { + "epoch": 0.3784100644812193, + "flos": 19252093782240.0, + "grad_norm": 1.5034041718472158, + "learning_rate": 2.856207528937382e-06, + "loss": 0.7738, + "num_input_tokens_seen": 136452940, + "step": 4196 + }, + { + "epoch": 0.37850024800468957, + "flos": 20598074256000.0, + "grad_norm": 4.2661338195271465, + "learning_rate": 2.855679523986759e-06, + "loss": 0.7496, + "num_input_tokens_seen": 136482840, + "step": 4197 + }, + { + "epoch": 0.3785904315281598, + "flos": 21290321295360.0, + "grad_norm": 4.0196662487199895, + "learning_rate": 2.8551514460248406e-06, + "loss": 0.814, + "num_input_tokens_seen": 136510265, + "step": 4198 + }, + { + "epoch": 0.3786806150516301, + "flos": 26465128888800.0, + "grad_norm": 1.8910708571643755, + "learning_rate": 2.8546232950966868e-06, + "loss": 0.7702, + "num_input_tokens_seen": 136540505, + "step": 4199 + }, + { + "epoch": 0.3787707985751003, + "flos": 25884445500000.0, + "grad_norm": 6.432472710371785, + "learning_rate": 2.85409507124736e-06, + "loss": 0.7145, + "num_input_tokens_seen": 136569160, + "step": 4200 + }, + { + "epoch": 0.3788609820985706, + "flos": 28580532290880.0, + "grad_norm": 3.40881796813505, + "learning_rate": 2.8535667745219324e-06, + "loss": 0.7569, + "num_input_tokens_seen": 136598170, + "step": 4201 + }, + { + "epoch": 0.37895116562204084, + "flos": 32224262507520.0, + "grad_norm": 1.7258230153353589, + "learning_rate": 2.853038404965481e-06, + "loss": 0.7563, + "num_input_tokens_seen": 136628305, + "step": 4202 + }, + { + "epoch": 0.37904134914551113, + "flos": 16297934382240.0, + "grad_norm": 1.952860944064325, + "learning_rate": 2.8525099626230894e-06, + "loss": 0.8368, + "num_input_tokens_seen": 136652740, + "step": 4203 + }, + { + "epoch": 0.37913153266898136, + "flos": 21914610546240.0, + "grad_norm": 2.7065036364096993, + "learning_rate": 2.8519814475398472e-06, + "loss": 0.6959, + "num_input_tokens_seen": 136681770, + "step": 4204 + }, + { + "epoch": 0.37922171619245165, + "flos": 26866328008320.0, + "grad_norm": 1.9114781585690812, + "learning_rate": 2.8514528597608502e-06, + "loss": 0.7424, + "num_input_tokens_seen": 136713545, + "step": 4205 + }, + { + "epoch": 0.3793118997159219, + "flos": 27377975685600.0, + "grad_norm": 1.6896213337505075, + "learning_rate": 2.8509241993312004e-06, + "loss": 0.7957, + "num_input_tokens_seen": 136745000, + "step": 4206 + }, + { + "epoch": 0.3794020832393922, + "flos": 22200937906560.0, + "grad_norm": 2.470827315153241, + "learning_rate": 2.850395466296006e-06, + "loss": 0.8178, + "num_input_tokens_seen": 136772525, + "step": 4207 + }, + { + "epoch": 0.3794922667628624, + "flos": 24059012094720.0, + "grad_norm": 2.115745250265205, + "learning_rate": 2.849866660700381e-06, + "loss": 0.7949, + "num_input_tokens_seen": 136801630, + "step": 4208 + }, + { + "epoch": 0.3795824502863327, + "flos": 30760176505440.0, + "grad_norm": 1.79866979774097, + "learning_rate": 2.8493377825894464e-06, + "loss": 0.8263, + "num_input_tokens_seen": 136831685, + "step": 4209 + }, + { + "epoch": 0.37967263380980293, + "flos": 25771506568320.0, + "grad_norm": 2.170086350546758, + "learning_rate": 2.848808832008329e-06, + "loss": 0.8256, + "num_input_tokens_seen": 136861035, + "step": 4210 + }, + { + "epoch": 0.3797628173332732, + "flos": 27924234143040.0, + "grad_norm": 1.7784739495515245, + "learning_rate": 2.848279809002162e-06, + "loss": 0.7941, + "num_input_tokens_seen": 136891870, + "step": 4211 + }, + { + "epoch": 0.37985300085674345, + "flos": 22315437968160.0, + "grad_norm": 2.1336178778586996, + "learning_rate": 2.8477507136160842e-06, + "loss": 0.782, + "num_input_tokens_seen": 136922305, + "step": 4212 + }, + { + "epoch": 0.37994318438021374, + "flos": 71144925962880.0, + "grad_norm": 0.6329129221790574, + "learning_rate": 2.847221545895241e-06, + "loss": 0.5884, + "num_input_tokens_seen": 137013925, + "step": 4213 + }, + { + "epoch": 0.38003336790368397, + "flos": 23115457342560.0, + "grad_norm": 1.499056439859635, + "learning_rate": 2.846692305884785e-06, + "loss": 0.7601, + "num_input_tokens_seen": 137044380, + "step": 4214 + }, + { + "epoch": 0.38012355142715426, + "flos": 23335202195520.0, + "grad_norm": 1.5502891240109806, + "learning_rate": 2.8461629936298718e-06, + "loss": 0.8078, + "num_input_tokens_seen": 137074985, + "step": 4215 + }, + { + "epoch": 0.3802137349506245, + "flos": 21767915738880.0, + "grad_norm": 1.8256050413321212, + "learning_rate": 2.845633609175666e-06, + "loss": 0.8201, + "num_input_tokens_seen": 137102360, + "step": 4216 + }, + { + "epoch": 0.3803039184740948, + "flos": 23954027491680.0, + "grad_norm": 1.8915471752030604, + "learning_rate": 2.8451041525673383e-06, + "loss": 0.7999, + "num_input_tokens_seen": 137132580, + "step": 4217 + }, + { + "epoch": 0.38039410199756507, + "flos": 19245849262560.0, + "grad_norm": 1.9050991774446464, + "learning_rate": 2.8445746238500647e-06, + "loss": 0.7975, + "num_input_tokens_seen": 137155350, + "step": 4218 + }, + { + "epoch": 0.3804842855210353, + "flos": 20638334839680.0, + "grad_norm": 1.8710358357874013, + "learning_rate": 2.844045023069027e-06, + "loss": 0.7471, + "num_input_tokens_seen": 137182935, + "step": 4219 + }, + { + "epoch": 0.3805744690445056, + "flos": 20820532501440.0, + "grad_norm": 1.6994273792091883, + "learning_rate": 2.8435153502694136e-06, + "loss": 0.8052, + "num_input_tokens_seen": 137212940, + "step": 4220 + }, + { + "epoch": 0.3806646525679758, + "flos": 17796222297120.0, + "grad_norm": 1.8577952365698778, + "learning_rate": 2.84298560549642e-06, + "loss": 0.7023, + "num_input_tokens_seen": 137239370, + "step": 4221 + }, + { + "epoch": 0.3807548360914461, + "flos": 23216724969600.0, + "grad_norm": 2.5444645620806603, + "learning_rate": 2.8424557887952462e-06, + "loss": 0.7939, + "num_input_tokens_seen": 137264210, + "step": 4222 + }, + { + "epoch": 0.38084501961491635, + "flos": 20565656491680.0, + "grad_norm": 4.470215243761929, + "learning_rate": 2.841925900211099e-06, + "loss": 0.6956, + "num_input_tokens_seen": 137292310, + "step": 4223 + }, + { + "epoch": 0.38093520313838664, + "flos": 25193239213920.0, + "grad_norm": 1.4380070994365033, + "learning_rate": 2.841395939789192e-06, + "loss": 0.8208, + "num_input_tokens_seen": 137323430, + "step": 4224 + }, + { + "epoch": 0.38102538666185687, + "flos": 22861584916320.0, + "grad_norm": 1.871139430243048, + "learning_rate": 2.8408659075747435e-06, + "loss": 0.7859, + "num_input_tokens_seen": 137352570, + "step": 4225 + }, + { + "epoch": 0.38111557018532716, + "flos": 18197830284000.0, + "grad_norm": 1.9892673407931782, + "learning_rate": 2.8403358036129796e-06, + "loss": 0.7667, + "num_input_tokens_seen": 137381350, + "step": 4226 + }, + { + "epoch": 0.3812057537087974, + "flos": 15246793143840.0, + "grad_norm": 1.7442792984703992, + "learning_rate": 2.839805627949132e-06, + "loss": 0.8203, + "num_input_tokens_seen": 137407925, + "step": 4227 + }, + { + "epoch": 0.3812959372322677, + "flos": 18124854577920.0, + "grad_norm": 1.905128586902324, + "learning_rate": 2.8392753806284367e-06, + "loss": 0.8222, + "num_input_tokens_seen": 137434410, + "step": 4228 + }, + { + "epoch": 0.3813861207557379, + "flos": 26322039548160.0, + "grad_norm": 4.3358631574426285, + "learning_rate": 2.838745061696139e-06, + "loss": 0.7178, + "num_input_tokens_seen": 137462780, + "step": 4229 + }, + { + "epoch": 0.3814763042792082, + "flos": 24423481757760.0, + "grad_norm": 1.6803620159564394, + "learning_rate": 2.838214671197487e-06, + "loss": 0.708, + "num_input_tokens_seen": 137493365, + "step": 4230 + }, + { + "epoch": 0.38156648780267843, + "flos": 22096436510400.0, + "grad_norm": 1.4969114454447665, + "learning_rate": 2.8376842091777377e-06, + "loss": 0.8748, + "num_input_tokens_seen": 137522585, + "step": 4231 + }, + { + "epoch": 0.3816566713261487, + "flos": 25225582638720.0, + "grad_norm": 1.7375652767516103, + "learning_rate": 2.8371536756821524e-06, + "loss": 0.678, + "num_input_tokens_seen": 137551455, + "step": 4232 + }, + { + "epoch": 0.38174685484961896, + "flos": 23844619687200.0, + "grad_norm": 1.7406370508403464, + "learning_rate": 2.836623070756e-06, + "loss": 0.7503, + "num_input_tokens_seen": 137582180, + "step": 4233 + }, + { + "epoch": 0.38183703837308924, + "flos": 18780372160800.0, + "grad_norm": 1.7384333511843386, + "learning_rate": 2.8360923944445542e-06, + "loss": 0.7957, + "num_input_tokens_seen": 137608005, + "step": 4234 + }, + { + "epoch": 0.3819272218965595, + "flos": 23587327643040.0, + "grad_norm": 1.6560524876702867, + "learning_rate": 2.8355616467930947e-06, + "loss": 0.7545, + "num_input_tokens_seen": 137637570, + "step": 4235 + }, + { + "epoch": 0.38201740542002977, + "flos": 20055718623360.0, + "grad_norm": 2.5234079270652483, + "learning_rate": 2.8350308278469085e-06, + "loss": 0.7494, + "num_input_tokens_seen": 137664670, + "step": 4236 + }, + { + "epoch": 0.3821075889435, + "flos": 24099532866720.0, + "grad_norm": 1.8339972449877897, + "learning_rate": 2.8344999376512877e-06, + "loss": 0.8083, + "num_input_tokens_seen": 137693060, + "step": 4237 + }, + { + "epoch": 0.3821977724669703, + "flos": 23334384460800.0, + "grad_norm": 1.5065436682387612, + "learning_rate": 2.8339689762515307e-06, + "loss": 0.761, + "num_input_tokens_seen": 137721640, + "step": 4238 + }, + { + "epoch": 0.3822879559904405, + "flos": 25957532715360.0, + "grad_norm": 2.3871533486722014, + "learning_rate": 2.8334379436929424e-06, + "loss": 0.7044, + "num_input_tokens_seen": 137748630, + "step": 4239 + }, + { + "epoch": 0.3823781395139108, + "flos": 24500546137440.0, + "grad_norm": 1.7473029061256107, + "learning_rate": 2.832906840020833e-06, + "loss": 0.7404, + "num_input_tokens_seen": 137779155, + "step": 4240 + }, + { + "epoch": 0.3824683230373811, + "flos": 22059707053920.0, + "grad_norm": 1.54123839882153, + "learning_rate": 2.83237566528052e-06, + "loss": 0.7839, + "num_input_tokens_seen": 137806835, + "step": 4241 + }, + { + "epoch": 0.38255850656085133, + "flos": 28835742828480.0, + "grad_norm": 1.7629071137468413, + "learning_rate": 2.831844419517325e-06, + "loss": 0.7488, + "num_input_tokens_seen": 137835255, + "step": 4242 + }, + { + "epoch": 0.3826486900843216, + "flos": 21148607235840.0, + "grad_norm": 1.5041488649929875, + "learning_rate": 2.8313131027765774e-06, + "loss": 0.7985, + "num_input_tokens_seen": 137863695, + "step": 4243 + }, + { + "epoch": 0.38273887360779185, + "flos": 20602014250560.0, + "grad_norm": 1.6236982266554019, + "learning_rate": 2.8307817151036124e-06, + "loss": 0.7433, + "num_input_tokens_seen": 137890905, + "step": 4244 + }, + { + "epoch": 0.38282905713126214, + "flos": 26539888743360.0, + "grad_norm": 1.675077775316712, + "learning_rate": 2.8302502565437704e-06, + "loss": 0.7455, + "num_input_tokens_seen": 137919385, + "step": 4245 + }, + { + "epoch": 0.3829192406547324, + "flos": 21473819898720.0, + "grad_norm": 1.9743014585124716, + "learning_rate": 2.829718727142398e-06, + "loss": 0.6986, + "num_input_tokens_seen": 137946620, + "step": 4246 + }, + { + "epoch": 0.38300942417820266, + "flos": 36085247203200.0, + "grad_norm": 1.7944960395171943, + "learning_rate": 2.829187126944849e-06, + "loss": 0.7795, + "num_input_tokens_seen": 137979565, + "step": 4247 + }, + { + "epoch": 0.3830996077016729, + "flos": 22642323270240.0, + "grad_norm": 2.465099097807575, + "learning_rate": 2.8286554559964826e-06, + "loss": 0.71, + "num_input_tokens_seen": 138006925, + "step": 4248 + }, + { + "epoch": 0.3831897912251432, + "flos": 24567723360960.0, + "grad_norm": 2.6572320356947032, + "learning_rate": 2.8281237143426637e-06, + "loss": 0.7893, + "num_input_tokens_seen": 138035325, + "step": 4249 + }, + { + "epoch": 0.3832799747486134, + "flos": 32841266485440.0, + "grad_norm": 2.5069310430683385, + "learning_rate": 2.8275919020287626e-06, + "loss": 0.6689, + "num_input_tokens_seen": 138068760, + "step": 4250 + }, + { + "epoch": 0.3833701582720837, + "flos": 24501029344320.0, + "grad_norm": 1.707704040489682, + "learning_rate": 2.827060019100158e-06, + "loss": 0.7982, + "num_input_tokens_seen": 138095885, + "step": 4251 + }, + { + "epoch": 0.38346034179555394, + "flos": 22565816436960.0, + "grad_norm": 1.627722953022118, + "learning_rate": 2.8265280656022315e-06, + "loss": 0.7633, + "num_input_tokens_seen": 138126100, + "step": 4252 + }, + { + "epoch": 0.3835505253190242, + "flos": 24274408085760.0, + "grad_norm": 1.8885228473997229, + "learning_rate": 2.825996041580373e-06, + "loss": 0.6601, + "num_input_tokens_seen": 138152170, + "step": 4253 + }, + { + "epoch": 0.38364070884249446, + "flos": 23042555976000.0, + "grad_norm": 1.7396097314472496, + "learning_rate": 2.825463947079978e-06, + "loss": 0.7913, + "num_input_tokens_seen": 138182645, + "step": 4254 + }, + { + "epoch": 0.38373089236596475, + "flos": 23402639607360.0, + "grad_norm": 1.6427461263808039, + "learning_rate": 2.8249317821464483e-06, + "loss": 0.7692, + "num_input_tokens_seen": 138211440, + "step": 4255 + }, + { + "epoch": 0.383821075889435, + "flos": 23771792660160.0, + "grad_norm": 1.6311484440435633, + "learning_rate": 2.824399546825189e-06, + "loss": 0.7016, + "num_input_tokens_seen": 138241930, + "step": 4256 + }, + { + "epoch": 0.38391125941290527, + "flos": 27303438849600.0, + "grad_norm": 1.474566609349861, + "learning_rate": 2.823867241161616e-06, + "loss": 0.7224, + "num_input_tokens_seen": 138272900, + "step": 4257 + }, + { + "epoch": 0.3840014429363755, + "flos": 23590078205280.0, + "grad_norm": 1.8758499762080512, + "learning_rate": 2.8233348652011456e-06, + "loss": 0.8228, + "num_input_tokens_seen": 138301860, + "step": 4258 + }, + { + "epoch": 0.3840916264598458, + "flos": 52080202172640.0, + "grad_norm": 1.8606442519099473, + "learning_rate": 2.8228024189892057e-06, + "loss": 0.6827, + "num_input_tokens_seen": 138334115, + "step": 4259 + }, + { + "epoch": 0.384181809983316, + "flos": 23224567788960.0, + "grad_norm": 1.3487051599807331, + "learning_rate": 2.822269902571226e-06, + "loss": 0.7732, + "num_input_tokens_seen": 138364055, + "step": 4260 + }, + { + "epoch": 0.3842719935067863, + "flos": 22893222115680.0, + "grad_norm": 1.5259678099100629, + "learning_rate": 2.8217373159926446e-06, + "loss": 0.75, + "num_input_tokens_seen": 138392630, + "step": 4261 + }, + { + "epoch": 0.38436217703025655, + "flos": 23802983822400.0, + "grad_norm": 1.292309042439524, + "learning_rate": 2.8212046592989046e-06, + "loss": 0.7089, + "num_input_tokens_seen": 138421240, + "step": 4262 + }, + { + "epoch": 0.38445236055372684, + "flos": 19756567695840.0, + "grad_norm": 1.8407058801675311, + "learning_rate": 2.820671932535455e-06, + "loss": 0.7769, + "num_input_tokens_seen": 138447575, + "step": 4263 + }, + { + "epoch": 0.3845425440771971, + "flos": 21221805960480.0, + "grad_norm": 3.438389056283213, + "learning_rate": 2.8201391357477506e-06, + "loss": 0.7517, + "num_input_tokens_seen": 138475110, + "step": 4264 + }, + { + "epoch": 0.38463272760066736, + "flos": 26030694270240.0, + "grad_norm": 4.287799364321035, + "learning_rate": 2.8196062689812525e-06, + "loss": 0.7457, + "num_input_tokens_seen": 138504285, + "step": 4265 + }, + { + "epoch": 0.38472291112413765, + "flos": 23334570309600.0, + "grad_norm": 1.917600141792467, + "learning_rate": 2.819073332281429e-06, + "loss": 0.7137, + "num_input_tokens_seen": 138533175, + "step": 4266 + }, + { + "epoch": 0.3848130946476079, + "flos": 70859044639680.0, + "grad_norm": 0.6157280772709415, + "learning_rate": 2.8185403256937524e-06, + "loss": 0.5494, + "num_input_tokens_seen": 138630810, + "step": 4267 + }, + { + "epoch": 0.38490327817107817, + "flos": 23258918380800.0, + "grad_norm": 1.7748452535522659, + "learning_rate": 2.8180072492637016e-06, + "loss": 0.7292, + "num_input_tokens_seen": 138659185, + "step": 4268 + }, + { + "epoch": 0.3849934616945484, + "flos": 18087790593600.0, + "grad_norm": 1.9666841961461976, + "learning_rate": 2.817474103036762e-06, + "loss": 0.7628, + "num_input_tokens_seen": 138686935, + "step": 4269 + }, + { + "epoch": 0.3850836452180187, + "flos": 24135741946560.0, + "grad_norm": 1.8938482882232548, + "learning_rate": 2.816940887058425e-06, + "loss": 0.7973, + "num_input_tokens_seen": 138715390, + "step": 4270 + }, + { + "epoch": 0.3851738287414889, + "flos": 19978988771520.0, + "grad_norm": 1.7537943033290084, + "learning_rate": 2.816407601374186e-06, + "loss": 0.8052, + "num_input_tokens_seen": 138741765, + "step": 4271 + }, + { + "epoch": 0.3852640122649592, + "flos": 48100703081280.0, + "grad_norm": 1.6948050165695914, + "learning_rate": 2.815874246029549e-06, + "loss": 0.7027, + "num_input_tokens_seen": 138773210, + "step": 4272 + }, + { + "epoch": 0.38535419578842944, + "flos": 30656381334720.0, + "grad_norm": 1.5863094551361105, + "learning_rate": 2.815340821070023e-06, + "loss": 0.717, + "num_input_tokens_seen": 138803785, + "step": 4273 + }, + { + "epoch": 0.38544437931189973, + "flos": 29272965179040.0, + "grad_norm": 1.769527814242836, + "learning_rate": 2.814807326541122e-06, + "loss": 0.7494, + "num_input_tokens_seen": 138829610, + "step": 4274 + }, + { + "epoch": 0.38553456283536996, + "flos": 23224010242560.0, + "grad_norm": 1.5300143959265722, + "learning_rate": 2.8142737624883676e-06, + "loss": 0.7027, + "num_input_tokens_seen": 138859090, + "step": 4275 + }, + { + "epoch": 0.38562474635884025, + "flos": 28102454640480.0, + "grad_norm": 1.6212247864491283, + "learning_rate": 2.8137401289572854e-06, + "loss": 0.7676, + "num_input_tokens_seen": 138888775, + "step": 4276 + }, + { + "epoch": 0.3857149298823105, + "flos": 24318162626880.0, + "grad_norm": 1.6962806995736925, + "learning_rate": 2.8132064259934086e-06, + "loss": 0.8118, + "num_input_tokens_seen": 138916840, + "step": 4277 + }, + { + "epoch": 0.3858051134057808, + "flos": 49778661114240.0, + "grad_norm": 1.597709524297061, + "learning_rate": 2.812672653642276e-06, + "loss": 0.7648, + "num_input_tokens_seen": 138949220, + "step": 4278 + }, + { + "epoch": 0.385895296929251, + "flos": 25410976899840.0, + "grad_norm": 1.4554097252188878, + "learning_rate": 2.812138811949431e-06, + "loss": 0.78, + "num_input_tokens_seen": 138979965, + "step": 4279 + }, + { + "epoch": 0.3859854804527213, + "flos": 20674692598560.0, + "grad_norm": 2.476974501169392, + "learning_rate": 2.8116049009604247e-06, + "loss": 0.8323, + "num_input_tokens_seen": 139007635, + "step": 4280 + }, + { + "epoch": 0.38607566397619153, + "flos": 28617559105440.0, + "grad_norm": 1.6193258549996694, + "learning_rate": 2.8110709207208132e-06, + "loss": 0.6502, + "num_input_tokens_seen": 139038735, + "step": 4281 + }, + { + "epoch": 0.3861658474996618, + "flos": 20529224393280.0, + "grad_norm": 2.712031338896049, + "learning_rate": 2.810536871276158e-06, + "loss": 0.7976, + "num_input_tokens_seen": 139065515, + "step": 4282 + }, + { + "epoch": 0.38625603102313205, + "flos": 59027353286880.0, + "grad_norm": 0.7161570923816054, + "learning_rate": 2.8100027526720283e-06, + "loss": 0.6289, + "num_input_tokens_seen": 139150325, + "step": 4283 + }, + { + "epoch": 0.38634621454660234, + "flos": 30693556828320.0, + "grad_norm": 1.6525398300380116, + "learning_rate": 2.8094685649539974e-06, + "loss": 0.6955, + "num_input_tokens_seen": 139179495, + "step": 4284 + }, + { + "epoch": 0.3864363980700726, + "flos": 64148364770400.0, + "grad_norm": 0.7123922617919866, + "learning_rate": 2.8089343081676455e-06, + "loss": 0.6005, + "num_input_tokens_seen": 139270760, + "step": 4285 + }, + { + "epoch": 0.38652658159354286, + "flos": 20711793752640.0, + "grad_norm": 1.7602719204175317, + "learning_rate": 2.8083999823585577e-06, + "loss": 0.7488, + "num_input_tokens_seen": 139298825, + "step": 4286 + }, + { + "epoch": 0.3866167651170131, + "flos": 22641691384320.0, + "grad_norm": 1.4257151967255144, + "learning_rate": 2.8078655875723254e-06, + "loss": 0.7929, + "num_input_tokens_seen": 139328110, + "step": 4287 + }, + { + "epoch": 0.3867069486404834, + "flos": 12841568424000.0, + "grad_norm": 2.394768508681906, + "learning_rate": 2.807331123854547e-06, + "loss": 0.851, + "num_input_tokens_seen": 139354575, + "step": 4288 + }, + { + "epoch": 0.38679713216395367, + "flos": 25229188105440.0, + "grad_norm": 2.367354887530212, + "learning_rate": 2.806796591250826e-06, + "loss": 0.6379, + "num_input_tokens_seen": 139382690, + "step": 4289 + }, + { + "epoch": 0.3868873156874239, + "flos": 18270359952960.0, + "grad_norm": 2.070114183824125, + "learning_rate": 2.8062619898067707e-06, + "loss": 0.8323, + "num_input_tokens_seen": 139410105, + "step": 4290 + }, + { + "epoch": 0.3869774992108942, + "flos": 28945039123680.0, + "grad_norm": 2.2679457781262653, + "learning_rate": 2.8057273195679963e-06, + "loss": 0.7804, + "num_input_tokens_seen": 139440115, + "step": 4291 + }, + { + "epoch": 0.3870676827343644, + "flos": 26426652453600.0, + "grad_norm": 2.3456139845351838, + "learning_rate": 2.8051925805801253e-06, + "loss": 0.783, + "num_input_tokens_seen": 139468280, + "step": 4292 + }, + { + "epoch": 0.3871578662578347, + "flos": 22679387254560.0, + "grad_norm": 1.6566874808008776, + "learning_rate": 2.804657772888783e-06, + "loss": 0.8001, + "num_input_tokens_seen": 139497595, + "step": 4293 + }, + { + "epoch": 0.38724804978130495, + "flos": 18889742795520.0, + "grad_norm": 1.9245220656166901, + "learning_rate": 2.804122896539602e-06, + "loss": 0.7647, + "num_input_tokens_seen": 139524315, + "step": 4294 + }, + { + "epoch": 0.38733823330477524, + "flos": 33353434539360.0, + "grad_norm": 1.5584520930495638, + "learning_rate": 2.8035879515782225e-06, + "loss": 0.7195, + "num_input_tokens_seen": 139553290, + "step": 4295 + }, + { + "epoch": 0.38742841682824547, + "flos": 18925245649920.0, + "grad_norm": 2.2006402037858335, + "learning_rate": 2.803052938050288e-06, + "loss": 0.82, + "num_input_tokens_seen": 139578460, + "step": 4296 + }, + { + "epoch": 0.38751860035171576, + "flos": 26795619657600.0, + "grad_norm": 2.615780398037893, + "learning_rate": 2.802517856001449e-06, + "loss": 0.8095, + "num_input_tokens_seen": 139608260, + "step": 4297 + }, + { + "epoch": 0.387608783875186, + "flos": 35093217350400.0, + "grad_norm": 2.1944544734269615, + "learning_rate": 2.801982705477361e-06, + "loss": 0.7299, + "num_input_tokens_seen": 139634915, + "step": 4298 + }, + { + "epoch": 0.3876989673986563, + "flos": 21548282395200.0, + "grad_norm": 1.5127023262073118, + "learning_rate": 2.8014474865236867e-06, + "loss": 0.749, + "num_input_tokens_seen": 139662890, + "step": 4299 + }, + { + "epoch": 0.3877891509221265, + "flos": 28723212764160.0, + "grad_norm": 1.727619037750093, + "learning_rate": 2.800912199186094e-06, + "loss": 0.7583, + "num_input_tokens_seen": 139691915, + "step": 4300 + }, + { + "epoch": 0.3878793344455968, + "flos": 34884363237120.0, + "grad_norm": 1.749162814296765, + "learning_rate": 2.800376843510256e-06, + "loss": 0.7758, + "num_input_tokens_seen": 139723715, + "step": 4301 + }, + { + "epoch": 0.38796951796906703, + "flos": 23661864479040.0, + "grad_norm": 1.6708059012173875, + "learning_rate": 2.799841419541852e-06, + "loss": 0.788, + "num_input_tokens_seen": 139753485, + "step": 4302 + }, + { + "epoch": 0.3880597014925373, + "flos": 69565224806400.0, + "grad_norm": 0.740754883077068, + "learning_rate": 2.799305927326568e-06, + "loss": 0.5842, + "num_input_tokens_seen": 139846160, + "step": 4303 + }, + { + "epoch": 0.38814988501600756, + "flos": 70475321040960.0, + "grad_norm": 0.7164630164366538, + "learning_rate": 2.7987703669100955e-06, + "loss": 0.5939, + "num_input_tokens_seen": 139934175, + "step": 4304 + }, + { + "epoch": 0.38824006853947785, + "flos": 21694791353760.0, + "grad_norm": 1.6269708566054304, + "learning_rate": 2.79823473833813e-06, + "loss": 0.7709, + "num_input_tokens_seen": 139963905, + "step": 4305 + }, + { + "epoch": 0.3883302520629481, + "flos": 22204654882560.0, + "grad_norm": 1.969315644401513, + "learning_rate": 2.797699041656376e-06, + "loss": 0.7127, + "num_input_tokens_seen": 139993050, + "step": 4306 + }, + { + "epoch": 0.38842043558641837, + "flos": 29636505598080.0, + "grad_norm": 2.051969754195063, + "learning_rate": 2.7971632769105412e-06, + "loss": 0.7807, + "num_input_tokens_seen": 140021255, + "step": 4307 + }, + { + "epoch": 0.3885106191098886, + "flos": 31859235298080.0, + "grad_norm": 1.8929342243920564, + "learning_rate": 2.79662744414634e-06, + "loss": 0.7289, + "num_input_tokens_seen": 140048590, + "step": 4308 + }, + { + "epoch": 0.3886008026333589, + "flos": 20273716497600.0, + "grad_norm": 1.852103946887181, + "learning_rate": 2.7960915434094923e-06, + "loss": 0.7159, + "num_input_tokens_seen": 140075785, + "step": 4309 + }, + { + "epoch": 0.3886909861568291, + "flos": 22678792538400.0, + "grad_norm": 1.8105993499887847, + "learning_rate": 2.7955555747457256e-06, + "loss": 0.741, + "num_input_tokens_seen": 140103145, + "step": 4310 + }, + { + "epoch": 0.3887811696802994, + "flos": 27888694118880.0, + "grad_norm": 1.5971666252669372, + "learning_rate": 2.79501953820077e-06, + "loss": 0.7591, + "num_input_tokens_seen": 140134780, + "step": 4311 + }, + { + "epoch": 0.3888713532037697, + "flos": 21366790958880.0, + "grad_norm": 1.6311390163039547, + "learning_rate": 2.7944834338203637e-06, + "loss": 0.7813, + "num_input_tokens_seen": 140161765, + "step": 4312 + }, + { + "epoch": 0.38896153672723993, + "flos": 26722012065600.0, + "grad_norm": 1.7161143121057978, + "learning_rate": 2.79394726165025e-06, + "loss": 0.7258, + "num_input_tokens_seen": 140193235, + "step": 4313 + }, + { + "epoch": 0.3890517202507102, + "flos": 21652560772800.0, + "grad_norm": 2.2457950293884523, + "learning_rate": 2.793411021736178e-06, + "loss": 0.7415, + "num_input_tokens_seen": 140216935, + "step": 4314 + }, + { + "epoch": 0.38914190377418045, + "flos": 23985590351520.0, + "grad_norm": 1.8477369617471309, + "learning_rate": 2.7928747141239027e-06, + "loss": 0.8528, + "num_input_tokens_seen": 140247535, + "step": 4315 + }, + { + "epoch": 0.38923208729765074, + "flos": 70222377858720.0, + "grad_norm": 0.6482798373192478, + "learning_rate": 2.7923383388591856e-06, + "loss": 0.5878, + "num_input_tokens_seen": 140338480, + "step": 4316 + }, + { + "epoch": 0.389322270821121, + "flos": 27412660805280.0, + "grad_norm": 2.142247163149019, + "learning_rate": 2.7918018959877923e-06, + "loss": 0.731, + "num_input_tokens_seen": 140368090, + "step": 4317 + }, + { + "epoch": 0.38941245434459126, + "flos": 25084017258240.0, + "grad_norm": 1.6636027386020242, + "learning_rate": 2.791265385555495e-06, + "loss": 0.7912, + "num_input_tokens_seen": 140395575, + "step": 4318 + }, + { + "epoch": 0.3895026378680615, + "flos": 22351052331840.0, + "grad_norm": 1.6803327873414242, + "learning_rate": 2.790728807608072e-06, + "loss": 0.828, + "num_input_tokens_seen": 140423215, + "step": 4319 + }, + { + "epoch": 0.3895928213915318, + "flos": 20383719018240.0, + "grad_norm": 2.025702427670454, + "learning_rate": 2.790192162191307e-06, + "loss": 0.7492, + "num_input_tokens_seen": 140449195, + "step": 4320 + }, + { + "epoch": 0.389683004915002, + "flos": 28362608756160.0, + "grad_norm": 2.4197433442770024, + "learning_rate": 2.78965544935099e-06, + "loss": 0.7936, + "num_input_tokens_seen": 140479090, + "step": 4321 + }, + { + "epoch": 0.3897731884384723, + "flos": 18743419685760.0, + "grad_norm": 1.8064212810259563, + "learning_rate": 2.789118669132916e-06, + "loss": 0.7445, + "num_input_tokens_seen": 140506225, + "step": 4322 + }, + { + "epoch": 0.38986337196194254, + "flos": 30292989594720.0, + "grad_norm": 3.2684400189028295, + "learning_rate": 2.7885818215828856e-06, + "loss": 0.8271, + "num_input_tokens_seen": 140537260, + "step": 4323 + }, + { + "epoch": 0.38995355548541283, + "flos": 18301290926880.0, + "grad_norm": 2.083812760184282, + "learning_rate": 2.7880449067467064e-06, + "loss": 0.7194, + "num_input_tokens_seen": 140560925, + "step": 4324 + }, + { + "epoch": 0.39004373900888306, + "flos": 23662570704480.0, + "grad_norm": 1.988742018035657, + "learning_rate": 2.78750792467019e-06, + "loss": 0.7196, + "num_input_tokens_seen": 140589840, + "step": 4325 + }, + { + "epoch": 0.39013392253235335, + "flos": 27122653638720.0, + "grad_norm": 1.8241522243875474, + "learning_rate": 2.786970875399156e-06, + "loss": 0.6318, + "num_input_tokens_seen": 140617065, + "step": 4326 + }, + { + "epoch": 0.3902241060558236, + "flos": 34700381426880.0, + "grad_norm": 2.001834385769846, + "learning_rate": 2.7864337589794267e-06, + "loss": 0.6982, + "num_input_tokens_seen": 140652455, + "step": 4327 + }, + { + "epoch": 0.39031428957929387, + "flos": 15094522872480.0, + "grad_norm": 1.7826618823018399, + "learning_rate": 2.7858965754568335e-06, + "loss": 0.7822, + "num_input_tokens_seen": 140677210, + "step": 4328 + }, + { + "epoch": 0.3904044731027641, + "flos": 21038864903520.0, + "grad_norm": 1.7418055778605959, + "learning_rate": 2.785359324877211e-06, + "loss": 0.6996, + "num_input_tokens_seen": 140706155, + "step": 4329 + }, + { + "epoch": 0.3904946566262344, + "flos": 24714492507840.0, + "grad_norm": 1.7467544978143796, + "learning_rate": 2.7848220072864e-06, + "loss": 0.7892, + "num_input_tokens_seen": 140735680, + "step": 4330 + }, + { + "epoch": 0.3905848401497046, + "flos": 33200866909920.0, + "grad_norm": 1.8292007541594044, + "learning_rate": 2.784284622730248e-06, + "loss": 0.7525, + "num_input_tokens_seen": 140764190, + "step": 4331 + }, + { + "epoch": 0.3906750236731749, + "flos": 22168222784160.0, + "grad_norm": 1.694502606288316, + "learning_rate": 2.7837471712546073e-06, + "loss": 0.7807, + "num_input_tokens_seen": 140792930, + "step": 4332 + }, + { + "epoch": 0.39076520719664515, + "flos": 20528815525920.0, + "grad_norm": 1.4731559411587885, + "learning_rate": 2.783209652905337e-06, + "loss": 0.7365, + "num_input_tokens_seen": 140823265, + "step": 4333 + }, + { + "epoch": 0.39085539072011544, + "flos": 24061651147680.0, + "grad_norm": 1.504532007598924, + "learning_rate": 2.7826720677283e-06, + "loss": 0.7255, + "num_input_tokens_seen": 140856355, + "step": 4334 + }, + { + "epoch": 0.39094557424358567, + "flos": 33498902744640.0, + "grad_norm": 6.403906714632623, + "learning_rate": 2.782134415769367e-06, + "loss": 0.726, + "num_input_tokens_seen": 140888260, + "step": 4335 + }, + { + "epoch": 0.39103575776705596, + "flos": 31531309242720.0, + "grad_norm": 2.02697091909888, + "learning_rate": 2.7815966970744126e-06, + "loss": 0.8081, + "num_input_tokens_seen": 140917760, + "step": 4336 + }, + { + "epoch": 0.39112594129052625, + "flos": 21003176200320.0, + "grad_norm": 1.9425256215373272, + "learning_rate": 2.7810589116893184e-06, + "loss": 0.8298, + "num_input_tokens_seen": 140945600, + "step": 4337 + }, + { + "epoch": 0.3912161248139965, + "flos": 21400137967200.0, + "grad_norm": 2.903329181368896, + "learning_rate": 2.780521059659972e-06, + "loss": 0.7648, + "num_input_tokens_seen": 140976245, + "step": 4338 + }, + { + "epoch": 0.39130630833746677, + "flos": 24639621144000.0, + "grad_norm": 2.712380412757311, + "learning_rate": 2.7799831410322637e-06, + "loss": 0.7478, + "num_input_tokens_seen": 141003800, + "step": 4339 + }, + { + "epoch": 0.391396491860937, + "flos": 21836468243520.0, + "grad_norm": 1.52818630228163, + "learning_rate": 2.779445155852094e-06, + "loss": 0.7181, + "num_input_tokens_seen": 141035190, + "step": 4340 + }, + { + "epoch": 0.3914866753844073, + "flos": 70190443301280.0, + "grad_norm": 0.6547709100070972, + "learning_rate": 2.7789071041653655e-06, + "loss": 0.6252, + "num_input_tokens_seen": 141132800, + "step": 4341 + }, + { + "epoch": 0.3915768589078775, + "flos": 23401784702880.0, + "grad_norm": 1.6772116223031515, + "learning_rate": 2.7783689860179875e-06, + "loss": 0.8147, + "num_input_tokens_seen": 141162675, + "step": 4342 + }, + { + "epoch": 0.3916670424313478, + "flos": 18379544738880.0, + "grad_norm": 2.074726624462525, + "learning_rate": 2.7778308014558767e-06, + "loss": 0.6996, + "num_input_tokens_seen": 141189720, + "step": 4343 + }, + { + "epoch": 0.39175722595481804, + "flos": 40384978209600.0, + "grad_norm": 2.011231239283507, + "learning_rate": 2.7772925505249524e-06, + "loss": 0.8359, + "num_input_tokens_seen": 141221350, + "step": 4344 + }, + { + "epoch": 0.39184740947828833, + "flos": 19722663141120.0, + "grad_norm": 1.6937500394138458, + "learning_rate": 2.7767542332711417e-06, + "loss": 0.7285, + "num_input_tokens_seen": 141247735, + "step": 4345 + }, + { + "epoch": 0.39193759300175857, + "flos": 21293071857600.0, + "grad_norm": 1.6673870993258924, + "learning_rate": 2.776215849740377e-06, + "loss": 0.7368, + "num_input_tokens_seen": 141275890, + "step": 4346 + }, + { + "epoch": 0.39202777652522885, + "flos": 24427273073280.0, + "grad_norm": 1.553777226940681, + "learning_rate": 2.775677399978596e-06, + "loss": 0.7767, + "num_input_tokens_seen": 141305700, + "step": 4347 + }, + { + "epoch": 0.3921179600486991, + "flos": 66685193375040.0, + "grad_norm": 0.5701360821574185, + "learning_rate": 2.775138884031742e-06, + "loss": 0.5723, + "num_input_tokens_seen": 141401215, + "step": 4348 + }, + { + "epoch": 0.3922081435721694, + "flos": 18888887891040.0, + "grad_norm": 1.8027582466636218, + "learning_rate": 2.774600301945764e-06, + "loss": 0.72, + "num_input_tokens_seen": 141429845, + "step": 4349 + }, + { + "epoch": 0.3922983270956396, + "flos": 24496791991680.0, + "grad_norm": 1.7773101468106045, + "learning_rate": 2.774061653766618e-06, + "loss": 0.7928, + "num_input_tokens_seen": 141459935, + "step": 4350 + }, + { + "epoch": 0.3923885106191099, + "flos": 23735434901280.0, + "grad_norm": 1.7027941041128805, + "learning_rate": 2.773522939540263e-06, + "loss": 0.7534, + "num_input_tokens_seen": 141486950, + "step": 4351 + }, + { + "epoch": 0.39247869414258013, + "flos": 23332823330880.0, + "grad_norm": 2.6691675578661846, + "learning_rate": 2.7729841593126663e-06, + "loss": 0.6873, + "num_input_tokens_seen": 141515420, + "step": 4352 + }, + { + "epoch": 0.3925688776660504, + "flos": 22680725365920.0, + "grad_norm": 2.124649018866208, + "learning_rate": 2.7724453131297988e-06, + "loss": 0.8253, + "num_input_tokens_seen": 141548865, + "step": 4353 + }, + { + "epoch": 0.39265906118952065, + "flos": 26468065299840.0, + "grad_norm": 2.126288416098879, + "learning_rate": 2.771906401037637e-06, + "loss": 0.7869, + "num_input_tokens_seen": 141578140, + "step": 4354 + }, + { + "epoch": 0.39274924471299094, + "flos": 18161100827520.0, + "grad_norm": 1.8550974263520326, + "learning_rate": 2.7713674230821664e-06, + "loss": 0.7238, + "num_input_tokens_seen": 141603885, + "step": 4355 + }, + { + "epoch": 0.3928394282364612, + "flos": 22022531560320.0, + "grad_norm": 1.6238454238304196, + "learning_rate": 2.7708283793093724e-06, + "loss": 0.7159, + "num_input_tokens_seen": 141634790, + "step": 4356 + }, + { + "epoch": 0.39292961175993146, + "flos": 35575606692960.0, + "grad_norm": 1.6343475864656913, + "learning_rate": 2.7702892697652514e-06, + "loss": 0.7337, + "num_input_tokens_seen": 141664120, + "step": 4357 + }, + { + "epoch": 0.3930197952834017, + "flos": 24753489319680.0, + "grad_norm": 1.4492206419240936, + "learning_rate": 2.7697500944958024e-06, + "loss": 0.7396, + "num_input_tokens_seen": 141695875, + "step": 4358 + }, + { + "epoch": 0.393109978806872, + "flos": 19976089530240.0, + "grad_norm": 1.8962286512547362, + "learning_rate": 2.7692108535470312e-06, + "loss": 0.8598, + "num_input_tokens_seen": 141722830, + "step": 4359 + }, + { + "epoch": 0.3932001623303423, + "flos": 18995322114720.0, + "grad_norm": 1.9243795930932126, + "learning_rate": 2.768671546964948e-06, + "loss": 0.7963, + "num_input_tokens_seen": 141747070, + "step": 4360 + }, + { + "epoch": 0.3932903458538125, + "flos": 13716236143680.0, + "grad_norm": 1.8302735964219665, + "learning_rate": 2.7681321747955713e-06, + "loss": 0.7864, + "num_input_tokens_seen": 141772000, + "step": 4361 + }, + { + "epoch": 0.3933805293772828, + "flos": 25626856097760.0, + "grad_norm": 1.6280766485433371, + "learning_rate": 2.767592737084921e-06, + "loss": 0.7794, + "num_input_tokens_seen": 141802795, + "step": 4362 + }, + { + "epoch": 0.39347071290075303, + "flos": 22605407964960.0, + "grad_norm": 2.0249801116720225, + "learning_rate": 2.767053233879026e-06, + "loss": 0.8584, + "num_input_tokens_seen": 141830340, + "step": 4363 + }, + { + "epoch": 0.3935608964242233, + "flos": 26139433019040.0, + "grad_norm": 1.68592223831354, + "learning_rate": 2.76651366522392e-06, + "loss": 0.8372, + "num_input_tokens_seen": 141858510, + "step": 4364 + }, + { + "epoch": 0.39365107994769355, + "flos": 22933965906240.0, + "grad_norm": 2.6052162700065957, + "learning_rate": 2.7659740311656413e-06, + "loss": 0.7312, + "num_input_tokens_seen": 141886355, + "step": 4365 + }, + { + "epoch": 0.39374126347116384, + "flos": 20854065358560.0, + "grad_norm": 2.103710225772498, + "learning_rate": 2.7654343317502352e-06, + "loss": 0.7855, + "num_input_tokens_seen": 141915125, + "step": 4366 + }, + { + "epoch": 0.39383144699463407, + "flos": 25259673042240.0, + "grad_norm": 1.785584179361904, + "learning_rate": 2.7648945670237502e-06, + "loss": 0.7446, + "num_input_tokens_seen": 141943630, + "step": 4367 + }, + { + "epoch": 0.39392163051810436, + "flos": 21731000433600.0, + "grad_norm": 3.394162631021135, + "learning_rate": 2.7643547370322446e-06, + "loss": 0.8207, + "num_input_tokens_seen": 141970615, + "step": 4368 + }, + { + "epoch": 0.3940118140415746, + "flos": 24901336389600.0, + "grad_norm": 2.267763168052818, + "learning_rate": 2.7638148418217775e-06, + "loss": 0.8012, + "num_input_tokens_seen": 142000000, + "step": 4369 + }, + { + "epoch": 0.3941019975650449, + "flos": 21658099067040.0, + "grad_norm": 1.9108586492357327, + "learning_rate": 2.7632748814384163e-06, + "loss": 0.8037, + "num_input_tokens_seen": 142023640, + "step": 4370 + }, + { + "epoch": 0.3941921810885151, + "flos": 26063260713600.0, + "grad_norm": 1.4093720852637759, + "learning_rate": 2.7627348559282335e-06, + "loss": 0.7874, + "num_input_tokens_seen": 142055130, + "step": 4371 + }, + { + "epoch": 0.3942823646119854, + "flos": 24134106477120.0, + "grad_norm": 1.6176681172943266, + "learning_rate": 2.7621947653373075e-06, + "loss": 0.7008, + "num_input_tokens_seen": 142084745, + "step": 4372 + }, + { + "epoch": 0.39437254813545564, + "flos": 64870279011840.0, + "grad_norm": 0.660464401104455, + "learning_rate": 2.7616546097117213e-06, + "loss": 0.5637, + "num_input_tokens_seen": 142178570, + "step": 4373 + }, + { + "epoch": 0.3944627316589259, + "flos": 20894437451520.0, + "grad_norm": 1.8974323515964642, + "learning_rate": 2.761114389097564e-06, + "loss": 0.7537, + "num_input_tokens_seen": 142207200, + "step": 4374 + }, + { + "epoch": 0.39455291518239616, + "flos": 22314248535840.0, + "grad_norm": 1.6952593395142965, + "learning_rate": 2.7605741035409305e-06, + "loss": 0.7968, + "num_input_tokens_seen": 142235205, + "step": 4375 + }, + { + "epoch": 0.39464309870586645, + "flos": 33536524275360.0, + "grad_norm": 1.532615519863372, + "learning_rate": 2.76003375308792e-06, + "loss": 0.7406, + "num_input_tokens_seen": 142267720, + "step": 4376 + }, + { + "epoch": 0.3947332822293367, + "flos": 22714667090400.0, + "grad_norm": 1.4711064042870505, + "learning_rate": 2.75949333778464e-06, + "loss": 0.726, + "num_input_tokens_seen": 142298560, + "step": 4377 + }, + { + "epoch": 0.39482346575280697, + "flos": 21949704533280.0, + "grad_norm": 1.593496190224849, + "learning_rate": 2.7589528576772e-06, + "loss": 0.6464, + "num_input_tokens_seen": 142328335, + "step": 4378 + }, + { + "epoch": 0.3949136492762772, + "flos": 40348360262400.0, + "grad_norm": 1.76937912714459, + "learning_rate": 2.758412312811717e-06, + "loss": 0.7486, + "num_input_tokens_seen": 142358535, + "step": 4379 + }, + { + "epoch": 0.3950038327997475, + "flos": 63282660696480.0, + "grad_norm": 0.7466359589923777, + "learning_rate": 2.7578717032343146e-06, + "loss": 0.6294, + "num_input_tokens_seen": 142440590, + "step": 4380 + }, + { + "epoch": 0.3950940163232177, + "flos": 23438402650080.0, + "grad_norm": 1.748117456130061, + "learning_rate": 2.757331028991119e-06, + "loss": 0.7845, + "num_input_tokens_seen": 142470680, + "step": 4381 + }, + { + "epoch": 0.395184199846688, + "flos": 22278076625760.0, + "grad_norm": 2.147885560317054, + "learning_rate": 2.7567902901282642e-06, + "loss": 0.7764, + "num_input_tokens_seen": 142498345, + "step": 4382 + }, + { + "epoch": 0.3952743833701583, + "flos": 21003176200320.0, + "grad_norm": 2.0758299810728826, + "learning_rate": 2.7562494866918892e-06, + "loss": 0.7506, + "num_input_tokens_seen": 142526475, + "step": 4383 + }, + { + "epoch": 0.39536456689362853, + "flos": 18416311365120.0, + "grad_norm": 1.5823535732123712, + "learning_rate": 2.7557086187281378e-06, + "loss": 0.7332, + "num_input_tokens_seen": 142554865, + "step": 4384 + }, + { + "epoch": 0.3954547504170988, + "flos": 22824149234400.0, + "grad_norm": 1.8749254847660728, + "learning_rate": 2.75516768628316e-06, + "loss": 0.8558, + "num_input_tokens_seen": 142583150, + "step": 4385 + }, + { + "epoch": 0.39554493394056905, + "flos": 24315337725120.0, + "grad_norm": 1.5972912658290213, + "learning_rate": 2.7546266894031114e-06, + "loss": 0.8297, + "num_input_tokens_seen": 142614415, + "step": 4386 + }, + { + "epoch": 0.39563511746403934, + "flos": 30798727280160.0, + "grad_norm": 1.63238826515375, + "learning_rate": 2.7540856281341526e-06, + "loss": 0.7035, + "num_input_tokens_seen": 142643215, + "step": 4387 + }, + { + "epoch": 0.3957253009875096, + "flos": 25809351117600.0, + "grad_norm": 2.18091485801714, + "learning_rate": 2.7535445025224506e-06, + "loss": 0.7233, + "num_input_tokens_seen": 142673250, + "step": 4388 + }, + { + "epoch": 0.39581548451097986, + "flos": 33207668976000.0, + "grad_norm": 1.9577467290085937, + "learning_rate": 2.753003312614176e-06, + "loss": 0.7618, + "num_input_tokens_seen": 142702590, + "step": 4389 + }, + { + "epoch": 0.3959056680344501, + "flos": 22350866483040.0, + "grad_norm": 2.5361408972363435, + "learning_rate": 2.7524620584555065e-06, + "loss": 0.7659, + "num_input_tokens_seen": 142729950, + "step": 4390 + }, + { + "epoch": 0.3959958515579204, + "flos": 19323768546720.0, + "grad_norm": 1.6923414644626178, + "learning_rate": 2.7519207400926253e-06, + "loss": 0.7951, + "num_input_tokens_seen": 142756405, + "step": 4391 + }, + { + "epoch": 0.3960860350813906, + "flos": 20527031377440.0, + "grad_norm": 1.640510337092448, + "learning_rate": 2.751379357571721e-06, + "loss": 0.7421, + "num_input_tokens_seen": 142782540, + "step": 4392 + }, + { + "epoch": 0.3961762186048609, + "flos": 24573744862080.0, + "grad_norm": 1.564988253125842, + "learning_rate": 2.7508379109389865e-06, + "loss": 0.8266, + "num_input_tokens_seen": 142812910, + "step": 4393 + }, + { + "epoch": 0.39626640212833114, + "flos": 23329626731520.0, + "grad_norm": 7.588861092621571, + "learning_rate": 2.750296400240622e-06, + "loss": 0.7733, + "num_input_tokens_seen": 142840480, + "step": 4394 + }, + { + "epoch": 0.39635658565180143, + "flos": 27633743769600.0, + "grad_norm": 1.421689031826775, + "learning_rate": 2.7497548255228305e-06, + "loss": 0.793, + "num_input_tokens_seen": 142871475, + "step": 4395 + }, + { + "epoch": 0.39644676917527166, + "flos": 22422429738240.0, + "grad_norm": 1.7512293993155492, + "learning_rate": 2.749213186831824e-06, + "loss": 0.7097, + "num_input_tokens_seen": 142899670, + "step": 4396 + }, + { + "epoch": 0.39653695269874195, + "flos": 23516619292320.0, + "grad_norm": 1.7940553518238267, + "learning_rate": 2.7486714842138173e-06, + "loss": 0.8091, + "num_input_tokens_seen": 142929245, + "step": 4397 + }, + { + "epoch": 0.3966271362222122, + "flos": 29815618169760.0, + "grad_norm": 13.971957322818591, + "learning_rate": 2.748129717715031e-06, + "loss": 0.7115, + "num_input_tokens_seen": 142958015, + "step": 4398 + }, + { + "epoch": 0.3967173197456825, + "flos": 29017828980960.0, + "grad_norm": 1.626038219669003, + "learning_rate": 2.747587887381692e-06, + "loss": 0.7781, + "num_input_tokens_seen": 142988100, + "step": 4399 + }, + { + "epoch": 0.3968075032691527, + "flos": 24207342371520.0, + "grad_norm": 1.5401747504082268, + "learning_rate": 2.7470459932600328e-06, + "loss": 0.7049, + "num_input_tokens_seen": 143017960, + "step": 4400 + }, + { + "epoch": 0.396897686792623, + "flos": 27122802317760.0, + "grad_norm": 1.78961692781487, + "learning_rate": 2.7465040353962897e-06, + "loss": 0.7597, + "num_input_tokens_seen": 143047315, + "step": 4401 + }, + { + "epoch": 0.3969878703160932, + "flos": 26466913037280.0, + "grad_norm": 1.8590178587467643, + "learning_rate": 2.745962013836706e-06, + "loss": 0.7437, + "num_input_tokens_seen": 143076125, + "step": 4402 + }, + { + "epoch": 0.3970780538395635, + "flos": 29163074167680.0, + "grad_norm": 1.9329323610410218, + "learning_rate": 2.74541992862753e-06, + "loss": 0.7349, + "num_input_tokens_seen": 143105015, + "step": 4403 + }, + { + "epoch": 0.39716823736303375, + "flos": 16521990927360.0, + "grad_norm": 2.0349311923829645, + "learning_rate": 2.744877779815016e-06, + "loss": 0.7594, + "num_input_tokens_seen": 143130345, + "step": 4404 + }, + { + "epoch": 0.39725842088650404, + "flos": 30762109332960.0, + "grad_norm": 1.382667189037613, + "learning_rate": 2.7443355674454234e-06, + "loss": 0.7531, + "num_input_tokens_seen": 143163395, + "step": 4405 + }, + { + "epoch": 0.39734860440997427, + "flos": 15391629463200.0, + "grad_norm": 2.1712209691032336, + "learning_rate": 2.743793291565015e-06, + "loss": 0.7967, + "num_input_tokens_seen": 143190850, + "step": 4406 + }, + { + "epoch": 0.39743878793344456, + "flos": 28283834567520.0, + "grad_norm": 2.688317995005976, + "learning_rate": 2.7432509522200617e-06, + "loss": 0.6803, + "num_input_tokens_seen": 143218755, + "step": 4407 + }, + { + "epoch": 0.39752897145691485, + "flos": 74630890517280.0, + "grad_norm": 2.4156966619484384, + "learning_rate": 2.7427085494568383e-06, + "loss": 0.7052, + "num_input_tokens_seen": 143254510, + "step": 4408 + }, + { + "epoch": 0.3976191549803851, + "flos": 70176727659840.0, + "grad_norm": 0.6590927857937298, + "learning_rate": 2.742166083321628e-06, + "loss": 0.6064, + "num_input_tokens_seen": 143346915, + "step": 4409 + }, + { + "epoch": 0.39770933850385537, + "flos": 21107751936000.0, + "grad_norm": 2.435904053757567, + "learning_rate": 2.7416235538607137e-06, + "loss": 0.7991, + "num_input_tokens_seen": 143374905, + "step": 4410 + }, + { + "epoch": 0.3977995220273256, + "flos": 38484264573120.0, + "grad_norm": 1.5592474080636538, + "learning_rate": 2.7410809611203894e-06, + "loss": 0.7388, + "num_input_tokens_seen": 143408620, + "step": 4411 + }, + { + "epoch": 0.3978897055507959, + "flos": 23913878417280.0, + "grad_norm": 1.7314000088810135, + "learning_rate": 2.7405383051469507e-06, + "loss": 0.719, + "num_input_tokens_seen": 143437530, + "step": 4412 + }, + { + "epoch": 0.3979798890742661, + "flos": 20966186555520.0, + "grad_norm": 2.0355317982884134, + "learning_rate": 2.7399955859867e-06, + "loss": 0.5741, + "num_input_tokens_seen": 143462670, + "step": 4413 + }, + { + "epoch": 0.3980700725977364, + "flos": 64970951922720.0, + "grad_norm": 0.6216651878859222, + "learning_rate": 2.7394528036859465e-06, + "loss": 0.5505, + "num_input_tokens_seen": 143552750, + "step": 4414 + }, + { + "epoch": 0.39816025612120665, + "flos": 23225050995840.0, + "grad_norm": 1.5660692266214802, + "learning_rate": 2.738909958291002e-06, + "loss": 0.758, + "num_input_tokens_seen": 143580720, + "step": 4415 + }, + { + "epoch": 0.39825043964467693, + "flos": 36814335208320.0, + "grad_norm": 1.8367707654434176, + "learning_rate": 2.7383670498481863e-06, + "loss": 0.6721, + "num_input_tokens_seen": 143610975, + "step": 4416 + }, + { + "epoch": 0.39834062316814717, + "flos": 64829869749120.0, + "grad_norm": 0.5771861572332619, + "learning_rate": 2.737824078403822e-06, + "loss": 0.5092, + "num_input_tokens_seen": 143702010, + "step": 4417 + }, + { + "epoch": 0.39843080669161746, + "flos": 17140630374720.0, + "grad_norm": 2.1520250364853823, + "learning_rate": 2.737281044004239e-06, + "loss": 0.7348, + "num_input_tokens_seen": 143728755, + "step": 4418 + }, + { + "epoch": 0.3985209902150877, + "flos": 30220088228160.0, + "grad_norm": 1.6844238925271116, + "learning_rate": 2.736737946695772e-06, + "loss": 0.7407, + "num_input_tokens_seen": 143758215, + "step": 4419 + }, + { + "epoch": 0.398611173738558, + "flos": 30216073894080.0, + "grad_norm": 1.9257618127426155, + "learning_rate": 2.736194786524761e-06, + "loss": 0.7543, + "num_input_tokens_seen": 143785255, + "step": 4420 + }, + { + "epoch": 0.3987013572620282, + "flos": 13862224725600.0, + "grad_norm": 1.9980009936959333, + "learning_rate": 2.7356515635375517e-06, + "loss": 0.7574, + "num_input_tokens_seen": 143811880, + "step": 4421 + }, + { + "epoch": 0.3987915407854985, + "flos": 21076003227360.0, + "grad_norm": 1.9581894787918441, + "learning_rate": 2.735108277780495e-06, + "loss": 0.6707, + "num_input_tokens_seen": 143839935, + "step": 4422 + }, + { + "epoch": 0.39888172430896873, + "flos": 25299710607360.0, + "grad_norm": 2.354122859130884, + "learning_rate": 2.7345649292999456e-06, + "loss": 0.7461, + "num_input_tokens_seen": 143869035, + "step": 4423 + }, + { + "epoch": 0.398971907832439, + "flos": 20819974955040.0, + "grad_norm": 1.9253455359731784, + "learning_rate": 2.734021518142267e-06, + "loss": 0.7631, + "num_input_tokens_seen": 143899265, + "step": 4424 + }, + { + "epoch": 0.39906209135590925, + "flos": 34810272438240.0, + "grad_norm": 1.692158802525303, + "learning_rate": 2.733478044353825e-06, + "loss": 0.596, + "num_input_tokens_seen": 143931125, + "step": 4425 + }, + { + "epoch": 0.39915227487937954, + "flos": 30949622270400.0, + "grad_norm": 1.8018845097518776, + "learning_rate": 2.7329345079809917e-06, + "loss": 0.7631, + "num_input_tokens_seen": 143963935, + "step": 4426 + }, + { + "epoch": 0.3992424584028498, + "flos": 23479964175360.0, + "grad_norm": 1.4625361055009203, + "learning_rate": 2.7323909090701447e-06, + "loss": 0.6967, + "num_input_tokens_seen": 143993625, + "step": 4427 + }, + { + "epoch": 0.39933264192632006, + "flos": 16776420900000.0, + "grad_norm": 1.875903039272898, + "learning_rate": 2.731847247667667e-06, + "loss": 0.8111, + "num_input_tokens_seen": 144020690, + "step": 4428 + }, + { + "epoch": 0.3994228254497903, + "flos": 25811581303200.0, + "grad_norm": 1.9900639005206635, + "learning_rate": 2.731303523819947e-06, + "loss": 0.7345, + "num_input_tokens_seen": 144052275, + "step": 4429 + }, + { + "epoch": 0.3995130089732606, + "flos": 16592773617600.0, + "grad_norm": 2.737288922398189, + "learning_rate": 2.7307597375733783e-06, + "loss": 0.8189, + "num_input_tokens_seen": 144077355, + "step": 4430 + }, + { + "epoch": 0.3996031924967309, + "flos": 26393788652160.0, + "grad_norm": 1.9333291229114504, + "learning_rate": 2.7302158889743587e-06, + "loss": 0.7559, + "num_input_tokens_seen": 144106920, + "step": 4431 + }, + { + "epoch": 0.3996933760202011, + "flos": 25009926459360.0, + "grad_norm": 1.7813770851174981, + "learning_rate": 2.7296719780692937e-06, + "loss": 0.7174, + "num_input_tokens_seen": 144135185, + "step": 4432 + }, + { + "epoch": 0.3997835595436714, + "flos": 29127757162080.0, + "grad_norm": 1.5947596773726154, + "learning_rate": 2.7291280049045916e-06, + "loss": 0.6664, + "num_input_tokens_seen": 144166270, + "step": 4433 + }, + { + "epoch": 0.39987374306714163, + "flos": 24385711548000.0, + "grad_norm": 1.5799008943162245, + "learning_rate": 2.7285839695266683e-06, + "loss": 0.7, + "num_input_tokens_seen": 144194980, + "step": 4434 + }, + { + "epoch": 0.3999639265906119, + "flos": 24460285553760.0, + "grad_norm": 1.9892418120559565, + "learning_rate": 2.7280398719819423e-06, + "loss": 0.7078, + "num_input_tokens_seen": 144224565, + "step": 4435 + }, + { + "epoch": 0.40005411011408215, + "flos": 21839627673120.0, + "grad_norm": 1.9228187662856702, + "learning_rate": 2.727495712316841e-06, + "loss": 0.8225, + "num_input_tokens_seen": 144250350, + "step": 4436 + }, + { + "epoch": 0.40014429363755244, + "flos": 29090581668480.0, + "grad_norm": 2.465106738489678, + "learning_rate": 2.7269514905777945e-06, + "loss": 0.7046, + "num_input_tokens_seen": 144276135, + "step": 4437 + }, + { + "epoch": 0.4002344771610227, + "flos": 25880840033280.0, + "grad_norm": 2.167651893898967, + "learning_rate": 2.7264072068112377e-06, + "loss": 0.8772, + "num_input_tokens_seen": 144305595, + "step": 4438 + }, + { + "epoch": 0.40032466068449296, + "flos": 33455817259200.0, + "grad_norm": 2.0755948604335583, + "learning_rate": 2.7258628610636133e-06, + "loss": 0.7196, + "num_input_tokens_seen": 144334885, + "step": 4439 + }, + { + "epoch": 0.4004148442079632, + "flos": 26759038880160.0, + "grad_norm": 1.8561974951514373, + "learning_rate": 2.7253184533813667e-06, + "loss": 0.8375, + "num_input_tokens_seen": 144363620, + "step": 4440 + }, + { + "epoch": 0.4005050277314335, + "flos": 19393138786080.0, + "grad_norm": 1.7729704657825611, + "learning_rate": 2.72477398381095e-06, + "loss": 0.7664, + "num_input_tokens_seen": 144388175, + "step": 4441 + }, + { + "epoch": 0.4005952112549037, + "flos": 33681509273760.0, + "grad_norm": 1.6175767365638285, + "learning_rate": 2.724229452398821e-06, + "loss": 0.7629, + "num_input_tokens_seen": 144419325, + "step": 4442 + }, + { + "epoch": 0.400685394778374, + "flos": 20310371614560.0, + "grad_norm": 2.122006608030509, + "learning_rate": 2.7236848591914422e-06, + "loss": 0.7782, + "num_input_tokens_seen": 144447525, + "step": 4443 + }, + { + "epoch": 0.40077557830184424, + "flos": 31712986527840.0, + "grad_norm": 1.873794676876009, + "learning_rate": 2.7231402042352803e-06, + "loss": 0.7404, + "num_input_tokens_seen": 144477320, + "step": 4444 + }, + { + "epoch": 0.4008657618253145, + "flos": 24242139000480.0, + "grad_norm": 1.4207491766802893, + "learning_rate": 2.722595487576809e-06, + "loss": 0.7863, + "num_input_tokens_seen": 144507215, + "step": 4445 + }, + { + "epoch": 0.40095594534878476, + "flos": 18014591868960.0, + "grad_norm": 1.78817453795117, + "learning_rate": 2.722050709262506e-06, + "loss": 0.8047, + "num_input_tokens_seen": 144533145, + "step": 4446 + }, + { + "epoch": 0.40104612887225505, + "flos": 23587104624480.0, + "grad_norm": 1.4362857500080788, + "learning_rate": 2.7215058693388557e-06, + "loss": 0.7363, + "num_input_tokens_seen": 144564000, + "step": 4447 + }, + { + "epoch": 0.4011363123957253, + "flos": 23225199674880.0, + "grad_norm": 1.7894402650688723, + "learning_rate": 2.720960967852346e-06, + "loss": 0.6565, + "num_input_tokens_seen": 144591425, + "step": 4448 + }, + { + "epoch": 0.40122649591919557, + "flos": 17505471735360.0, + "grad_norm": 1.5304050060353858, + "learning_rate": 2.720416004849471e-06, + "loss": 0.689, + "num_input_tokens_seen": 144619400, + "step": 4449 + }, + { + "epoch": 0.4013166794426658, + "flos": 24536495028960.0, + "grad_norm": 1.9342273637330352, + "learning_rate": 2.7198709803767304e-06, + "loss": 0.7077, + "num_input_tokens_seen": 144648825, + "step": 4450 + }, + { + "epoch": 0.4014068629661361, + "flos": 39109037030880.0, + "grad_norm": 1.4178793802813332, + "learning_rate": 2.7193258944806286e-06, + "loss": 0.7263, + "num_input_tokens_seen": 144681860, + "step": 4451 + }, + { + "epoch": 0.4014970464896063, + "flos": 43480963178400.0, + "grad_norm": 1.8504417021009187, + "learning_rate": 2.718780747207675e-06, + "loss": 0.7593, + "num_input_tokens_seen": 144711500, + "step": 4452 + }, + { + "epoch": 0.4015872300130766, + "flos": 22716376899360.0, + "grad_norm": 1.9530903317235664, + "learning_rate": 2.7182355386043847e-06, + "loss": 0.8172, + "num_input_tokens_seen": 144739195, + "step": 4453 + }, + { + "epoch": 0.40167741353654685, + "flos": 21110985705120.0, + "grad_norm": 1.5546135812947204, + "learning_rate": 2.717690268717278e-06, + "loss": 0.8231, + "num_input_tokens_seen": 144767550, + "step": 4454 + }, + { + "epoch": 0.40176759706001713, + "flos": 26173226064480.0, + "grad_norm": 1.4614208054386142, + "learning_rate": 2.7171449375928803e-06, + "loss": 0.7792, + "num_input_tokens_seen": 144797305, + "step": 4455 + }, + { + "epoch": 0.4018577805834874, + "flos": 16111833895680.0, + "grad_norm": 2.1252620759806455, + "learning_rate": 2.716599545277722e-06, + "loss": 0.7843, + "num_input_tokens_seen": 144820420, + "step": 4456 + }, + { + "epoch": 0.40194796410695766, + "flos": 22058963658720.0, + "grad_norm": 1.688590015595388, + "learning_rate": 2.7160540918183394e-06, + "loss": 0.7181, + "num_input_tokens_seen": 144847070, + "step": 4457 + }, + { + "epoch": 0.40203814763042794, + "flos": 24169609331520.0, + "grad_norm": 2.2927496576347424, + "learning_rate": 2.715508577261273e-06, + "loss": 0.6775, + "num_input_tokens_seen": 144876640, + "step": 4458 + }, + { + "epoch": 0.4021283311538982, + "flos": 20707370551200.0, + "grad_norm": 1.8344836383213667, + "learning_rate": 2.7149630016530702e-06, + "loss": 0.7377, + "num_input_tokens_seen": 144903810, + "step": 4459 + }, + { + "epoch": 0.40221851467736847, + "flos": 22820395088640.0, + "grad_norm": 1.433114618795321, + "learning_rate": 2.7144173650402815e-06, + "loss": 0.7756, + "num_input_tokens_seen": 144931710, + "step": 4460 + }, + { + "epoch": 0.4023086982008387, + "flos": 32513749297440.0, + "grad_norm": 10.396182944331622, + "learning_rate": 2.7138716674694636e-06, + "loss": 0.7798, + "num_input_tokens_seen": 144960205, + "step": 4461 + }, + { + "epoch": 0.402398881724309, + "flos": 32440104535680.0, + "grad_norm": 1.9131602126342016, + "learning_rate": 2.7133259089871795e-06, + "loss": 0.7054, + "num_input_tokens_seen": 144990570, + "step": 4462 + }, + { + "epoch": 0.4024890652477792, + "flos": 16149529765920.0, + "grad_norm": 1.7516209043894435, + "learning_rate": 2.712780089639995e-06, + "loss": 0.7541, + "num_input_tokens_seen": 145015530, + "step": 4463 + }, + { + "epoch": 0.4025792487712495, + "flos": 26941831258080.0, + "grad_norm": 1.7313799730118906, + "learning_rate": 2.712234209474483e-06, + "loss": 0.6364, + "num_input_tokens_seen": 145047955, + "step": 4464 + }, + { + "epoch": 0.40266943229471974, + "flos": 25956417622560.0, + "grad_norm": 1.7236485812841014, + "learning_rate": 2.7116882685372218e-06, + "loss": 0.7967, + "num_input_tokens_seen": 145077725, + "step": 4465 + }, + { + "epoch": 0.40275961581819003, + "flos": 41222693454240.0, + "grad_norm": 2.4452877348111954, + "learning_rate": 2.7111422668747927e-06, + "loss": 0.7228, + "num_input_tokens_seen": 145108515, + "step": 4466 + }, + { + "epoch": 0.40284979934166026, + "flos": 27083247959520.0, + "grad_norm": 1.7641553585727388, + "learning_rate": 2.7105962045337846e-06, + "loss": 0.7555, + "num_input_tokens_seen": 145138600, + "step": 4467 + }, + { + "epoch": 0.40293998286513055, + "flos": 23370705049920.0, + "grad_norm": 2.119202615560759, + "learning_rate": 2.7100500815607898e-06, + "loss": 0.7511, + "num_input_tokens_seen": 145167705, + "step": 4468 + }, + { + "epoch": 0.4030301663886008, + "flos": 26868112156800.0, + "grad_norm": 1.6354882615407837, + "learning_rate": 2.709503898002407e-06, + "loss": 0.7765, + "num_input_tokens_seen": 145199665, + "step": 4469 + }, + { + "epoch": 0.4031203499120711, + "flos": 22853407569120.0, + "grad_norm": 1.7149118049342955, + "learning_rate": 2.708957653905239e-06, + "loss": 0.753, + "num_input_tokens_seen": 145225665, + "step": 4470 + }, + { + "epoch": 0.4032105334355413, + "flos": 21585829586400.0, + "grad_norm": 1.5989172699301653, + "learning_rate": 2.7084113493158956e-06, + "loss": 0.7678, + "num_input_tokens_seen": 145255410, + "step": 4471 + }, + { + "epoch": 0.4033007169590116, + "flos": 21655943220960.0, + "grad_norm": 1.5554507121827374, + "learning_rate": 2.7078649842809888e-06, + "loss": 0.7674, + "num_input_tokens_seen": 145285635, + "step": 4472 + }, + { + "epoch": 0.40339090048248183, + "flos": 20416285461600.0, + "grad_norm": 6.2388077276666385, + "learning_rate": 2.707318558847139e-06, + "loss": 0.8054, + "num_input_tokens_seen": 145310610, + "step": 4473 + }, + { + "epoch": 0.4034810840059521, + "flos": 58420942690560.0, + "grad_norm": 0.7165589730350221, + "learning_rate": 2.7067720730609697e-06, + "loss": 0.579, + "num_input_tokens_seen": 145402360, + "step": 4474 + }, + { + "epoch": 0.40357126752942235, + "flos": 27228567485760.0, + "grad_norm": 1.691626342831536, + "learning_rate": 2.70622552696911e-06, + "loss": 0.7294, + "num_input_tokens_seen": 145431650, + "step": 4475 + }, + { + "epoch": 0.40366145105289264, + "flos": 26063149204320.0, + "grad_norm": 1.853576787777584, + "learning_rate": 2.7056789206181943e-06, + "loss": 0.7074, + "num_input_tokens_seen": 145460535, + "step": 4476 + }, + { + "epoch": 0.40375163457636287, + "flos": 17978828826240.0, + "grad_norm": 1.6546970500436313, + "learning_rate": 2.7051322540548615e-06, + "loss": 0.8122, + "num_input_tokens_seen": 145487490, + "step": 4477 + }, + { + "epoch": 0.40384181809983316, + "flos": 24536829556800.0, + "grad_norm": 2.600857017561627, + "learning_rate": 2.704585527325757e-06, + "loss": 0.727, + "num_input_tokens_seen": 145517005, + "step": 4478 + }, + { + "epoch": 0.40393200162330345, + "flos": 25301346076800.0, + "grad_norm": 1.9941473203242157, + "learning_rate": 2.7040387404775303e-06, + "loss": 0.7125, + "num_input_tokens_seen": 145547260, + "step": 4479 + }, + { + "epoch": 0.4040221851467737, + "flos": 27305297337600.0, + "grad_norm": 2.41503067107696, + "learning_rate": 2.703491893556837e-06, + "loss": 0.7871, + "num_input_tokens_seen": 145577265, + "step": 4480 + }, + { + "epoch": 0.40411236867024397, + "flos": 22022977597440.0, + "grad_norm": 1.4453863176620103, + "learning_rate": 2.702944986610335e-06, + "loss": 0.7265, + "num_input_tokens_seen": 145605605, + "step": 4481 + }, + { + "epoch": 0.4042025521937142, + "flos": 29634461261280.0, + "grad_norm": 1.6751628170306347, + "learning_rate": 2.7023980196846917e-06, + "loss": 0.767, + "num_input_tokens_seen": 145634770, + "step": 4482 + }, + { + "epoch": 0.4042927357171845, + "flos": 61163980621920.0, + "grad_norm": 0.7040983456605607, + "learning_rate": 2.7018509928265763e-06, + "loss": 0.5983, + "num_input_tokens_seen": 145724470, + "step": 4483 + }, + { + "epoch": 0.4043829192406547, + "flos": 20638706537280.0, + "grad_norm": 1.6791441282490398, + "learning_rate": 2.7013039060826635e-06, + "loss": 0.7827, + "num_input_tokens_seen": 145754585, + "step": 4484 + }, + { + "epoch": 0.404473102764125, + "flos": 23917000677120.0, + "grad_norm": 2.09773115027911, + "learning_rate": 2.7007567594996347e-06, + "loss": 0.8282, + "num_input_tokens_seen": 145780590, + "step": 4485 + }, + { + "epoch": 0.40456328628759525, + "flos": 29592453698880.0, + "grad_norm": 1.640207073932004, + "learning_rate": 2.7002095531241757e-06, + "loss": 0.6601, + "num_input_tokens_seen": 145810025, + "step": 4486 + }, + { + "epoch": 0.40465346981106554, + "flos": 21911413946880.0, + "grad_norm": 1.8151595893035144, + "learning_rate": 2.6996622870029767e-06, + "loss": 0.7286, + "num_input_tokens_seen": 145839170, + "step": 4487 + }, + { + "epoch": 0.40474365333453577, + "flos": 20019063506400.0, + "grad_norm": 1.9620372226578915, + "learning_rate": 2.6991149611827335e-06, + "loss": 0.8674, + "num_input_tokens_seen": 145866000, + "step": 4488 + }, + { + "epoch": 0.40483383685800606, + "flos": 31166839579680.0, + "grad_norm": 2.064860062926244, + "learning_rate": 2.6985675757101466e-06, + "loss": 0.797, + "num_input_tokens_seen": 145894970, + "step": 4489 + }, + { + "epoch": 0.4049240203814763, + "flos": 20346729373440.0, + "grad_norm": 1.8901761996167272, + "learning_rate": 2.698020130631922e-06, + "loss": 0.7144, + "num_input_tokens_seen": 145922610, + "step": 4490 + }, + { + "epoch": 0.4050142039049466, + "flos": 25848719627040.0, + "grad_norm": 1.9187116563177808, + "learning_rate": 2.6974726259947713e-06, + "loss": 0.6901, + "num_input_tokens_seen": 145954630, + "step": 4491 + }, + { + "epoch": 0.4051043874284168, + "flos": 20820718350240.0, + "grad_norm": 2.181038167991192, + "learning_rate": 2.6969250618454106e-06, + "loss": 0.732, + "num_input_tokens_seen": 145979445, + "step": 4492 + }, + { + "epoch": 0.4051945709518871, + "flos": 25445773528800.0, + "grad_norm": 1.5055446666982226, + "learning_rate": 2.696377438230561e-06, + "loss": 0.8296, + "num_input_tokens_seen": 146009475, + "step": 4493 + }, + { + "epoch": 0.40528475447535733, + "flos": 48286320360960.0, + "grad_norm": 1.7927003214906638, + "learning_rate": 2.6958297551969484e-06, + "loss": 0.6969, + "num_input_tokens_seen": 146040450, + "step": 4494 + }, + { + "epoch": 0.4053749379988276, + "flos": 22456334292960.0, + "grad_norm": 2.15414879187049, + "learning_rate": 2.695282012791304e-06, + "loss": 0.7246, + "num_input_tokens_seen": 146066120, + "step": 4495 + }, + { + "epoch": 0.40546512152229786, + "flos": 27487866696960.0, + "grad_norm": 12.837361549886332, + "learning_rate": 2.6947342110603646e-06, + "loss": 0.813, + "num_input_tokens_seen": 146094790, + "step": 4496 + }, + { + "epoch": 0.40555530504576814, + "flos": 22788274682400.0, + "grad_norm": 2.4155185098604135, + "learning_rate": 2.6941863500508717e-06, + "loss": 0.8096, + "num_input_tokens_seen": 146124130, + "step": 4497 + }, + { + "epoch": 0.4056454885692384, + "flos": 23480596061280.0, + "grad_norm": 1.723377725458622, + "learning_rate": 2.693638429809572e-06, + "loss": 0.8088, + "num_input_tokens_seen": 146151290, + "step": 4498 + }, + { + "epoch": 0.40573567209270867, + "flos": 20820272313120.0, + "grad_norm": 1.827856060362737, + "learning_rate": 2.6930904503832167e-06, + "loss": 0.7413, + "num_input_tokens_seen": 146177815, + "step": 4499 + }, + { + "epoch": 0.4058258556161789, + "flos": 23403866209440.0, + "grad_norm": 9.404110078947031, + "learning_rate": 2.692542411818562e-06, + "loss": 0.7425, + "num_input_tokens_seen": 146206880, + "step": 4500 + }, + { + "epoch": 0.4059160391396492, + "flos": 21075073983360.0, + "grad_norm": 1.7348625941104747, + "learning_rate": 2.69199431416237e-06, + "loss": 0.7683, + "num_input_tokens_seen": 146235800, + "step": 4501 + }, + { + "epoch": 0.4060062226631194, + "flos": 23043373710720.0, + "grad_norm": 1.6773337972884959, + "learning_rate": 2.691446157461408e-06, + "loss": 0.713, + "num_input_tokens_seen": 146265325, + "step": 4502 + }, + { + "epoch": 0.4060964061865897, + "flos": 24390952484160.0, + "grad_norm": 1.7115636925086353, + "learning_rate": 2.690897941762447e-06, + "loss": 0.8169, + "num_input_tokens_seen": 146295630, + "step": 4503 + }, + { + "epoch": 0.40618658971006, + "flos": 21585792416640.0, + "grad_norm": 1.6170376325815476, + "learning_rate": 2.6903496671122642e-06, + "loss": 0.734, + "num_input_tokens_seen": 146324010, + "step": 4504 + }, + { + "epoch": 0.40627677323353023, + "flos": 63674673151680.0, + "grad_norm": 0.6351980057207626, + "learning_rate": 2.689801333557641e-06, + "loss": 0.5282, + "num_input_tokens_seen": 146410010, + "step": 4505 + }, + { + "epoch": 0.4063669567570005, + "flos": 14335358797920.0, + "grad_norm": 2.190721265389135, + "learning_rate": 2.689252941145365e-06, + "loss": 0.7576, + "num_input_tokens_seen": 146433955, + "step": 4506 + }, + { + "epoch": 0.40645714028047075, + "flos": 28214241309600.0, + "grad_norm": 1.7433601846079791, + "learning_rate": 2.6887044899222277e-06, + "loss": 0.7534, + "num_input_tokens_seen": 146465385, + "step": 4507 + }, + { + "epoch": 0.40654732380394104, + "flos": 22456222783680.0, + "grad_norm": 1.7114839421509986, + "learning_rate": 2.688155979935025e-06, + "loss": 0.8032, + "num_input_tokens_seen": 146494165, + "step": 4508 + }, + { + "epoch": 0.4066375073274113, + "flos": 22674369336960.0, + "grad_norm": 1.6688900978057475, + "learning_rate": 2.68760741123056e-06, + "loss": 0.7165, + "num_input_tokens_seen": 146523530, + "step": 4509 + }, + { + "epoch": 0.40672769085088156, + "flos": 23077426944480.0, + "grad_norm": 2.1793881009022344, + "learning_rate": 2.6870587838556394e-06, + "loss": 0.7747, + "num_input_tokens_seen": 146549960, + "step": 4510 + }, + { + "epoch": 0.4068178743743518, + "flos": 22635967241280.0, + "grad_norm": 1.9340406729791255, + "learning_rate": 2.686510097857075e-06, + "loss": 0.7765, + "num_input_tokens_seen": 146577410, + "step": 4511 + }, + { + "epoch": 0.4069080578978221, + "flos": 32299431229440.0, + "grad_norm": 1.4241628998955291, + "learning_rate": 2.685961353281683e-06, + "loss": 0.7523, + "num_input_tokens_seen": 146610425, + "step": 4512 + }, + { + "epoch": 0.4069982414212923, + "flos": 12003853179360.0, + "grad_norm": 2.0448000342388477, + "learning_rate": 2.6854125501762863e-06, + "loss": 0.735, + "num_input_tokens_seen": 146637245, + "step": 4513 + }, + { + "epoch": 0.4070884249447626, + "flos": 21731780998560.0, + "grad_norm": 1.946235774679756, + "learning_rate": 2.684863688587712e-06, + "loss": 0.7598, + "num_input_tokens_seen": 146668135, + "step": 4514 + }, + { + "epoch": 0.40717860846823284, + "flos": 28178738455200.0, + "grad_norm": 7.024882913611527, + "learning_rate": 2.6843147685627916e-06, + "loss": 0.7288, + "num_input_tokens_seen": 146700610, + "step": 4515 + }, + { + "epoch": 0.4072687919917031, + "flos": 40748853156480.0, + "grad_norm": 6.0359425054179345, + "learning_rate": 2.683765790148361e-06, + "loss": 0.73, + "num_input_tokens_seen": 146731035, + "step": 4516 + }, + { + "epoch": 0.40735897551517336, + "flos": 26431707540960.0, + "grad_norm": 1.9081730757253699, + "learning_rate": 2.6832167533912637e-06, + "loss": 0.8141, + "num_input_tokens_seen": 146759080, + "step": 4517 + }, + { + "epoch": 0.40744915903864365, + "flos": 18379210211040.0, + "grad_norm": 2.664314772667679, + "learning_rate": 2.682667658338345e-06, + "loss": 0.8314, + "num_input_tokens_seen": 146784870, + "step": 4518 + }, + { + "epoch": 0.4075393425621139, + "flos": 23877669337440.0, + "grad_norm": 1.6869890357782227, + "learning_rate": 2.682118505036458e-06, + "loss": 0.7305, + "num_input_tokens_seen": 146812850, + "step": 4519 + }, + { + "epoch": 0.40762952608558417, + "flos": 66434591887680.0, + "grad_norm": 0.6129673683059527, + "learning_rate": 2.681569293532459e-06, + "loss": 0.5487, + "num_input_tokens_seen": 146905535, + "step": 4520 + }, + { + "epoch": 0.4077197096090544, + "flos": 68627431367040.0, + "grad_norm": 0.6508891626395588, + "learning_rate": 2.6810200238732102e-06, + "loss": 0.577, + "num_input_tokens_seen": 147005110, + "step": 4521 + }, + { + "epoch": 0.4078098931325247, + "flos": 18378838513440.0, + "grad_norm": 2.7917652017844428, + "learning_rate": 2.6804706961055776e-06, + "loss": 0.8011, + "num_input_tokens_seen": 147033325, + "step": 4522 + }, + { + "epoch": 0.4079000766559949, + "flos": 27268419202080.0, + "grad_norm": 1.8711086583190653, + "learning_rate": 2.6799213102764326e-06, + "loss": 0.7894, + "num_input_tokens_seen": 147063630, + "step": 4523 + }, + { + "epoch": 0.4079902601794652, + "flos": 15465348564480.0, + "grad_norm": 1.8253190767587375, + "learning_rate": 2.679371866432653e-06, + "loss": 0.8489, + "num_input_tokens_seen": 147089170, + "step": 4524 + }, + { + "epoch": 0.40808044370293545, + "flos": 21034330192800.0, + "grad_norm": 1.6816644434288812, + "learning_rate": 2.6788223646211194e-06, + "loss": 0.728, + "num_input_tokens_seen": 147116845, + "step": 4525 + }, + { + "epoch": 0.40817062722640574, + "flos": 21075928887840.0, + "grad_norm": 1.758221659195037, + "learning_rate": 2.6782728048887183e-06, + "loss": 0.6521, + "num_input_tokens_seen": 147145890, + "step": 4526 + }, + { + "epoch": 0.408260810749876, + "flos": 42278518082400.0, + "grad_norm": 1.590216140811198, + "learning_rate": 2.6777231872823416e-06, + "loss": 0.7421, + "num_input_tokens_seen": 147179850, + "step": 4527 + }, + { + "epoch": 0.40835099427334626, + "flos": 26905287650400.0, + "grad_norm": 2.326879488202689, + "learning_rate": 2.6771735118488864e-06, + "loss": 0.7582, + "num_input_tokens_seen": 147208615, + "step": 4528 + }, + { + "epoch": 0.40844117779681655, + "flos": 15793274619840.0, + "grad_norm": 1.6743209629658757, + "learning_rate": 2.6766237786352523e-06, + "loss": 0.7942, + "num_input_tokens_seen": 147235315, + "step": 4529 + }, + { + "epoch": 0.4085313613202868, + "flos": 41585081610720.0, + "grad_norm": 1.4638609984650688, + "learning_rate": 2.676073987688347e-06, + "loss": 0.6824, + "num_input_tokens_seen": 147269170, + "step": 4530 + }, + { + "epoch": 0.40862154484375707, + "flos": 40129767672000.0, + "grad_norm": 1.720340478376809, + "learning_rate": 2.6755241390550818e-06, + "loss": 0.7598, + "num_input_tokens_seen": 147301120, + "step": 4531 + }, + { + "epoch": 0.4087117283672273, + "flos": 27522180119040.0, + "grad_norm": 2.50344658737238, + "learning_rate": 2.6749742327823716e-06, + "loss": 0.761, + "num_input_tokens_seen": 147332000, + "step": 4532 + }, + { + "epoch": 0.4088019118906976, + "flos": 23477325122400.0, + "grad_norm": 2.077785951878227, + "learning_rate": 2.674424268917138e-06, + "loss": 0.7705, + "num_input_tokens_seen": 147359035, + "step": 4533 + }, + { + "epoch": 0.4088920954141678, + "flos": 17500602496800.0, + "grad_norm": 1.8477528972228827, + "learning_rate": 2.6738742475063074e-06, + "loss": 0.8195, + "num_input_tokens_seen": 147384525, + "step": 4534 + }, + { + "epoch": 0.4089822789376381, + "flos": 12841902951840.0, + "grad_norm": 3.7863917425807503, + "learning_rate": 2.6733241685968104e-06, + "loss": 0.7988, + "num_input_tokens_seen": 147410690, + "step": 4535 + }, + { + "epoch": 0.40907246246110834, + "flos": 12987036629280.0, + "grad_norm": 2.209569501071627, + "learning_rate": 2.6727740322355826e-06, + "loss": 0.7733, + "num_input_tokens_seen": 147436705, + "step": 4536 + }, + { + "epoch": 0.40916264598457863, + "flos": 23115494512320.0, + "grad_norm": 2.7210335623267032, + "learning_rate": 2.6722238384695644e-06, + "loss": 0.7437, + "num_input_tokens_seen": 147464385, + "step": 4537 + }, + { + "epoch": 0.40925282950804887, + "flos": 22678346501280.0, + "grad_norm": 3.56325027863856, + "learning_rate": 2.671673587345702e-06, + "loss": 0.7676, + "num_input_tokens_seen": 147492910, + "step": 4538 + }, + { + "epoch": 0.40934301303151915, + "flos": 64938013781760.0, + "grad_norm": 0.6502865544816421, + "learning_rate": 2.6711232789109455e-06, + "loss": 0.5673, + "num_input_tokens_seen": 147579150, + "step": 4539 + }, + { + "epoch": 0.4094331965549894, + "flos": 27597014313120.0, + "grad_norm": 2.0768280571342794, + "learning_rate": 2.6705729132122497e-06, + "loss": 0.7625, + "num_input_tokens_seen": 147611535, + "step": 4540 + }, + { + "epoch": 0.4095233800784597, + "flos": 23297580664800.0, + "grad_norm": 2.1813640094681537, + "learning_rate": 2.670022490296576e-06, + "loss": 0.832, + "num_input_tokens_seen": 147636245, + "step": 4541 + }, + { + "epoch": 0.4096135636019299, + "flos": 21622336024320.0, + "grad_norm": 1.9016586417325017, + "learning_rate": 2.669472010210889e-06, + "loss": 0.8163, + "num_input_tokens_seen": 147664525, + "step": 4542 + }, + { + "epoch": 0.4097037471254002, + "flos": 16484629584960.0, + "grad_norm": 1.9995053991047873, + "learning_rate": 2.668921473002159e-06, + "loss": 0.7339, + "num_input_tokens_seen": 147692040, + "step": 4543 + }, + { + "epoch": 0.40979393064887043, + "flos": 38380246383840.0, + "grad_norm": 1.5466008569566796, + "learning_rate": 2.6683708787173596e-06, + "loss": 0.8123, + "num_input_tokens_seen": 147723475, + "step": 4544 + }, + { + "epoch": 0.4098841141723407, + "flos": 28177586192640.0, + "grad_norm": 1.7121022061740316, + "learning_rate": 2.6678202274034718e-06, + "loss": 0.7768, + "num_input_tokens_seen": 147754080, + "step": 4545 + }, + { + "epoch": 0.40997429769581095, + "flos": 24900853182720.0, + "grad_norm": 1.934482562844573, + "learning_rate": 2.66726951910748e-06, + "loss": 0.8455, + "num_input_tokens_seen": 147784340, + "step": 4546 + }, + { + "epoch": 0.41006448121928124, + "flos": 23037835416480.0, + "grad_norm": 1.9062074124941473, + "learning_rate": 2.6667187538763737e-06, + "loss": 0.7608, + "num_input_tokens_seen": 147813545, + "step": 4547 + }, + { + "epoch": 0.4101546647427515, + "flos": 24208754822400.0, + "grad_norm": 1.564138621365227, + "learning_rate": 2.6661679317571473e-06, + "loss": 0.7408, + "num_input_tokens_seen": 147842260, + "step": 4548 + }, + { + "epoch": 0.41024484826622176, + "flos": 26795619657600.0, + "grad_norm": 1.7035001289136975, + "learning_rate": 2.665617052796799e-06, + "loss": 0.7371, + "num_input_tokens_seen": 147873255, + "step": 4549 + }, + { + "epoch": 0.41033503178969205, + "flos": 20747928492960.0, + "grad_norm": 1.9893896464660483, + "learning_rate": 2.6650661170423346e-06, + "loss": 0.6409, + "num_input_tokens_seen": 147900460, + "step": 4550 + }, + { + "epoch": 0.4104252153131623, + "flos": 31457813160000.0, + "grad_norm": 1.8015856487444506, + "learning_rate": 2.6645151245407614e-06, + "loss": 0.8362, + "num_input_tokens_seen": 147928620, + "step": 4551 + }, + { + "epoch": 0.4105153988366326, + "flos": 20383310150880.0, + "grad_norm": 1.7137980477061856, + "learning_rate": 2.6639640753390936e-06, + "loss": 0.7529, + "num_input_tokens_seen": 147956200, + "step": 4552 + }, + { + "epoch": 0.4106055823601028, + "flos": 29599404444000.0, + "grad_norm": 1.469427074449094, + "learning_rate": 2.66341296948435e-06, + "loss": 0.661, + "num_input_tokens_seen": 147986780, + "step": 4553 + }, + { + "epoch": 0.4106957658835731, + "flos": 18234819928800.0, + "grad_norm": 1.3931286166500554, + "learning_rate": 2.6628618070235534e-06, + "loss": 0.7646, + "num_input_tokens_seen": 148016140, + "step": 4554 + }, + { + "epoch": 0.4107859494070433, + "flos": 26103149599680.0, + "grad_norm": 1.7705819418990825, + "learning_rate": 2.662310588003733e-06, + "loss": 0.7619, + "num_input_tokens_seen": 148045215, + "step": 4555 + }, + { + "epoch": 0.4108761329305136, + "flos": 11749125848640.0, + "grad_norm": 2.4066395630019164, + "learning_rate": 2.6617593124719205e-06, + "loss": 0.7342, + "num_input_tokens_seen": 148072925, + "step": 4556 + }, + { + "epoch": 0.41096631645398385, + "flos": 25408077658560.0, + "grad_norm": 1.8905780250451367, + "learning_rate": 2.661207980475155e-06, + "loss": 0.7621, + "num_input_tokens_seen": 148105735, + "step": 4557 + }, + { + "epoch": 0.41105649997745414, + "flos": 22641654214560.0, + "grad_norm": 1.6850201080888312, + "learning_rate": 2.6606565920604793e-06, + "loss": 0.7136, + "num_input_tokens_seen": 148133300, + "step": 4558 + }, + { + "epoch": 0.41114668350092437, + "flos": 22714667090400.0, + "grad_norm": 1.571788341036978, + "learning_rate": 2.66010514727494e-06, + "loss": 0.8344, + "num_input_tokens_seen": 148160460, + "step": 4559 + }, + { + "epoch": 0.41123686702439466, + "flos": 22529384338560.0, + "grad_norm": 1.8231580374636975, + "learning_rate": 2.659553646165589e-06, + "loss": 0.6537, + "num_input_tokens_seen": 148189330, + "step": 4560 + }, + { + "epoch": 0.4113270505478649, + "flos": 23079508451040.0, + "grad_norm": 1.8172598197105978, + "learning_rate": 2.659002088779485e-06, + "loss": 0.7266, + "num_input_tokens_seen": 148218455, + "step": 4561 + }, + { + "epoch": 0.4114172340713352, + "flos": 70394279496960.0, + "grad_norm": 0.5846384270696016, + "learning_rate": 2.6584504751636888e-06, + "loss": 0.5273, + "num_input_tokens_seen": 148308060, + "step": 4562 + }, + { + "epoch": 0.4115074175948054, + "flos": 24318162626880.0, + "grad_norm": 1.7433732449550141, + "learning_rate": 2.657898805365268e-06, + "loss": 0.7221, + "num_input_tokens_seen": 148339080, + "step": 4563 + }, + { + "epoch": 0.4115976011182757, + "flos": 26320181060160.0, + "grad_norm": 3.8892130913397676, + "learning_rate": 2.657347079431293e-06, + "loss": 0.7239, + "num_input_tokens_seen": 148369895, + "step": 4564 + }, + { + "epoch": 0.41168778464174594, + "flos": 31271638333920.0, + "grad_norm": 1.434107051912944, + "learning_rate": 2.6567952974088403e-06, + "loss": 0.7302, + "num_input_tokens_seen": 148403230, + "step": 4565 + }, + { + "epoch": 0.4117779681652162, + "flos": 24936393206880.0, + "grad_norm": 1.5024955176922907, + "learning_rate": 2.6562434593449917e-06, + "loss": 0.8069, + "num_input_tokens_seen": 148433475, + "step": 4566 + }, + { + "epoch": 0.41186815168868646, + "flos": 22715484825120.0, + "grad_norm": 1.8962456251381867, + "learning_rate": 2.6556915652868325e-06, + "loss": 0.782, + "num_input_tokens_seen": 148460475, + "step": 4567 + }, + { + "epoch": 0.41195833521215675, + "flos": 28140447868800.0, + "grad_norm": 1.7909382810709096, + "learning_rate": 2.6551396152814534e-06, + "loss": 0.766, + "num_input_tokens_seen": 148491855, + "step": 4568 + }, + { + "epoch": 0.412048518735627, + "flos": 22168111274880.0, + "grad_norm": 5.080776238496333, + "learning_rate": 2.65458760937595e-06, + "loss": 0.7505, + "num_input_tokens_seen": 148520300, + "step": 4569 + }, + { + "epoch": 0.41213870225909727, + "flos": 38016148418400.0, + "grad_norm": 1.446748834303984, + "learning_rate": 2.654035547617423e-06, + "loss": 0.7081, + "num_input_tokens_seen": 148553930, + "step": 4570 + }, + { + "epoch": 0.4122288857825675, + "flos": 24755459316960.0, + "grad_norm": 2.002332638379662, + "learning_rate": 2.653483430052976e-06, + "loss": 0.6593, + "num_input_tokens_seen": 148582970, + "step": 4571 + }, + { + "epoch": 0.4123190693060378, + "flos": 21112063628160.0, + "grad_norm": 1.7651631924449718, + "learning_rate": 2.6529312567297197e-06, + "loss": 0.7391, + "num_input_tokens_seen": 148611295, + "step": 4572 + }, + { + "epoch": 0.412409252829508, + "flos": 20784137572800.0, + "grad_norm": 4.927938934084058, + "learning_rate": 2.652379027694768e-06, + "loss": 0.8372, + "num_input_tokens_seen": 148638850, + "step": 4573 + }, + { + "epoch": 0.4124994363529783, + "flos": 21731706659040.0, + "grad_norm": 1.9417626650522322, + "learning_rate": 2.651826742995241e-06, + "loss": 0.8157, + "num_input_tokens_seen": 148666920, + "step": 4574 + }, + { + "epoch": 0.4125896198764486, + "flos": 21109573254240.0, + "grad_norm": 1.5450191732408252, + "learning_rate": 2.651274402678262e-06, + "loss": 0.7531, + "num_input_tokens_seen": 148695270, + "step": 4575 + }, + { + "epoch": 0.41267980339991883, + "flos": 27816015770880.0, + "grad_norm": 1.775573116781139, + "learning_rate": 2.6507220067909597e-06, + "loss": 0.7823, + "num_input_tokens_seen": 148726630, + "step": 4576 + }, + { + "epoch": 0.4127699869233891, + "flos": 30182280848640.0, + "grad_norm": 1.848286495022318, + "learning_rate": 2.650169555380468e-06, + "loss": 0.7121, + "num_input_tokens_seen": 148756110, + "step": 4577 + }, + { + "epoch": 0.41286017044685935, + "flos": 20747482455840.0, + "grad_norm": 3.766509779191878, + "learning_rate": 2.6496170484939253e-06, + "loss": 0.7509, + "num_input_tokens_seen": 148786370, + "step": 4578 + }, + { + "epoch": 0.41295035397032964, + "flos": 22824520932000.0, + "grad_norm": 1.5822350641522998, + "learning_rate": 2.6490644861784735e-06, + "loss": 0.7685, + "num_input_tokens_seen": 148815715, + "step": 4579 + }, + { + "epoch": 0.4130405374937999, + "flos": 23115754700640.0, + "grad_norm": 2.7850654234127443, + "learning_rate": 2.648511868481261e-06, + "loss": 0.778, + "num_input_tokens_seen": 148845045, + "step": 4580 + }, + { + "epoch": 0.41313072101727016, + "flos": 20965889197440.0, + "grad_norm": 1.6487469857212187, + "learning_rate": 2.6479591954494397e-06, + "loss": 0.7562, + "num_input_tokens_seen": 148873900, + "step": 4581 + }, + { + "epoch": 0.4132209045407404, + "flos": 26613384826080.0, + "grad_norm": 1.8824496097583645, + "learning_rate": 2.647406467130167e-06, + "loss": 0.6552, + "num_input_tokens_seen": 148903550, + "step": 4582 + }, + { + "epoch": 0.4133110880642107, + "flos": 20019137845920.0, + "grad_norm": 1.630704550332418, + "learning_rate": 2.646853683570605e-06, + "loss": 0.7774, + "num_input_tokens_seen": 148932505, + "step": 4583 + }, + { + "epoch": 0.4134012715876809, + "flos": 27160349508960.0, + "grad_norm": 1.9094885534805945, + "learning_rate": 2.6463008448179196e-06, + "loss": 0.6952, + "num_input_tokens_seen": 148963710, + "step": 4584 + }, + { + "epoch": 0.4134914551111512, + "flos": 50878128774240.0, + "grad_norm": 1.8083638978241452, + "learning_rate": 2.6457479509192828e-06, + "loss": 0.7071, + "num_input_tokens_seen": 148998905, + "step": 4585 + }, + { + "epoch": 0.41358163863462144, + "flos": 24787616892960.0, + "grad_norm": 4.546016651423478, + "learning_rate": 2.645195001921871e-06, + "loss": 0.7491, + "num_input_tokens_seen": 149027365, + "step": 4586 + }, + { + "epoch": 0.41367182215809173, + "flos": 25301234567520.0, + "grad_norm": 1.9383193098803544, + "learning_rate": 2.644641997872863e-06, + "loss": 0.7368, + "num_input_tokens_seen": 149056750, + "step": 4587 + }, + { + "epoch": 0.41376200568156196, + "flos": 21586201284000.0, + "grad_norm": 2.62373466876141, + "learning_rate": 2.644088938819445e-06, + "loss": 0.7585, + "num_input_tokens_seen": 149083445, + "step": 4588 + }, + { + "epoch": 0.41385218920503225, + "flos": 16776197881440.0, + "grad_norm": 1.8726953329719407, + "learning_rate": 2.6435358248088077e-06, + "loss": 0.8394, + "num_input_tokens_seen": 149106645, + "step": 4589 + }, + { + "epoch": 0.4139423727285025, + "flos": 19217780360160.0, + "grad_norm": 1.816135181176603, + "learning_rate": 2.642982655888146e-06, + "loss": 0.7964, + "num_input_tokens_seen": 149133440, + "step": 4590 + }, + { + "epoch": 0.41403255625197277, + "flos": 18015595452480.0, + "grad_norm": 2.161271592005313, + "learning_rate": 2.6424294321046585e-06, + "loss": 0.7197, + "num_input_tokens_seen": 149161075, + "step": 4591 + }, + { + "epoch": 0.414122739775443, + "flos": 70218289185120.0, + "grad_norm": 0.7322946259903202, + "learning_rate": 2.641876153505549e-06, + "loss": 0.5735, + "num_input_tokens_seen": 149246800, + "step": 4592 + }, + { + "epoch": 0.4142129232989133, + "flos": 23910198611040.0, + "grad_norm": 2.450844062678789, + "learning_rate": 2.641322820138027e-06, + "loss": 0.7671, + "num_input_tokens_seen": 149274855, + "step": 4593 + }, + { + "epoch": 0.4143031068223835, + "flos": 27086704747200.0, + "grad_norm": 1.9123613544149074, + "learning_rate": 2.640769432049306e-06, + "loss": 0.7064, + "num_input_tokens_seen": 149303920, + "step": 4594 + }, + { + "epoch": 0.4143932903458538, + "flos": 20566251207840.0, + "grad_norm": 2.315763902294851, + "learning_rate": 2.6402159892866038e-06, + "loss": 0.6752, + "num_input_tokens_seen": 149333055, + "step": 4595 + }, + { + "epoch": 0.41448347386932405, + "flos": 23953544284800.0, + "grad_norm": 1.5753565006357955, + "learning_rate": 2.639662491897143e-06, + "loss": 0.738, + "num_input_tokens_seen": 149364430, + "step": 4596 + }, + { + "epoch": 0.41457365739279434, + "flos": 21841114463520.0, + "grad_norm": 1.9668360898484765, + "learning_rate": 2.639108939928152e-06, + "loss": 0.7711, + "num_input_tokens_seen": 149393655, + "step": 4597 + }, + { + "epoch": 0.4146638409162646, + "flos": 24645456796320.0, + "grad_norm": 1.839440433169211, + "learning_rate": 2.638555333426862e-06, + "loss": 0.7699, + "num_input_tokens_seen": 149420530, + "step": 4598 + }, + { + "epoch": 0.41475402443973486, + "flos": 21767469701760.0, + "grad_norm": 1.4816775256267842, + "learning_rate": 2.6380016724405093e-06, + "loss": 0.79, + "num_input_tokens_seen": 149448270, + "step": 4599 + }, + { + "epoch": 0.41484420796320515, + "flos": 18270805990080.0, + "grad_norm": 1.7053587377821797, + "learning_rate": 2.637447957016336e-06, + "loss": 0.7139, + "num_input_tokens_seen": 149475715, + "step": 4600 + }, + { + "epoch": 0.4149343914866754, + "flos": 23589334810080.0, + "grad_norm": 1.9293088966268506, + "learning_rate": 2.636894187201589e-06, + "loss": 0.8121, + "num_input_tokens_seen": 149500250, + "step": 4601 + }, + { + "epoch": 0.41502457501014567, + "flos": 27410430619680.0, + "grad_norm": 2.8458215902224904, + "learning_rate": 2.6363403630435176e-06, + "loss": 0.6647, + "num_input_tokens_seen": 149530805, + "step": 4602 + }, + { + "epoch": 0.4151147585336159, + "flos": 24865201649280.0, + "grad_norm": 1.7320462650789035, + "learning_rate": 2.635786484589378e-06, + "loss": 0.7482, + "num_input_tokens_seen": 149561595, + "step": 4603 + }, + { + "epoch": 0.4152049420570862, + "flos": 15136864962720.0, + "grad_norm": 1.7122160432500384, + "learning_rate": 2.63523255188643e-06, + "loss": 0.6578, + "num_input_tokens_seen": 149588750, + "step": 4604 + }, + { + "epoch": 0.4152951255805564, + "flos": 25119334263840.0, + "grad_norm": 1.4558576096139433, + "learning_rate": 2.6346785649819375e-06, + "loss": 0.7821, + "num_input_tokens_seen": 149618235, + "step": 4605 + }, + { + "epoch": 0.4153853091040267, + "flos": 27997395697920.0, + "grad_norm": 1.7279013711500553, + "learning_rate": 2.6341245239231706e-06, + "loss": 0.7805, + "num_input_tokens_seen": 149647515, + "step": 4606 + }, + { + "epoch": 0.41547549262749695, + "flos": 26829189684480.0, + "grad_norm": 1.2940224217885217, + "learning_rate": 2.6335704287574024e-06, + "loss": 0.7565, + "num_input_tokens_seen": 149678335, + "step": 4607 + }, + { + "epoch": 0.41556567615096723, + "flos": 26062257130080.0, + "grad_norm": 2.5115682933589145, + "learning_rate": 2.6330162795319124e-06, + "loss": 0.8304, + "num_input_tokens_seen": 149706295, + "step": 4608 + }, + { + "epoch": 0.41565585967443747, + "flos": 24317902438560.0, + "grad_norm": 2.260632506113666, + "learning_rate": 2.632462076293983e-06, + "loss": 0.8102, + "num_input_tokens_seen": 149735255, + "step": 4609 + }, + { + "epoch": 0.41574604319790776, + "flos": 24099123999360.0, + "grad_norm": 1.5843953118160299, + "learning_rate": 2.6319078190909017e-06, + "loss": 0.7539, + "num_input_tokens_seen": 149765305, + "step": 4610 + }, + { + "epoch": 0.415836226721378, + "flos": 16959399126720.0, + "grad_norm": 2.498076939047735, + "learning_rate": 2.6313535079699606e-06, + "loss": 0.6739, + "num_input_tokens_seen": 149792710, + "step": 4611 + }, + { + "epoch": 0.4159264102448483, + "flos": 25519715648640.0, + "grad_norm": 2.0796029707813855, + "learning_rate": 2.6307991429784572e-06, + "loss": 0.7327, + "num_input_tokens_seen": 149818510, + "step": 4612 + }, + { + "epoch": 0.4160165937683185, + "flos": 24533038241280.0, + "grad_norm": 1.5436081732338571, + "learning_rate": 2.6302447241636924e-06, + "loss": 0.8163, + "num_input_tokens_seen": 149847520, + "step": 4613 + }, + { + "epoch": 0.4161067772917888, + "flos": 19397227459680.0, + "grad_norm": 1.8026688420213932, + "learning_rate": 2.629690251572973e-06, + "loss": 0.8327, + "num_input_tokens_seen": 149875095, + "step": 4614 + }, + { + "epoch": 0.41619696081525903, + "flos": 19472916558240.0, + "grad_norm": 1.6548268663828964, + "learning_rate": 2.629135725253609e-06, + "loss": 0.7316, + "num_input_tokens_seen": 149901475, + "step": 4615 + }, + { + "epoch": 0.4162871443387293, + "flos": 26430629617920.0, + "grad_norm": 1.6684903260688762, + "learning_rate": 2.6285811452529162e-06, + "loss": 0.7437, + "num_input_tokens_seen": 149930345, + "step": 4616 + }, + { + "epoch": 0.41637732786219955, + "flos": 34987043315040.0, + "grad_norm": 1.6297902148932444, + "learning_rate": 2.6280265116182136e-06, + "loss": 0.7723, + "num_input_tokens_seen": 149961380, + "step": 4617 + }, + { + "epoch": 0.41646751138566984, + "flos": 31235317744800.0, + "grad_norm": 1.5547144135105697, + "learning_rate": 2.6274718243968266e-06, + "loss": 0.7568, + "num_input_tokens_seen": 149991980, + "step": 4618 + }, + { + "epoch": 0.4165576949091401, + "flos": 23152632836160.0, + "grad_norm": 1.8314331640097106, + "learning_rate": 2.626917083636084e-06, + "loss": 0.6732, + "num_input_tokens_seen": 150019035, + "step": 4619 + }, + { + "epoch": 0.41664787843261036, + "flos": 23443197549120.0, + "grad_norm": 1.4831519458893403, + "learning_rate": 2.6263622893833183e-06, + "loss": 0.6898, + "num_input_tokens_seen": 150049055, + "step": 4620 + }, + { + "epoch": 0.4167380619560806, + "flos": 24063546805440.0, + "grad_norm": 1.5947343367079336, + "learning_rate": 2.625807441685869e-06, + "loss": 0.7014, + "num_input_tokens_seen": 150079930, + "step": 4621 + }, + { + "epoch": 0.4168282454795509, + "flos": 69703667927040.0, + "grad_norm": 0.6082362024503867, + "learning_rate": 2.625252540591078e-06, + "loss": 0.5608, + "num_input_tokens_seen": 150168685, + "step": 4622 + }, + { + "epoch": 0.4169184290030212, + "flos": 27227006355840.0, + "grad_norm": 1.7603699115510878, + "learning_rate": 2.6246975861462927e-06, + "loss": 0.8281, + "num_input_tokens_seen": 150196690, + "step": 4623 + }, + { + "epoch": 0.4170086125264914, + "flos": 22532878296000.0, + "grad_norm": 1.8989727380331427, + "learning_rate": 2.624142578398864e-06, + "loss": 0.8005, + "num_input_tokens_seen": 150223570, + "step": 4624 + }, + { + "epoch": 0.4170987960499617, + "flos": 29418210365760.0, + "grad_norm": 3.2081870669096935, + "learning_rate": 2.6235875173961498e-06, + "loss": 0.7387, + "num_input_tokens_seen": 150255570, + "step": 4625 + }, + { + "epoch": 0.41718897957343193, + "flos": 26867554610400.0, + "grad_norm": 1.773438844386192, + "learning_rate": 2.62303240318551e-06, + "loss": 0.686, + "num_input_tokens_seen": 150286780, + "step": 4626 + }, + { + "epoch": 0.4172791630969022, + "flos": 28070594422560.0, + "grad_norm": 2.3279838775260013, + "learning_rate": 2.62247723581431e-06, + "loss": 0.8316, + "num_input_tokens_seen": 150317115, + "step": 4627 + }, + { + "epoch": 0.41736934662037245, + "flos": 20347026731520.0, + "grad_norm": 3.436456001576194, + "learning_rate": 2.62192201532992e-06, + "loss": 0.7576, + "num_input_tokens_seen": 150341930, + "step": 4628 + }, + { + "epoch": 0.41745953014384274, + "flos": 23006830103040.0, + "grad_norm": 1.6791128195686684, + "learning_rate": 2.6213667417797145e-06, + "loss": 0.773, + "num_input_tokens_seen": 150372425, + "step": 4629 + }, + { + "epoch": 0.41754971366731297, + "flos": 17211561744000.0, + "grad_norm": 2.268615012448571, + "learning_rate": 2.6208114152110725e-06, + "loss": 0.7822, + "num_input_tokens_seen": 150399695, + "step": 4630 + }, + { + "epoch": 0.41763989719078326, + "flos": 16519500553440.0, + "grad_norm": 1.7563796922033128, + "learning_rate": 2.6202560356713774e-06, + "loss": 0.7709, + "num_input_tokens_seen": 150425530, + "step": 4631 + }, + { + "epoch": 0.4177300807142535, + "flos": 67156877826720.0, + "grad_norm": 0.6377957368672943, + "learning_rate": 2.619700603208017e-06, + "loss": 0.6384, + "num_input_tokens_seen": 150517955, + "step": 4632 + }, + { + "epoch": 0.4178202642377238, + "flos": 23330630315040.0, + "grad_norm": 1.6009054185138796, + "learning_rate": 2.6191451178683842e-06, + "loss": 0.6683, + "num_input_tokens_seen": 150547845, + "step": 4633 + }, + { + "epoch": 0.417910447761194, + "flos": 24570919960320.0, + "grad_norm": 1.877042420929661, + "learning_rate": 2.6185895796998764e-06, + "loss": 0.7851, + "num_input_tokens_seen": 150575835, + "step": 4634 + }, + { + "epoch": 0.4180006312846643, + "flos": 18451888559040.0, + "grad_norm": 1.8241876440234732, + "learning_rate": 2.6180339887498946e-06, + "loss": 0.8254, + "num_input_tokens_seen": 150600865, + "step": 4635 + }, + { + "epoch": 0.41809081480813454, + "flos": 24500211609600.0, + "grad_norm": 2.17266172801359, + "learning_rate": 2.617478345065846e-06, + "loss": 0.6978, + "num_input_tokens_seen": 150629415, + "step": 4636 + }, + { + "epoch": 0.4181809983316048, + "flos": 29054521267680.0, + "grad_norm": 1.8519943956862943, + "learning_rate": 2.616922648695139e-06, + "loss": 0.7765, + "num_input_tokens_seen": 150656955, + "step": 4637 + }, + { + "epoch": 0.41827118185507506, + "flos": 29345197489920.0, + "grad_norm": 1.9504893104712087, + "learning_rate": 2.61636689968519e-06, + "loss": 0.6548, + "num_input_tokens_seen": 150685815, + "step": 4638 + }, + { + "epoch": 0.41836136537854535, + "flos": 12622975833600.0, + "grad_norm": 2.0271511096607786, + "learning_rate": 2.6158110980834186e-06, + "loss": 0.8197, + "num_input_tokens_seen": 150711745, + "step": 4639 + }, + { + "epoch": 0.4184515489020156, + "flos": 18161658373920.0, + "grad_norm": 3.094348888036634, + "learning_rate": 2.615255243937249e-06, + "loss": 0.8096, + "num_input_tokens_seen": 150737930, + "step": 4640 + }, + { + "epoch": 0.41854173242548587, + "flos": 18779963293440.0, + "grad_norm": 2.5348362889744687, + "learning_rate": 2.61469933729411e-06, + "loss": 0.7822, + "num_input_tokens_seen": 150763560, + "step": 4641 + }, + { + "epoch": 0.4186319159489561, + "flos": 29309137089120.0, + "grad_norm": 1.8895362682712784, + "learning_rate": 2.614143378201433e-06, + "loss": 0.7414, + "num_input_tokens_seen": 150791305, + "step": 4642 + }, + { + "epoch": 0.4187220994724264, + "flos": 23549222905440.0, + "grad_norm": 1.6982581920429003, + "learning_rate": 2.6135873667066567e-06, + "loss": 0.7371, + "num_input_tokens_seen": 150821120, + "step": 4643 + }, + { + "epoch": 0.4188122829958966, + "flos": 22819688863200.0, + "grad_norm": 12.912612639800935, + "learning_rate": 2.613031302857224e-06, + "loss": 0.716, + "num_input_tokens_seen": 150848950, + "step": 4644 + }, + { + "epoch": 0.4189024665193669, + "flos": 20491788711360.0, + "grad_norm": 1.941828476490018, + "learning_rate": 2.6124751867005792e-06, + "loss": 0.7835, + "num_input_tokens_seen": 150875255, + "step": 4645 + }, + { + "epoch": 0.4189926500428372, + "flos": 22054726306080.0, + "grad_norm": 1.3136365601610562, + "learning_rate": 2.611919018284175e-06, + "loss": 0.807, + "num_input_tokens_seen": 150904875, + "step": 4646 + }, + { + "epoch": 0.41908283356630743, + "flos": 24026743009440.0, + "grad_norm": 2.0815031659452474, + "learning_rate": 2.611362797655466e-06, + "loss": 0.8077, + "num_input_tokens_seen": 150932955, + "step": 4647 + }, + { + "epoch": 0.4191730170897777, + "flos": 24099272678400.0, + "grad_norm": 1.9216207320564749, + "learning_rate": 2.6108065248619124e-06, + "loss": 0.8153, + "num_input_tokens_seen": 150962550, + "step": 4648 + }, + { + "epoch": 0.41926320061324795, + "flos": 27924643010400.0, + "grad_norm": 1.4293950516562386, + "learning_rate": 2.610250199950978e-06, + "loss": 0.6915, + "num_input_tokens_seen": 150994760, + "step": 4649 + }, + { + "epoch": 0.41935338413671824, + "flos": 18962346804000.0, + "grad_norm": 1.9888916277851778, + "learning_rate": 2.609693822970131e-06, + "loss": 0.728, + "num_input_tokens_seen": 151023165, + "step": 4650 + }, + { + "epoch": 0.4194435676601885, + "flos": 19104358221600.0, + "grad_norm": 2.17641452086361, + "learning_rate": 2.609137393966846e-06, + "loss": 0.7199, + "num_input_tokens_seen": 151051175, + "step": 4651 + }, + { + "epoch": 0.41953375118365877, + "flos": 23042407296960.0, + "grad_norm": 1.8370667576130377, + "learning_rate": 2.6085809129886e-06, + "loss": 0.713, + "num_input_tokens_seen": 151079205, + "step": 4652 + }, + { + "epoch": 0.419623934707129, + "flos": 17469262655520.0, + "grad_norm": 2.5379571474694678, + "learning_rate": 2.608024380082874e-06, + "loss": 0.767, + "num_input_tokens_seen": 151104720, + "step": 4653 + }, + { + "epoch": 0.4197141182305993, + "flos": 25334284217760.0, + "grad_norm": 1.6347044428561923, + "learning_rate": 2.6074677952971554e-06, + "loss": 0.7552, + "num_input_tokens_seen": 151131535, + "step": 4654 + }, + { + "epoch": 0.4198043017540695, + "flos": 34446509000640.0, + "grad_norm": 2.1062485873163848, + "learning_rate": 2.606911158678935e-06, + "loss": 0.6845, + "num_input_tokens_seen": 151164885, + "step": 4655 + }, + { + "epoch": 0.4198944852775398, + "flos": 18743754213600.0, + "grad_norm": 2.1717746629110493, + "learning_rate": 2.606354470275708e-06, + "loss": 0.7463, + "num_input_tokens_seen": 151191755, + "step": 4656 + }, + { + "epoch": 0.41998466880101004, + "flos": 39947681519520.0, + "grad_norm": 3.969831424468335, + "learning_rate": 2.6057977301349744e-06, + "loss": 0.7261, + "num_input_tokens_seen": 151222925, + "step": 4657 + }, + { + "epoch": 0.42007485232448033, + "flos": 28031560440960.0, + "grad_norm": 1.7067818535860377, + "learning_rate": 2.6052409383042383e-06, + "loss": 0.7162, + "num_input_tokens_seen": 151250860, + "step": 4658 + }, + { + "epoch": 0.42016503584795056, + "flos": 23258881211040.0, + "grad_norm": 1.6151751100306808, + "learning_rate": 2.6046840948310074e-06, + "loss": 0.7351, + "num_input_tokens_seen": 151280210, + "step": 4659 + }, + { + "epoch": 0.42025521937142085, + "flos": 28143495789120.0, + "grad_norm": 1.6258863086109197, + "learning_rate": 2.6041271997627962e-06, + "loss": 0.7566, + "num_input_tokens_seen": 151310565, + "step": 4660 + }, + { + "epoch": 0.4203454028948911, + "flos": 20493461350560.0, + "grad_norm": 2.1804055839308916, + "learning_rate": 2.6035702531471202e-06, + "loss": 0.7434, + "num_input_tokens_seen": 151338195, + "step": 4661 + }, + { + "epoch": 0.4204355864183614, + "flos": 20782502103360.0, + "grad_norm": 1.581885135637757, + "learning_rate": 2.6030132550315035e-06, + "loss": 0.7528, + "num_input_tokens_seen": 151366495, + "step": 4662 + }, + { + "epoch": 0.4205257699418316, + "flos": 58100116059360.0, + "grad_norm": 0.6480982249071569, + "learning_rate": 2.60245620546347e-06, + "loss": 0.613, + "num_input_tokens_seen": 151454865, + "step": 4663 + }, + { + "epoch": 0.4206159534653019, + "flos": 25483023361920.0, + "grad_norm": 2.603684668121926, + "learning_rate": 2.6018991044905517e-06, + "loss": 0.7337, + "num_input_tokens_seen": 151481925, + "step": 4664 + }, + { + "epoch": 0.42070613698877213, + "flos": 21221434262880.0, + "grad_norm": 1.651036604463376, + "learning_rate": 2.6013419521602825e-06, + "loss": 0.7934, + "num_input_tokens_seen": 151508030, + "step": 4665 + }, + { + "epoch": 0.4207963205122424, + "flos": 26207427977280.0, + "grad_norm": 1.887110256857361, + "learning_rate": 2.600784748520202e-06, + "loss": 0.8829, + "num_input_tokens_seen": 151537895, + "step": 4666 + }, + { + "epoch": 0.42088650403571265, + "flos": 20128471310880.0, + "grad_norm": 1.586219649749743, + "learning_rate": 2.6002274936178544e-06, + "loss": 0.6929, + "num_input_tokens_seen": 151566385, + "step": 4667 + }, + { + "epoch": 0.42097668755918294, + "flos": 17105201859840.0, + "grad_norm": 1.9155728978699382, + "learning_rate": 2.5996701875007873e-06, + "loss": 0.7705, + "num_input_tokens_seen": 151593000, + "step": 4668 + }, + { + "epoch": 0.4210668710826532, + "flos": 27451657617120.0, + "grad_norm": 1.6493892724278651, + "learning_rate": 2.5991128302165533e-06, + "loss": 0.743, + "num_input_tokens_seen": 151621310, + "step": 4669 + }, + { + "epoch": 0.42115705460612346, + "flos": 23371113917280.0, + "grad_norm": 2.291644102043266, + "learning_rate": 2.5985554218127094e-06, + "loss": 0.8196, + "num_input_tokens_seen": 151651090, + "step": 4670 + }, + { + "epoch": 0.42124723812959375, + "flos": 17686702983360.0, + "grad_norm": 2.1320736901464836, + "learning_rate": 2.597997962336816e-06, + "loss": 0.6895, + "num_input_tokens_seen": 151677990, + "step": 4671 + }, + { + "epoch": 0.421337421653064, + "flos": 21658210576320.0, + "grad_norm": 2.229783789754952, + "learning_rate": 2.5974404518364393e-06, + "loss": 0.7453, + "num_input_tokens_seen": 151707295, + "step": 4672 + }, + { + "epoch": 0.42142760517653427, + "flos": 21075817378560.0, + "grad_norm": 1.821025463183886, + "learning_rate": 2.596882890359149e-06, + "loss": 0.8763, + "num_input_tokens_seen": 151732995, + "step": 4673 + }, + { + "epoch": 0.4215177887000045, + "flos": 18451777049760.0, + "grad_norm": 2.3844771777016915, + "learning_rate": 2.5963252779525196e-06, + "loss": 0.6779, + "num_input_tokens_seen": 151758990, + "step": 4674 + }, + { + "epoch": 0.4216079722234748, + "flos": 29636505598080.0, + "grad_norm": 1.7522920176556676, + "learning_rate": 2.595767614664129e-06, + "loss": 0.7579, + "num_input_tokens_seen": 151788455, + "step": 4675 + }, + { + "epoch": 0.421698155746945, + "flos": 18015149415360.0, + "grad_norm": 1.6903744693510698, + "learning_rate": 2.5952099005415607e-06, + "loss": 0.7359, + "num_input_tokens_seen": 151815765, + "step": 4676 + }, + { + "epoch": 0.4217883392704153, + "flos": 23152000950240.0, + "grad_norm": 3.1091526277856176, + "learning_rate": 2.594652135632402e-06, + "loss": 0.7205, + "num_input_tokens_seen": 151841960, + "step": 4677 + }, + { + "epoch": 0.42187852279388555, + "flos": 25115096911200.0, + "grad_norm": 1.3617578865244495, + "learning_rate": 2.594094319984244e-06, + "loss": 0.7115, + "num_input_tokens_seen": 151871430, + "step": 4678 + }, + { + "epoch": 0.42196870631735584, + "flos": 24136485341760.0, + "grad_norm": 1.9270178362739014, + "learning_rate": 2.5935364536446825e-06, + "loss": 0.8142, + "num_input_tokens_seen": 151900920, + "step": 4679 + }, + { + "epoch": 0.42205888984082607, + "flos": 21941118318720.0, + "grad_norm": 2.2787146915749164, + "learning_rate": 2.5929785366613185e-06, + "loss": 0.7568, + "num_input_tokens_seen": 151925880, + "step": 4680 + }, + { + "epoch": 0.42214907336429636, + "flos": 27669692661120.0, + "grad_norm": 2.070030682654276, + "learning_rate": 2.592420569081756e-06, + "loss": 0.7914, + "num_input_tokens_seen": 151953920, + "step": 4681 + }, + { + "epoch": 0.4222392568877666, + "flos": 30547828434720.0, + "grad_norm": 3.040549103435567, + "learning_rate": 2.5918625509536037e-06, + "loss": 0.7141, + "num_input_tokens_seen": 151984740, + "step": 4682 + }, + { + "epoch": 0.4223294404112369, + "flos": 67684471331040.0, + "grad_norm": 0.6219766207913199, + "learning_rate": 2.591304482324475e-06, + "loss": 0.5776, + "num_input_tokens_seen": 152078360, + "step": 4683 + }, + { + "epoch": 0.4224196239347071, + "flos": 23146760014080.0, + "grad_norm": 2.1097705611460955, + "learning_rate": 2.5907463632419878e-06, + "loss": 0.7681, + "num_input_tokens_seen": 152107305, + "step": 4684 + }, + { + "epoch": 0.4225098074581774, + "flos": 26828446289280.0, + "grad_norm": 2.3058207166643903, + "learning_rate": 2.5901881937537632e-06, + "loss": 0.7199, + "num_input_tokens_seen": 152134930, + "step": 4685 + }, + { + "epoch": 0.42259999098164763, + "flos": 67126355720160.0, + "grad_norm": 0.7125988506878522, + "learning_rate": 2.589629973907428e-06, + "loss": 0.6117, + "num_input_tokens_seen": 152221295, + "step": 4686 + }, + { + "epoch": 0.4226901745051179, + "flos": 26795470978560.0, + "grad_norm": 2.2100643954669277, + "learning_rate": 2.589071703750612e-06, + "loss": 0.7083, + "num_input_tokens_seen": 152249900, + "step": 4687 + }, + { + "epoch": 0.42278035802858815, + "flos": 19542955853280.0, + "grad_norm": 1.9530887447806906, + "learning_rate": 2.5885133833309504e-06, + "loss": 0.7939, + "num_input_tokens_seen": 152278070, + "step": 4688 + }, + { + "epoch": 0.42287054155205844, + "flos": 21693564751680.0, + "grad_norm": 1.6190037502695418, + "learning_rate": 2.5879550126960814e-06, + "loss": 0.8337, + "num_input_tokens_seen": 152305740, + "step": 4689 + }, + { + "epoch": 0.4229607250755287, + "flos": 20456323026720.0, + "grad_norm": 1.514651743075455, + "learning_rate": 2.5873965918936494e-06, + "loss": 0.812, + "num_input_tokens_seen": 152333975, + "step": 4690 + }, + { + "epoch": 0.42305090859899896, + "flos": 30614187923520.0, + "grad_norm": 1.7783510811923233, + "learning_rate": 2.586838120971301e-06, + "loss": 0.7826, + "num_input_tokens_seen": 152363365, + "step": 4691 + }, + { + "epoch": 0.4231410921224692, + "flos": 25513248110400.0, + "grad_norm": 1.5208140942475792, + "learning_rate": 2.586279599976689e-06, + "loss": 0.8173, + "num_input_tokens_seen": 152391850, + "step": 4692 + }, + { + "epoch": 0.4232312756459395, + "flos": 23953432775520.0, + "grad_norm": 1.6297289191127708, + "learning_rate": 2.585721028957468e-06, + "loss": 0.7542, + "num_input_tokens_seen": 152421705, + "step": 4693 + }, + { + "epoch": 0.4233214591694098, + "flos": 24645382456800.0, + "grad_norm": 1.6208467226366163, + "learning_rate": 2.585162407961299e-06, + "loss": 0.7661, + "num_input_tokens_seen": 152450285, + "step": 4694 + }, + { + "epoch": 0.42341164269288, + "flos": 19290384368640.0, + "grad_norm": 2.025850955515638, + "learning_rate": 2.584603737035847e-06, + "loss": 0.697, + "num_input_tokens_seen": 152476915, + "step": 4695 + }, + { + "epoch": 0.4235018262163503, + "flos": 23950236176160.0, + "grad_norm": 2.059946733915667, + "learning_rate": 2.5840450162287806e-06, + "loss": 0.7506, + "num_input_tokens_seen": 152503885, + "step": 4696 + }, + { + "epoch": 0.42359200973982053, + "flos": 26358397307040.0, + "grad_norm": 8.933501089238518, + "learning_rate": 2.583486245587774e-06, + "loss": 0.669, + "num_input_tokens_seen": 152533520, + "step": 4697 + }, + { + "epoch": 0.4236821932632908, + "flos": 19654482334080.0, + "grad_norm": 1.97918274186951, + "learning_rate": 2.5829274251605023e-06, + "loss": 0.8607, + "num_input_tokens_seen": 152560855, + "step": 4698 + }, + { + "epoch": 0.42377237678676105, + "flos": 18708214189440.0, + "grad_norm": 1.7530518895644513, + "learning_rate": 2.582368554994649e-06, + "loss": 0.7318, + "num_input_tokens_seen": 152588575, + "step": 4699 + }, + { + "epoch": 0.42386256031023134, + "flos": 65856436042560.0, + "grad_norm": 0.7477637967973452, + "learning_rate": 2.5818096351378994e-06, + "loss": 0.6132, + "num_input_tokens_seen": 152670405, + "step": 4700 + }, + { + "epoch": 0.4239527438337016, + "flos": 31713804262560.0, + "grad_norm": 2.6999416239043863, + "learning_rate": 2.5812506656379435e-06, + "loss": 0.8117, + "num_input_tokens_seen": 152700175, + "step": 4701 + }, + { + "epoch": 0.42404292735717186, + "flos": 24172545742560.0, + "grad_norm": 2.2839114197558135, + "learning_rate": 2.580691646542476e-06, + "loss": 0.7059, + "num_input_tokens_seen": 152727735, + "step": 4702 + }, + { + "epoch": 0.4241331108806421, + "flos": 22966234991520.0, + "grad_norm": 1.5738810515433967, + "learning_rate": 2.5801325778991958e-06, + "loss": 0.7894, + "num_input_tokens_seen": 152754355, + "step": 4703 + }, + { + "epoch": 0.4242232944041124, + "flos": 20747296607040.0, + "grad_norm": 4.045910812458446, + "learning_rate": 2.5795734597558043e-06, + "loss": 0.613, + "num_input_tokens_seen": 152781360, + "step": 4704 + }, + { + "epoch": 0.4243134779275826, + "flos": 24501029344320.0, + "grad_norm": 3.4731586380394135, + "learning_rate": 2.579014292160011e-06, + "loss": 0.7564, + "num_input_tokens_seen": 152809930, + "step": 4705 + }, + { + "epoch": 0.4244036614510529, + "flos": 26722272253920.0, + "grad_norm": 1.46145514634923, + "learning_rate": 2.5784550751595236e-06, + "loss": 0.739, + "num_input_tokens_seen": 152842505, + "step": 4706 + }, + { + "epoch": 0.42449384497452314, + "flos": 20129326215360.0, + "grad_norm": 1.4445551357079098, + "learning_rate": 2.577895808802061e-06, + "loss": 0.8154, + "num_input_tokens_seen": 152869360, + "step": 4707 + }, + { + "epoch": 0.4245840284979934, + "flos": 28837303958400.0, + "grad_norm": 1.813468937197889, + "learning_rate": 2.577336493135341e-06, + "loss": 0.673, + "num_input_tokens_seen": 152901580, + "step": 4708 + }, + { + "epoch": 0.42467421202146366, + "flos": 21550252392480.0, + "grad_norm": 2.0295984449811426, + "learning_rate": 2.576777128207088e-06, + "loss": 0.8522, + "num_input_tokens_seen": 152930890, + "step": 4709 + }, + { + "epoch": 0.42476439554493395, + "flos": 31091596518240.0, + "grad_norm": 1.5441208799312311, + "learning_rate": 2.5762177140650306e-06, + "loss": 0.7162, + "num_input_tokens_seen": 152962895, + "step": 4710 + }, + { + "epoch": 0.4248545790684042, + "flos": 22897236449760.0, + "grad_norm": 2.1042635552285938, + "learning_rate": 2.5756582507569003e-06, + "loss": 0.7656, + "num_input_tokens_seen": 152989435, + "step": 4711 + }, + { + "epoch": 0.42494476259187447, + "flos": 27086741916960.0, + "grad_norm": 1.7842018698162658, + "learning_rate": 2.5750987383304335e-06, + "loss": 0.7807, + "num_input_tokens_seen": 153020345, + "step": 4712 + }, + { + "epoch": 0.4250349461153447, + "flos": 28507853942880.0, + "grad_norm": 2.3203442503943537, + "learning_rate": 2.574539176833372e-06, + "loss": 0.743, + "num_input_tokens_seen": 153047230, + "step": 4713 + }, + { + "epoch": 0.425125129638815, + "flos": 22384473679680.0, + "grad_norm": 1.9399689356292964, + "learning_rate": 2.5739795663134594e-06, + "loss": 0.7056, + "num_input_tokens_seen": 153075355, + "step": 4714 + }, + { + "epoch": 0.4252153131622852, + "flos": 25042418563200.0, + "grad_norm": 1.5737656161235694, + "learning_rate": 2.5734199068184454e-06, + "loss": 0.8413, + "num_input_tokens_seen": 153105125, + "step": 4715 + }, + { + "epoch": 0.4253054966857555, + "flos": 31380191233920.0, + "grad_norm": 1.8032286060523288, + "learning_rate": 2.572860198396083e-06, + "loss": 0.6793, + "num_input_tokens_seen": 153132935, + "step": 4716 + }, + { + "epoch": 0.4253956802092258, + "flos": 27742928555520.0, + "grad_norm": 1.814874409200274, + "learning_rate": 2.57230044109413e-06, + "loss": 0.7799, + "num_input_tokens_seen": 153161415, + "step": 4717 + }, + { + "epoch": 0.42548586373269603, + "flos": 20455728310560.0, + "grad_norm": 2.2221653401194352, + "learning_rate": 2.5717406349603483e-06, + "loss": 0.7395, + "num_input_tokens_seen": 153186460, + "step": 4718 + }, + { + "epoch": 0.4255760472561663, + "flos": 27848730893280.0, + "grad_norm": 1.7583365104369377, + "learning_rate": 2.5711807800425026e-06, + "loss": 0.773, + "num_input_tokens_seen": 153216010, + "step": 4719 + }, + { + "epoch": 0.42566623077963656, + "flos": 31634249508960.0, + "grad_norm": 2.3402054495974673, + "learning_rate": 2.5706208763883633e-06, + "loss": 0.5921, + "num_input_tokens_seen": 153245795, + "step": 4720 + }, + { + "epoch": 0.42575641430310684, + "flos": 22679164236000.0, + "grad_norm": 1.7841856339913562, + "learning_rate": 2.570060924045704e-06, + "loss": 0.7132, + "num_input_tokens_seen": 153273050, + "step": 4721 + }, + { + "epoch": 0.4258465978265771, + "flos": 22635930071520.0, + "grad_norm": 1.6175381929820436, + "learning_rate": 2.569500923062304e-06, + "loss": 0.788, + "num_input_tokens_seen": 153301185, + "step": 4722 + }, + { + "epoch": 0.42593678135004737, + "flos": 43590779850240.0, + "grad_norm": 2.023607990638051, + "learning_rate": 2.5689408734859445e-06, + "loss": 0.6782, + "num_input_tokens_seen": 153332010, + "step": 4723 + }, + { + "epoch": 0.4260269648735176, + "flos": 22023163446240.0, + "grad_norm": 1.8404039311704248, + "learning_rate": 2.5683807753644127e-06, + "loss": 0.7001, + "num_input_tokens_seen": 153361205, + "step": 4724 + }, + { + "epoch": 0.4261171483969879, + "flos": 16849582454880.0, + "grad_norm": 1.544817085425681, + "learning_rate": 2.5678206287454996e-06, + "loss": 0.7333, + "num_input_tokens_seen": 153387520, + "step": 4725 + }, + { + "epoch": 0.4262073319204581, + "flos": 22095581605920.0, + "grad_norm": 2.2583486586985644, + "learning_rate": 2.567260433676999e-06, + "loss": 0.7886, + "num_input_tokens_seen": 153416550, + "step": 4726 + }, + { + "epoch": 0.4262975154439284, + "flos": 27487866696960.0, + "grad_norm": 2.0143274670093105, + "learning_rate": 2.5667001902067107e-06, + "loss": 0.7291, + "num_input_tokens_seen": 153444485, + "step": 4727 + }, + { + "epoch": 0.42638769896739864, + "flos": 24646051512480.0, + "grad_norm": 2.244134675525186, + "learning_rate": 2.566139898382437e-06, + "loss": 0.8076, + "num_input_tokens_seen": 153472530, + "step": 4728 + }, + { + "epoch": 0.42647788249086893, + "flos": 20199365510400.0, + "grad_norm": 3.728627910092256, + "learning_rate": 2.5655795582519853e-06, + "loss": 0.771, + "num_input_tokens_seen": 153500885, + "step": 4729 + }, + { + "epoch": 0.42656806601433916, + "flos": 24241655793600.0, + "grad_norm": 2.0218239024572857, + "learning_rate": 2.565019169863168e-06, + "loss": 0.8283, + "num_input_tokens_seen": 153529520, + "step": 4730 + }, + { + "epoch": 0.42665824953780945, + "flos": 64977456630720.0, + "grad_norm": 0.586500177420795, + "learning_rate": 2.5644587332637994e-06, + "loss": 0.547, + "num_input_tokens_seen": 153629635, + "step": 4731 + }, + { + "epoch": 0.4267484330612797, + "flos": 26176125305760.0, + "grad_norm": 1.9331393237713324, + "learning_rate": 2.5638982485016994e-06, + "loss": 0.7865, + "num_input_tokens_seen": 153658595, + "step": 4732 + }, + { + "epoch": 0.42683861658475, + "flos": 70773691403520.0, + "grad_norm": 0.7081816361244886, + "learning_rate": 2.5633377156246917e-06, + "loss": 0.5516, + "num_input_tokens_seen": 153748455, + "step": 4733 + }, + { + "epoch": 0.4269288001082202, + "flos": 19254175288800.0, + "grad_norm": 2.533639415855223, + "learning_rate": 2.562777134680603e-06, + "loss": 0.7368, + "num_input_tokens_seen": 153776115, + "step": 4734 + }, + { + "epoch": 0.4270189836316905, + "flos": 26030545591200.0, + "grad_norm": 1.8006957855134615, + "learning_rate": 2.562216505717267e-06, + "loss": 0.7627, + "num_input_tokens_seen": 153803830, + "step": 4735 + }, + { + "epoch": 0.42710916715516073, + "flos": 20274162534720.0, + "grad_norm": 1.723495936159589, + "learning_rate": 2.561655828782518e-06, + "loss": 0.7811, + "num_input_tokens_seen": 153831430, + "step": 4736 + }, + { + "epoch": 0.427199350678631, + "flos": 23331448049760.0, + "grad_norm": 2.4370131373186363, + "learning_rate": 2.561095103924197e-06, + "loss": 0.7412, + "num_input_tokens_seen": 153857850, + "step": 4737 + }, + { + "epoch": 0.42728953420210125, + "flos": 25372240276320.0, + "grad_norm": 1.6925441094965745, + "learning_rate": 2.560534331190148e-06, + "loss": 0.7045, + "num_input_tokens_seen": 153888625, + "step": 4738 + }, + { + "epoch": 0.42737971772557154, + "flos": 70838006555520.0, + "grad_norm": 0.5834601610683537, + "learning_rate": 2.559973510628218e-06, + "loss": 0.5486, + "num_input_tokens_seen": 153984575, + "step": 4739 + }, + { + "epoch": 0.4274699012490418, + "flos": 21331325274240.0, + "grad_norm": 4.085193573603762, + "learning_rate": 2.5594126422862615e-06, + "loss": 0.7043, + "num_input_tokens_seen": 154013640, + "step": 4740 + }, + { + "epoch": 0.42756008477251206, + "flos": 24426938545440.0, + "grad_norm": 2.1872999917283176, + "learning_rate": 2.558851726212134e-06, + "loss": 0.7042, + "num_input_tokens_seen": 154041865, + "step": 4741 + }, + { + "epoch": 0.42765026829598235, + "flos": 28288629466560.0, + "grad_norm": 1.5401319476982487, + "learning_rate": 2.5582907624536953e-06, + "loss": 0.7238, + "num_input_tokens_seen": 154072345, + "step": 4742 + }, + { + "epoch": 0.4277404518194526, + "flos": 20893322358720.0, + "grad_norm": 3.009518464056675, + "learning_rate": 2.557729751058811e-06, + "loss": 0.7722, + "num_input_tokens_seen": 154099015, + "step": 4743 + }, + { + "epoch": 0.42783063534292287, + "flos": 17210595330240.0, + "grad_norm": 1.8432816379304513, + "learning_rate": 2.557168692075348e-06, + "loss": 0.7289, + "num_input_tokens_seen": 154125975, + "step": 4744 + }, + { + "epoch": 0.4279208188663931, + "flos": 26249026672320.0, + "grad_norm": 1.6132536878840715, + "learning_rate": 2.556607585551181e-06, + "loss": 0.6968, + "num_input_tokens_seen": 154156575, + "step": 4745 + }, + { + "epoch": 0.4280110023898634, + "flos": 66301947249600.0, + "grad_norm": 0.7153684014458144, + "learning_rate": 2.5560464315341844e-06, + "loss": 0.6767, + "num_input_tokens_seen": 154241995, + "step": 4746 + }, + { + "epoch": 0.4281011859133336, + "flos": 20893619716800.0, + "grad_norm": 1.8792280845445144, + "learning_rate": 2.555485230072242e-06, + "loss": 0.7472, + "num_input_tokens_seen": 154271175, + "step": 4747 + }, + { + "epoch": 0.4281913694368039, + "flos": 25484064115200.0, + "grad_norm": 1.8215900571279524, + "learning_rate": 2.5549239812132354e-06, + "loss": 0.7994, + "num_input_tokens_seen": 154297860, + "step": 4748 + }, + { + "epoch": 0.42828155296027415, + "flos": 26030619930720.0, + "grad_norm": 1.5932439860959642, + "learning_rate": 2.5543626850050556e-06, + "loss": 0.8176, + "num_input_tokens_seen": 154328115, + "step": 4749 + }, + { + "epoch": 0.42837173648374444, + "flos": 27159791962560.0, + "grad_norm": 2.5610561489994, + "learning_rate": 2.5538013414955944e-06, + "loss": 0.808, + "num_input_tokens_seen": 154357375, + "step": 4750 + }, + { + "epoch": 0.42846192000721467, + "flos": 24719473255680.0, + "grad_norm": 2.585474716219292, + "learning_rate": 2.5532399507327494e-06, + "loss": 0.7115, + "num_input_tokens_seen": 154388025, + "step": 4751 + }, + { + "epoch": 0.42855210353068496, + "flos": 67245093134400.0, + "grad_norm": 0.6006901844390559, + "learning_rate": 2.552678512764421e-06, + "loss": 0.5855, + "num_input_tokens_seen": 154488055, + "step": 4752 + }, + { + "epoch": 0.4286422870541552, + "flos": 28872026247840.0, + "grad_norm": 2.055509569610241, + "learning_rate": 2.5521170276385147e-06, + "loss": 0.8105, + "num_input_tokens_seen": 154515085, + "step": 4753 + }, + { + "epoch": 0.4287324705776255, + "flos": 24318088287360.0, + "grad_norm": 1.5773061762014617, + "learning_rate": 2.5515554954029394e-06, + "loss": 0.8035, + "num_input_tokens_seen": 154543170, + "step": 4754 + }, + { + "epoch": 0.4288226541010957, + "flos": 24974089077120.0, + "grad_norm": 1.8025958337584782, + "learning_rate": 2.550993916105608e-06, + "loss": 0.7832, + "num_input_tokens_seen": 154574360, + "step": 4755 + }, + { + "epoch": 0.428912837624566, + "flos": 24391101163200.0, + "grad_norm": 1.4691307710592494, + "learning_rate": 2.550432289794437e-06, + "loss": 0.712, + "num_input_tokens_seen": 154602965, + "step": 4756 + }, + { + "epoch": 0.42900302114803623, + "flos": 26941125032640.0, + "grad_norm": 3.2726266079650963, + "learning_rate": 2.5498706165173483e-06, + "loss": 0.7633, + "num_input_tokens_seen": 154629290, + "step": 4757 + }, + { + "epoch": 0.4290932046715065, + "flos": 24790181606400.0, + "grad_norm": 1.4508707753272199, + "learning_rate": 2.5493088963222668e-06, + "loss": 0.7292, + "num_input_tokens_seen": 154659080, + "step": 4758 + }, + { + "epoch": 0.42918338819497676, + "flos": 22314397214880.0, + "grad_norm": 1.8948942171027383, + "learning_rate": 2.548747129257121e-06, + "loss": 0.8055, + "num_input_tokens_seen": 154689535, + "step": 4759 + }, + { + "epoch": 0.42927357171844704, + "flos": 21292142613600.0, + "grad_norm": 1.7558852417666788, + "learning_rate": 2.548185315369845e-06, + "loss": 0.6767, + "num_input_tokens_seen": 154718305, + "step": 4760 + }, + { + "epoch": 0.4293637552419173, + "flos": 25120114828800.0, + "grad_norm": 1.7045331960295103, + "learning_rate": 2.5476234547083746e-06, + "loss": 0.8432, + "num_input_tokens_seen": 154748270, + "step": 4761 + }, + { + "epoch": 0.42945393876538757, + "flos": 26102889411360.0, + "grad_norm": 1.7189470351593814, + "learning_rate": 2.547061547320652e-06, + "loss": 0.8234, + "num_input_tokens_seen": 154776170, + "step": 4762 + }, + { + "epoch": 0.4295441222888578, + "flos": 24312921690720.0, + "grad_norm": 1.33224628759931, + "learning_rate": 2.5464995932546217e-06, + "loss": 0.7382, + "num_input_tokens_seen": 154806485, + "step": 4763 + }, + { + "epoch": 0.4296343058123281, + "flos": 25554846805440.0, + "grad_norm": 2.024968808722754, + "learning_rate": 2.545937592558232e-06, + "loss": 0.7077, + "num_input_tokens_seen": 154836855, + "step": 4764 + }, + { + "epoch": 0.4297244893357984, + "flos": 44392471863840.0, + "grad_norm": 1.57174882628875, + "learning_rate": 2.5453755452794374e-06, + "loss": 0.7015, + "num_input_tokens_seen": 154867815, + "step": 4765 + }, + { + "epoch": 0.4298146728592686, + "flos": 25117698794400.0, + "grad_norm": 1.487962984111766, + "learning_rate": 2.5448134514661938e-06, + "loss": 0.7967, + "num_input_tokens_seen": 154897175, + "step": 4766 + }, + { + "epoch": 0.4299048563827389, + "flos": 23836628188800.0, + "grad_norm": 1.4690001254461453, + "learning_rate": 2.5442513111664623e-06, + "loss": 0.7489, + "num_input_tokens_seen": 154924160, + "step": 4767 + }, + { + "epoch": 0.42999503990620913, + "flos": 14153049626880.0, + "grad_norm": 2.1441893591670227, + "learning_rate": 2.5436891244282084e-06, + "loss": 0.7465, + "num_input_tokens_seen": 154948990, + "step": 4768 + }, + { + "epoch": 0.4300852234296794, + "flos": 19503104136960.0, + "grad_norm": 1.7075164594178116, + "learning_rate": 2.5431268912994004e-06, + "loss": 0.8521, + "num_input_tokens_seen": 154974905, + "step": 4769 + }, + { + "epoch": 0.43017540695314965, + "flos": 20821313066400.0, + "grad_norm": 1.7347773179826453, + "learning_rate": 2.5425646118280108e-06, + "loss": 0.725, + "num_input_tokens_seen": 155002980, + "step": 4770 + }, + { + "epoch": 0.43026559047661994, + "flos": 66624409350240.0, + "grad_norm": 0.591291129675719, + "learning_rate": 2.5420022860620172e-06, + "loss": 0.5617, + "num_input_tokens_seen": 155101895, + "step": 4771 + }, + { + "epoch": 0.4303557740000902, + "flos": 15391852481760.0, + "grad_norm": 1.8522322204041324, + "learning_rate": 2.5414399140493995e-06, + "loss": 0.6852, + "num_input_tokens_seen": 155128155, + "step": 4772 + }, + { + "epoch": 0.43044595752356046, + "flos": 37615544015040.0, + "grad_norm": 3.188460411042553, + "learning_rate": 2.5408774958381436e-06, + "loss": 0.6947, + "num_input_tokens_seen": 155159160, + "step": 4773 + }, + { + "epoch": 0.4305361410470307, + "flos": 22424028037920.0, + "grad_norm": 1.8807972613543096, + "learning_rate": 2.540315031476237e-06, + "loss": 0.789, + "num_input_tokens_seen": 155189675, + "step": 4774 + }, + { + "epoch": 0.430626324570501, + "flos": 21331139425440.0, + "grad_norm": 2.0885012298499994, + "learning_rate": 2.5397525210116737e-06, + "loss": 0.7619, + "num_input_tokens_seen": 155217390, + "step": 4775 + }, + { + "epoch": 0.4307165080939712, + "flos": 35394115256640.0, + "grad_norm": 1.7324782301070336, + "learning_rate": 2.539189964492448e-06, + "loss": 0.708, + "num_input_tokens_seen": 155249060, + "step": 4776 + }, + { + "epoch": 0.4308066916174415, + "flos": 28136842402080.0, + "grad_norm": 1.7217459490094054, + "learning_rate": 2.5386273619665613e-06, + "loss": 0.7936, + "num_input_tokens_seen": 155277410, + "step": 4777 + }, + { + "epoch": 0.43089687514091174, + "flos": 56329222159200.0, + "grad_norm": 0.6791125529521442, + "learning_rate": 2.5380647134820186e-06, + "loss": 0.5833, + "num_input_tokens_seen": 155356575, + "step": 4778 + }, + { + "epoch": 0.43098705866438203, + "flos": 19288488710880.0, + "grad_norm": 1.4127474829672102, + "learning_rate": 2.5375020190868277e-06, + "loss": 0.8046, + "num_input_tokens_seen": 155384640, + "step": 4779 + }, + { + "epoch": 0.43107724218785226, + "flos": 21767172343680.0, + "grad_norm": 1.5660932042580409, + "learning_rate": 2.536939278829001e-06, + "loss": 0.7482, + "num_input_tokens_seen": 155416165, + "step": 4780 + }, + { + "epoch": 0.43116742571132255, + "flos": 22861287558240.0, + "grad_norm": 2.1894750126094302, + "learning_rate": 2.5363764927565536e-06, + "loss": 0.8153, + "num_input_tokens_seen": 155443940, + "step": 4781 + }, + { + "epoch": 0.4312576092347928, + "flos": 26102257525440.0, + "grad_norm": 2.171538621938893, + "learning_rate": 2.5358136609175064e-06, + "loss": 0.7079, + "num_input_tokens_seen": 155476015, + "step": 4782 + }, + { + "epoch": 0.43134779275826307, + "flos": 20274868760160.0, + "grad_norm": 1.4815973896584471, + "learning_rate": 2.535250783359884e-06, + "loss": 0.7567, + "num_input_tokens_seen": 155503270, + "step": 4783 + }, + { + "epoch": 0.4314379762817333, + "flos": 29600296518240.0, + "grad_norm": 1.5215562208496198, + "learning_rate": 2.5346878601317124e-06, + "loss": 0.6495, + "num_input_tokens_seen": 155533775, + "step": 4784 + }, + { + "epoch": 0.4315281598052036, + "flos": 70669227177120.0, + "grad_norm": 0.5730813627780046, + "learning_rate": 2.534124891281025e-06, + "loss": 0.5461, + "num_input_tokens_seen": 155633555, + "step": 4785 + }, + { + "epoch": 0.4316183433286738, + "flos": 22022903257920.0, + "grad_norm": 1.7786224795623964, + "learning_rate": 2.533561876855857e-06, + "loss": 0.7644, + "num_input_tokens_seen": 155661525, + "step": 4786 + }, + { + "epoch": 0.4317085268521441, + "flos": 22095916133760.0, + "grad_norm": 7.455020303679359, + "learning_rate": 2.532998816904247e-06, + "loss": 0.7038, + "num_input_tokens_seen": 155690765, + "step": 4787 + }, + { + "epoch": 0.4317987103756144, + "flos": 40858149451680.0, + "grad_norm": 1.7787754203390032, + "learning_rate": 2.53243571147424e-06, + "loss": 0.6444, + "num_input_tokens_seen": 155723360, + "step": 4788 + }, + { + "epoch": 0.43188889389908464, + "flos": 22422206719680.0, + "grad_norm": 1.9835988126181578, + "learning_rate": 2.5318725606138815e-06, + "loss": 0.6619, + "num_input_tokens_seen": 155751080, + "step": 4789 + }, + { + "epoch": 0.4319790774225549, + "flos": 25297926458880.0, + "grad_norm": 1.8769770689200194, + "learning_rate": 2.5313093643712235e-06, + "loss": 0.7614, + "num_input_tokens_seen": 155781350, + "step": 4790 + }, + { + "epoch": 0.43206926094602516, + "flos": 22351424029440.0, + "grad_norm": 1.8440934200826669, + "learning_rate": 2.530746122794321e-06, + "loss": 0.802, + "num_input_tokens_seen": 155811185, + "step": 4791 + }, + { + "epoch": 0.43215944446949545, + "flos": 35866097066400.0, + "grad_norm": 1.5393729114711903, + "learning_rate": 2.5301828359312323e-06, + "loss": 0.6852, + "num_input_tokens_seen": 155843245, + "step": 4792 + }, + { + "epoch": 0.4322496279929657, + "flos": 29229656675040.0, + "grad_norm": 1.8341135907787702, + "learning_rate": 2.529619503830021e-06, + "loss": 0.712, + "num_input_tokens_seen": 155872665, + "step": 4793 + }, + { + "epoch": 0.43233981151643597, + "flos": 25808273194560.0, + "grad_norm": 2.5796816663485487, + "learning_rate": 2.529056126538753e-06, + "loss": 0.6819, + "num_input_tokens_seen": 155901205, + "step": 4794 + }, + { + "epoch": 0.4324299950399062, + "flos": 27488089715520.0, + "grad_norm": 1.8081626955145151, + "learning_rate": 2.5284927041054995e-06, + "loss": 0.7839, + "num_input_tokens_seen": 155931455, + "step": 4795 + }, + { + "epoch": 0.4325201785633765, + "flos": 55483658361600.0, + "grad_norm": 0.610270136170064, + "learning_rate": 2.5279292365783348e-06, + "loss": 0.5244, + "num_input_tokens_seen": 156026240, + "step": 4796 + }, + { + "epoch": 0.4326103620868467, + "flos": 27342212642880.0, + "grad_norm": 1.6531046384772516, + "learning_rate": 2.527365724005336e-06, + "loss": 0.8424, + "num_input_tokens_seen": 156054105, + "step": 4797 + }, + { + "epoch": 0.432700545610317, + "flos": 21148347047520.0, + "grad_norm": 1.8808102546845566, + "learning_rate": 2.526802166434586e-06, + "loss": 0.807, + "num_input_tokens_seen": 156081845, + "step": 4798 + }, + { + "epoch": 0.43279072913378724, + "flos": 23404944132480.0, + "grad_norm": 1.6113899727713787, + "learning_rate": 2.5262385639141708e-06, + "loss": 0.7214, + "num_input_tokens_seen": 156112060, + "step": 4799 + }, + { + "epoch": 0.43288091265725753, + "flos": 26500668912960.0, + "grad_norm": 1.549844300233159, + "learning_rate": 2.525674916492179e-06, + "loss": 0.824, + "num_input_tokens_seen": 156143460, + "step": 4800 + }, + { + "epoch": 0.43297109618072777, + "flos": 29018683885440.0, + "grad_norm": 1.3667591840938889, + "learning_rate": 2.5251112242167056e-06, + "loss": 0.759, + "num_input_tokens_seen": 156175880, + "step": 4801 + }, + { + "epoch": 0.43306127970419805, + "flos": 19506746773440.0, + "grad_norm": 1.8767151934830046, + "learning_rate": 2.5245474871358464e-06, + "loss": 0.7107, + "num_input_tokens_seen": 156204320, + "step": 4802 + }, + { + "epoch": 0.4331514632276683, + "flos": 17031334079520.0, + "grad_norm": 2.831557914377374, + "learning_rate": 2.5239837052977037e-06, + "loss": 0.6682, + "num_input_tokens_seen": 156231260, + "step": 4803 + }, + { + "epoch": 0.4332416467511386, + "flos": 18958927186080.0, + "grad_norm": 2.54967424518531, + "learning_rate": 2.523419878750381e-06, + "loss": 0.8644, + "num_input_tokens_seen": 156258460, + "step": 4804 + }, + { + "epoch": 0.4333318302746088, + "flos": 23586472738560.0, + "grad_norm": 2.071636424418046, + "learning_rate": 2.522856007541989e-06, + "loss": 0.706, + "num_input_tokens_seen": 156288530, + "step": 4805 + }, + { + "epoch": 0.4334220137980791, + "flos": 20412494146080.0, + "grad_norm": 1.8550476881411189, + "learning_rate": 2.5222920917206397e-06, + "loss": 0.6643, + "num_input_tokens_seen": 156315845, + "step": 4806 + }, + { + "epoch": 0.43351219732154933, + "flos": 70608002848800.0, + "grad_norm": 0.6446617225600716, + "learning_rate": 2.5217281313344493e-06, + "loss": 0.5616, + "num_input_tokens_seen": 156402375, + "step": 4807 + }, + { + "epoch": 0.4336023808450196, + "flos": 25119631621920.0, + "grad_norm": 1.678909192408863, + "learning_rate": 2.5211641264315372e-06, + "loss": 0.7996, + "num_input_tokens_seen": 156429820, + "step": 4808 + }, + { + "epoch": 0.43369256436848985, + "flos": 26246053091520.0, + "grad_norm": 1.7356065381407206, + "learning_rate": 2.5206000770600286e-06, + "loss": 0.7331, + "num_input_tokens_seen": 156460315, + "step": 4809 + }, + { + "epoch": 0.43378274789196014, + "flos": 25921695333120.0, + "grad_norm": 2.8629178470999066, + "learning_rate": 2.520035983268051e-06, + "loss": 0.7101, + "num_input_tokens_seen": 156490385, + "step": 4810 + }, + { + "epoch": 0.4338729314154304, + "flos": 33675562112160.0, + "grad_norm": 2.5752115144245256, + "learning_rate": 2.5194718451037357e-06, + "loss": 0.7524, + "num_input_tokens_seen": 156521165, + "step": 4811 + }, + { + "epoch": 0.43396311493890066, + "flos": 24464597245920.0, + "grad_norm": 1.8058980869547758, + "learning_rate": 2.518907662615218e-06, + "loss": 0.7804, + "num_input_tokens_seen": 156548485, + "step": 4812 + }, + { + "epoch": 0.43405329846237095, + "flos": 22059223847040.0, + "grad_norm": 1.879560709519555, + "learning_rate": 2.5183434358506373e-06, + "loss": 0.7139, + "num_input_tokens_seen": 156577000, + "step": 4813 + }, + { + "epoch": 0.4341434819858412, + "flos": 25265025487680.0, + "grad_norm": 3.2424616318596633, + "learning_rate": 2.5177791648581368e-06, + "loss": 0.7359, + "num_input_tokens_seen": 156607405, + "step": 4814 + }, + { + "epoch": 0.4342336655093115, + "flos": 22234656612480.0, + "grad_norm": 1.7997324241334245, + "learning_rate": 2.517214849685863e-06, + "loss": 0.7568, + "num_input_tokens_seen": 156633545, + "step": 4815 + }, + { + "epoch": 0.4343238490327817, + "flos": 18160914978720.0, + "grad_norm": 1.7538303282348056, + "learning_rate": 2.5166504903819663e-06, + "loss": 0.7703, + "num_input_tokens_seen": 156658705, + "step": 4816 + }, + { + "epoch": 0.434414032556252, + "flos": 22897310789280.0, + "grad_norm": 1.9335970791874748, + "learning_rate": 2.5160860869946014e-06, + "loss": 0.8191, + "num_input_tokens_seen": 156685135, + "step": 4817 + }, + { + "epoch": 0.4345042160797222, + "flos": 23698593935520.0, + "grad_norm": 1.6200191015129939, + "learning_rate": 2.5155216395719253e-06, + "loss": 0.8356, + "num_input_tokens_seen": 156715640, + "step": 4818 + }, + { + "epoch": 0.4345943996031925, + "flos": 26175196061760.0, + "grad_norm": 3.838267564584166, + "learning_rate": 2.5149571481621e-06, + "loss": 0.7146, + "num_input_tokens_seen": 156745880, + "step": 4819 + }, + { + "epoch": 0.43468458312666275, + "flos": 71544489612960.0, + "grad_norm": 0.6112401490695113, + "learning_rate": 2.514392612813292e-06, + "loss": 0.5435, + "num_input_tokens_seen": 156842310, + "step": 4820 + }, + { + "epoch": 0.43477476665013304, + "flos": 22715745013440.0, + "grad_norm": 4.177368661675299, + "learning_rate": 2.5138280335736695e-06, + "loss": 0.8054, + "num_input_tokens_seen": 156872625, + "step": 4821 + }, + { + "epoch": 0.43486495017360327, + "flos": 28831133778240.0, + "grad_norm": 3.4137444598985134, + "learning_rate": 2.5132634104914064e-06, + "loss": 0.8362, + "num_input_tokens_seen": 156902300, + "step": 4822 + }, + { + "epoch": 0.43495513369707356, + "flos": 27450728373120.0, + "grad_norm": 1.5655285285877167, + "learning_rate": 2.5126987436146794e-06, + "loss": 0.7661, + "num_input_tokens_seen": 156934210, + "step": 4823 + }, + { + "epoch": 0.4350453172205438, + "flos": 22388041976640.0, + "grad_norm": 1.6935015731997936, + "learning_rate": 2.5121340329916675e-06, + "loss": 0.7332, + "num_input_tokens_seen": 156961585, + "step": 4824 + }, + { + "epoch": 0.4351355007440141, + "flos": 20164866239520.0, + "grad_norm": 1.7033971866476492, + "learning_rate": 2.5115692786705566e-06, + "loss": 0.6735, + "num_input_tokens_seen": 156988740, + "step": 4825 + }, + { + "epoch": 0.4352256842674843, + "flos": 21803976139680.0, + "grad_norm": 1.8971935325279081, + "learning_rate": 2.511004480699534e-06, + "loss": 0.7916, + "num_input_tokens_seen": 157015530, + "step": 4826 + }, + { + "epoch": 0.4353158677909546, + "flos": 69773409864000.0, + "grad_norm": 0.7210246457002589, + "learning_rate": 2.510439639126791e-06, + "loss": 0.6189, + "num_input_tokens_seen": 157104015, + "step": 4827 + }, + { + "epoch": 0.43540605131442484, + "flos": 24537312763680.0, + "grad_norm": 1.5629922472907352, + "learning_rate": 2.509874754000524e-06, + "loss": 0.7525, + "num_input_tokens_seen": 157134110, + "step": 4828 + }, + { + "epoch": 0.4354962348378951, + "flos": 39363727191840.0, + "grad_norm": 2.0694518405631372, + "learning_rate": 2.509309825368932e-06, + "loss": 0.7377, + "num_input_tokens_seen": 157161655, + "step": 4829 + }, + { + "epoch": 0.43558641836136536, + "flos": 23006235386880.0, + "grad_norm": 2.3987641251919927, + "learning_rate": 2.5087448532802173e-06, + "loss": 0.7502, + "num_input_tokens_seen": 157191010, + "step": 4830 + }, + { + "epoch": 0.43567660188483565, + "flos": 23254829707200.0, + "grad_norm": 2.0558616851265996, + "learning_rate": 2.508179837782586e-06, + "loss": 0.7324, + "num_input_tokens_seen": 157221085, + "step": 4831 + }, + { + "epoch": 0.4357667854083059, + "flos": 26248506295680.0, + "grad_norm": 1.6189314427554253, + "learning_rate": 2.5076147789242493e-06, + "loss": 0.7238, + "num_input_tokens_seen": 157249905, + "step": 4832 + }, + { + "epoch": 0.43585696893177617, + "flos": 25520644892640.0, + "grad_norm": 1.6584547950442303, + "learning_rate": 2.5070496767534202e-06, + "loss": 0.7696, + "num_input_tokens_seen": 157279500, + "step": 4833 + }, + { + "epoch": 0.4359471524552464, + "flos": 22132199553120.0, + "grad_norm": 1.8620990705675138, + "learning_rate": 2.506484531318317e-06, + "loss": 0.7777, + "num_input_tokens_seen": 157307390, + "step": 4834 + }, + { + "epoch": 0.4360373359787167, + "flos": 66476079073440.0, + "grad_norm": 0.7607958166295229, + "learning_rate": 2.5059193426671613e-06, + "loss": 0.5907, + "num_input_tokens_seen": 157404450, + "step": 4835 + }, + { + "epoch": 0.436127519502187, + "flos": 28762544103840.0, + "grad_norm": 1.746795445031337, + "learning_rate": 2.5053541108481772e-06, + "loss": 0.8849, + "num_input_tokens_seen": 157433370, + "step": 4836 + }, + { + "epoch": 0.4362177030256572, + "flos": 17577927064800.0, + "grad_norm": 2.1278941697285365, + "learning_rate": 2.5047888359095935e-06, + "loss": 0.7356, + "num_input_tokens_seen": 157460195, + "step": 4837 + }, + { + "epoch": 0.4363078865491275, + "flos": 20273902346400.0, + "grad_norm": 2.3389296744327606, + "learning_rate": 2.5042235178996436e-06, + "loss": 0.7344, + "num_input_tokens_seen": 157484915, + "step": 4838 + }, + { + "epoch": 0.43639807007259773, + "flos": 19436224271520.0, + "grad_norm": 1.708468881517354, + "learning_rate": 2.5036581568665627e-06, + "loss": 0.8157, + "num_input_tokens_seen": 157511900, + "step": 4839 + }, + { + "epoch": 0.436488253596068, + "flos": 22460534475840.0, + "grad_norm": 1.9497745310207066, + "learning_rate": 2.503092752858591e-06, + "loss": 0.8178, + "num_input_tokens_seen": 157537035, + "step": 4840 + }, + { + "epoch": 0.43657843711953825, + "flos": 69642326355840.0, + "grad_norm": 0.7781243167246075, + "learning_rate": 2.502527305923971e-06, + "loss": 0.6074, + "num_input_tokens_seen": 157626770, + "step": 4841 + }, + { + "epoch": 0.43666862064300854, + "flos": 22453323542400.0, + "grad_norm": 2.0716962688333083, + "learning_rate": 2.5019618161109506e-06, + "loss": 0.7374, + "num_input_tokens_seen": 157648065, + "step": 4842 + }, + { + "epoch": 0.4367588041664788, + "flos": 28981396882560.0, + "grad_norm": 1.801478054585274, + "learning_rate": 2.5013962834677804e-06, + "loss": 0.7712, + "num_input_tokens_seen": 157678475, + "step": 4843 + }, + { + "epoch": 0.43684898768994906, + "flos": 23844322329120.0, + "grad_norm": 4.33857012033768, + "learning_rate": 2.500830708042715e-06, + "loss": 0.7395, + "num_input_tokens_seen": 157705115, + "step": 4844 + }, + { + "epoch": 0.4369391712134193, + "flos": 23589780847200.0, + "grad_norm": 1.8920192779179883, + "learning_rate": 2.500265089884011e-06, + "loss": 0.8165, + "num_input_tokens_seen": 157734925, + "step": 4845 + }, + { + "epoch": 0.4370293547368896, + "flos": 30547382397600.0, + "grad_norm": 1.5253527217752052, + "learning_rate": 2.499699429039932e-06, + "loss": 0.7517, + "num_input_tokens_seen": 157765485, + "step": 4846 + }, + { + "epoch": 0.4371195382603598, + "flos": 27706459287360.0, + "grad_norm": 1.707131738563808, + "learning_rate": 2.4991337255587425e-06, + "loss": 0.7042, + "num_input_tokens_seen": 157795115, + "step": 4847 + }, + { + "epoch": 0.4372097217838301, + "flos": 28937828190240.0, + "grad_norm": 1.8740401036175935, + "learning_rate": 2.4985679794887106e-06, + "loss": 0.6237, + "num_input_tokens_seen": 157823740, + "step": 4848 + }, + { + "epoch": 0.43729990530730034, + "flos": 23006830103040.0, + "grad_norm": 1.9076112358287096, + "learning_rate": 2.49800219087811e-06, + "loss": 0.7874, + "num_input_tokens_seen": 157850755, + "step": 4849 + }, + { + "epoch": 0.43739008883077063, + "flos": 14954332773120.0, + "grad_norm": 2.3993165831274865, + "learning_rate": 2.4974363597752163e-06, + "loss": 0.7869, + "num_input_tokens_seen": 157875695, + "step": 4850 + }, + { + "epoch": 0.43748027235424086, + "flos": 21257903531040.0, + "grad_norm": 2.333682806092791, + "learning_rate": 2.4968704862283097e-06, + "loss": 0.7977, + "num_input_tokens_seen": 157904955, + "step": 4851 + }, + { + "epoch": 0.43757045587771115, + "flos": 24281321661120.0, + "grad_norm": 1.651173856078404, + "learning_rate": 2.4963045702856737e-06, + "loss": 0.8215, + "num_input_tokens_seen": 157935370, + "step": 4852 + }, + { + "epoch": 0.4376606394011814, + "flos": 28471310335200.0, + "grad_norm": 1.5483202107465055, + "learning_rate": 2.4957386119955946e-06, + "loss": 0.697, + "num_input_tokens_seen": 157964375, + "step": 4853 + }, + { + "epoch": 0.4377508229246517, + "flos": 68445605402880.0, + "grad_norm": 0.686778969234594, + "learning_rate": 2.495172611406364e-06, + "loss": 0.5679, + "num_input_tokens_seen": 158064570, + "step": 4854 + }, + { + "epoch": 0.4378410064481219, + "flos": 23770937755680.0, + "grad_norm": 7.2189798690065, + "learning_rate": 2.4946065685662757e-06, + "loss": 0.7789, + "num_input_tokens_seen": 158092870, + "step": 4855 + }, + { + "epoch": 0.4379311899715922, + "flos": 27192915952320.0, + "grad_norm": 2.274580939965341, + "learning_rate": 2.4940404835236283e-06, + "loss": 0.7552, + "num_input_tokens_seen": 158121580, + "step": 4856 + }, + { + "epoch": 0.4380213734950624, + "flos": 32806469856480.0, + "grad_norm": 1.7428510283311527, + "learning_rate": 2.4934743563267223e-06, + "loss": 0.7423, + "num_input_tokens_seen": 158153885, + "step": 4857 + }, + { + "epoch": 0.4381115570185327, + "flos": 21546981453600.0, + "grad_norm": 1.9945246133421215, + "learning_rate": 2.4929081870238635e-06, + "loss": 0.7158, + "num_input_tokens_seen": 158183215, + "step": 4858 + }, + { + "epoch": 0.43820174054200295, + "flos": 16411765388160.0, + "grad_norm": 1.7004429408446375, + "learning_rate": 2.49234197566336e-06, + "loss": 0.7917, + "num_input_tokens_seen": 158209735, + "step": 4859 + }, + { + "epoch": 0.43829192406547324, + "flos": 28833140945280.0, + "grad_norm": 1.9193083208650954, + "learning_rate": 2.4917757222935247e-06, + "loss": 0.7718, + "num_input_tokens_seen": 158240435, + "step": 4860 + }, + { + "epoch": 0.4383821075889435, + "flos": 21875948262240.0, + "grad_norm": 1.782030837925684, + "learning_rate": 2.4912094269626725e-06, + "loss": 0.7747, + "num_input_tokens_seen": 158269300, + "step": 4861 + }, + { + "epoch": 0.43847229111241376, + "flos": 13533852633120.0, + "grad_norm": 1.9444438419643861, + "learning_rate": 2.4906430897191245e-06, + "loss": 0.7928, + "num_input_tokens_seen": 158296000, + "step": 4862 + }, + { + "epoch": 0.43856247463588405, + "flos": 25625109119040.0, + "grad_norm": 1.641803136791384, + "learning_rate": 2.490076710611202e-06, + "loss": 0.7874, + "num_input_tokens_seen": 158323560, + "step": 4863 + }, + { + "epoch": 0.4386526581593543, + "flos": 19400163870720.0, + "grad_norm": 2.1378126395395136, + "learning_rate": 2.4895102896872326e-06, + "loss": 0.7999, + "num_input_tokens_seen": 158349350, + "step": 4864 + }, + { + "epoch": 0.43874284168282457, + "flos": 23624354457600.0, + "grad_norm": 1.9215785472458506, + "learning_rate": 2.4889438269955457e-06, + "loss": 0.6843, + "num_input_tokens_seen": 158379205, + "step": 4865 + }, + { + "epoch": 0.4388330252062948, + "flos": 17360003530080.0, + "grad_norm": 1.5009945909935094, + "learning_rate": 2.4883773225844755e-06, + "loss": 0.7327, + "num_input_tokens_seen": 158407090, + "step": 4866 + }, + { + "epoch": 0.4389232087297651, + "flos": 15902905442880.0, + "grad_norm": 1.7381248146559072, + "learning_rate": 2.48781077650236e-06, + "loss": 0.7445, + "num_input_tokens_seen": 158432555, + "step": 4867 + }, + { + "epoch": 0.4390133922532353, + "flos": 28977940094880.0, + "grad_norm": 2.905285911212718, + "learning_rate": 2.4872441887975386e-06, + "loss": 0.7242, + "num_input_tokens_seen": 158463810, + "step": 4868 + }, + { + "epoch": 0.4391035757767056, + "flos": 26905138971360.0, + "grad_norm": 1.3448609040527926, + "learning_rate": 2.486677559518356e-06, + "loss": 0.6923, + "num_input_tokens_seen": 158491760, + "step": 4869 + }, + { + "epoch": 0.43919375930017585, + "flos": 24824643707520.0, + "grad_norm": 1.3417394367567708, + "learning_rate": 2.4861108887131614e-06, + "loss": 0.7398, + "num_input_tokens_seen": 158522095, + "step": 4870 + }, + { + "epoch": 0.43928394282364613, + "flos": 33130678935840.0, + "grad_norm": 2.018286197635912, + "learning_rate": 2.485544176430305e-06, + "loss": 0.7197, + "num_input_tokens_seen": 158554235, + "step": 4871 + }, + { + "epoch": 0.43937412634711637, + "flos": 25551687375840.0, + "grad_norm": 1.7914981836751067, + "learning_rate": 2.4849774227181425e-06, + "loss": 0.8312, + "num_input_tokens_seen": 158580950, + "step": 4872 + }, + { + "epoch": 0.43946430987058666, + "flos": 25302089472000.0, + "grad_norm": 1.501382984938273, + "learning_rate": 2.484410627625032e-06, + "loss": 0.751, + "num_input_tokens_seen": 158610440, + "step": 4873 + }, + { + "epoch": 0.4395544933940569, + "flos": 17099552056320.0, + "grad_norm": 1.8870859361415688, + "learning_rate": 2.4838437911993356e-06, + "loss": 0.7907, + "num_input_tokens_seen": 158635225, + "step": 4874 + }, + { + "epoch": 0.4396446769175272, + "flos": 23147057372160.0, + "grad_norm": 4.570012635395104, + "learning_rate": 2.483276913489419e-06, + "loss": 0.7922, + "num_input_tokens_seen": 158662320, + "step": 4875 + }, + { + "epoch": 0.4397348604409974, + "flos": 37870643043360.0, + "grad_norm": 1.8587256828518344, + "learning_rate": 2.4827099945436516e-06, + "loss": 0.6506, + "num_input_tokens_seen": 158690140, + "step": 4876 + }, + { + "epoch": 0.4398250439644677, + "flos": 13460468059680.0, + "grad_norm": 2.40702371419676, + "learning_rate": 2.482143034410405e-06, + "loss": 0.8342, + "num_input_tokens_seen": 158713075, + "step": 4877 + }, + { + "epoch": 0.43991522748793793, + "flos": 24714678356640.0, + "grad_norm": 1.5738440888921454, + "learning_rate": 2.4815760331380573e-06, + "loss": 0.7816, + "num_input_tokens_seen": 158743330, + "step": 4878 + }, + { + "epoch": 0.4400054110114082, + "flos": 27160423848480.0, + "grad_norm": 1.4825516594929113, + "learning_rate": 2.481008990774987e-06, + "loss": 0.7118, + "num_input_tokens_seen": 158774000, + "step": 4879 + }, + { + "epoch": 0.44009559453487845, + "flos": 14584250476320.0, + "grad_norm": 1.7956392766037752, + "learning_rate": 2.480441907369577e-06, + "loss": 0.7505, + "num_input_tokens_seen": 158799025, + "step": 4880 + }, + { + "epoch": 0.44018577805834874, + "flos": 32368727129280.0, + "grad_norm": 1.6533608338673542, + "learning_rate": 2.479874782970214e-06, + "loss": 0.6953, + "num_input_tokens_seen": 158832335, + "step": 4881 + }, + { + "epoch": 0.440275961581819, + "flos": 21912900737280.0, + "grad_norm": 2.1877523004264536, + "learning_rate": 2.4793076176252887e-06, + "loss": 0.7706, + "num_input_tokens_seen": 158861830, + "step": 4882 + }, + { + "epoch": 0.44036614510528926, + "flos": 25156063720320.0, + "grad_norm": 1.7896959940657315, + "learning_rate": 2.478740411383195e-06, + "loss": 0.6362, + "num_input_tokens_seen": 158894485, + "step": 4883 + }, + { + "epoch": 0.44045632862875955, + "flos": 26248431956160.0, + "grad_norm": 1.7652006061225298, + "learning_rate": 2.4781731642923296e-06, + "loss": 0.7434, + "num_input_tokens_seen": 158924945, + "step": 4884 + }, + { + "epoch": 0.4405465121522298, + "flos": 24866391081600.0, + "grad_norm": 2.6608421572921137, + "learning_rate": 2.477605876401093e-06, + "loss": 0.737, + "num_input_tokens_seen": 158955045, + "step": 4885 + }, + { + "epoch": 0.4406366956757001, + "flos": 25521983004000.0, + "grad_norm": 1.7654990092198852, + "learning_rate": 2.4770385477578894e-06, + "loss": 0.7097, + "num_input_tokens_seen": 158988725, + "step": 4886 + }, + { + "epoch": 0.4407268791991703, + "flos": 26537770067040.0, + "grad_norm": 1.4008678011688411, + "learning_rate": 2.476471178411127e-06, + "loss": 0.7066, + "num_input_tokens_seen": 159020085, + "step": 4887 + }, + { + "epoch": 0.4408170627226406, + "flos": 23877037451520.0, + "grad_norm": 3.1323048472827537, + "learning_rate": 2.475903768409216e-06, + "loss": 0.7712, + "num_input_tokens_seen": 159048495, + "step": 4888 + }, + { + "epoch": 0.44090724624611083, + "flos": 20419965267840.0, + "grad_norm": 1.8252725763433388, + "learning_rate": 2.475336317800572e-06, + "loss": 0.728, + "num_input_tokens_seen": 159076850, + "step": 4889 + }, + { + "epoch": 0.4409974297695811, + "flos": 23006161047360.0, + "grad_norm": 1.670204254725284, + "learning_rate": 2.4747688266336118e-06, + "loss": 0.7386, + "num_input_tokens_seen": 159103130, + "step": 4890 + }, + { + "epoch": 0.44108761329305135, + "flos": 20887152178560.0, + "grad_norm": 2.1610307630602628, + "learning_rate": 2.4742012949567574e-06, + "loss": 0.8544, + "num_input_tokens_seen": 159130610, + "step": 4891 + }, + { + "epoch": 0.44117779681652164, + "flos": 23188507388160.0, + "grad_norm": 4.697000463539704, + "learning_rate": 2.4736337228184338e-06, + "loss": 0.7333, + "num_input_tokens_seen": 159159525, + "step": 4892 + }, + { + "epoch": 0.44126798033999187, + "flos": 22022903257920.0, + "grad_norm": 11.80182855953402, + "learning_rate": 2.4730661102670692e-06, + "loss": 0.8504, + "num_input_tokens_seen": 159186735, + "step": 4893 + }, + { + "epoch": 0.44135816386346216, + "flos": 28978386132000.0, + "grad_norm": 1.8766894041121551, + "learning_rate": 2.472498457351096e-06, + "loss": 0.7567, + "num_input_tokens_seen": 159217090, + "step": 4894 + }, + { + "epoch": 0.4414483473869324, + "flos": 25447037300640.0, + "grad_norm": 2.0834393919333527, + "learning_rate": 2.4719307641189495e-06, + "loss": 0.7386, + "num_input_tokens_seen": 159244090, + "step": 4895 + }, + { + "epoch": 0.4415385309104027, + "flos": 26644687497600.0, + "grad_norm": 2.9711181832210904, + "learning_rate": 2.4713630306190673e-06, + "loss": 0.7152, + "num_input_tokens_seen": 159271710, + "step": 4896 + }, + { + "epoch": 0.4416287144338729, + "flos": 20783803044960.0, + "grad_norm": 2.0818095420182607, + "learning_rate": 2.4707952568998923e-06, + "loss": 0.8335, + "num_input_tokens_seen": 159298645, + "step": 4897 + }, + { + "epoch": 0.4417188979573432, + "flos": 20456806233600.0, + "grad_norm": 2.5283519968321766, + "learning_rate": 2.4702274430098703e-06, + "loss": 0.7241, + "num_input_tokens_seen": 159322820, + "step": 4898 + }, + { + "epoch": 0.44180908148081344, + "flos": 19509125638080.0, + "grad_norm": 2.289043856079418, + "learning_rate": 2.4696595889974497e-06, + "loss": 0.7051, + "num_input_tokens_seen": 159350770, + "step": 4899 + }, + { + "epoch": 0.4418992650042837, + "flos": 23765399461440.0, + "grad_norm": 1.62917437279221, + "learning_rate": 2.469091694911084e-06, + "loss": 0.8085, + "num_input_tokens_seen": 159377715, + "step": 4900 + }, + { + "epoch": 0.44198944852775396, + "flos": 22491465449760.0, + "grad_norm": 1.660493847797395, + "learning_rate": 2.4685237607992276e-06, + "loss": 0.751, + "num_input_tokens_seen": 159406980, + "step": 4901 + }, + { + "epoch": 0.44207963205122425, + "flos": 21695088711840.0, + "grad_norm": 1.6821388986232602, + "learning_rate": 2.4679557867103416e-06, + "loss": 0.8163, + "num_input_tokens_seen": 159432355, + "step": 4902 + }, + { + "epoch": 0.4421698155746945, + "flos": 22312315708320.0, + "grad_norm": 3.076238683550317, + "learning_rate": 2.4673877726928865e-06, + "loss": 0.7596, + "num_input_tokens_seen": 159462470, + "step": 4903 + }, + { + "epoch": 0.44225999909816477, + "flos": 24318125457120.0, + "grad_norm": 2.161930838418232, + "learning_rate": 2.46681971879533e-06, + "loss": 0.6414, + "num_input_tokens_seen": 159491960, + "step": 4904 + }, + { + "epoch": 0.442350182621635, + "flos": 14984446012320.0, + "grad_norm": 1.7844165296114338, + "learning_rate": 2.4662516250661407e-06, + "loss": 0.6752, + "num_input_tokens_seen": 159517365, + "step": 4905 + }, + { + "epoch": 0.4424403661451053, + "flos": 20346617864160.0, + "grad_norm": 1.873961478953949, + "learning_rate": 2.465683491553792e-06, + "loss": 0.7255, + "num_input_tokens_seen": 159542230, + "step": 4906 + }, + { + "epoch": 0.4425305496685755, + "flos": 20565619321920.0, + "grad_norm": 1.80836449111371, + "learning_rate": 2.4651153183067604e-06, + "loss": 0.7565, + "num_input_tokens_seen": 159570630, + "step": 4907 + }, + { + "epoch": 0.4426207331920458, + "flos": 23842017804000.0, + "grad_norm": 1.5306012862001341, + "learning_rate": 2.4645471053735245e-06, + "loss": 0.6929, + "num_input_tokens_seen": 159600955, + "step": 4908 + }, + { + "epoch": 0.4427109167155161, + "flos": 22388116316160.0, + "grad_norm": 1.6916494569075302, + "learning_rate": 2.4639788528025684e-06, + "loss": 0.6896, + "num_input_tokens_seen": 159630040, + "step": 4909 + }, + { + "epoch": 0.44280110023898633, + "flos": 26649519566400.0, + "grad_norm": 1.2753016862922575, + "learning_rate": 2.463410560642378e-06, + "loss": 0.7561, + "num_input_tokens_seen": 159662465, + "step": 4910 + }, + { + "epoch": 0.4428912837624566, + "flos": 62651675155200.0, + "grad_norm": 0.631435332144186, + "learning_rate": 2.4628422289414448e-06, + "loss": 0.588, + "num_input_tokens_seen": 159757325, + "step": 4911 + }, + { + "epoch": 0.44298146728592686, + "flos": 70892763345600.0, + "grad_norm": 0.66563648607973, + "learning_rate": 2.4622738577482592e-06, + "loss": 0.563, + "num_input_tokens_seen": 159850890, + "step": 4912 + }, + { + "epoch": 0.44307165080939714, + "flos": 38161505114400.0, + "grad_norm": 1.550399387341586, + "learning_rate": 2.461705447111319e-06, + "loss": 0.7385, + "num_input_tokens_seen": 159884660, + "step": 4913 + }, + { + "epoch": 0.4431618343328674, + "flos": 25884742858080.0, + "grad_norm": 2.112442047407693, + "learning_rate": 2.4611369970791246e-06, + "loss": 0.7253, + "num_input_tokens_seen": 159914805, + "step": 4914 + }, + { + "epoch": 0.44325201785633767, + "flos": 19873186433760.0, + "grad_norm": 1.66048070990693, + "learning_rate": 2.460568507700179e-06, + "loss": 0.748, + "num_input_tokens_seen": 159943880, + "step": 4915 + }, + { + "epoch": 0.4433422013798079, + "flos": 19143094845120.0, + "grad_norm": 1.7454641004819225, + "learning_rate": 2.4599999790229887e-06, + "loss": 0.796, + "num_input_tokens_seen": 159969565, + "step": 4916 + }, + { + "epoch": 0.4434323849032782, + "flos": 20638966725600.0, + "grad_norm": 2.5121312492366146, + "learning_rate": 2.459431411096064e-06, + "loss": 0.7137, + "num_input_tokens_seen": 159997235, + "step": 4917 + }, + { + "epoch": 0.4435225684267484, + "flos": 29598029162880.0, + "grad_norm": 1.7368000706463997, + "learning_rate": 2.458862803967918e-06, + "loss": 0.7196, + "num_input_tokens_seen": 160027885, + "step": 4918 + }, + { + "epoch": 0.4436127519502187, + "flos": 18525533320800.0, + "grad_norm": 1.693897694134259, + "learning_rate": 2.4582941576870667e-06, + "loss": 0.7688, + "num_input_tokens_seen": 160055290, + "step": 4919 + }, + { + "epoch": 0.44370293547368894, + "flos": 20346915222240.0, + "grad_norm": 1.7402830106401668, + "learning_rate": 2.4577254723020315e-06, + "loss": 0.7328, + "num_input_tokens_seen": 160083870, + "step": 4920 + }, + { + "epoch": 0.44379311899715923, + "flos": 21982196637120.0, + "grad_norm": 1.7642125919847096, + "learning_rate": 2.457156747861335e-06, + "loss": 0.7701, + "num_input_tokens_seen": 160112250, + "step": 4921 + }, + { + "epoch": 0.44388330252062946, + "flos": 20492643615840.0, + "grad_norm": 2.8735676390076756, + "learning_rate": 2.456587984413504e-06, + "loss": 0.8179, + "num_input_tokens_seen": 160135330, + "step": 4922 + }, + { + "epoch": 0.44397348604409975, + "flos": 25228704898560.0, + "grad_norm": 1.6936694502592096, + "learning_rate": 2.4560191820070683e-06, + "loss": 0.6841, + "num_input_tokens_seen": 160167355, + "step": 4923 + }, + { + "epoch": 0.44406366956757, + "flos": 24828658041600.0, + "grad_norm": 2.1507225508046246, + "learning_rate": 2.4554503406905617e-06, + "loss": 0.7332, + "num_input_tokens_seen": 160196715, + "step": 4924 + }, + { + "epoch": 0.4441538530910403, + "flos": 28033864966080.0, + "grad_norm": 2.3808205204735633, + "learning_rate": 2.454881460512521e-06, + "loss": 0.7901, + "num_input_tokens_seen": 160223640, + "step": 4925 + }, + { + "epoch": 0.4442440366145105, + "flos": 22896790412640.0, + "grad_norm": 1.7750461357460825, + "learning_rate": 2.4543125415214856e-06, + "loss": 0.7174, + "num_input_tokens_seen": 160255785, + "step": 4926 + }, + { + "epoch": 0.4443342201379808, + "flos": 26394829405440.0, + "grad_norm": 1.726845489371295, + "learning_rate": 2.4537435837659996e-06, + "loss": 0.7769, + "num_input_tokens_seen": 160285025, + "step": 4927 + }, + { + "epoch": 0.44442440366145103, + "flos": 24347495301120.0, + "grad_norm": 2.5229477540003247, + "learning_rate": 2.4531745872946085e-06, + "loss": 0.7072, + "num_input_tokens_seen": 160313490, + "step": 4928 + }, + { + "epoch": 0.4445145871849213, + "flos": 13639023084960.0, + "grad_norm": 2.01871850429094, + "learning_rate": 2.4526055521558632e-06, + "loss": 0.8157, + "num_input_tokens_seen": 160339420, + "step": 4929 + }, + { + "epoch": 0.44460477070839155, + "flos": 22205138089440.0, + "grad_norm": 1.6992228190055447, + "learning_rate": 2.4520364783983164e-06, + "loss": 0.7698, + "num_input_tokens_seen": 160368355, + "step": 4930 + }, + { + "epoch": 0.44469495423186184, + "flos": 23298361229760.0, + "grad_norm": 1.6234422700167532, + "learning_rate": 2.451467366070525e-06, + "loss": 0.7333, + "num_input_tokens_seen": 160397370, + "step": 4931 + }, + { + "epoch": 0.4447851377553321, + "flos": 14734959617760.0, + "grad_norm": 2.852290815595162, + "learning_rate": 2.450898215221048e-06, + "loss": 0.8214, + "num_input_tokens_seen": 160422110, + "step": 4932 + }, + { + "epoch": 0.44487532127880236, + "flos": 19909395513600.0, + "grad_norm": 2.4511925928220557, + "learning_rate": 2.4503290258984498e-06, + "loss": 0.8951, + "num_input_tokens_seen": 160447540, + "step": 4933 + }, + { + "epoch": 0.44496550480227265, + "flos": 25448486921280.0, + "grad_norm": 1.969754220279453, + "learning_rate": 2.4497597981512952e-06, + "loss": 0.6164, + "num_input_tokens_seen": 160476445, + "step": 4934 + }, + { + "epoch": 0.4450556883257429, + "flos": 66213954960480.0, + "grad_norm": 0.6410584495322006, + "learning_rate": 2.4491905320281555e-06, + "loss": 0.5915, + "num_input_tokens_seen": 160570715, + "step": 4935 + }, + { + "epoch": 0.44514587184921317, + "flos": 21803864630400.0, + "grad_norm": 1.9180906921281797, + "learning_rate": 2.448621227577602e-06, + "loss": 0.7258, + "num_input_tokens_seen": 160600450, + "step": 4936 + }, + { + "epoch": 0.4452360553726834, + "flos": 26212260046080.0, + "grad_norm": 1.7897274330870225, + "learning_rate": 2.4480518848482123e-06, + "loss": 0.6979, + "num_input_tokens_seen": 160629490, + "step": 4937 + }, + { + "epoch": 0.4453262388961537, + "flos": 22749463719360.0, + "grad_norm": 1.6902626876275875, + "learning_rate": 2.447482503888565e-06, + "loss": 0.7639, + "num_input_tokens_seen": 160658800, + "step": 4938 + }, + { + "epoch": 0.4454164224196239, + "flos": 19837200372480.0, + "grad_norm": 2.1401245200392416, + "learning_rate": 2.4469130847472434e-06, + "loss": 0.728, + "num_input_tokens_seen": 160685645, + "step": 4939 + }, + { + "epoch": 0.4455066059430942, + "flos": 22460162778240.0, + "grad_norm": 1.8643633181659638, + "learning_rate": 2.4463436274728326e-06, + "loss": 0.7924, + "num_input_tokens_seen": 160714010, + "step": 4940 + }, + { + "epoch": 0.44559678946656445, + "flos": 30874490718240.0, + "grad_norm": 1.6144827370428054, + "learning_rate": 2.4457741321139227e-06, + "loss": 0.5989, + "num_input_tokens_seen": 160747190, + "step": 4941 + }, + { + "epoch": 0.44568697299003474, + "flos": 19108409725440.0, + "grad_norm": 1.9624813734980784, + "learning_rate": 2.4452045987191063e-06, + "loss": 0.7748, + "num_input_tokens_seen": 160773950, + "step": 4942 + }, + { + "epoch": 0.44577715651350497, + "flos": 21876877506240.0, + "grad_norm": 1.4784312054055209, + "learning_rate": 2.4446350273369776e-06, + "loss": 0.7242, + "num_input_tokens_seen": 160804830, + "step": 4943 + }, + { + "epoch": 0.44586734003697526, + "flos": 26212408725120.0, + "grad_norm": 1.5282294395657494, + "learning_rate": 2.4440654180161374e-06, + "loss": 0.7894, + "num_input_tokens_seen": 160834590, + "step": 4944 + }, + { + "epoch": 0.4459575235604455, + "flos": 33096588532320.0, + "grad_norm": 1.4714758718562388, + "learning_rate": 2.4434957708051875e-06, + "loss": 0.7066, + "num_input_tokens_seen": 160868905, + "step": 4945 + }, + { + "epoch": 0.4460477070839158, + "flos": 29526205719360.0, + "grad_norm": 1.827427657713773, + "learning_rate": 2.4429260857527324e-06, + "loss": 0.6942, + "num_input_tokens_seen": 160900205, + "step": 4946 + }, + { + "epoch": 0.446137890607386, + "flos": 33025731502560.0, + "grad_norm": 2.2988252210315134, + "learning_rate": 2.4423563629073815e-06, + "loss": 0.6611, + "num_input_tokens_seen": 160930805, + "step": 4947 + }, + { + "epoch": 0.4462280741308563, + "flos": 26211925518240.0, + "grad_norm": 1.5998755525829005, + "learning_rate": 2.4417866023177466e-06, + "loss": 0.7542, + "num_input_tokens_seen": 160959910, + "step": 4948 + }, + { + "epoch": 0.44631825765432653, + "flos": 22825078478400.0, + "grad_norm": 2.7342078457558276, + "learning_rate": 2.441216804032443e-06, + "loss": 0.786, + "num_input_tokens_seen": 160988635, + "step": 4949 + }, + { + "epoch": 0.4464084411777968, + "flos": 22716079541280.0, + "grad_norm": 1.7254203367190406, + "learning_rate": 2.440646968100089e-06, + "loss": 0.7662, + "num_input_tokens_seen": 161018890, + "step": 4950 + }, + { + "epoch": 0.44649862470126706, + "flos": 59170293948480.0, + "grad_norm": 0.6847879278856355, + "learning_rate": 2.4400770945693055e-06, + "loss": 0.5732, + "num_input_tokens_seen": 161105095, + "step": 4951 + }, + { + "epoch": 0.44658880822473734, + "flos": 20674246561440.0, + "grad_norm": 2.0926844604203976, + "learning_rate": 2.4395071834887177e-06, + "loss": 0.7829, + "num_input_tokens_seen": 161133205, + "step": 4952 + }, + { + "epoch": 0.4466789917482076, + "flos": 61221828235680.0, + "grad_norm": 0.6970486557135614, + "learning_rate": 2.438937234906954e-06, + "loss": 0.6346, + "num_input_tokens_seen": 161224805, + "step": 4953 + }, + { + "epoch": 0.44676917527167787, + "flos": 22708273891680.0, + "grad_norm": 1.63424240362168, + "learning_rate": 2.4383672488726447e-06, + "loss": 0.739, + "num_input_tokens_seen": 161254250, + "step": 4954 + }, + { + "epoch": 0.44685935879514815, + "flos": 18996734565600.0, + "grad_norm": 1.9836748945684897, + "learning_rate": 2.4377972254344256e-06, + "loss": 0.6735, + "num_input_tokens_seen": 161281465, + "step": 4955 + }, + { + "epoch": 0.4469495423186184, + "flos": 26248766484000.0, + "grad_norm": 1.8088695076964572, + "learning_rate": 2.437227164640932e-06, + "loss": 0.6378, + "num_input_tokens_seen": 161308870, + "step": 4956 + }, + { + "epoch": 0.4470397258420887, + "flos": 23553646106880.0, + "grad_norm": 1.6914776441205754, + "learning_rate": 2.436657066540807e-06, + "loss": 0.6937, + "num_input_tokens_seen": 161338095, + "step": 4957 + }, + { + "epoch": 0.4471299093655589, + "flos": 18853496545920.0, + "grad_norm": 2.368557675085215, + "learning_rate": 2.4360869311826927e-06, + "loss": 0.6287, + "num_input_tokens_seen": 161364535, + "step": 4958 + }, + { + "epoch": 0.4472200928890292, + "flos": 19144767484320.0, + "grad_norm": 3.5363187617240452, + "learning_rate": 2.4355167586152367e-06, + "loss": 0.8434, + "num_input_tokens_seen": 161390290, + "step": 4959 + }, + { + "epoch": 0.44731027641249943, + "flos": 25155952211040.0, + "grad_norm": 1.9781564309273207, + "learning_rate": 2.4349465488870896e-06, + "loss": 0.718, + "num_input_tokens_seen": 161419085, + "step": 4960 + }, + { + "epoch": 0.4474004599359697, + "flos": 26867443101120.0, + "grad_norm": 6.377897613823235, + "learning_rate": 2.434376302046905e-06, + "loss": 0.8393, + "num_input_tokens_seen": 161450875, + "step": 4961 + }, + { + "epoch": 0.44749064345943995, + "flos": 16813187526240.0, + "grad_norm": 1.8401403491769832, + "learning_rate": 2.433806018143339e-06, + "loss": 0.7794, + "num_input_tokens_seen": 161477940, + "step": 4962 + }, + { + "epoch": 0.44758082698291024, + "flos": 65037274241760.0, + "grad_norm": 0.6724514373631191, + "learning_rate": 2.433235697225051e-06, + "loss": 0.6153, + "num_input_tokens_seen": 161561410, + "step": 4963 + }, + { + "epoch": 0.4476710105063805, + "flos": 20165089258080.0, + "grad_norm": 2.014783106790214, + "learning_rate": 2.4326653393407048e-06, + "loss": 0.7666, + "num_input_tokens_seen": 161588845, + "step": 4964 + }, + { + "epoch": 0.44776119402985076, + "flos": 20966335234560.0, + "grad_norm": 2.16666462482454, + "learning_rate": 2.432094944538966e-06, + "loss": 0.7204, + "num_input_tokens_seen": 161618190, + "step": 4965 + }, + { + "epoch": 0.447851377553321, + "flos": 24865164479520.0, + "grad_norm": 1.4779308780359368, + "learning_rate": 2.4315245128685047e-06, + "loss": 0.7195, + "num_input_tokens_seen": 161647095, + "step": 4966 + }, + { + "epoch": 0.4479415610767913, + "flos": 23298361229760.0, + "grad_norm": 1.6355080872594947, + "learning_rate": 2.4309540443779925e-06, + "loss": 0.7517, + "num_input_tokens_seen": 161678720, + "step": 4967 + }, + { + "epoch": 0.4480317446002615, + "flos": 31197473195520.0, + "grad_norm": 1.5955836649410717, + "learning_rate": 2.4303835391161047e-06, + "loss": 0.6873, + "num_input_tokens_seen": 161709160, + "step": 4968 + }, + { + "epoch": 0.4481219281237318, + "flos": 19870584550560.0, + "grad_norm": 2.079596321291127, + "learning_rate": 2.42981299713152e-06, + "loss": 0.7011, + "num_input_tokens_seen": 161739915, + "step": 4969 + }, + { + "epoch": 0.44821211164720204, + "flos": 32110914708480.0, + "grad_norm": 6.635831220084286, + "learning_rate": 2.4292424184729204e-06, + "loss": 0.7888, + "num_input_tokens_seen": 161770490, + "step": 4970 + }, + { + "epoch": 0.4483022951706723, + "flos": 23298324060000.0, + "grad_norm": 1.7050360360909962, + "learning_rate": 2.4286718031889913e-06, + "loss": 0.7453, + "num_input_tokens_seen": 161800405, + "step": 4971 + }, + { + "epoch": 0.44839247869414256, + "flos": 27922784522400.0, + "grad_norm": 1.4332713250490592, + "learning_rate": 2.4281011513284202e-06, + "loss": 0.725, + "num_input_tokens_seen": 161830380, + "step": 4972 + }, + { + "epoch": 0.44848266221761285, + "flos": 26103261108960.0, + "grad_norm": 2.2531106427247143, + "learning_rate": 2.4275304629398985e-06, + "loss": 0.686, + "num_input_tokens_seen": 161858865, + "step": 4973 + }, + { + "epoch": 0.4485728457410831, + "flos": 25331496485760.0, + "grad_norm": 1.8206399778567801, + "learning_rate": 2.4269597380721194e-06, + "loss": 0.8269, + "num_input_tokens_seen": 161887675, + "step": 4974 + }, + { + "epoch": 0.44866302926455337, + "flos": 21510549355200.0, + "grad_norm": 1.5808511736152544, + "learning_rate": 2.426388976773782e-06, + "loss": 0.7453, + "num_input_tokens_seen": 161914920, + "step": 4975 + }, + { + "epoch": 0.4487532127880236, + "flos": 17578410271680.0, + "grad_norm": 2.541922311485576, + "learning_rate": 2.425818179093586e-06, + "loss": 0.8554, + "num_input_tokens_seen": 161941385, + "step": 4976 + }, + { + "epoch": 0.4488433963114939, + "flos": 25700612368800.0, + "grad_norm": 1.534844985739284, + "learning_rate": 2.4252473450802346e-06, + "loss": 0.6608, + "num_input_tokens_seen": 161971525, + "step": 4977 + }, + { + "epoch": 0.4489335798349641, + "flos": 23186017014240.0, + "grad_norm": 1.9311993194152215, + "learning_rate": 2.4246764747824355e-06, + "loss": 0.7818, + "num_input_tokens_seen": 162000565, + "step": 4978 + }, + { + "epoch": 0.4490237633584344, + "flos": 23516693631840.0, + "grad_norm": 1.712806203712621, + "learning_rate": 2.424105568248897e-06, + "loss": 0.7612, + "num_input_tokens_seen": 162028520, + "step": 4979 + }, + { + "epoch": 0.4491139468819047, + "flos": 28216322816160.0, + "grad_norm": 1.6005825859625908, + "learning_rate": 2.4235346255283337e-06, + "loss": 0.6906, + "num_input_tokens_seen": 162060300, + "step": 4980 + }, + { + "epoch": 0.44920413040537494, + "flos": 20347286919840.0, + "grad_norm": 2.4120845654404786, + "learning_rate": 2.42296364666946e-06, + "loss": 0.7603, + "num_input_tokens_seen": 162085955, + "step": 4981 + }, + { + "epoch": 0.4492943139288452, + "flos": 22965305747520.0, + "grad_norm": 1.7416675972213571, + "learning_rate": 2.4223926317209965e-06, + "loss": 0.6971, + "num_input_tokens_seen": 162115400, + "step": 4982 + }, + { + "epoch": 0.44938449745231546, + "flos": 31856893603200.0, + "grad_norm": 1.5727952906112173, + "learning_rate": 2.4218215807316647e-06, + "loss": 0.7521, + "num_input_tokens_seen": 162146975, + "step": 4983 + }, + { + "epoch": 0.44947468097578575, + "flos": 20274013855680.0, + "grad_norm": 1.8498889425811575, + "learning_rate": 2.4212504937501894e-06, + "loss": 0.8356, + "num_input_tokens_seen": 162173725, + "step": 4984 + }, + { + "epoch": 0.449564864499256, + "flos": 21658024727520.0, + "grad_norm": 1.4266838665546338, + "learning_rate": 2.4206793708253e-06, + "loss": 0.8081, + "num_input_tokens_seen": 162204080, + "step": 4985 + }, + { + "epoch": 0.44965504802272627, + "flos": 18525198792960.0, + "grad_norm": 2.0563382671855805, + "learning_rate": 2.420108212005726e-06, + "loss": 0.7056, + "num_input_tokens_seen": 162229605, + "step": 4986 + }, + { + "epoch": 0.4497452315461965, + "flos": 17615325576960.0, + "grad_norm": 1.9178005540987608, + "learning_rate": 2.4195370173402034e-06, + "loss": 0.7468, + "num_input_tokens_seen": 162254360, + "step": 4987 + }, + { + "epoch": 0.4498354150696668, + "flos": 20420485644480.0, + "grad_norm": 2.1316016256725994, + "learning_rate": 2.4189657868774696e-06, + "loss": 0.8713, + "num_input_tokens_seen": 162282510, + "step": 4988 + }, + { + "epoch": 0.449925598593137, + "flos": 22605147776640.0, + "grad_norm": 1.9636273420211712, + "learning_rate": 2.418394520666264e-06, + "loss": 0.6888, + "num_input_tokens_seen": 162312480, + "step": 4989 + }, + { + "epoch": 0.4500157821166073, + "flos": 24976913978880.0, + "grad_norm": 1.5816110311892404, + "learning_rate": 2.4178232187553307e-06, + "loss": 0.7362, + "num_input_tokens_seen": 162345815, + "step": 4990 + }, + { + "epoch": 0.45010596564007754, + "flos": 24281953547040.0, + "grad_norm": 1.5689044921913782, + "learning_rate": 2.417251881193417e-06, + "loss": 0.7805, + "num_input_tokens_seen": 162373780, + "step": 4991 + }, + { + "epoch": 0.45019614916354783, + "flos": 23949901648320.0, + "grad_norm": 1.6024104409030517, + "learning_rate": 2.4166805080292723e-06, + "loss": 0.8082, + "num_input_tokens_seen": 162399175, + "step": 4992 + }, + { + "epoch": 0.45028633268701806, + "flos": 25008105141120.0, + "grad_norm": 1.5779985433844936, + "learning_rate": 2.4161090993116485e-06, + "loss": 0.7022, + "num_input_tokens_seen": 162430125, + "step": 4993 + }, + { + "epoch": 0.45037651621048835, + "flos": 23620600311840.0, + "grad_norm": 2.1342034467839293, + "learning_rate": 2.4155376550893026e-06, + "loss": 0.6337, + "num_input_tokens_seen": 162456985, + "step": 4994 + }, + { + "epoch": 0.4504666997339586, + "flos": 19400089531200.0, + "grad_norm": 1.659832373049253, + "learning_rate": 2.4149661754109926e-06, + "loss": 0.763, + "num_input_tokens_seen": 162483160, + "step": 4995 + }, + { + "epoch": 0.4505568832574289, + "flos": 24245409939360.0, + "grad_norm": 1.8902012373748562, + "learning_rate": 2.41439466032548e-06, + "loss": 0.8096, + "num_input_tokens_seen": 162507650, + "step": 4996 + }, + { + "epoch": 0.4506470667808991, + "flos": 30616269430080.0, + "grad_norm": 14.891909436831957, + "learning_rate": 2.41382310988153e-06, + "loss": 0.7027, + "num_input_tokens_seen": 162536025, + "step": 4997 + }, + { + "epoch": 0.4507372503043694, + "flos": 20747222267520.0, + "grad_norm": 2.305860524772604, + "learning_rate": 2.413251524127911e-06, + "loss": 0.8079, + "num_input_tokens_seen": 162564125, + "step": 4998 + }, + { + "epoch": 0.45082743382783963, + "flos": 18816952938240.0, + "grad_norm": 1.6605252921622857, + "learning_rate": 2.412679903113393e-06, + "loss": 0.7225, + "num_input_tokens_seen": 162590325, + "step": 4999 + }, + { + "epoch": 0.4509176173513099, + "flos": 16916759678400.0, + "grad_norm": 2.3648087538318414, + "learning_rate": 2.4121082468867505e-06, + "loss": 0.7674, + "num_input_tokens_seen": 162610250, + "step": 5000 + }, + { + "epoch": 0.45100780087478015, + "flos": 37797481488480.0, + "grad_norm": 4.138424846627234, + "learning_rate": 2.4115365554967597e-06, + "loss": 0.6883, + "num_input_tokens_seen": 162637580, + "step": 5001 + }, + { + "epoch": 0.45109798439825044, + "flos": 21695237390880.0, + "grad_norm": 1.717945951177805, + "learning_rate": 2.4109648289922006e-06, + "loss": 0.752, + "num_input_tokens_seen": 162665165, + "step": 5002 + }, + { + "epoch": 0.45118816792172073, + "flos": 28249260957120.0, + "grad_norm": 1.9651849348096808, + "learning_rate": 2.4103930674218565e-06, + "loss": 0.7614, + "num_input_tokens_seen": 162696235, + "step": 5003 + }, + { + "epoch": 0.45127835144519096, + "flos": 21182511790560.0, + "grad_norm": 1.5645654378477947, + "learning_rate": 2.409821270834513e-06, + "loss": 0.7295, + "num_input_tokens_seen": 162724240, + "step": 5004 + }, + { + "epoch": 0.45136853496866125, + "flos": 25666410456000.0, + "grad_norm": 1.7950080792768566, + "learning_rate": 2.409249439278959e-06, + "loss": 0.8072, + "num_input_tokens_seen": 162754575, + "step": 5005 + }, + { + "epoch": 0.4514587184921315, + "flos": 62953836833280.0, + "grad_norm": 0.6262304829508627, + "learning_rate": 2.408677572803986e-06, + "loss": 0.5519, + "num_input_tokens_seen": 162850010, + "step": 5006 + }, + { + "epoch": 0.45154890201560177, + "flos": 23152186799040.0, + "grad_norm": 1.7645926369065619, + "learning_rate": 2.408105671458389e-06, + "loss": 0.7533, + "num_input_tokens_seen": 162877875, + "step": 5007 + }, + { + "epoch": 0.451639085539072, + "flos": 29964580332480.0, + "grad_norm": 1.3969286658711868, + "learning_rate": 2.4075337352909663e-06, + "loss": 0.6956, + "num_input_tokens_seen": 162911225, + "step": 5008 + }, + { + "epoch": 0.4517292690625423, + "flos": 23401375835520.0, + "grad_norm": 1.5354225762442015, + "learning_rate": 2.4069617643505177e-06, + "loss": 0.8801, + "num_input_tokens_seen": 162938770, + "step": 5009 + }, + { + "epoch": 0.4518194525860125, + "flos": 22678160652480.0, + "grad_norm": 1.5574033393072586, + "learning_rate": 2.406389758685848e-06, + "loss": 0.8369, + "num_input_tokens_seen": 162968280, + "step": 5010 + }, + { + "epoch": 0.4519096361094828, + "flos": 25079928584640.0, + "grad_norm": 1.6418796646406217, + "learning_rate": 2.405817718345763e-06, + "loss": 0.7894, + "num_input_tokens_seen": 162996635, + "step": 5011 + }, + { + "epoch": 0.45199981963295305, + "flos": 23553088560480.0, + "grad_norm": 2.3795978812301417, + "learning_rate": 2.4052456433790726e-06, + "loss": 0.7083, + "num_input_tokens_seen": 163028545, + "step": 5012 + }, + { + "epoch": 0.45209000315642334, + "flos": 31561794179520.0, + "grad_norm": 1.7661568844838618, + "learning_rate": 2.4046735338345897e-06, + "loss": 0.8105, + "num_input_tokens_seen": 163060040, + "step": 5013 + }, + { + "epoch": 0.45218018667989357, + "flos": 22859391900480.0, + "grad_norm": 3.122385228096347, + "learning_rate": 2.404101389761129e-06, + "loss": 0.7679, + "num_input_tokens_seen": 163087770, + "step": 5014 + }, + { + "epoch": 0.45227037020336386, + "flos": 62577026809920.0, + "grad_norm": 0.7094921889113729, + "learning_rate": 2.4035292112075097e-06, + "loss": 0.5631, + "num_input_tokens_seen": 163179625, + "step": 5015 + }, + { + "epoch": 0.4523605537268341, + "flos": 15861901464000.0, + "grad_norm": 1.5629846965506768, + "learning_rate": 2.4029569982225534e-06, + "loss": 0.811, + "num_input_tokens_seen": 163205775, + "step": 5016 + }, + { + "epoch": 0.4524507372503044, + "flos": 57732784324800.0, + "grad_norm": 0.6981271502291796, + "learning_rate": 2.402384750855084e-06, + "loss": 0.6351, + "num_input_tokens_seen": 163293765, + "step": 5017 + }, + { + "epoch": 0.4525409207737746, + "flos": 27233473894080.0, + "grad_norm": 2.5138010556476993, + "learning_rate": 2.4018124691539286e-06, + "loss": 0.7552, + "num_input_tokens_seen": 163323175, + "step": 5018 + }, + { + "epoch": 0.4526311042972449, + "flos": 25989207084480.0, + "grad_norm": 4.7413297125634815, + "learning_rate": 2.4012401531679178e-06, + "loss": 0.7409, + "num_input_tokens_seen": 163350860, + "step": 5019 + }, + { + "epoch": 0.45272128782071513, + "flos": 27086816256480.0, + "grad_norm": 2.035682888314716, + "learning_rate": 2.4006678029458847e-06, + "loss": 0.7495, + "num_input_tokens_seen": 163379380, + "step": 5020 + }, + { + "epoch": 0.4528114713441854, + "flos": 21731780998560.0, + "grad_norm": 1.626172816257722, + "learning_rate": 2.400095418536666e-06, + "loss": 0.6592, + "num_input_tokens_seen": 163406695, + "step": 5021 + }, + { + "epoch": 0.45290165486765566, + "flos": 22422615587040.0, + "grad_norm": 2.272687860493951, + "learning_rate": 2.3995229999890996e-06, + "loss": 0.73, + "num_input_tokens_seen": 163436285, + "step": 5022 + }, + { + "epoch": 0.45299183839112594, + "flos": 23116014888960.0, + "grad_norm": 2.4612708456339436, + "learning_rate": 2.398950547352028e-06, + "loss": 0.6952, + "num_input_tokens_seen": 163465160, + "step": 5023 + }, + { + "epoch": 0.4530820219145962, + "flos": 20784249082080.0, + "grad_norm": 2.2785627658780117, + "learning_rate": 2.398378060674295e-06, + "loss": 0.6794, + "num_input_tokens_seen": 163494110, + "step": 5024 + }, + { + "epoch": 0.45317220543806647, + "flos": 24318088287360.0, + "grad_norm": 1.5884660379719433, + "learning_rate": 2.39780554000475e-06, + "loss": 0.7538, + "num_input_tokens_seen": 163524515, + "step": 5025 + }, + { + "epoch": 0.4532623889615367, + "flos": 17869495361280.0, + "grad_norm": 1.9562511602526262, + "learning_rate": 2.3972329853922434e-06, + "loss": 0.782, + "num_input_tokens_seen": 163551310, + "step": 5026 + }, + { + "epoch": 0.453352572485007, + "flos": 67686255479520.0, + "grad_norm": 0.5923153711522896, + "learning_rate": 2.3966603968856278e-06, + "loss": 0.5704, + "num_input_tokens_seen": 163654700, + "step": 5027 + }, + { + "epoch": 0.4534427560084773, + "flos": 22713291809280.0, + "grad_norm": 1.7515926606901973, + "learning_rate": 2.39608777453376e-06, + "loss": 0.8243, + "num_input_tokens_seen": 163682240, + "step": 5028 + }, + { + "epoch": 0.4535329395319475, + "flos": 21330619048800.0, + "grad_norm": 1.7907806541868583, + "learning_rate": 2.3955151183854993e-06, + "loss": 0.7783, + "num_input_tokens_seen": 163708690, + "step": 5029 + }, + { + "epoch": 0.4536231230554178, + "flos": 25079928584640.0, + "grad_norm": 2.2322148905439008, + "learning_rate": 2.3949424284897073e-06, + "loss": 0.7157, + "num_input_tokens_seen": 163737270, + "step": 5030 + }, + { + "epoch": 0.45371330657888803, + "flos": 33787088592960.0, + "grad_norm": 1.9173765181031768, + "learning_rate": 2.39436970489525e-06, + "loss": 0.6223, + "num_input_tokens_seen": 163769065, + "step": 5031 + }, + { + "epoch": 0.4538034901023583, + "flos": 22532803956480.0, + "grad_norm": 1.7324220814323088, + "learning_rate": 2.3937969476509955e-06, + "loss": 0.7527, + "num_input_tokens_seen": 163797435, + "step": 5032 + }, + { + "epoch": 0.45389367362582855, + "flos": 20492569276320.0, + "grad_norm": 1.6103085523127962, + "learning_rate": 2.393224156805813e-06, + "loss": 0.826, + "num_input_tokens_seen": 163824415, + "step": 5033 + }, + { + "epoch": 0.45398385714929884, + "flos": 20747222267520.0, + "grad_norm": 1.9363163439600253, + "learning_rate": 2.392651332408578e-06, + "loss": 0.7288, + "num_input_tokens_seen": 163850850, + "step": 5034 + }, + { + "epoch": 0.4540740406727691, + "flos": 20273679327840.0, + "grad_norm": 2.5080706977431735, + "learning_rate": 2.3920784745081655e-06, + "loss": 0.7352, + "num_input_tokens_seen": 163877440, + "step": 5035 + }, + { + "epoch": 0.45416422419623936, + "flos": 13570544919840.0, + "grad_norm": 1.6435599097510616, + "learning_rate": 2.391505583153456e-06, + "loss": 0.7904, + "num_input_tokens_seen": 163904330, + "step": 5036 + }, + { + "epoch": 0.4542544077197096, + "flos": 22314360045120.0, + "grad_norm": 1.6655332128487912, + "learning_rate": 2.3909326583933315e-06, + "loss": 0.7558, + "num_input_tokens_seen": 163931095, + "step": 5037 + }, + { + "epoch": 0.4543445912431799, + "flos": 37791125459520.0, + "grad_norm": 1.8058676555888455, + "learning_rate": 2.3903597002766777e-06, + "loss": 0.7177, + "num_input_tokens_seen": 163965055, + "step": 5038 + }, + { + "epoch": 0.4544347747666501, + "flos": 25483989775680.0, + "grad_norm": 1.6155685208333976, + "learning_rate": 2.389786708852381e-06, + "loss": 0.7128, + "num_input_tokens_seen": 163993870, + "step": 5039 + }, + { + "epoch": 0.4545249582901204, + "flos": 67126169871360.0, + "grad_norm": 0.9229860238105104, + "learning_rate": 2.389213684169333e-06, + "loss": 0.6944, + "num_input_tokens_seen": 164090375, + "step": 5040 + }, + { + "epoch": 0.45461514181359064, + "flos": 22788274682400.0, + "grad_norm": 1.6310739707240849, + "learning_rate": 2.388640626276428e-06, + "loss": 0.7175, + "num_input_tokens_seen": 164121140, + "step": 5041 + }, + { + "epoch": 0.45470532533706093, + "flos": 63961372209600.0, + "grad_norm": 0.6799492825148993, + "learning_rate": 2.388067535222561e-06, + "loss": 0.5352, + "num_input_tokens_seen": 164218035, + "step": 5042 + }, + { + "epoch": 0.45479550886053116, + "flos": 28250710577760.0, + "grad_norm": 1.8462091630204325, + "learning_rate": 2.3874944110566332e-06, + "loss": 0.7671, + "num_input_tokens_seen": 164248495, + "step": 5043 + }, + { + "epoch": 0.45488569238400145, + "flos": 70356249098880.0, + "grad_norm": 0.7592672169491473, + "learning_rate": 2.3869212538275447e-06, + "loss": 0.6775, + "num_input_tokens_seen": 164334055, + "step": 5044 + }, + { + "epoch": 0.4549758759074717, + "flos": 20420597153760.0, + "grad_norm": 1.991183038939858, + "learning_rate": 2.386348063584202e-06, + "loss": 0.7689, + "num_input_tokens_seen": 164364695, + "step": 5045 + }, + { + "epoch": 0.45506605943094197, + "flos": 19618459103040.0, + "grad_norm": 1.776082289185491, + "learning_rate": 2.385774840375511e-06, + "loss": 0.8017, + "num_input_tokens_seen": 164391490, + "step": 5046 + }, + { + "epoch": 0.4551562429544122, + "flos": 19071717438720.0, + "grad_norm": 1.9378044750562358, + "learning_rate": 2.385201584250385e-06, + "loss": 0.7042, + "num_input_tokens_seen": 164419590, + "step": 5047 + }, + { + "epoch": 0.4552464264778825, + "flos": 62073667989120.0, + "grad_norm": 0.6362285013819832, + "learning_rate": 2.3846282952577346e-06, + "loss": 0.6127, + "num_input_tokens_seen": 164514555, + "step": 5048 + }, + { + "epoch": 0.4553366100013527, + "flos": 23731271888160.0, + "grad_norm": 1.6433583333252655, + "learning_rate": 2.3840549734464785e-06, + "loss": 0.8435, + "num_input_tokens_seen": 164543935, + "step": 5049 + }, + { + "epoch": 0.455426793524823, + "flos": 17979386372640.0, + "grad_norm": 2.4564060695462016, + "learning_rate": 2.3834816188655336e-06, + "loss": 0.6865, + "num_input_tokens_seen": 164572615, + "step": 5050 + }, + { + "epoch": 0.4555169770482933, + "flos": 36595891296960.0, + "grad_norm": 1.6049330065759517, + "learning_rate": 2.3829082315638224e-06, + "loss": 0.7002, + "num_input_tokens_seen": 164604080, + "step": 5051 + }, + { + "epoch": 0.45560716057176354, + "flos": 26467284734880.0, + "grad_norm": 2.025445479383144, + "learning_rate": 2.3823348115902695e-06, + "loss": 0.6123, + "num_input_tokens_seen": 164631455, + "step": 5052 + }, + { + "epoch": 0.4556973440952338, + "flos": 21549694846080.0, + "grad_norm": 1.6897062078999128, + "learning_rate": 2.3817613589938026e-06, + "loss": 0.7742, + "num_input_tokens_seen": 164657725, + "step": 5053 + }, + { + "epoch": 0.45578752761870406, + "flos": 22240640943840.0, + "grad_norm": 1.6202632385240452, + "learning_rate": 2.3811878738233517e-06, + "loss": 0.7815, + "num_input_tokens_seen": 164687615, + "step": 5054 + }, + { + "epoch": 0.45587771114217435, + "flos": 35029013707680.0, + "grad_norm": 2.0368263582014925, + "learning_rate": 2.380614356127849e-06, + "loss": 0.6979, + "num_input_tokens_seen": 164717365, + "step": 5055 + }, + { + "epoch": 0.4559678946656446, + "flos": 22420719929280.0, + "grad_norm": 1.8411804656387056, + "learning_rate": 2.3800408059562318e-06, + "loss": 0.7486, + "num_input_tokens_seen": 164746075, + "step": 5056 + }, + { + "epoch": 0.45605807818911487, + "flos": 22532469428640.0, + "grad_norm": 1.815026528345865, + "learning_rate": 2.3794672233574365e-06, + "loss": 0.7703, + "num_input_tokens_seen": 164777440, + "step": 5057 + }, + { + "epoch": 0.4561482617125851, + "flos": 23917446714240.0, + "grad_norm": 2.0733139449722766, + "learning_rate": 2.3788936083804058e-06, + "loss": 0.7674, + "num_input_tokens_seen": 164805780, + "step": 5058 + }, + { + "epoch": 0.4562384452360554, + "flos": 23331894086880.0, + "grad_norm": 1.5545921104726628, + "learning_rate": 2.378319961074083e-06, + "loss": 0.7546, + "num_input_tokens_seen": 164835290, + "step": 5059 + }, + { + "epoch": 0.4563286287595256, + "flos": 24135556097760.0, + "grad_norm": 1.7437620018986957, + "learning_rate": 2.377746281487415e-06, + "loss": 0.79, + "num_input_tokens_seen": 164864500, + "step": 5060 + }, + { + "epoch": 0.4564188122829959, + "flos": 21367497184320.0, + "grad_norm": 2.1503009275045444, + "learning_rate": 2.377172569669352e-06, + "loss": 0.8441, + "num_input_tokens_seen": 164892480, + "step": 5061 + }, + { + "epoch": 0.45650899580646614, + "flos": 23480298703200.0, + "grad_norm": 1.7333573113519147, + "learning_rate": 2.376598825668845e-06, + "loss": 0.6025, + "num_input_tokens_seen": 164922385, + "step": 5062 + }, + { + "epoch": 0.45659917932993643, + "flos": 20889493873440.0, + "grad_norm": 1.7904024395263864, + "learning_rate": 2.3760250495348495e-06, + "loss": 0.7415, + "num_input_tokens_seen": 164949985, + "step": 5063 + }, + { + "epoch": 0.45668936285340667, + "flos": 24097042492800.0, + "grad_norm": 1.7851488376188083, + "learning_rate": 2.3754512413163236e-06, + "loss": 0.7986, + "num_input_tokens_seen": 164980225, + "step": 5064 + }, + { + "epoch": 0.45677954637687695, + "flos": 28871580210720.0, + "grad_norm": 1.4704649632342028, + "learning_rate": 2.3748774010622285e-06, + "loss": 0.7319, + "num_input_tokens_seen": 165010770, + "step": 5065 + }, + { + "epoch": 0.4568697299003472, + "flos": 15465014036640.0, + "grad_norm": 1.8873072746864705, + "learning_rate": 2.3743035288215254e-06, + "loss": 0.7575, + "num_input_tokens_seen": 165037375, + "step": 5066 + }, + { + "epoch": 0.4569599134238175, + "flos": 22168334293440.0, + "grad_norm": 1.6231607151618945, + "learning_rate": 2.3737296246431815e-06, + "loss": 0.7622, + "num_input_tokens_seen": 165063380, + "step": 5067 + }, + { + "epoch": 0.4570500969472877, + "flos": 26249695728000.0, + "grad_norm": 2.2878000990165863, + "learning_rate": 2.3731556885761656e-06, + "loss": 0.7239, + "num_input_tokens_seen": 165092215, + "step": 5068 + }, + { + "epoch": 0.457140280470758, + "flos": 70419895195200.0, + "grad_norm": 0.6390624645578239, + "learning_rate": 2.372581720669449e-06, + "loss": 0.5792, + "num_input_tokens_seen": 165167985, + "step": 5069 + }, + { + "epoch": 0.45723046399422823, + "flos": 17392198275840.0, + "grad_norm": 2.125280586118611, + "learning_rate": 2.3720077209720046e-06, + "loss": 0.7815, + "num_input_tokens_seen": 165194045, + "step": 5070 + }, + { + "epoch": 0.4573206475176985, + "flos": 70690828541280.0, + "grad_norm": 0.6940605593967231, + "learning_rate": 2.3714336895328112e-06, + "loss": 0.5724, + "num_input_tokens_seen": 165285845, + "step": 5071 + }, + { + "epoch": 0.45741083104116875, + "flos": 22205175259200.0, + "grad_norm": 1.7769410735682025, + "learning_rate": 2.370859626400847e-06, + "loss": 0.7089, + "num_input_tokens_seen": 165313160, + "step": 5072 + }, + { + "epoch": 0.45750101456463904, + "flos": 19434142764960.0, + "grad_norm": 1.5177018398686222, + "learning_rate": 2.3702855316250943e-06, + "loss": 0.7296, + "num_input_tokens_seen": 165341520, + "step": 5073 + }, + { + "epoch": 0.45759119808810933, + "flos": 25337889684480.0, + "grad_norm": 1.6160530866080334, + "learning_rate": 2.369711405254539e-06, + "loss": 0.7537, + "num_input_tokens_seen": 165369130, + "step": 5074 + }, + { + "epoch": 0.45768138161157956, + "flos": 22277927946720.0, + "grad_norm": 1.7402593779988698, + "learning_rate": 2.3691372473381673e-06, + "loss": 0.8062, + "num_input_tokens_seen": 165397545, + "step": 5075 + }, + { + "epoch": 0.45777156513504985, + "flos": 24172954609920.0, + "grad_norm": 1.6744516072791202, + "learning_rate": 2.3685630579249708e-06, + "loss": 0.8122, + "num_input_tokens_seen": 165427700, + "step": 5076 + }, + { + "epoch": 0.4578617486585201, + "flos": 27520433140320.0, + "grad_norm": 1.602431196619275, + "learning_rate": 2.367988837063942e-06, + "loss": 0.7288, + "num_input_tokens_seen": 165455505, + "step": 5077 + }, + { + "epoch": 0.4579519321819904, + "flos": 23042853334080.0, + "grad_norm": 1.6853518765526927, + "learning_rate": 2.367414584804076e-06, + "loss": 0.8298, + "num_input_tokens_seen": 165482765, + "step": 5078 + }, + { + "epoch": 0.4580421157054606, + "flos": 22897124940480.0, + "grad_norm": 2.244096959737562, + "learning_rate": 2.366840301194372e-06, + "loss": 0.7871, + "num_input_tokens_seen": 165509870, + "step": 5079 + }, + { + "epoch": 0.4581322992289309, + "flos": 22058629130880.0, + "grad_norm": 1.600411499274639, + "learning_rate": 2.3662659862838308e-06, + "loss": 0.7408, + "num_input_tokens_seen": 165538895, + "step": 5080 + }, + { + "epoch": 0.45822248275240113, + "flos": 19508345073120.0, + "grad_norm": 2.2904219512665676, + "learning_rate": 2.365691640121456e-06, + "loss": 0.8112, + "num_input_tokens_seen": 165566835, + "step": 5081 + }, + { + "epoch": 0.4583126662758714, + "flos": 34118174077920.0, + "grad_norm": 1.8909540441769945, + "learning_rate": 2.365117262756254e-06, + "loss": 0.6857, + "num_input_tokens_seen": 165601860, + "step": 5082 + }, + { + "epoch": 0.45840284979934165, + "flos": 20457735477600.0, + "grad_norm": 1.7418833924780162, + "learning_rate": 2.3645428542372342e-06, + "loss": 0.7479, + "num_input_tokens_seen": 165630260, + "step": 5083 + }, + { + "epoch": 0.45849303332281194, + "flos": 70922393377920.0, + "grad_norm": 0.6169102202837902, + "learning_rate": 2.3639684146134083e-06, + "loss": 0.5315, + "num_input_tokens_seen": 165725155, + "step": 5084 + }, + { + "epoch": 0.45858321684628217, + "flos": 28906488348960.0, + "grad_norm": 2.2945123055403034, + "learning_rate": 2.3633939439337897e-06, + "loss": 0.7156, + "num_input_tokens_seen": 165753525, + "step": 5085 + }, + { + "epoch": 0.45867340036975246, + "flos": 18051693023040.0, + "grad_norm": 6.06898639191054, + "learning_rate": 2.362819442247396e-06, + "loss": 0.7266, + "num_input_tokens_seen": 165781325, + "step": 5086 + }, + { + "epoch": 0.4587635838932227, + "flos": 68469325443360.0, + "grad_norm": 0.636441995584808, + "learning_rate": 2.3622449096032477e-06, + "loss": 0.542, + "num_input_tokens_seen": 165878115, + "step": 5087 + }, + { + "epoch": 0.458853767416693, + "flos": 18014368850400.0, + "grad_norm": 3.884072051506504, + "learning_rate": 2.361670346050366e-06, + "loss": 0.7891, + "num_input_tokens_seen": 165906100, + "step": 5088 + }, + { + "epoch": 0.4589439509401632, + "flos": 16922372312160.0, + "grad_norm": 2.0139719964195284, + "learning_rate": 2.3610957516377757e-06, + "loss": 0.6906, + "num_input_tokens_seen": 165933405, + "step": 5089 + }, + { + "epoch": 0.4590341344636335, + "flos": 28796522998080.0, + "grad_norm": 2.449309084259146, + "learning_rate": 2.3605211264145048e-06, + "loss": 0.7552, + "num_input_tokens_seen": 165962635, + "step": 5090 + }, + { + "epoch": 0.45912431798710374, + "flos": 23078876565120.0, + "grad_norm": 1.6342550230026374, + "learning_rate": 2.3599464704295836e-06, + "loss": 0.7191, + "num_input_tokens_seen": 165993440, + "step": 5091 + }, + { + "epoch": 0.459214501510574, + "flos": 24099793055040.0, + "grad_norm": 1.8651585748259654, + "learning_rate": 2.359371783732045e-06, + "loss": 0.6454, + "num_input_tokens_seen": 166022635, + "step": 5092 + }, + { + "epoch": 0.45930468503404426, + "flos": 22091121234720.0, + "grad_norm": 1.5939053478449696, + "learning_rate": 2.358797066370924e-06, + "loss": 0.7331, + "num_input_tokens_seen": 166051565, + "step": 5093 + }, + { + "epoch": 0.45939486855751455, + "flos": 23662013158080.0, + "grad_norm": 1.744920033150122, + "learning_rate": 2.3582223183952594e-06, + "loss": 0.7529, + "num_input_tokens_seen": 166081445, + "step": 5094 + }, + { + "epoch": 0.4594850520809848, + "flos": 21585903925920.0, + "grad_norm": 1.701616371500003, + "learning_rate": 2.357647539854091e-06, + "loss": 0.7864, + "num_input_tokens_seen": 166109410, + "step": 5095 + }, + { + "epoch": 0.45957523560445507, + "flos": 23986742614080.0, + "grad_norm": 2.0809075792678997, + "learning_rate": 2.3570727307964624e-06, + "loss": 0.7558, + "num_input_tokens_seen": 166136905, + "step": 5096 + }, + { + "epoch": 0.4596654191279253, + "flos": 28137437118240.0, + "grad_norm": 1.8044184364452636, + "learning_rate": 2.35649789127142e-06, + "loss": 0.8002, + "num_input_tokens_seen": 166163085, + "step": 5097 + }, + { + "epoch": 0.4597556026513956, + "flos": 16957986675840.0, + "grad_norm": 1.9142762356749161, + "learning_rate": 2.3559230213280115e-06, + "loss": 0.716, + "num_input_tokens_seen": 166190460, + "step": 5098 + }, + { + "epoch": 0.4598457861748659, + "flos": 23516359104000.0, + "grad_norm": 1.7155078479672945, + "learning_rate": 2.3553481210152886e-06, + "loss": 0.6652, + "num_input_tokens_seen": 166220880, + "step": 5099 + }, + { + "epoch": 0.4599359696983361, + "flos": 55843667653440.0, + "grad_norm": 0.8452517249634999, + "learning_rate": 2.3547731903823043e-06, + "loss": 0.5969, + "num_input_tokens_seen": 166301880, + "step": 5100 + }, + { + "epoch": 0.4600261532218064, + "flos": 25265025487680.0, + "grad_norm": 1.637176740931025, + "learning_rate": 2.3541982294781155e-06, + "loss": 0.7284, + "num_input_tokens_seen": 166328330, + "step": 5101 + }, + { + "epoch": 0.46011633674527663, + "flos": 26866476687360.0, + "grad_norm": 3.247798027007061, + "learning_rate": 2.3536232383517804e-06, + "loss": 0.7515, + "num_input_tokens_seen": 166358480, + "step": 5102 + }, + { + "epoch": 0.4602065202687469, + "flos": 32224634205120.0, + "grad_norm": 2.2868700170970926, + "learning_rate": 2.3530482170523602e-06, + "loss": 0.7463, + "num_input_tokens_seen": 166388160, + "step": 5103 + }, + { + "epoch": 0.46029670379221715, + "flos": 22782513369600.0, + "grad_norm": 1.500977515553536, + "learning_rate": 2.3524731656289206e-06, + "loss": 0.8461, + "num_input_tokens_seen": 166416475, + "step": 5104 + }, + { + "epoch": 0.46038688731568744, + "flos": 25517299614240.0, + "grad_norm": 2.0229565847945485, + "learning_rate": 2.351898084130526e-06, + "loss": 0.8379, + "num_input_tokens_seen": 166446210, + "step": 5105 + }, + { + "epoch": 0.4604770708391577, + "flos": 23619448049280.0, + "grad_norm": 1.7293308900116362, + "learning_rate": 2.351322972606247e-06, + "loss": 0.8183, + "num_input_tokens_seen": 166473355, + "step": 5106 + }, + { + "epoch": 0.46056725436262796, + "flos": 24026222632800.0, + "grad_norm": 1.8752342713549124, + "learning_rate": 2.350747831105155e-06, + "loss": 0.903, + "num_input_tokens_seen": 166499005, + "step": 5107 + }, + { + "epoch": 0.4606574378860982, + "flos": 24201692568000.0, + "grad_norm": 1.897709962187567, + "learning_rate": 2.350172659676323e-06, + "loss": 0.8316, + "num_input_tokens_seen": 166525520, + "step": 5108 + }, + { + "epoch": 0.4607476214095685, + "flos": 21585569398080.0, + "grad_norm": 1.8874469875950013, + "learning_rate": 2.3495974583688306e-06, + "loss": 0.7889, + "num_input_tokens_seen": 166552610, + "step": 5109 + }, + { + "epoch": 0.4608378049330387, + "flos": 23694468092160.0, + "grad_norm": 2.123928585346182, + "learning_rate": 2.3490222272317543e-06, + "loss": 0.8239, + "num_input_tokens_seen": 166580530, + "step": 5110 + }, + { + "epoch": 0.460927988456509, + "flos": 17541346287360.0, + "grad_norm": 1.8482881719485578, + "learning_rate": 2.348446966314177e-06, + "loss": 0.7998, + "num_input_tokens_seen": 166604380, + "step": 5111 + }, + { + "epoch": 0.46101817197997924, + "flos": 23224976656320.0, + "grad_norm": 1.9967823968229088, + "learning_rate": 2.3478716756651837e-06, + "loss": 0.7262, + "num_input_tokens_seen": 166632415, + "step": 5112 + }, + { + "epoch": 0.46110835550344953, + "flos": 21222028979040.0, + "grad_norm": 1.8358532176049849, + "learning_rate": 2.347296355333861e-06, + "loss": 0.7466, + "num_input_tokens_seen": 166659490, + "step": 5113 + }, + { + "epoch": 0.46119853902691976, + "flos": 24536346349920.0, + "grad_norm": 1.8055545986205606, + "learning_rate": 2.3467210053692972e-06, + "loss": 0.7608, + "num_input_tokens_seen": 166689070, + "step": 5114 + }, + { + "epoch": 0.46128872255039005, + "flos": 21039459619680.0, + "grad_norm": 1.7739176499306137, + "learning_rate": 2.3461456258205866e-06, + "loss": 0.757, + "num_input_tokens_seen": 166716705, + "step": 5115 + }, + { + "epoch": 0.4613789060738603, + "flos": 37211928861120.0, + "grad_norm": 1.971825150768043, + "learning_rate": 2.345570216736822e-06, + "loss": 0.6935, + "num_input_tokens_seen": 166749330, + "step": 5116 + }, + { + "epoch": 0.4614690895973306, + "flos": 24755905354080.0, + "grad_norm": 1.8889158736277332, + "learning_rate": 2.3449947781671013e-06, + "loss": 0.7621, + "num_input_tokens_seen": 166779465, + "step": 5117 + }, + { + "epoch": 0.4615592731208008, + "flos": 20853693660960.0, + "grad_norm": 1.776404163883812, + "learning_rate": 2.3444193101605237e-06, + "loss": 0.7827, + "num_input_tokens_seen": 166807130, + "step": 5118 + }, + { + "epoch": 0.4616494566442711, + "flos": 26063000525280.0, + "grad_norm": 1.5930444988446093, + "learning_rate": 2.3438438127661913e-06, + "loss": 0.7719, + "num_input_tokens_seen": 166833565, + "step": 5119 + }, + { + "epoch": 0.4617396401677413, + "flos": 35390546959680.0, + "grad_norm": 1.929608115598114, + "learning_rate": 2.3432682860332096e-06, + "loss": 0.5846, + "num_input_tokens_seen": 166864605, + "step": 5120 + }, + { + "epoch": 0.4618298236912116, + "flos": 28939277810880.0, + "grad_norm": 1.8287992619667333, + "learning_rate": 2.342692730010684e-06, + "loss": 0.7047, + "num_input_tokens_seen": 166894605, + "step": 5121 + }, + { + "epoch": 0.4619200072146819, + "flos": 19799244313920.0, + "grad_norm": 2.2299530448588802, + "learning_rate": 2.342117144747726e-06, + "loss": 0.7639, + "num_input_tokens_seen": 166917000, + "step": 5122 + }, + { + "epoch": 0.46201019073815214, + "flos": 22747753910400.0, + "grad_norm": 1.8601284904603603, + "learning_rate": 2.3415415302934457e-06, + "loss": 0.7547, + "num_input_tokens_seen": 166943800, + "step": 5123 + }, + { + "epoch": 0.4621003742616224, + "flos": 25264988317920.0, + "grad_norm": 1.6193514733793704, + "learning_rate": 2.340965886696959e-06, + "loss": 0.7838, + "num_input_tokens_seen": 166970000, + "step": 5124 + }, + { + "epoch": 0.46219055778509266, + "flos": 19142760317280.0, + "grad_norm": 1.6233205920247475, + "learning_rate": 2.340390214007384e-06, + "loss": 0.7021, + "num_input_tokens_seen": 166998665, + "step": 5125 + }, + { + "epoch": 0.46228074130856295, + "flos": 21109350235680.0, + "grad_norm": 1.6684003713720326, + "learning_rate": 2.339814512273838e-06, + "loss": 0.6352, + "num_input_tokens_seen": 167025555, + "step": 5126 + }, + { + "epoch": 0.4623709248320332, + "flos": 24500620476960.0, + "grad_norm": 1.6421429832224725, + "learning_rate": 2.3392387815454447e-06, + "loss": 0.8585, + "num_input_tokens_seen": 167053175, + "step": 5127 + }, + { + "epoch": 0.46246110835550347, + "flos": 39288335451360.0, + "grad_norm": 2.168865726252592, + "learning_rate": 2.3386630218713273e-06, + "loss": 0.7486, + "num_input_tokens_seen": 167084905, + "step": 5128 + }, + { + "epoch": 0.4625512918789737, + "flos": 21840296728800.0, + "grad_norm": 1.716182108270436, + "learning_rate": 2.3380872333006135e-06, + "loss": 0.7769, + "num_input_tokens_seen": 167112505, + "step": 5129 + }, + { + "epoch": 0.462641475402444, + "flos": 23147131711680.0, + "grad_norm": 1.8410966175053454, + "learning_rate": 2.3375114158824335e-06, + "loss": 0.7985, + "num_input_tokens_seen": 167141720, + "step": 5130 + }, + { + "epoch": 0.4627316589259142, + "flos": 27670027188960.0, + "grad_norm": 1.5373617156574724, + "learning_rate": 2.3369355696659184e-06, + "loss": 0.7898, + "num_input_tokens_seen": 167172960, + "step": 5131 + }, + { + "epoch": 0.4628218424493845, + "flos": 21002916012000.0, + "grad_norm": 1.8110500323768912, + "learning_rate": 2.336359694700202e-06, + "loss": 0.7412, + "num_input_tokens_seen": 167201030, + "step": 5132 + }, + { + "epoch": 0.46291202597285475, + "flos": 18774313489920.0, + "grad_norm": 1.7973308482582127, + "learning_rate": 2.335783791034422e-06, + "loss": 0.7574, + "num_input_tokens_seen": 167226490, + "step": 5133 + }, + { + "epoch": 0.46300220949632503, + "flos": 20820644010720.0, + "grad_norm": 1.583550597225355, + "learning_rate": 2.3352078587177173e-06, + "loss": 0.7358, + "num_input_tokens_seen": 167255310, + "step": 5134 + }, + { + "epoch": 0.46309239301979527, + "flos": 13970963474400.0, + "grad_norm": 1.8704116629283403, + "learning_rate": 2.33463189779923e-06, + "loss": 0.7623, + "num_input_tokens_seen": 167279810, + "step": 5135 + }, + { + "epoch": 0.46318257654326556, + "flos": 21731260621920.0, + "grad_norm": 1.5107084145990681, + "learning_rate": 2.334055908328104e-06, + "loss": 0.8291, + "num_input_tokens_seen": 167310495, + "step": 5136 + }, + { + "epoch": 0.4632727600667358, + "flos": 22568901527040.0, + "grad_norm": 1.8967961252521275, + "learning_rate": 2.3334798903534866e-06, + "loss": 0.6629, + "num_input_tokens_seen": 167340375, + "step": 5137 + }, + { + "epoch": 0.4633629435902061, + "flos": 19545557736480.0, + "grad_norm": 1.4081776017456629, + "learning_rate": 2.3329038439245252e-06, + "loss": 0.6503, + "num_input_tokens_seen": 167368580, + "step": 5138 + }, + { + "epoch": 0.4634531271136763, + "flos": 21439506476640.0, + "grad_norm": 2.315496925590845, + "learning_rate": 2.3323277690903724e-06, + "loss": 0.7544, + "num_input_tokens_seen": 167394540, + "step": 5139 + }, + { + "epoch": 0.4635433106371466, + "flos": 33533587864320.0, + "grad_norm": 1.6921011045878975, + "learning_rate": 2.3317516659001827e-06, + "loss": 0.7304, + "num_input_tokens_seen": 167424855, + "step": 5140 + }, + { + "epoch": 0.46363349416061683, + "flos": 21072249081600.0, + "grad_norm": 1.575651742419978, + "learning_rate": 2.331175534403111e-06, + "loss": 0.7673, + "num_input_tokens_seen": 167452370, + "step": 5141 + }, + { + "epoch": 0.4637236776840871, + "flos": 28831914343200.0, + "grad_norm": 1.5070094999382229, + "learning_rate": 2.3305993746483167e-06, + "loss": 0.8383, + "num_input_tokens_seen": 167478925, + "step": 5142 + }, + { + "epoch": 0.46381386120755735, + "flos": 15683346438720.0, + "grad_norm": 2.3007561726004946, + "learning_rate": 2.3300231866849606e-06, + "loss": 0.7327, + "num_input_tokens_seen": 167503855, + "step": 5143 + }, + { + "epoch": 0.46390404473102764, + "flos": 34773542981760.0, + "grad_norm": 1.9318568587893779, + "learning_rate": 2.3294469705622067e-06, + "loss": 0.7945, + "num_input_tokens_seen": 167535730, + "step": 5144 + }, + { + "epoch": 0.4639942282544979, + "flos": 23953395605760.0, + "grad_norm": 1.8246172647340326, + "learning_rate": 2.3288707263292203e-06, + "loss": 0.7307, + "num_input_tokens_seen": 167563520, + "step": 5145 + }, + { + "epoch": 0.46408441177796816, + "flos": 32405047718400.0, + "grad_norm": 1.4694774530633825, + "learning_rate": 2.3282944540351707e-06, + "loss": 0.7712, + "num_input_tokens_seen": 167595480, + "step": 5146 + }, + { + "epoch": 0.46417459530143845, + "flos": 21913235265120.0, + "grad_norm": 2.106456744181138, + "learning_rate": 2.327718153729228e-06, + "loss": 0.7294, + "num_input_tokens_seen": 167624490, + "step": 5147 + }, + { + "epoch": 0.4642647788249087, + "flos": 24172768761120.0, + "grad_norm": 1.7558263790437674, + "learning_rate": 2.327141825460566e-06, + "loss": 0.7942, + "num_input_tokens_seen": 167652680, + "step": 5148 + }, + { + "epoch": 0.464354962348379, + "flos": 57606018242400.0, + "grad_norm": 0.6705694821958026, + "learning_rate": 2.326565469278358e-06, + "loss": 0.5844, + "num_input_tokens_seen": 167736310, + "step": 5149 + }, + { + "epoch": 0.4644451458718492, + "flos": 55101235704480.0, + "grad_norm": 0.7127051543300827, + "learning_rate": 2.3259890852317846e-06, + "loss": 0.5938, + "num_input_tokens_seen": 167816400, + "step": 5150 + }, + { + "epoch": 0.4645353293953195, + "flos": 21542149384800.0, + "grad_norm": 1.9797396850933817, + "learning_rate": 2.3254126733700246e-06, + "loss": 0.7534, + "num_input_tokens_seen": 167843545, + "step": 5151 + }, + { + "epoch": 0.46462551291878973, + "flos": 20451825485760.0, + "grad_norm": 1.6837794650642481, + "learning_rate": 2.324836233742262e-06, + "loss": 0.7852, + "num_input_tokens_seen": 167869230, + "step": 5152 + }, + { + "epoch": 0.46471569644226, + "flos": 20346803712960.0, + "grad_norm": 1.9424247909773793, + "learning_rate": 2.3242597663976793e-06, + "loss": 0.7922, + "num_input_tokens_seen": 167896315, + "step": 5153 + }, + { + "epoch": 0.46480587996573025, + "flos": 19363806111840.0, + "grad_norm": 1.8166898772597209, + "learning_rate": 2.3236832713854663e-06, + "loss": 0.7355, + "num_input_tokens_seen": 167924545, + "step": 5154 + }, + { + "epoch": 0.46489606348920054, + "flos": 21075557190240.0, + "grad_norm": 1.7936425366260513, + "learning_rate": 2.323106748754812e-06, + "loss": 0.7512, + "num_input_tokens_seen": 167952550, + "step": 5155 + }, + { + "epoch": 0.4649862470126708, + "flos": 18307275258240.0, + "grad_norm": 15.717114221235379, + "learning_rate": 2.3225301985549077e-06, + "loss": 0.8465, + "num_input_tokens_seen": 167981015, + "step": 5156 + }, + { + "epoch": 0.46507643053614106, + "flos": 22095321417600.0, + "grad_norm": 1.925625808354115, + "learning_rate": 2.321953620834948e-06, + "loss": 0.686, + "num_input_tokens_seen": 168009200, + "step": 5157 + }, + { + "epoch": 0.4651666140596113, + "flos": 23220887982720.0, + "grad_norm": 1.5024851240034167, + "learning_rate": 2.3213770156441314e-06, + "loss": 0.721, + "num_input_tokens_seen": 168037305, + "step": 5158 + }, + { + "epoch": 0.4652567975830816, + "flos": 26102963750880.0, + "grad_norm": 2.313140883433724, + "learning_rate": 2.3208003830316554e-06, + "loss": 0.8037, + "num_input_tokens_seen": 168068730, + "step": 5159 + }, + { + "epoch": 0.4653469811065518, + "flos": 18925877535840.0, + "grad_norm": 1.6610630565710844, + "learning_rate": 2.3202237230467215e-06, + "loss": 0.8435, + "num_input_tokens_seen": 168093940, + "step": 5160 + }, + { + "epoch": 0.4654371646300221, + "flos": 22203428280480.0, + "grad_norm": 2.5195448764166404, + "learning_rate": 2.3196470357385338e-06, + "loss": 0.7107, + "num_input_tokens_seen": 168123615, + "step": 5161 + }, + { + "epoch": 0.46552734815349234, + "flos": 31676851787520.0, + "grad_norm": 2.6215996289599577, + "learning_rate": 2.319070321156299e-06, + "loss": 0.7474, + "num_input_tokens_seen": 168155530, + "step": 5162 + }, + { + "epoch": 0.4656175316769626, + "flos": 27304442433120.0, + "grad_norm": 1.5060501472564005, + "learning_rate": 2.318493579349224e-06, + "loss": 0.886, + "num_input_tokens_seen": 168187475, + "step": 5163 + }, + { + "epoch": 0.46570771520043286, + "flos": 18525347472000.0, + "grad_norm": 1.9428256888307778, + "learning_rate": 2.317916810366522e-06, + "loss": 0.7433, + "num_input_tokens_seen": 168215555, + "step": 5164 + }, + { + "epoch": 0.46579789872390315, + "flos": 25916454396960.0, + "grad_norm": 1.7474497878210438, + "learning_rate": 2.317340014257404e-06, + "loss": 0.6998, + "num_input_tokens_seen": 168244645, + "step": 5165 + }, + { + "epoch": 0.4658880822473734, + "flos": 23040251450880.0, + "grad_norm": 1.4254782994979185, + "learning_rate": 2.316763191071086e-06, + "loss": 0.6825, + "num_input_tokens_seen": 168273475, + "step": 5166 + }, + { + "epoch": 0.46597826577084367, + "flos": 43700893880160.0, + "grad_norm": 1.739789398536609, + "learning_rate": 2.316186340856787e-06, + "loss": 0.7132, + "num_input_tokens_seen": 168305460, + "step": 5167 + }, + { + "epoch": 0.4660684492943139, + "flos": 25950581970240.0, + "grad_norm": 2.0750894963830517, + "learning_rate": 2.315609463663725e-06, + "loss": 0.7056, + "num_input_tokens_seen": 168332040, + "step": 5168 + }, + { + "epoch": 0.4661586328177842, + "flos": 37834285284480.0, + "grad_norm": 2.0296646974382493, + "learning_rate": 2.315032559541123e-06, + "loss": 0.7297, + "num_input_tokens_seen": 168361730, + "step": 5169 + }, + { + "epoch": 0.4662488163412545, + "flos": 28616667031200.0, + "grad_norm": 1.8825615680044228, + "learning_rate": 2.314455628538207e-06, + "loss": 0.7692, + "num_input_tokens_seen": 168388950, + "step": 5170 + }, + { + "epoch": 0.4663389998647247, + "flos": 46717435604640.0, + "grad_norm": 1.5625805643292545, + "learning_rate": 2.3138786707042023e-06, + "loss": 0.7251, + "num_input_tokens_seen": 168421265, + "step": 5171 + }, + { + "epoch": 0.466429183388195, + "flos": 24317121873600.0, + "grad_norm": 1.5786499812081, + "learning_rate": 2.3133016860883387e-06, + "loss": 0.764, + "num_input_tokens_seen": 168451895, + "step": 5172 + }, + { + "epoch": 0.46651936691166523, + "flos": 22678681029120.0, + "grad_norm": 1.7435599767041288, + "learning_rate": 2.3127246747398475e-06, + "loss": 0.7284, + "num_input_tokens_seen": 168481405, + "step": 5173 + }, + { + "epoch": 0.4666095504351355, + "flos": 31239852455520.0, + "grad_norm": 1.3674576410710484, + "learning_rate": 2.312147636707963e-06, + "loss": 0.75, + "num_input_tokens_seen": 168513230, + "step": 5174 + }, + { + "epoch": 0.46669973395860576, + "flos": 18233667666240.0, + "grad_norm": 1.764915728948183, + "learning_rate": 2.3115705720419214e-06, + "loss": 0.7369, + "num_input_tokens_seen": 168539535, + "step": 5175 + }, + { + "epoch": 0.46678991748207604, + "flos": 33754410640320.0, + "grad_norm": 1.9459028458959335, + "learning_rate": 2.31099348079096e-06, + "loss": 0.7723, + "num_input_tokens_seen": 168571170, + "step": 5176 + }, + { + "epoch": 0.4668801010055463, + "flos": 22424102377440.0, + "grad_norm": 1.4480237944054748, + "learning_rate": 2.31041636300432e-06, + "loss": 0.7956, + "num_input_tokens_seen": 168598840, + "step": 5177 + }, + { + "epoch": 0.46697028452901657, + "flos": 21805388590560.0, + "grad_norm": 1.835574499727285, + "learning_rate": 2.3098392187312445e-06, + "loss": 0.8375, + "num_input_tokens_seen": 168626260, + "step": 5178 + }, + { + "epoch": 0.4670604680524868, + "flos": 21513337087200.0, + "grad_norm": 2.0114203074288595, + "learning_rate": 2.309262048020978e-06, + "loss": 0.7307, + "num_input_tokens_seen": 168656080, + "step": 5179 + }, + { + "epoch": 0.4671506515759571, + "flos": 16446338998560.0, + "grad_norm": 1.6618675118715571, + "learning_rate": 2.308684850922769e-06, + "loss": 0.7832, + "num_input_tokens_seen": 168681865, + "step": 5180 + }, + { + "epoch": 0.4672408350994273, + "flos": 17978977505280.0, + "grad_norm": 1.8341542775649795, + "learning_rate": 2.3081076274858664e-06, + "loss": 0.7475, + "num_input_tokens_seen": 168708920, + "step": 5181 + }, + { + "epoch": 0.4673310186228976, + "flos": 19284920413920.0, + "grad_norm": 1.466643425728429, + "learning_rate": 2.307530377759522e-06, + "loss": 0.7781, + "num_input_tokens_seen": 168736525, + "step": 5182 + }, + { + "epoch": 0.46742120214636784, + "flos": 20200926640320.0, + "grad_norm": 2.1249279683190285, + "learning_rate": 2.30695310179299e-06, + "loss": 0.6837, + "num_input_tokens_seen": 168762705, + "step": 5183 + }, + { + "epoch": 0.46751138566983813, + "flos": 23478811912800.0, + "grad_norm": 1.5128693846795591, + "learning_rate": 2.3063757996355267e-06, + "loss": 0.7518, + "num_input_tokens_seen": 168791995, + "step": 5184 + }, + { + "epoch": 0.46760156919330836, + "flos": 17723915646720.0, + "grad_norm": 2.01518516768827, + "learning_rate": 2.3057984713363903e-06, + "loss": 0.783, + "num_input_tokens_seen": 168818045, + "step": 5185 + }, + { + "epoch": 0.46769175271677865, + "flos": 71176972029600.0, + "grad_norm": 0.6954944672518728, + "learning_rate": 2.3052211169448436e-06, + "loss": 0.556, + "num_input_tokens_seen": 168917620, + "step": 5186 + }, + { + "epoch": 0.4677819362402489, + "flos": 32661373348800.0, + "grad_norm": 1.7790695363436118, + "learning_rate": 2.3046437365101474e-06, + "loss": 0.6954, + "num_input_tokens_seen": 168949460, + "step": 5187 + }, + { + "epoch": 0.4678721197637192, + "flos": 21439580816160.0, + "grad_norm": 1.787639601798066, + "learning_rate": 2.3040663300815673e-06, + "loss": 0.7281, + "num_input_tokens_seen": 168976355, + "step": 5188 + }, + { + "epoch": 0.4679623032871894, + "flos": 65410404458880.0, + "grad_norm": 0.7275193999020828, + "learning_rate": 2.3034888977083723e-06, + "loss": 0.5646, + "num_input_tokens_seen": 169064405, + "step": 5189 + }, + { + "epoch": 0.4680524868106597, + "flos": 28362348567840.0, + "grad_norm": 2.180625686273661, + "learning_rate": 2.30291143943983e-06, + "loss": 0.6635, + "num_input_tokens_seen": 169094435, + "step": 5190 + }, + { + "epoch": 0.46814267033412993, + "flos": 32771227190400.0, + "grad_norm": 2.1512043839826704, + "learning_rate": 2.3023339553252145e-06, + "loss": 0.7145, + "num_input_tokens_seen": 169126230, + "step": 5191 + }, + { + "epoch": 0.4682328538576002, + "flos": 23881534992480.0, + "grad_norm": 2.366846544745927, + "learning_rate": 2.301756445413799e-06, + "loss": 0.7862, + "num_input_tokens_seen": 169155865, + "step": 5192 + }, + { + "epoch": 0.4683230373810705, + "flos": 25993704625440.0, + "grad_norm": 1.8618553991909086, + "learning_rate": 2.3011789097548585e-06, + "loss": 0.7595, + "num_input_tokens_seen": 169184640, + "step": 5193 + }, + { + "epoch": 0.46841322090454074, + "flos": 24934609058400.0, + "grad_norm": 1.8473292377930837, + "learning_rate": 2.3006013483976738e-06, + "loss": 0.7489, + "num_input_tokens_seen": 169214480, + "step": 5194 + }, + { + "epoch": 0.468503404428011, + "flos": 19685041610400.0, + "grad_norm": 2.2873831046832387, + "learning_rate": 2.300023761391524e-06, + "loss": 0.7492, + "num_input_tokens_seen": 169240275, + "step": 5195 + }, + { + "epoch": 0.46859358795148126, + "flos": 25629346471680.0, + "grad_norm": 2.09092430629396, + "learning_rate": 2.299446148785693e-06, + "loss": 0.8046, + "num_input_tokens_seen": 169266910, + "step": 5196 + }, + { + "epoch": 0.46868377147495155, + "flos": 15063703407840.0, + "grad_norm": 2.107005618246778, + "learning_rate": 2.2988685106294654e-06, + "loss": 0.7385, + "num_input_tokens_seen": 169291910, + "step": 5197 + }, + { + "epoch": 0.4687739549984218, + "flos": 32442743588640.0, + "grad_norm": 2.0399613427258063, + "learning_rate": 2.2982908469721284e-06, + "loss": 0.6514, + "num_input_tokens_seen": 169319720, + "step": 5198 + }, + { + "epoch": 0.46886413852189207, + "flos": 23989604685600.0, + "grad_norm": 1.8474079634643994, + "learning_rate": 2.2977131578629714e-06, + "loss": 0.8223, + "num_input_tokens_seen": 169348640, + "step": 5199 + }, + { + "epoch": 0.4689543220453623, + "flos": 23007164630880.0, + "grad_norm": 1.6641571304565093, + "learning_rate": 2.297135443351286e-06, + "loss": 0.7759, + "num_input_tokens_seen": 169379545, + "step": 5200 + }, + { + "epoch": 0.4690445055688326, + "flos": 25995005567040.0, + "grad_norm": 1.3696200718994274, + "learning_rate": 2.296557703486367e-06, + "loss": 0.7653, + "num_input_tokens_seen": 169410540, + "step": 5201 + }, + { + "epoch": 0.4691346890923028, + "flos": 22781435446560.0, + "grad_norm": 1.9289164373314986, + "learning_rate": 2.295979938317509e-06, + "loss": 0.7799, + "num_input_tokens_seen": 169438630, + "step": 5202 + }, + { + "epoch": 0.4692248726157731, + "flos": 15647025849600.0, + "grad_norm": 2.3053744052157508, + "learning_rate": 2.295402147894011e-06, + "loss": 0.7561, + "num_input_tokens_seen": 169467115, + "step": 5203 + }, + { + "epoch": 0.46931505613924335, + "flos": 36483472741920.0, + "grad_norm": 1.5329121501856418, + "learning_rate": 2.2948243322651723e-06, + "loss": 0.6789, + "num_input_tokens_seen": 169501380, + "step": 5204 + }, + { + "epoch": 0.46940523966271364, + "flos": 32296792176480.0, + "grad_norm": 1.5943129050299059, + "learning_rate": 2.2942464914802962e-06, + "loss": 0.7584, + "num_input_tokens_seen": 169531930, + "step": 5205 + }, + { + "epoch": 0.46949542318618387, + "flos": 17468928127680.0, + "grad_norm": 1.9328526165067115, + "learning_rate": 2.293668625588687e-06, + "loss": 0.7339, + "num_input_tokens_seen": 169557170, + "step": 5206 + }, + { + "epoch": 0.46958560670965416, + "flos": 15282816374880.0, + "grad_norm": 2.1813447729469853, + "learning_rate": 2.293090734639651e-06, + "loss": 0.799, + "num_input_tokens_seen": 169581985, + "step": 5207 + }, + { + "epoch": 0.4696757902331244, + "flos": 23406579601920.0, + "grad_norm": 2.390419745516134, + "learning_rate": 2.2925128186824983e-06, + "loss": 0.8713, + "num_input_tokens_seen": 169608980, + "step": 5208 + }, + { + "epoch": 0.4697659737565947, + "flos": 25738196729760.0, + "grad_norm": 2.312148608754386, + "learning_rate": 2.2919348777665384e-06, + "loss": 0.8568, + "num_input_tokens_seen": 169636780, + "step": 5209 + }, + { + "epoch": 0.4698561572800649, + "flos": 25046730255360.0, + "grad_norm": 1.748005888900614, + "learning_rate": 2.2913569119410856e-06, + "loss": 0.7609, + "num_input_tokens_seen": 169665205, + "step": 5210 + }, + { + "epoch": 0.4699463408035352, + "flos": 27123174015360.0, + "grad_norm": 1.966194371178006, + "learning_rate": 2.290778921255454e-06, + "loss": 0.7801, + "num_input_tokens_seen": 169694810, + "step": 5211 + }, + { + "epoch": 0.47003652432700543, + "flos": 21803307084000.0, + "grad_norm": 1.5270498632047826, + "learning_rate": 2.2902009057589613e-06, + "loss": 0.7896, + "num_input_tokens_seen": 169725625, + "step": 5212 + }, + { + "epoch": 0.4701267078504757, + "flos": 27158007814080.0, + "grad_norm": 1.4154918043787996, + "learning_rate": 2.2896228655009276e-06, + "loss": 0.711, + "num_input_tokens_seen": 169758395, + "step": 5213 + }, + { + "epoch": 0.47021689137394596, + "flos": 20565619321920.0, + "grad_norm": 5.884778589638889, + "learning_rate": 2.289044800530674e-06, + "loss": 0.7236, + "num_input_tokens_seen": 169786175, + "step": 5214 + }, + { + "epoch": 0.47030707489741624, + "flos": 13898991351840.0, + "grad_norm": 1.9902559378432003, + "learning_rate": 2.2884667108975245e-06, + "loss": 0.7205, + "num_input_tokens_seen": 169812690, + "step": 5215 + }, + { + "epoch": 0.4703972584208865, + "flos": 20966521083360.0, + "grad_norm": 1.632102118195657, + "learning_rate": 2.287888596650804e-06, + "loss": 0.7922, + "num_input_tokens_seen": 169841755, + "step": 5216 + }, + { + "epoch": 0.47048744194435677, + "flos": 22200714888000.0, + "grad_norm": 1.926502580167229, + "learning_rate": 2.287310457839841e-06, + "loss": 0.7522, + "num_input_tokens_seen": 169869880, + "step": 5217 + }, + { + "epoch": 0.47057762546782705, + "flos": 26465203228320.0, + "grad_norm": 1.881862669885049, + "learning_rate": 2.286732294513966e-06, + "loss": 0.8268, + "num_input_tokens_seen": 169896445, + "step": 5218 + }, + { + "epoch": 0.4706678089912973, + "flos": 19801139971680.0, + "grad_norm": 1.683468913607883, + "learning_rate": 2.2861541067225106e-06, + "loss": 0.7668, + "num_input_tokens_seen": 169924545, + "step": 5219 + }, + { + "epoch": 0.4707579925147676, + "flos": 21513151238400.0, + "grad_norm": 2.8319118711369233, + "learning_rate": 2.2855758945148095e-06, + "loss": 0.7396, + "num_input_tokens_seen": 169953460, + "step": 5220 + }, + { + "epoch": 0.4708481760382378, + "flos": 16594111728960.0, + "grad_norm": 2.729347446268502, + "learning_rate": 2.2849976579401977e-06, + "loss": 0.8152, + "num_input_tokens_seen": 169980315, + "step": 5221 + }, + { + "epoch": 0.4709383595617081, + "flos": 19180195999200.0, + "grad_norm": 3.332772462070513, + "learning_rate": 2.284419397048014e-06, + "loss": 0.7467, + "num_input_tokens_seen": 170008290, + "step": 5222 + }, + { + "epoch": 0.47102854308517833, + "flos": 26831977416480.0, + "grad_norm": 1.688106992838072, + "learning_rate": 2.2838411118875997e-06, + "loss": 0.7632, + "num_input_tokens_seen": 170037375, + "step": 5223 + }, + { + "epoch": 0.4711187266086486, + "flos": 24645977172960.0, + "grad_norm": 1.5898740423442417, + "learning_rate": 2.283262802508296e-06, + "loss": 0.679, + "num_input_tokens_seen": 170065765, + "step": 5224 + }, + { + "epoch": 0.47120891013211885, + "flos": 25082939335200.0, + "grad_norm": 1.3943489874546895, + "learning_rate": 2.2826844689594492e-06, + "loss": 0.7342, + "num_input_tokens_seen": 170097600, + "step": 5225 + }, + { + "epoch": 0.47129909365558914, + "flos": 19545892264320.0, + "grad_norm": 1.7677962869295516, + "learning_rate": 2.282106111290404e-06, + "loss": 0.8288, + "num_input_tokens_seen": 170127965, + "step": 5226 + }, + { + "epoch": 0.4713892771790594, + "flos": 64635554745600.0, + "grad_norm": 0.7008727771947619, + "learning_rate": 2.2815277295505098e-06, + "loss": 0.6009, + "num_input_tokens_seen": 170215995, + "step": 5227 + }, + { + "epoch": 0.47147946070252966, + "flos": 22857830770560.0, + "grad_norm": 4.365842772549651, + "learning_rate": 2.2809493237891174e-06, + "loss": 0.7009, + "num_input_tokens_seen": 170243620, + "step": 5228 + }, + { + "epoch": 0.4715696442259999, + "flos": 24282585432960.0, + "grad_norm": 1.7004045229553502, + "learning_rate": 2.2803708940555796e-06, + "loss": 0.783, + "num_input_tokens_seen": 170272805, + "step": 5229 + }, + { + "epoch": 0.4716598277494702, + "flos": 25192421479200.0, + "grad_norm": 1.4212313022910157, + "learning_rate": 2.2797924403992514e-06, + "loss": 0.8269, + "num_input_tokens_seen": 170303480, + "step": 5230 + }, + { + "epoch": 0.4717500112729404, + "flos": 19101756338400.0, + "grad_norm": 1.9961786121462093, + "learning_rate": 2.2792139628694892e-06, + "loss": 0.7747, + "num_input_tokens_seen": 170330585, + "step": 5231 + }, + { + "epoch": 0.4718401947964107, + "flos": 25264876808640.0, + "grad_norm": 1.6872535808218319, + "learning_rate": 2.2786354615156524e-06, + "loss": 0.7151, + "num_input_tokens_seen": 170358500, + "step": 5232 + }, + { + "epoch": 0.47193037831988094, + "flos": 17755143978720.0, + "grad_norm": 1.9354535647167892, + "learning_rate": 2.2780569363871016e-06, + "loss": 0.8198, + "num_input_tokens_seen": 170384960, + "step": 5233 + }, + { + "epoch": 0.4720205618433512, + "flos": 67630086238560.0, + "grad_norm": 0.6158620145732211, + "learning_rate": 2.277478387533199e-06, + "loss": 0.5719, + "num_input_tokens_seen": 170471740, + "step": 5234 + }, + { + "epoch": 0.47211074536682146, + "flos": 22201235264640.0, + "grad_norm": 1.6941274553156047, + "learning_rate": 2.276899815003311e-06, + "loss": 0.7912, + "num_input_tokens_seen": 170499720, + "step": 5235 + }, + { + "epoch": 0.47220092889029175, + "flos": 18853347866880.0, + "grad_norm": 1.5871979914267966, + "learning_rate": 2.2763212188468045e-06, + "loss": 0.7476, + "num_input_tokens_seen": 170527025, + "step": 5236 + }, + { + "epoch": 0.472291112413762, + "flos": 24280355247360.0, + "grad_norm": 1.4533651010004438, + "learning_rate": 2.2757425991130473e-06, + "loss": 0.7682, + "num_input_tokens_seen": 170557175, + "step": 5237 + }, + { + "epoch": 0.47238129593723227, + "flos": 67819978040640.0, + "grad_norm": 0.7596039587775469, + "learning_rate": 2.2751639558514117e-06, + "loss": 0.5851, + "num_input_tokens_seen": 170646095, + "step": 5238 + }, + { + "epoch": 0.4724714794607025, + "flos": 25738568427360.0, + "grad_norm": 3.1567228359661783, + "learning_rate": 2.2745852891112697e-06, + "loss": 0.7944, + "num_input_tokens_seen": 170677400, + "step": 5239 + }, + { + "epoch": 0.4725616629841728, + "flos": 23584688590080.0, + "grad_norm": 1.9957896260338357, + "learning_rate": 2.274006598941997e-06, + "loss": 0.7592, + "num_input_tokens_seen": 170707220, + "step": 5240 + }, + { + "epoch": 0.4726518465076431, + "flos": 31711871435040.0, + "grad_norm": 1.7922043843625561, + "learning_rate": 2.27342788539297e-06, + "loss": 0.7339, + "num_input_tokens_seen": 170740945, + "step": 5241 + }, + { + "epoch": 0.4727420300311133, + "flos": 29673978449760.0, + "grad_norm": 1.527449270052203, + "learning_rate": 2.2728491485135684e-06, + "loss": 0.712, + "num_input_tokens_seen": 170769385, + "step": 5242 + }, + { + "epoch": 0.4728322135545836, + "flos": 26212260046080.0, + "grad_norm": 2.040785839524558, + "learning_rate": 2.272270388353173e-06, + "loss": 0.8116, + "num_input_tokens_seen": 170797885, + "step": 5243 + }, + { + "epoch": 0.47292239707805384, + "flos": 24973828888800.0, + "grad_norm": 1.7937486070368338, + "learning_rate": 2.2716916049611666e-06, + "loss": 0.7286, + "num_input_tokens_seen": 170828350, + "step": 5244 + }, + { + "epoch": 0.4730125806015241, + "flos": 24095258344320.0, + "grad_norm": 1.8021098302168446, + "learning_rate": 2.2711127983869346e-06, + "loss": 0.7367, + "num_input_tokens_seen": 170855935, + "step": 5245 + }, + { + "epoch": 0.47310276412499436, + "flos": 27811704078720.0, + "grad_norm": 1.5090952738419885, + "learning_rate": 2.270533968679864e-06, + "loss": 0.7648, + "num_input_tokens_seen": 170884810, + "step": 5246 + }, + { + "epoch": 0.47319294764846465, + "flos": 18307052239680.0, + "grad_norm": 3.3253655254250694, + "learning_rate": 2.269955115889343e-06, + "loss": 0.7615, + "num_input_tokens_seen": 170913095, + "step": 5247 + }, + { + "epoch": 0.4732831311719349, + "flos": 20857187618400.0, + "grad_norm": 2.171977088779283, + "learning_rate": 2.269376240064763e-06, + "loss": 0.7152, + "num_input_tokens_seen": 170941290, + "step": 5248 + }, + { + "epoch": 0.47337331469540517, + "flos": 26170326823200.0, + "grad_norm": 1.7566791865260372, + "learning_rate": 2.268797341255517e-06, + "loss": 0.7739, + "num_input_tokens_seen": 170972965, + "step": 5249 + }, + { + "epoch": 0.4734634982188754, + "flos": 21621109422240.0, + "grad_norm": 1.988848232526229, + "learning_rate": 2.268218419511e-06, + "loss": 0.6712, + "num_input_tokens_seen": 170999720, + "step": 5250 + }, + { + "epoch": 0.4735536817423457, + "flos": 23657329768320.0, + "grad_norm": 1.733335234873291, + "learning_rate": 2.267639474880608e-06, + "loss": 0.6794, + "num_input_tokens_seen": 171027920, + "step": 5251 + }, + { + "epoch": 0.4736438652658159, + "flos": 24973345681920.0, + "grad_norm": 1.5152961924865895, + "learning_rate": 2.2670605074137407e-06, + "loss": 0.7356, + "num_input_tokens_seen": 171058530, + "step": 5252 + }, + { + "epoch": 0.4737340487892862, + "flos": 24209015010720.0, + "grad_norm": 1.7990738393357635, + "learning_rate": 2.2664815171597983e-06, + "loss": 0.7381, + "num_input_tokens_seen": 171085920, + "step": 5253 + }, + { + "epoch": 0.47382423231275644, + "flos": 19982594238240.0, + "grad_norm": 1.6466201237550075, + "learning_rate": 2.265902504168183e-06, + "loss": 0.6996, + "num_input_tokens_seen": 171112800, + "step": 5254 + }, + { + "epoch": 0.47391441583622673, + "flos": 19909692871680.0, + "grad_norm": 1.6402624547107887, + "learning_rate": 2.2653234684883007e-06, + "loss": 0.7099, + "num_input_tokens_seen": 171138295, + "step": 5255 + }, + { + "epoch": 0.47400459935969697, + "flos": 13420095966720.0, + "grad_norm": 2.329412731441313, + "learning_rate": 2.264744410169556e-06, + "loss": 0.6752, + "num_input_tokens_seen": 171163570, + "step": 5256 + }, + { + "epoch": 0.47409478288316725, + "flos": 30183804808800.0, + "grad_norm": 2.038013641581303, + "learning_rate": 2.264165329261359e-06, + "loss": 0.7353, + "num_input_tokens_seen": 171196865, + "step": 5257 + }, + { + "epoch": 0.4741849664066375, + "flos": 67058063403840.0, + "grad_norm": 0.5738187421702872, + "learning_rate": 2.26358622581312e-06, + "loss": 0.5746, + "num_input_tokens_seen": 171298160, + "step": 5258 + }, + { + "epoch": 0.4742751499301078, + "flos": 28285618716000.0, + "grad_norm": 1.880838649552381, + "learning_rate": 2.2630070998742504e-06, + "loss": 0.817, + "num_input_tokens_seen": 171325235, + "step": 5259 + }, + { + "epoch": 0.474365333453578, + "flos": 22787754305760.0, + "grad_norm": 1.6696840708099627, + "learning_rate": 2.262427951494165e-06, + "loss": 0.7778, + "num_input_tokens_seen": 171355660, + "step": 5260 + }, + { + "epoch": 0.4744555169770483, + "flos": 67534728603360.0, + "grad_norm": 0.6498629975822026, + "learning_rate": 2.2618487807222794e-06, + "loss": 0.6142, + "num_input_tokens_seen": 171451225, + "step": 5261 + }, + { + "epoch": 0.47454570050051853, + "flos": 26686286192640.0, + "grad_norm": 1.5236610444212135, + "learning_rate": 2.261269587608012e-06, + "loss": 0.8366, + "num_input_tokens_seen": 171480765, + "step": 5262 + }, + { + "epoch": 0.4746358840239888, + "flos": 25957607054880.0, + "grad_norm": 1.5193966999646284, + "learning_rate": 2.260690372200783e-06, + "loss": 0.7601, + "num_input_tokens_seen": 171509425, + "step": 5263 + }, + { + "epoch": 0.47472606754745905, + "flos": 21328463202720.0, + "grad_norm": 1.699267507270129, + "learning_rate": 2.2601111345500138e-06, + "loss": 0.8413, + "num_input_tokens_seen": 171539710, + "step": 5264 + }, + { + "epoch": 0.47481625107092934, + "flos": 24643300950240.0, + "grad_norm": 1.6136417309455648, + "learning_rate": 2.2595318747051286e-06, + "loss": 0.7452, + "num_input_tokens_seen": 171569675, + "step": 5265 + }, + { + "epoch": 0.47490643459439963, + "flos": 21366939637920.0, + "grad_norm": 1.6089551804003837, + "learning_rate": 2.258952592715553e-06, + "loss": 0.7775, + "num_input_tokens_seen": 171597320, + "step": 5266 + }, + { + "epoch": 0.47499661811786986, + "flos": 64023977552640.0, + "grad_norm": 0.7054054155783758, + "learning_rate": 2.2583732886307142e-06, + "loss": 0.6171, + "num_input_tokens_seen": 171689825, + "step": 5267 + }, + { + "epoch": 0.47508680164134015, + "flos": 23152186799040.0, + "grad_norm": 1.660101891917988, + "learning_rate": 2.2577939625000414e-06, + "loss": 0.703, + "num_input_tokens_seen": 171719790, + "step": 5268 + }, + { + "epoch": 0.4751769851648104, + "flos": 66473291341440.0, + "grad_norm": 0.67215292749032, + "learning_rate": 2.257214614372967e-06, + "loss": 0.5768, + "num_input_tokens_seen": 171806735, + "step": 5269 + }, + { + "epoch": 0.4752671686882807, + "flos": 23152223968800.0, + "grad_norm": 1.5566079245163549, + "learning_rate": 2.2566352442989227e-06, + "loss": 0.7796, + "num_input_tokens_seen": 171835125, + "step": 5270 + }, + { + "epoch": 0.4753573522117509, + "flos": 22423990868160.0, + "grad_norm": 1.9423189224621598, + "learning_rate": 2.256055852327344e-06, + "loss": 0.7603, + "num_input_tokens_seen": 171864715, + "step": 5271 + }, + { + "epoch": 0.4754475357352212, + "flos": 34151558256000.0, + "grad_norm": 1.5620275927706213, + "learning_rate": 2.2554764385076685e-06, + "loss": 0.7018, + "num_input_tokens_seen": 171894605, + "step": 5272 + }, + { + "epoch": 0.4755377192586914, + "flos": 35134927554720.0, + "grad_norm": 1.5183183001553224, + "learning_rate": 2.2548970028893348e-06, + "loss": 0.767, + "num_input_tokens_seen": 171925230, + "step": 5273 + }, + { + "epoch": 0.4756279027821617, + "flos": 34082039337600.0, + "grad_norm": 2.1206304841626955, + "learning_rate": 2.254317545521783e-06, + "loss": 0.7658, + "num_input_tokens_seen": 171952730, + "step": 5274 + }, + { + "epoch": 0.47571808630563195, + "flos": 20602162929600.0, + "grad_norm": 2.355367427437002, + "learning_rate": 2.253738066454457e-06, + "loss": 0.7855, + "num_input_tokens_seen": 171980090, + "step": 5275 + }, + { + "epoch": 0.47580826982910224, + "flos": 24536457859200.0, + "grad_norm": 1.6803774813771557, + "learning_rate": 2.2531585657367986e-06, + "loss": 0.7304, + "num_input_tokens_seen": 172008490, + "step": 5276 + }, + { + "epoch": 0.47589845335257247, + "flos": 25518414707040.0, + "grad_norm": 3.0721354035597415, + "learning_rate": 2.252579043418256e-06, + "loss": 0.7112, + "num_input_tokens_seen": 172038255, + "step": 5277 + }, + { + "epoch": 0.47598863687604276, + "flos": 21545569002720.0, + "grad_norm": 2.6036006973687633, + "learning_rate": 2.251999499548277e-06, + "loss": 0.7277, + "num_input_tokens_seen": 172066475, + "step": 5278 + }, + { + "epoch": 0.476078820399513, + "flos": 37797927525600.0, + "grad_norm": 1.9443594194926512, + "learning_rate": 2.251419934176311e-06, + "loss": 0.7167, + "num_input_tokens_seen": 172096630, + "step": 5279 + }, + { + "epoch": 0.4761690039229833, + "flos": 22237221325920.0, + "grad_norm": 1.4855048447269619, + "learning_rate": 2.25084034735181e-06, + "loss": 0.8036, + "num_input_tokens_seen": 172124370, + "step": 5280 + }, + { + "epoch": 0.4762591874464535, + "flos": 22783145255520.0, + "grad_norm": 1.7296471292114342, + "learning_rate": 2.2502607391242274e-06, + "loss": 0.8019, + "num_input_tokens_seen": 172154415, + "step": 5281 + }, + { + "epoch": 0.4763493709699238, + "flos": 57703191462240.0, + "grad_norm": 0.6596517045552386, + "learning_rate": 2.2496811095430182e-06, + "loss": 0.5815, + "num_input_tokens_seen": 172241715, + "step": 5282 + }, + { + "epoch": 0.47643955449339404, + "flos": 21400732683360.0, + "grad_norm": 2.385359006848389, + "learning_rate": 2.249101458657641e-06, + "loss": 0.7525, + "num_input_tokens_seen": 172268410, + "step": 5283 + }, + { + "epoch": 0.4765297380168643, + "flos": 18124297031520.0, + "grad_norm": 1.9650473517616636, + "learning_rate": 2.2485217865175526e-06, + "loss": 0.8239, + "num_input_tokens_seen": 172297040, + "step": 5284 + }, + { + "epoch": 0.47661992154033456, + "flos": 21585680907360.0, + "grad_norm": 2.5516471853411615, + "learning_rate": 2.2479420931722156e-06, + "loss": 0.8919, + "num_input_tokens_seen": 172325595, + "step": 5285 + }, + { + "epoch": 0.47671010506380485, + "flos": 64725963069120.0, + "grad_norm": 0.6980127771350337, + "learning_rate": 2.2473623786710923e-06, + "loss": 0.6186, + "num_input_tokens_seen": 172425125, + "step": 5286 + }, + { + "epoch": 0.4768002885872751, + "flos": 26650262961600.0, + "grad_norm": 1.9561233698898877, + "learning_rate": 2.2467826430636465e-06, + "loss": 0.67, + "num_input_tokens_seen": 172456115, + "step": 5287 + }, + { + "epoch": 0.47689047211074537, + "flos": 24493521052800.0, + "grad_norm": 1.4561103590128435, + "learning_rate": 2.246202886399345e-06, + "loss": 0.779, + "num_input_tokens_seen": 172486000, + "step": 5288 + }, + { + "epoch": 0.47698065563421566, + "flos": 22819428674880.0, + "grad_norm": 1.9770800851434114, + "learning_rate": 2.2456231087276556e-06, + "loss": 0.7505, + "num_input_tokens_seen": 172514390, + "step": 5289 + }, + { + "epoch": 0.4770708391576859, + "flos": 28583617380960.0, + "grad_norm": 1.5281278738919246, + "learning_rate": 2.245043310098048e-06, + "loss": 0.6834, + "num_input_tokens_seen": 172548390, + "step": 5290 + }, + { + "epoch": 0.4771610226811562, + "flos": 22387372920960.0, + "grad_norm": 1.9916247002983214, + "learning_rate": 2.244463490559995e-06, + "loss": 0.795, + "num_input_tokens_seen": 172574230, + "step": 5291 + }, + { + "epoch": 0.4772512062046264, + "flos": 21257866361280.0, + "grad_norm": 1.6188341685914496, + "learning_rate": 2.2438836501629683e-06, + "loss": 0.7657, + "num_input_tokens_seen": 172602440, + "step": 5292 + }, + { + "epoch": 0.4773413897280967, + "flos": 25410642372000.0, + "grad_norm": 1.4883456479054724, + "learning_rate": 2.2433037889564437e-06, + "loss": 0.7649, + "num_input_tokens_seen": 172635125, + "step": 5293 + }, + { + "epoch": 0.47743157325156693, + "flos": 22380756703680.0, + "grad_norm": 1.9927784361192882, + "learning_rate": 2.242723906989899e-06, + "loss": 0.6996, + "num_input_tokens_seen": 172667940, + "step": 5294 + }, + { + "epoch": 0.4775217567750372, + "flos": 27779472163200.0, + "grad_norm": 2.087637676335975, + "learning_rate": 2.2421440043128114e-06, + "loss": 0.6885, + "num_input_tokens_seen": 172699895, + "step": 5295 + }, + { + "epoch": 0.47761194029850745, + "flos": 13052504043840.0, + "grad_norm": 4.106501165456965, + "learning_rate": 2.241564080974662e-06, + "loss": 0.6711, + "num_input_tokens_seen": 172723370, + "step": 5296 + }, + { + "epoch": 0.47770212382197774, + "flos": 22606114190400.0, + "grad_norm": 1.9530010336635373, + "learning_rate": 2.2409841370249343e-06, + "loss": 0.7062, + "num_input_tokens_seen": 172751185, + "step": 5297 + }, + { + "epoch": 0.477792307345448, + "flos": 16339235719200.0, + "grad_norm": 1.9244816168811223, + "learning_rate": 2.2404041725131106e-06, + "loss": 0.8406, + "num_input_tokens_seen": 172776460, + "step": 5298 + }, + { + "epoch": 0.47788249086891826, + "flos": 21075557190240.0, + "grad_norm": 3.6448014170445475, + "learning_rate": 2.239824187488677e-06, + "loss": 0.7316, + "num_input_tokens_seen": 172802915, + "step": 5299 + }, + { + "epoch": 0.4779726743923885, + "flos": 64071857937120.0, + "grad_norm": 0.7239560371929686, + "learning_rate": 2.239244182001122e-06, + "loss": 0.6178, + "num_input_tokens_seen": 172902115, + "step": 5300 + }, + { + "epoch": 0.4780628579158588, + "flos": 28032229496640.0, + "grad_norm": 1.7560077812439518, + "learning_rate": 2.2386641560999336e-06, + "loss": 0.7714, + "num_input_tokens_seen": 172931815, + "step": 5301 + }, + { + "epoch": 0.478153041439329, + "flos": 22714964448480.0, + "grad_norm": 1.783441767857072, + "learning_rate": 2.238084109834604e-06, + "loss": 0.792, + "num_input_tokens_seen": 172959790, + "step": 5302 + }, + { + "epoch": 0.4782432249627993, + "flos": 22345253849280.0, + "grad_norm": 1.900465584479213, + "learning_rate": 2.237504043254625e-06, + "loss": 0.7596, + "num_input_tokens_seen": 172987035, + "step": 5303 + }, + { + "epoch": 0.47833340848626954, + "flos": 31746296366400.0, + "grad_norm": 1.454015889407602, + "learning_rate": 2.2369239564094915e-06, + "loss": 0.7274, + "num_input_tokens_seen": 173017595, + "step": 5304 + }, + { + "epoch": 0.47842359200973983, + "flos": 23329255033920.0, + "grad_norm": 1.6352505528668215, + "learning_rate": 2.2363438493486995e-06, + "loss": 0.6856, + "num_input_tokens_seen": 173047705, + "step": 5305 + }, + { + "epoch": 0.47851377553321006, + "flos": 23225757221280.0, + "grad_norm": 2.5010347132418653, + "learning_rate": 2.235763722121747e-06, + "loss": 0.6444, + "num_input_tokens_seen": 173076555, + "step": 5306 + }, + { + "epoch": 0.47860395905668035, + "flos": 27305854884000.0, + "grad_norm": 2.957028347401977, + "learning_rate": 2.2351835747781346e-06, + "loss": 0.7757, + "num_input_tokens_seen": 173106660, + "step": 5307 + }, + { + "epoch": 0.4786941425801506, + "flos": 23661269762880.0, + "grad_norm": 1.605409644308143, + "learning_rate": 2.234603407367362e-06, + "loss": 0.8001, + "num_input_tokens_seen": 173136135, + "step": 5308 + }, + { + "epoch": 0.47878432610362087, + "flos": 15824911819200.0, + "grad_norm": 1.7318995224790423, + "learning_rate": 2.2340232199389337e-06, + "loss": 0.7696, + "num_input_tokens_seen": 173161865, + "step": 5309 + }, + { + "epoch": 0.4788745096270911, + "flos": 24824457858720.0, + "grad_norm": 1.5927524810853122, + "learning_rate": 2.2334430125423538e-06, + "loss": 0.6379, + "num_input_tokens_seen": 173192150, + "step": 5310 + }, + { + "epoch": 0.4789646931505614, + "flos": 29448100586400.0, + "grad_norm": 1.858861323334212, + "learning_rate": 2.232862785227128e-06, + "loss": 0.804, + "num_input_tokens_seen": 173218910, + "step": 5311 + }, + { + "epoch": 0.4790548766740317, + "flos": 21111989288640.0, + "grad_norm": 1.6554106258829726, + "learning_rate": 2.232282538042766e-06, + "loss": 0.7915, + "num_input_tokens_seen": 173248245, + "step": 5312 + }, + { + "epoch": 0.4791450601975019, + "flos": 17533131770400.0, + "grad_norm": 2.1859116373447707, + "learning_rate": 2.231702271038777e-06, + "loss": 0.8045, + "num_input_tokens_seen": 173273150, + "step": 5313 + }, + { + "epoch": 0.4792352437209722, + "flos": 26794467395040.0, + "grad_norm": 2.0488210265233917, + "learning_rate": 2.231121984264673e-06, + "loss": 0.7296, + "num_input_tokens_seen": 173302510, + "step": 5314 + }, + { + "epoch": 0.47932542724444244, + "flos": 19977464811360.0, + "grad_norm": 1.694731371025253, + "learning_rate": 2.2305416777699665e-06, + "loss": 0.7872, + "num_input_tokens_seen": 173330775, + "step": 5315 + }, + { + "epoch": 0.4794156107679127, + "flos": 28981619901120.0, + "grad_norm": 1.574712406119703, + "learning_rate": 2.229961351604173e-06, + "loss": 0.755, + "num_input_tokens_seen": 173359490, + "step": 5316 + }, + { + "epoch": 0.47950579429138296, + "flos": 28068550085760.0, + "grad_norm": 1.6538611404545176, + "learning_rate": 2.2293810058168085e-06, + "loss": 0.714, + "num_input_tokens_seen": 173389725, + "step": 5317 + }, + { + "epoch": 0.47959597781485325, + "flos": 19726157098560.0, + "grad_norm": 2.2091653683933323, + "learning_rate": 2.2288006404573922e-06, + "loss": 0.7785, + "num_input_tokens_seen": 173416655, + "step": 5318 + }, + { + "epoch": 0.4796861613383235, + "flos": 20854362716640.0, + "grad_norm": 1.9553425521455516, + "learning_rate": 2.228220255575444e-06, + "loss": 0.7446, + "num_input_tokens_seen": 173444090, + "step": 5319 + }, + { + "epoch": 0.47977634486179377, + "flos": 25993853304480.0, + "grad_norm": 1.7488529669637707, + "learning_rate": 2.2276398512204847e-06, + "loss": 0.7013, + "num_input_tokens_seen": 173473140, + "step": 5320 + }, + { + "epoch": 0.479866528385264, + "flos": 22530908298720.0, + "grad_norm": 1.7969472455967208, + "learning_rate": 2.2270594274420382e-06, + "loss": 0.7632, + "num_input_tokens_seen": 173501620, + "step": 5321 + }, + { + "epoch": 0.4799567119087343, + "flos": 64154386271520.0, + "grad_norm": 0.6751350223635711, + "learning_rate": 2.22647898428963e-06, + "loss": 0.6075, + "num_input_tokens_seen": 173595145, + "step": 5322 + }, + { + "epoch": 0.4800468954322045, + "flos": 22275363233280.0, + "grad_norm": 1.6153613113287366, + "learning_rate": 2.225898521812785e-06, + "loss": 0.8007, + "num_input_tokens_seen": 173623170, + "step": 5323 + }, + { + "epoch": 0.4801370789556748, + "flos": 25812324698400.0, + "grad_norm": 1.7277156379471998, + "learning_rate": 2.2253180400610337e-06, + "loss": 0.8089, + "num_input_tokens_seen": 173651200, + "step": 5324 + }, + { + "epoch": 0.48022726247914505, + "flos": 25774591658400.0, + "grad_norm": 1.7894038245424053, + "learning_rate": 2.2247375390839037e-06, + "loss": 0.7373, + "num_input_tokens_seen": 173679485, + "step": 5325 + }, + { + "epoch": 0.48031744600261533, + "flos": 37397211612960.0, + "grad_norm": 1.737413374024377, + "learning_rate": 2.224157018930928e-06, + "loss": 0.7231, + "num_input_tokens_seen": 173707845, + "step": 5326 + }, + { + "epoch": 0.48040762952608557, + "flos": 24792002924640.0, + "grad_norm": 1.6028039354202246, + "learning_rate": 2.2235764796516395e-06, + "loss": 0.756, + "num_input_tokens_seen": 173737940, + "step": 5327 + }, + { + "epoch": 0.48049781304955586, + "flos": 25046247048480.0, + "grad_norm": 1.4861971462852237, + "learning_rate": 2.222995921295573e-06, + "loss": 0.753, + "num_input_tokens_seen": 173770155, + "step": 5328 + }, + { + "epoch": 0.4805879965730261, + "flos": 26174155308480.0, + "grad_norm": 1.7518846717483116, + "learning_rate": 2.222415343912265e-06, + "loss": 0.734, + "num_input_tokens_seen": 173798205, + "step": 5329 + }, + { + "epoch": 0.4806781800964964, + "flos": 14225542126080.0, + "grad_norm": 2.159300042934061, + "learning_rate": 2.221834747551254e-06, + "loss": 0.8192, + "num_input_tokens_seen": 173823705, + "step": 5330 + }, + { + "epoch": 0.4807683636199666, + "flos": 21658433594880.0, + "grad_norm": 2.347136936732531, + "learning_rate": 2.221254132262078e-06, + "loss": 0.7278, + "num_input_tokens_seen": 173851080, + "step": 5331 + }, + { + "epoch": 0.4808585471434369, + "flos": 28143793147200.0, + "grad_norm": 2.875568665365332, + "learning_rate": 2.2206734980942802e-06, + "loss": 0.8005, + "num_input_tokens_seen": 173879480, + "step": 5332 + }, + { + "epoch": 0.48094873066690713, + "flos": 38453259259680.0, + "grad_norm": 2.4555206264219094, + "learning_rate": 2.2200928450974024e-06, + "loss": 0.7342, + "num_input_tokens_seen": 173909015, + "step": 5333 + }, + { + "epoch": 0.4810389141903774, + "flos": 28576666635840.0, + "grad_norm": 1.7354092650085848, + "learning_rate": 2.21951217332099e-06, + "loss": 0.7779, + "num_input_tokens_seen": 173936020, + "step": 5334 + }, + { + "epoch": 0.48112909771384765, + "flos": 14439488496480.0, + "grad_norm": 1.9450639313141904, + "learning_rate": 2.2189314828145883e-06, + "loss": 0.6749, + "num_input_tokens_seen": 173960830, + "step": 5335 + }, + { + "epoch": 0.48121928123731794, + "flos": 27670361716800.0, + "grad_norm": 1.6252596721047439, + "learning_rate": 2.2183507736277453e-06, + "loss": 0.738, + "num_input_tokens_seen": 173992790, + "step": 5336 + }, + { + "epoch": 0.48130946476078823, + "flos": 24646237361280.0, + "grad_norm": 1.8980494601052895, + "learning_rate": 2.2177700458100107e-06, + "loss": 0.7729, + "num_input_tokens_seen": 174019440, + "step": 5337 + }, + { + "epoch": 0.48139964828425846, + "flos": 24241209756480.0, + "grad_norm": 1.7407894255471221, + "learning_rate": 2.2171892994109346e-06, + "loss": 0.7318, + "num_input_tokens_seen": 174047300, + "step": 5338 + }, + { + "epoch": 0.48148983180772875, + "flos": 27629394907680.0, + "grad_norm": 1.514239513651245, + "learning_rate": 2.21660853448007e-06, + "loss": 0.8213, + "num_input_tokens_seen": 174077520, + "step": 5339 + }, + { + "epoch": 0.481580015331199, + "flos": 20893694056320.0, + "grad_norm": 3.637714581534536, + "learning_rate": 2.2160277510669703e-06, + "loss": 0.7143, + "num_input_tokens_seen": 174106535, + "step": 5340 + }, + { + "epoch": 0.4816701988546693, + "flos": 30111237970080.0, + "grad_norm": 1.4543506826507941, + "learning_rate": 2.215446949221193e-06, + "loss": 0.7478, + "num_input_tokens_seen": 174139410, + "step": 5341 + }, + { + "epoch": 0.4817603823781395, + "flos": 21075222662400.0, + "grad_norm": 1.7707350385724012, + "learning_rate": 2.2148661289922924e-06, + "loss": 0.7661, + "num_input_tokens_seen": 174167040, + "step": 5342 + }, + { + "epoch": 0.4818505659016098, + "flos": 20091927703200.0, + "grad_norm": 1.7722323856175273, + "learning_rate": 2.21428529042983e-06, + "loss": 0.8191, + "num_input_tokens_seen": 174193560, + "step": 5343 + }, + { + "epoch": 0.48194074942508003, + "flos": 16776866937120.0, + "grad_norm": 1.782941450157902, + "learning_rate": 2.2137044335833647e-06, + "loss": 0.8035, + "num_input_tokens_seen": 174218275, + "step": 5344 + }, + { + "epoch": 0.4820309329485503, + "flos": 25629680999520.0, + "grad_norm": 2.2413830526912624, + "learning_rate": 2.213123558502459e-06, + "loss": 0.7411, + "num_input_tokens_seen": 174246110, + "step": 5345 + }, + { + "epoch": 0.48212111647202055, + "flos": 23731086039360.0, + "grad_norm": 1.8581023148205023, + "learning_rate": 2.2125426652366763e-06, + "loss": 0.7713, + "num_input_tokens_seen": 174274055, + "step": 5346 + }, + { + "epoch": 0.48221129999549084, + "flos": 23916889167840.0, + "grad_norm": 1.8103009399487482, + "learning_rate": 2.211961753835581e-06, + "loss": 0.8531, + "num_input_tokens_seen": 174300430, + "step": 5347 + }, + { + "epoch": 0.48230148351896107, + "flos": 23408140731840.0, + "grad_norm": 1.3525781508763959, + "learning_rate": 2.21138082434874e-06, + "loss": 0.7494, + "num_input_tokens_seen": 174332265, + "step": 5348 + }, + { + "epoch": 0.48239166704243136, + "flos": 20565842340480.0, + "grad_norm": 1.7280596287867789, + "learning_rate": 2.210799876825722e-06, + "loss": 0.7833, + "num_input_tokens_seen": 174356005, + "step": 5349 + }, + { + "epoch": 0.4824818505659016, + "flos": 18707582303520.0, + "grad_norm": 3.712805733470275, + "learning_rate": 2.210218911316096e-06, + "loss": 0.7159, + "num_input_tokens_seen": 174383280, + "step": 5350 + }, + { + "epoch": 0.4825720340893719, + "flos": 25410530862720.0, + "grad_norm": 1.9252867769920092, + "learning_rate": 2.2096379278694336e-06, + "loss": 0.7046, + "num_input_tokens_seen": 174412890, + "step": 5351 + }, + { + "epoch": 0.4826622176128421, + "flos": 23807332684320.0, + "grad_norm": 1.5984503304571582, + "learning_rate": 2.2090569265353074e-06, + "loss": 0.702, + "num_input_tokens_seen": 174441280, + "step": 5352 + }, + { + "epoch": 0.4827524011363124, + "flos": 20747705474400.0, + "grad_norm": 2.1843509350762207, + "learning_rate": 2.2084759073632912e-06, + "loss": 0.7534, + "num_input_tokens_seen": 174465410, + "step": 5353 + }, + { + "epoch": 0.48284258465978264, + "flos": 26067423726720.0, + "grad_norm": 1.7408876231219266, + "learning_rate": 2.2078948704029606e-06, + "loss": 0.7137, + "num_input_tokens_seen": 174495880, + "step": 5354 + }, + { + "epoch": 0.4829327681832529, + "flos": 36157516683840.0, + "grad_norm": 1.947062857735378, + "learning_rate": 2.2073138157038935e-06, + "loss": 0.679, + "num_input_tokens_seen": 174527515, + "step": 5355 + }, + { + "epoch": 0.48302295170672316, + "flos": 17214089287680.0, + "grad_norm": 1.68236877768582, + "learning_rate": 2.2067327433156687e-06, + "loss": 0.7422, + "num_input_tokens_seen": 174555265, + "step": 5356 + }, + { + "epoch": 0.48311313523019345, + "flos": 27302881303200.0, + "grad_norm": 1.5483252922536415, + "learning_rate": 2.2061516532878667e-06, + "loss": 0.7712, + "num_input_tokens_seen": 174585130, + "step": 5357 + }, + { + "epoch": 0.4832033187536637, + "flos": 17541643645440.0, + "grad_norm": 1.780589198632607, + "learning_rate": 2.2055705456700686e-06, + "loss": 0.8028, + "num_input_tokens_seen": 174613510, + "step": 5358 + }, + { + "epoch": 0.48329350227713397, + "flos": 26649779754720.0, + "grad_norm": 2.040878714756153, + "learning_rate": 2.204989420511858e-06, + "loss": 0.6539, + "num_input_tokens_seen": 174643920, + "step": 5359 + }, + { + "epoch": 0.48338368580060426, + "flos": 23517288348000.0, + "grad_norm": 1.50832591432605, + "learning_rate": 2.20440827786282e-06, + "loss": 0.7009, + "num_input_tokens_seen": 174674540, + "step": 5360 + }, + { + "epoch": 0.4834738693240745, + "flos": 23583350478720.0, + "grad_norm": 1.9605674850496317, + "learning_rate": 2.20382711777254e-06, + "loss": 0.76, + "num_input_tokens_seen": 174702390, + "step": 5361 + }, + { + "epoch": 0.4835640528475448, + "flos": 34404947475360.0, + "grad_norm": 1.552144256351917, + "learning_rate": 2.203245940290607e-06, + "loss": 0.7686, + "num_input_tokens_seen": 174733645, + "step": 5362 + }, + { + "epoch": 0.483654236371015, + "flos": 17505508905120.0, + "grad_norm": 2.4143738777302213, + "learning_rate": 2.2026647454666097e-06, + "loss": 0.7483, + "num_input_tokens_seen": 174761570, + "step": 5363 + }, + { + "epoch": 0.4837444198944853, + "flos": 17936561075520.0, + "grad_norm": 2.035331264592675, + "learning_rate": 2.2020835333501384e-06, + "loss": 0.7871, + "num_input_tokens_seen": 174785895, + "step": 5364 + }, + { + "epoch": 0.48383460341795553, + "flos": 23662645044000.0, + "grad_norm": 1.5611431333863373, + "learning_rate": 2.2015023039907863e-06, + "loss": 0.8133, + "num_input_tokens_seen": 174813850, + "step": 5365 + }, + { + "epoch": 0.4839247869414258, + "flos": 29564273287200.0, + "grad_norm": 1.5100517445976163, + "learning_rate": 2.2009210574381464e-06, + "loss": 0.7416, + "num_input_tokens_seen": 174845345, + "step": 5366 + }, + { + "epoch": 0.48401497046489605, + "flos": 26244900828960.0, + "grad_norm": 1.6398286612490294, + "learning_rate": 2.2003397937418134e-06, + "loss": 0.7011, + "num_input_tokens_seen": 174873715, + "step": 5367 + }, + { + "epoch": 0.48410515398836634, + "flos": 62255270934720.0, + "grad_norm": 0.6707057315195899, + "learning_rate": 2.1997585129513852e-06, + "loss": 0.6021, + "num_input_tokens_seen": 174967045, + "step": 5368 + }, + { + "epoch": 0.4841953375118366, + "flos": 24792188773440.0, + "grad_norm": 1.6411984122798342, + "learning_rate": 2.1991772151164595e-06, + "loss": 0.7815, + "num_input_tokens_seen": 174997705, + "step": 5369 + }, + { + "epoch": 0.48428552103530687, + "flos": 22891326457920.0, + "grad_norm": 1.431854356994085, + "learning_rate": 2.1985959002866346e-06, + "loss": 0.7997, + "num_input_tokens_seen": 175029565, + "step": 5370 + }, + { + "epoch": 0.4843757045587771, + "flos": 16921145710080.0, + "grad_norm": 2.841330033067529, + "learning_rate": 2.198014568511513e-06, + "loss": 0.8196, + "num_input_tokens_seen": 175051825, + "step": 5371 + }, + { + "epoch": 0.4844658880822474, + "flos": 67027504127520.0, + "grad_norm": 0.6633462296682746, + "learning_rate": 2.1974332198406965e-06, + "loss": 0.5747, + "num_input_tokens_seen": 175143690, + "step": 5372 + }, + { + "epoch": 0.4845560716057176, + "flos": 26610188226720.0, + "grad_norm": 1.6691902922800745, + "learning_rate": 2.196851854323789e-06, + "loss": 0.7384, + "num_input_tokens_seen": 175173150, + "step": 5373 + }, + { + "epoch": 0.4846462551291879, + "flos": 37907409669600.0, + "grad_norm": 1.8603488071050789, + "learning_rate": 2.196270472010396e-06, + "loss": 0.7385, + "num_input_tokens_seen": 175203325, + "step": 5374 + }, + { + "epoch": 0.48473643865265814, + "flos": 23333901253920.0, + "grad_norm": 1.7248001425139698, + "learning_rate": 2.195689072950124e-06, + "loss": 0.7501, + "num_input_tokens_seen": 175232935, + "step": 5375 + }, + { + "epoch": 0.48482662217612843, + "flos": 68779775977920.0, + "grad_norm": 0.6099664556551121, + "learning_rate": 2.195107657192581e-06, + "loss": 0.5621, + "num_input_tokens_seen": 175334085, + "step": 5376 + }, + { + "epoch": 0.48491680569959866, + "flos": 68732490309600.0, + "grad_norm": 0.5865889170677573, + "learning_rate": 2.194526224787378e-06, + "loss": 0.5164, + "num_input_tokens_seen": 175429845, + "step": 5377 + }, + { + "epoch": 0.48500698922306895, + "flos": 50730876420480.0, + "grad_norm": 1.6500356439150596, + "learning_rate": 2.1939447757841236e-06, + "loss": 0.7228, + "num_input_tokens_seen": 175462835, + "step": 5378 + }, + { + "epoch": 0.4850971727465392, + "flos": 29200100982240.0, + "grad_norm": 1.864525465537922, + "learning_rate": 2.193363310232432e-06, + "loss": 0.8148, + "num_input_tokens_seen": 175489470, + "step": 5379 + }, + { + "epoch": 0.4851873562700095, + "flos": 19431280693440.0, + "grad_norm": 1.3560435810428215, + "learning_rate": 2.192781828181917e-06, + "loss": 0.854, + "num_input_tokens_seen": 175518285, + "step": 5380 + }, + { + "epoch": 0.4852775397934797, + "flos": 27814603320000.0, + "grad_norm": 1.3644274508520455, + "learning_rate": 2.192200329682193e-06, + "loss": 0.777, + "num_input_tokens_seen": 175550090, + "step": 5381 + }, + { + "epoch": 0.48536772331695, + "flos": 21547501830240.0, + "grad_norm": 1.3936052426829744, + "learning_rate": 2.1916188147828767e-06, + "loss": 0.7517, + "num_input_tokens_seen": 175579580, + "step": 5382 + }, + { + "epoch": 0.48545790684042023, + "flos": 19727160682080.0, + "grad_norm": 1.9104884820195067, + "learning_rate": 2.191037283533587e-06, + "loss": 0.7696, + "num_input_tokens_seen": 175607695, + "step": 5383 + }, + { + "epoch": 0.4855480903638905, + "flos": 20383124302080.0, + "grad_norm": 2.4433458115942983, + "learning_rate": 2.1904557359839428e-06, + "loss": 0.7515, + "num_input_tokens_seen": 175633245, + "step": 5384 + }, + { + "epoch": 0.4856382738873608, + "flos": 19071903287520.0, + "grad_norm": 1.8470660632039124, + "learning_rate": 2.189874172183565e-06, + "loss": 0.7317, + "num_input_tokens_seen": 175660630, + "step": 5385 + }, + { + "epoch": 0.48572845741083104, + "flos": 23335313704800.0, + "grad_norm": 1.6293248066489148, + "learning_rate": 2.1892925921820763e-06, + "loss": 0.7255, + "num_input_tokens_seen": 175690475, + "step": 5386 + }, + { + "epoch": 0.4858186409343013, + "flos": 14736855275520.0, + "grad_norm": 1.9768200129852525, + "learning_rate": 2.1887109960290994e-06, + "loss": 0.7495, + "num_input_tokens_seen": 175717355, + "step": 5387 + }, + { + "epoch": 0.48590882445777156, + "flos": 18671150205120.0, + "grad_norm": 1.6717971355809356, + "learning_rate": 2.18812938377426e-06, + "loss": 0.8503, + "num_input_tokens_seen": 175744810, + "step": 5388 + }, + { + "epoch": 0.48599900798124185, + "flos": 23516730801600.0, + "grad_norm": 1.6346262663914886, + "learning_rate": 2.187547755467184e-06, + "loss": 0.7928, + "num_input_tokens_seen": 175772330, + "step": 5389 + }, + { + "epoch": 0.4860891915047121, + "flos": 34810829984640.0, + "grad_norm": 1.537348300924388, + "learning_rate": 2.1869661111574994e-06, + "loss": 0.6467, + "num_input_tokens_seen": 175802470, + "step": 5390 + }, + { + "epoch": 0.48617937502818237, + "flos": 25848236420160.0, + "grad_norm": 2.0312116179141126, + "learning_rate": 2.1863844508948353e-06, + "loss": 0.8099, + "num_input_tokens_seen": 175830610, + "step": 5391 + }, + { + "epoch": 0.4862695585516526, + "flos": 21869629403040.0, + "grad_norm": 1.699291920456743, + "learning_rate": 2.185802774728823e-06, + "loss": 0.769, + "num_input_tokens_seen": 175857205, + "step": 5392 + }, + { + "epoch": 0.4863597420751229, + "flos": 20528704016640.0, + "grad_norm": 1.5302213210802702, + "learning_rate": 2.1852210827090927e-06, + "loss": 0.7091, + "num_input_tokens_seen": 175885975, + "step": 5393 + }, + { + "epoch": 0.4864499255985931, + "flos": 24427830619680.0, + "grad_norm": 1.6774509887812472, + "learning_rate": 2.184639374885278e-06, + "loss": 0.6977, + "num_input_tokens_seen": 175914765, + "step": 5394 + }, + { + "epoch": 0.4865401091220634, + "flos": 20274385553280.0, + "grad_norm": 2.261310336276444, + "learning_rate": 2.184057651307014e-06, + "loss": 0.8461, + "num_input_tokens_seen": 175942255, + "step": 5395 + }, + { + "epoch": 0.48663029264553365, + "flos": 20820086464320.0, + "grad_norm": 3.3549930825993592, + "learning_rate": 2.183475912023937e-06, + "loss": 0.8217, + "num_input_tokens_seen": 175964860, + "step": 5396 + }, + { + "epoch": 0.48672047616900393, + "flos": 19502918288160.0, + "grad_norm": 1.7385179679936071, + "learning_rate": 2.1828941570856826e-06, + "loss": 0.7681, + "num_input_tokens_seen": 175992110, + "step": 5397 + }, + { + "epoch": 0.48681065969247417, + "flos": 68138202788640.0, + "grad_norm": 0.6352070608613191, + "learning_rate": 2.1823123865418903e-06, + "loss": 0.5718, + "num_input_tokens_seen": 176083760, + "step": 5398 + }, + { + "epoch": 0.48690084321594446, + "flos": 24499988591040.0, + "grad_norm": 2.4516362817965955, + "learning_rate": 2.1817306004422e-06, + "loss": 0.7314, + "num_input_tokens_seen": 176113315, + "step": 5399 + }, + { + "epoch": 0.4869910267394147, + "flos": 29923055976960.0, + "grad_norm": 1.7152949893034697, + "learning_rate": 2.1811487988362527e-06, + "loss": 0.7767, + "num_input_tokens_seen": 176141810, + "step": 5400 + }, + { + "epoch": 0.487081210262885, + "flos": 24645531135840.0, + "grad_norm": 2.1430971669692287, + "learning_rate": 2.1805669817736917e-06, + "loss": 0.8256, + "num_input_tokens_seen": 176166965, + "step": 5401 + }, + { + "epoch": 0.4871713937863552, + "flos": 23881126125120.0, + "grad_norm": 7.015905293937901, + "learning_rate": 2.17998514930416e-06, + "loss": 0.5797, + "num_input_tokens_seen": 176197245, + "step": 5402 + }, + { + "epoch": 0.4872615773098255, + "flos": 27087410972640.0, + "grad_norm": 2.079286867182496, + "learning_rate": 2.1794033014773025e-06, + "loss": 0.7513, + "num_input_tokens_seen": 176224290, + "step": 5403 + }, + { + "epoch": 0.48735176083329573, + "flos": 21622410363840.0, + "grad_norm": 1.7152008866524895, + "learning_rate": 2.178821438342766e-06, + "loss": 0.8215, + "num_input_tokens_seen": 176252075, + "step": 5404 + }, + { + "epoch": 0.487441944356766, + "flos": 26794393055520.0, + "grad_norm": 1.9693755109829303, + "learning_rate": 2.1782395599501996e-06, + "loss": 0.7491, + "num_input_tokens_seen": 176283200, + "step": 5405 + }, + { + "epoch": 0.48753212788023625, + "flos": 70364909652960.0, + "grad_norm": 0.648771108272896, + "learning_rate": 2.1776576663492498e-06, + "loss": 0.5841, + "num_input_tokens_seen": 176374880, + "step": 5406 + }, + { + "epoch": 0.48762231140370654, + "flos": 24057079267200.0, + "grad_norm": 2.2551417410682526, + "learning_rate": 2.177075757589569e-06, + "loss": 0.8423, + "num_input_tokens_seen": 176402115, + "step": 5407 + }, + { + "epoch": 0.48771249492717683, + "flos": 24680104746240.0, + "grad_norm": 1.5891822732293444, + "learning_rate": 2.176493833720808e-06, + "loss": 0.6782, + "num_input_tokens_seen": 176432690, + "step": 5408 + }, + { + "epoch": 0.48780267845064706, + "flos": 25083162353760.0, + "grad_norm": 3.3599066734976817, + "learning_rate": 2.1759118947926195e-06, + "loss": 0.7227, + "num_input_tokens_seen": 176461820, + "step": 5409 + }, + { + "epoch": 0.48789286197411735, + "flos": 32295788592960.0, + "grad_norm": 1.4584653612953125, + "learning_rate": 2.1753299408546587e-06, + "loss": 0.763, + "num_input_tokens_seen": 176493555, + "step": 5410 + }, + { + "epoch": 0.4879830454975876, + "flos": 33750136117920.0, + "grad_norm": 1.3788534267779267, + "learning_rate": 2.1747479719565803e-06, + "loss": 0.7591, + "num_input_tokens_seen": 176527020, + "step": 5411 + }, + { + "epoch": 0.4880732290210579, + "flos": 26176013796480.0, + "grad_norm": 1.3896701406257184, + "learning_rate": 2.174165988148042e-06, + "loss": 0.8171, + "num_input_tokens_seen": 176558135, + "step": 5412 + }, + { + "epoch": 0.4881634125445281, + "flos": 22565110211520.0, + "grad_norm": 1.5817847542584045, + "learning_rate": 2.1735839894787003e-06, + "loss": 0.7253, + "num_input_tokens_seen": 176584760, + "step": 5413 + }, + { + "epoch": 0.4882535960679984, + "flos": 64627520343840.0, + "grad_norm": 0.8039442166481799, + "learning_rate": 2.1730019759982163e-06, + "loss": 0.6286, + "num_input_tokens_seen": 176674455, + "step": 5414 + }, + { + "epoch": 0.48834377959146863, + "flos": 20565953849760.0, + "grad_norm": 1.9931006638771802, + "learning_rate": 2.172419947756249e-06, + "loss": 0.8084, + "num_input_tokens_seen": 176702365, + "step": 5415 + }, + { + "epoch": 0.4884339631149389, + "flos": 19690840092960.0, + "grad_norm": 2.045471175925725, + "learning_rate": 2.171837904802461e-06, + "loss": 0.8012, + "num_input_tokens_seen": 176729615, + "step": 5416 + }, + { + "epoch": 0.48852414663840915, + "flos": 27445821964800.0, + "grad_norm": 1.4727966889217274, + "learning_rate": 2.171255847186516e-06, + "loss": 0.7538, + "num_input_tokens_seen": 176760685, + "step": 5417 + }, + { + "epoch": 0.48861433016187944, + "flos": 18452780633280.0, + "grad_norm": 2.591569466521144, + "learning_rate": 2.1706737749580783e-06, + "loss": 0.7416, + "num_input_tokens_seen": 176786635, + "step": 5418 + }, + { + "epoch": 0.4887045136853497, + "flos": 22751545225920.0, + "grad_norm": 1.3786911837222822, + "learning_rate": 2.1700916881668127e-06, + "loss": 0.8283, + "num_input_tokens_seen": 176816840, + "step": 5419 + }, + { + "epoch": 0.48879469720881996, + "flos": 20601568213440.0, + "grad_norm": 1.733018155358907, + "learning_rate": 2.1695095868623862e-06, + "loss": 0.8457, + "num_input_tokens_seen": 176845070, + "step": 5420 + }, + { + "epoch": 0.4888848807322902, + "flos": 34992878967360.0, + "grad_norm": 1.6674349285737755, + "learning_rate": 2.168927471094467e-06, + "loss": 0.6877, + "num_input_tokens_seen": 176877590, + "step": 5421 + }, + { + "epoch": 0.4889750642557605, + "flos": 21184667636640.0, + "grad_norm": 2.3035457563184276, + "learning_rate": 2.168345340912725e-06, + "loss": 0.7759, + "num_input_tokens_seen": 176904090, + "step": 5422 + }, + { + "epoch": 0.4890652477792307, + "flos": 23806254761280.0, + "grad_norm": 1.5391064942554105, + "learning_rate": 2.1677631963668298e-06, + "loss": 0.7727, + "num_input_tokens_seen": 176936260, + "step": 5423 + }, + { + "epoch": 0.489155431302701, + "flos": 20492569276320.0, + "grad_norm": 3.273372676817513, + "learning_rate": 2.167181037506453e-06, + "loss": 0.7151, + "num_input_tokens_seen": 176963040, + "step": 5424 + }, + { + "epoch": 0.48924561482617124, + "flos": 22788163173120.0, + "grad_norm": 1.67437770727949, + "learning_rate": 2.1665988643812693e-06, + "loss": 0.6918, + "num_input_tokens_seen": 176990880, + "step": 5425 + }, + { + "epoch": 0.4893357983496415, + "flos": 26503605324000.0, + "grad_norm": 3.7608372813217525, + "learning_rate": 2.166016677040951e-06, + "loss": 0.7531, + "num_input_tokens_seen": 177019820, + "step": 5426 + }, + { + "epoch": 0.48942598187311176, + "flos": 27123285524640.0, + "grad_norm": 1.9135032615344758, + "learning_rate": 2.165434475535175e-06, + "loss": 0.7428, + "num_input_tokens_seen": 177047510, + "step": 5427 + }, + { + "epoch": 0.48951616539658205, + "flos": 22489383943200.0, + "grad_norm": 1.69906601493434, + "learning_rate": 2.1648522599136173e-06, + "loss": 0.7948, + "num_input_tokens_seen": 177077170, + "step": 5428 + }, + { + "epoch": 0.4896063489200523, + "flos": 25992180665280.0, + "grad_norm": 1.6150454280357383, + "learning_rate": 2.164270030225956e-06, + "loss": 0.7271, + "num_input_tokens_seen": 177107915, + "step": 5429 + }, + { + "epoch": 0.48969653244352257, + "flos": 22313170612800.0, + "grad_norm": 2.558827629756673, + "learning_rate": 2.16368778652187e-06, + "loss": 0.6675, + "num_input_tokens_seen": 177137215, + "step": 5430 + }, + { + "epoch": 0.4897867159669928, + "flos": 24828100495200.0, + "grad_norm": 1.7671818578817218, + "learning_rate": 2.163105528851039e-06, + "loss": 0.7929, + "num_input_tokens_seen": 177166590, + "step": 5431 + }, + { + "epoch": 0.4898768994904631, + "flos": 24790962171360.0, + "grad_norm": 1.656228407233214, + "learning_rate": 2.1625232572631448e-06, + "loss": 0.7603, + "num_input_tokens_seen": 177197730, + "step": 5432 + }, + { + "epoch": 0.4899670830139334, + "flos": 70840496929440.0, + "grad_norm": 0.7307862188607892, + "learning_rate": 2.161940971807871e-06, + "loss": 0.5901, + "num_input_tokens_seen": 177296250, + "step": 5433 + }, + { + "epoch": 0.4900572665374036, + "flos": 25301903623200.0, + "grad_norm": 2.318333870700067, + "learning_rate": 2.1613586725348994e-06, + "loss": 0.7398, + "num_input_tokens_seen": 177323770, + "step": 5434 + }, + { + "epoch": 0.4901474500608739, + "flos": 23581752179040.0, + "grad_norm": 1.5382438844992894, + "learning_rate": 2.1607763594939176e-06, + "loss": 0.8317, + "num_input_tokens_seen": 177351335, + "step": 5435 + }, + { + "epoch": 0.49023763358434413, + "flos": 31605214192800.0, + "grad_norm": 1.9271477422174623, + "learning_rate": 2.1601940327346093e-06, + "loss": 0.6682, + "num_input_tokens_seen": 177383955, + "step": 5436 + }, + { + "epoch": 0.4903278171078144, + "flos": 26032143890880.0, + "grad_norm": 1.4876496984984418, + "learning_rate": 2.159611692306663e-06, + "loss": 0.8286, + "num_input_tokens_seen": 177415750, + "step": 5437 + }, + { + "epoch": 0.49041800063128466, + "flos": 29527804019040.0, + "grad_norm": 1.9196230092083966, + "learning_rate": 2.1590293382597667e-06, + "loss": 0.7209, + "num_input_tokens_seen": 177445650, + "step": 5438 + }, + { + "epoch": 0.49050818415475494, + "flos": 26322336906240.0, + "grad_norm": 1.650461048677767, + "learning_rate": 2.1584469706436102e-06, + "loss": 0.6956, + "num_input_tokens_seen": 177475670, + "step": 5439 + }, + { + "epoch": 0.4905983676782252, + "flos": 23589371979840.0, + "grad_norm": 2.0239167695253233, + "learning_rate": 2.1578645895078855e-06, + "loss": 0.6738, + "num_input_tokens_seen": 177505220, + "step": 5440 + }, + { + "epoch": 0.49068855120169547, + "flos": 26613421995840.0, + "grad_norm": 1.7054249344150092, + "learning_rate": 2.157282194902283e-06, + "loss": 0.8126, + "num_input_tokens_seen": 177532000, + "step": 5441 + }, + { + "epoch": 0.4907787347251657, + "flos": 12550780692480.0, + "grad_norm": 1.852824943656396, + "learning_rate": 2.1566997868764965e-06, + "loss": 0.6984, + "num_input_tokens_seen": 177559220, + "step": 5442 + }, + { + "epoch": 0.490868918248636, + "flos": 18962272464480.0, + "grad_norm": 3.2538413308068503, + "learning_rate": 2.15611736548022e-06, + "loss": 0.7149, + "num_input_tokens_seen": 177584350, + "step": 5443 + }, + { + "epoch": 0.4909591017721062, + "flos": 20310222935520.0, + "grad_norm": 2.01488286039948, + "learning_rate": 2.155534930763149e-06, + "loss": 0.7608, + "num_input_tokens_seen": 177611760, + "step": 5444 + }, + { + "epoch": 0.4910492852955765, + "flos": 19836977353920.0, + "grad_norm": 1.711963781232165, + "learning_rate": 2.1549524827749804e-06, + "loss": 0.6363, + "num_input_tokens_seen": 177638210, + "step": 5445 + }, + { + "epoch": 0.49113946881904674, + "flos": 41185369281600.0, + "grad_norm": 2.747171681497105, + "learning_rate": 2.1543700215654115e-06, + "loss": 0.6051, + "num_input_tokens_seen": 177671850, + "step": 5446 + }, + { + "epoch": 0.49122965234251703, + "flos": 22205361108000.0, + "grad_norm": 1.6926226393698185, + "learning_rate": 2.153787547184141e-06, + "loss": 0.7454, + "num_input_tokens_seen": 177700650, + "step": 5447 + }, + { + "epoch": 0.49131983586598726, + "flos": 36085916258880.0, + "grad_norm": 1.984678530569926, + "learning_rate": 2.1532050596808695e-06, + "loss": 0.683, + "num_input_tokens_seen": 177732320, + "step": 5448 + }, + { + "epoch": 0.49141001938945755, + "flos": 24937694148480.0, + "grad_norm": 1.7783754142256125, + "learning_rate": 2.152622559105297e-06, + "loss": 0.8013, + "num_input_tokens_seen": 177761725, + "step": 5449 + }, + { + "epoch": 0.4915002029129278, + "flos": 22824037725120.0, + "grad_norm": 2.118940127683997, + "learning_rate": 2.152040045507126e-06, + "loss": 0.7966, + "num_input_tokens_seen": 177788595, + "step": 5450 + }, + { + "epoch": 0.4915903864363981, + "flos": 21257866361280.0, + "grad_norm": 1.8334005155680604, + "learning_rate": 2.1514575189360607e-06, + "loss": 0.7677, + "num_input_tokens_seen": 177816120, + "step": 5451 + }, + { + "epoch": 0.4916805699598683, + "flos": 17614396332960.0, + "grad_norm": 1.6095881644991226, + "learning_rate": 2.1508749794418043e-06, + "loss": 0.7113, + "num_input_tokens_seen": 177844445, + "step": 5452 + }, + { + "epoch": 0.4917707534833386, + "flos": 26282076322560.0, + "grad_norm": 1.8201722185166476, + "learning_rate": 2.1502924270740626e-06, + "loss": 0.725, + "num_input_tokens_seen": 177873370, + "step": 5453 + }, + { + "epoch": 0.49186093700680883, + "flos": 20450413034880.0, + "grad_norm": 2.2866473166904715, + "learning_rate": 2.1497098618825427e-06, + "loss": 0.7617, + "num_input_tokens_seen": 177900795, + "step": 5454 + }, + { + "epoch": 0.4919511205302791, + "flos": 22350903652800.0, + "grad_norm": 3.252700124179811, + "learning_rate": 2.1491272839169516e-06, + "loss": 0.7945, + "num_input_tokens_seen": 177928945, + "step": 5455 + }, + { + "epoch": 0.4920413040537494, + "flos": 21104369487840.0, + "grad_norm": 1.6585972545584957, + "learning_rate": 2.1485446932269986e-06, + "loss": 0.8646, + "num_input_tokens_seen": 177957085, + "step": 5456 + }, + { + "epoch": 0.49213148757721964, + "flos": 22675410090240.0, + "grad_norm": 1.5839792573707403, + "learning_rate": 2.147962089862393e-06, + "loss": 0.7909, + "num_input_tokens_seen": 177987380, + "step": 5457 + }, + { + "epoch": 0.49222167110068993, + "flos": 23370110333760.0, + "grad_norm": 2.3935780954014017, + "learning_rate": 2.1473794738728462e-06, + "loss": 0.6628, + "num_input_tokens_seen": 178015115, + "step": 5458 + }, + { + "epoch": 0.49231185462416016, + "flos": 18452631954240.0, + "grad_norm": 1.7796047040721537, + "learning_rate": 2.14679684530807e-06, + "loss": 0.7887, + "num_input_tokens_seen": 178040150, + "step": 5459 + }, + { + "epoch": 0.49240203814763045, + "flos": 25262312095200.0, + "grad_norm": 1.8514144130027497, + "learning_rate": 2.1462142042177774e-06, + "loss": 0.7364, + "num_input_tokens_seen": 178070100, + "step": 5460 + }, + { + "epoch": 0.4924922216711007, + "flos": 22423842189120.0, + "grad_norm": 1.6540217255222966, + "learning_rate": 2.145631550651683e-06, + "loss": 0.7467, + "num_input_tokens_seen": 178098305, + "step": 5461 + }, + { + "epoch": 0.49258240519457097, + "flos": 21112100797920.0, + "grad_norm": 1.564449924651455, + "learning_rate": 2.1450488846595016e-06, + "loss": 0.8562, + "num_input_tokens_seen": 178126050, + "step": 5462 + }, + { + "epoch": 0.4926725887180412, + "flos": 17100852997920.0, + "grad_norm": 2.035258753679121, + "learning_rate": 2.14446620629095e-06, + "loss": 0.732, + "num_input_tokens_seen": 178152790, + "step": 5463 + }, + { + "epoch": 0.4927627722415115, + "flos": 24388870977600.0, + "grad_norm": 1.6690355235228302, + "learning_rate": 2.1438835155957445e-06, + "loss": 0.8187, + "num_input_tokens_seen": 178182430, + "step": 5464 + }, + { + "epoch": 0.4928529557649817, + "flos": 23734505657280.0, + "grad_norm": 2.5218160516025185, + "learning_rate": 2.143300812623604e-06, + "loss": 0.7945, + "num_input_tokens_seen": 178210295, + "step": 5465 + }, + { + "epoch": 0.492943139288452, + "flos": 21513114068640.0, + "grad_norm": 1.7133011885916585, + "learning_rate": 2.1427180974242485e-06, + "loss": 0.7373, + "num_input_tokens_seen": 178237640, + "step": 5466 + }, + { + "epoch": 0.49303332281192225, + "flos": 21725759497440.0, + "grad_norm": 1.847976818059041, + "learning_rate": 2.142135370047398e-06, + "loss": 0.7811, + "num_input_tokens_seen": 178264835, + "step": 5467 + }, + { + "epoch": 0.49312350633539254, + "flos": 17577889895040.0, + "grad_norm": 1.9161460279910656, + "learning_rate": 2.1415526305427735e-06, + "loss": 0.7863, + "num_input_tokens_seen": 178289040, + "step": 5468 + }, + { + "epoch": 0.49321368985886277, + "flos": 30182912734560.0, + "grad_norm": 1.4179610154962954, + "learning_rate": 2.140969878960098e-06, + "loss": 0.6781, + "num_input_tokens_seen": 178319185, + "step": 5469 + }, + { + "epoch": 0.49330387338233306, + "flos": 26318099553600.0, + "grad_norm": 1.5589505124315433, + "learning_rate": 2.1403871153490956e-06, + "loss": 0.8103, + "num_input_tokens_seen": 178348615, + "step": 5470 + }, + { + "epoch": 0.4933940569058033, + "flos": 20527031377440.0, + "grad_norm": 1.5648212737870368, + "learning_rate": 2.13980433975949e-06, + "loss": 0.8747, + "num_input_tokens_seen": 178373110, + "step": 5471 + }, + { + "epoch": 0.4934842404292736, + "flos": 24935761320960.0, + "grad_norm": 1.937345129406839, + "learning_rate": 2.1392215522410076e-06, + "loss": 0.7979, + "num_input_tokens_seen": 178404085, + "step": 5472 + }, + { + "epoch": 0.4935744239527438, + "flos": 20165052088320.0, + "grad_norm": 2.2378625927589533, + "learning_rate": 2.1386387528433743e-06, + "loss": 0.8671, + "num_input_tokens_seen": 178430335, + "step": 5473 + }, + { + "epoch": 0.4936646074762141, + "flos": 20383310150880.0, + "grad_norm": 1.8827787530810913, + "learning_rate": 2.1380559416163186e-06, + "loss": 0.781, + "num_input_tokens_seen": 178456965, + "step": 5474 + }, + { + "epoch": 0.49375479099968433, + "flos": 17431864143360.0, + "grad_norm": 1.7808261835611945, + "learning_rate": 2.1374731186095685e-06, + "loss": 0.8, + "num_input_tokens_seen": 178484990, + "step": 5475 + }, + { + "epoch": 0.4938449745231546, + "flos": 21360397760160.0, + "grad_norm": 1.585232507517435, + "learning_rate": 2.136890283872854e-06, + "loss": 0.7084, + "num_input_tokens_seen": 178512085, + "step": 5476 + }, + { + "epoch": 0.49393515804662486, + "flos": 24567277323840.0, + "grad_norm": 2.5361922253100504, + "learning_rate": 2.136307437455906e-06, + "loss": 0.7359, + "num_input_tokens_seen": 178540495, + "step": 5477 + }, + { + "epoch": 0.49402534157009514, + "flos": 25190228463360.0, + "grad_norm": 1.4893304125490523, + "learning_rate": 2.135724579408456e-06, + "loss": 0.7019, + "num_input_tokens_seen": 178570115, + "step": 5478 + }, + { + "epoch": 0.49411552509356543, + "flos": 18926286403200.0, + "grad_norm": 1.8317612569032762, + "learning_rate": 2.1351417097802356e-06, + "loss": 0.8474, + "num_input_tokens_seen": 178598075, + "step": 5479 + }, + { + "epoch": 0.49420570861703567, + "flos": 23698445256480.0, + "grad_norm": 2.0307710523085882, + "learning_rate": 2.1345588286209798e-06, + "loss": 0.6382, + "num_input_tokens_seen": 178624545, + "step": 5480 + }, + { + "epoch": 0.49429589214050595, + "flos": 36303839793600.0, + "grad_norm": 2.009850443689718, + "learning_rate": 2.1339759359804227e-06, + "loss": 0.6036, + "num_input_tokens_seen": 178653425, + "step": 5481 + }, + { + "epoch": 0.4943860756639762, + "flos": 34407103321440.0, + "grad_norm": 1.561851519484101, + "learning_rate": 2.1333930319082997e-06, + "loss": 0.7277, + "num_input_tokens_seen": 178683720, + "step": 5482 + }, + { + "epoch": 0.4944762591874465, + "flos": 25845485857920.0, + "grad_norm": 1.340458987256387, + "learning_rate": 2.132810116454348e-06, + "loss": 0.7574, + "num_input_tokens_seen": 178713595, + "step": 5483 + }, + { + "epoch": 0.4945664427109167, + "flos": 25665518381760.0, + "grad_norm": 2.251844391923161, + "learning_rate": 2.132227189668305e-06, + "loss": 0.7502, + "num_input_tokens_seen": 178742570, + "step": 5484 + }, + { + "epoch": 0.494656626234387, + "flos": 32402891872320.0, + "grad_norm": 1.7276229708635986, + "learning_rate": 2.1316442515999096e-06, + "loss": 0.6984, + "num_input_tokens_seen": 178772505, + "step": 5485 + }, + { + "epoch": 0.49474680975785723, + "flos": 28252346047200.0, + "grad_norm": 1.4577901464385972, + "learning_rate": 2.1310613022989e-06, + "loss": 0.7848, + "num_input_tokens_seen": 178803605, + "step": 5486 + }, + { + "epoch": 0.4948369932813275, + "flos": 33353062841760.0, + "grad_norm": 2.079146858177034, + "learning_rate": 2.130478341815017e-06, + "loss": 0.6043, + "num_input_tokens_seen": 178834455, + "step": 5487 + }, + { + "epoch": 0.49492717680479775, + "flos": 25153387497600.0, + "grad_norm": 1.6600134215903888, + "learning_rate": 2.1298953701980033e-06, + "loss": 0.6478, + "num_input_tokens_seen": 178864715, + "step": 5488 + }, + { + "epoch": 0.49501736032826804, + "flos": 38744455858560.0, + "grad_norm": 2.1008313123044386, + "learning_rate": 2.1293123874976003e-06, + "loss": 0.7613, + "num_input_tokens_seen": 178892105, + "step": 5489 + }, + { + "epoch": 0.4951075438517383, + "flos": 24682037573760.0, + "grad_norm": 1.7159711394784083, + "learning_rate": 2.1287293937635513e-06, + "loss": 0.7018, + "num_input_tokens_seen": 178920435, + "step": 5490 + }, + { + "epoch": 0.49519772737520856, + "flos": 26468288318400.0, + "grad_norm": 1.7403434265424655, + "learning_rate": 2.1281463890456005e-06, + "loss": 0.8052, + "num_input_tokens_seen": 178949765, + "step": 5491 + }, + { + "epoch": 0.4952879108986788, + "flos": 19540651328160.0, + "grad_norm": 1.7586366882911912, + "learning_rate": 2.127563373393493e-06, + "loss": 0.7474, + "num_input_tokens_seen": 178976775, + "step": 5492 + }, + { + "epoch": 0.4953780944221491, + "flos": 24245075411520.0, + "grad_norm": 2.1718287840392474, + "learning_rate": 2.1269803468569756e-06, + "loss": 0.7464, + "num_input_tokens_seen": 179006650, + "step": 5493 + }, + { + "epoch": 0.4954682779456193, + "flos": 47704744897920.0, + "grad_norm": 1.6948619191013603, + "learning_rate": 2.126397309485794e-06, + "loss": 0.7899, + "num_input_tokens_seen": 179039490, + "step": 5494 + }, + { + "epoch": 0.4955584614690896, + "flos": 30330350937120.0, + "grad_norm": 1.667869118209608, + "learning_rate": 2.1258142613296983e-06, + "loss": 0.7146, + "num_input_tokens_seen": 179072545, + "step": 5495 + }, + { + "epoch": 0.49564864499255984, + "flos": 33097406267040.0, + "grad_norm": 1.4113205517550622, + "learning_rate": 2.125231202438435e-06, + "loss": 0.7278, + "num_input_tokens_seen": 179105895, + "step": 5496 + }, + { + "epoch": 0.49573882851603013, + "flos": 20346097487520.0, + "grad_norm": 1.4517620627018877, + "learning_rate": 2.1246481328617553e-06, + "loss": 0.7882, + "num_input_tokens_seen": 179135065, + "step": 5497 + }, + { + "epoch": 0.49582901203950036, + "flos": 22932999492480.0, + "grad_norm": 2.009554571074694, + "learning_rate": 2.1240650526494096e-06, + "loss": 0.7353, + "num_input_tokens_seen": 179160930, + "step": 5498 + }, + { + "epoch": 0.49591919556297065, + "flos": 24755050449600.0, + "grad_norm": 1.9573495571744326, + "learning_rate": 2.1234819618511493e-06, + "loss": 0.7219, + "num_input_tokens_seen": 179189305, + "step": 5499 + }, + { + "epoch": 0.4960093790864409, + "flos": 67518894285600.0, + "grad_norm": 0.681736477558639, + "learning_rate": 2.122898860516728e-06, + "loss": 0.6178, + "num_input_tokens_seen": 179276915, + "step": 5500 + }, + { + "epoch": 0.49609956260991117, + "flos": 24208829161920.0, + "grad_norm": 1.5290527148842905, + "learning_rate": 2.1223157486958976e-06, + "loss": 0.7384, + "num_input_tokens_seen": 179306145, + "step": 5501 + }, + { + "epoch": 0.4961897461333814, + "flos": 26794616074080.0, + "grad_norm": 1.514484408933398, + "learning_rate": 2.1217326264384127e-06, + "loss": 0.6436, + "num_input_tokens_seen": 179336470, + "step": 5502 + }, + { + "epoch": 0.4962799296568517, + "flos": 14546889133920.0, + "grad_norm": 2.1032180788208477, + "learning_rate": 2.1211494937940296e-06, + "loss": 0.7762, + "num_input_tokens_seen": 179357275, + "step": 5503 + }, + { + "epoch": 0.496370113180322, + "flos": 27159160076640.0, + "grad_norm": 1.3982821756552313, + "learning_rate": 2.1205663508125034e-06, + "loss": 0.7686, + "num_input_tokens_seen": 179386315, + "step": 5504 + }, + { + "epoch": 0.4964602967037922, + "flos": 25555404351840.0, + "grad_norm": 1.5173010456564064, + "learning_rate": 2.1199831975435914e-06, + "loss": 0.7158, + "num_input_tokens_seen": 179415145, + "step": 5505 + }, + { + "epoch": 0.4965504802272625, + "flos": 26174675685120.0, + "grad_norm": 1.5839773758852378, + "learning_rate": 2.1194000340370517e-06, + "loss": 0.8029, + "num_input_tokens_seen": 179447630, + "step": 5506 + }, + { + "epoch": 0.49664066375073274, + "flos": 24464448566880.0, + "grad_norm": 2.238891837836196, + "learning_rate": 2.1188168603426423e-06, + "loss": 0.6927, + "num_input_tokens_seen": 179475780, + "step": 5507 + }, + { + "epoch": 0.496730847274203, + "flos": 19909246834560.0, + "grad_norm": 3.065196971195979, + "learning_rate": 2.118233676510123e-06, + "loss": 0.7178, + "num_input_tokens_seen": 179503005, + "step": 5508 + }, + { + "epoch": 0.49682103079767326, + "flos": 24208829161920.0, + "grad_norm": 1.7209685832053654, + "learning_rate": 2.117650482589255e-06, + "loss": 0.7636, + "num_input_tokens_seen": 179531220, + "step": 5509 + }, + { + "epoch": 0.49691121432114355, + "flos": 13315371552000.0, + "grad_norm": 1.9360277673645256, + "learning_rate": 2.1170672786297988e-06, + "loss": 0.7929, + "num_input_tokens_seen": 179556170, + "step": 5510 + }, + { + "epoch": 0.4970013978446138, + "flos": 23731197548640.0, + "grad_norm": 1.5231978007791576, + "learning_rate": 2.1164840646815174e-06, + "loss": 0.7225, + "num_input_tokens_seen": 179586540, + "step": 5511 + }, + { + "epoch": 0.49709158136808407, + "flos": 21581480724480.0, + "grad_norm": 1.8209843521283786, + "learning_rate": 2.1159008407941726e-06, + "loss": 0.8366, + "num_input_tokens_seen": 179614225, + "step": 5512 + }, + { + "epoch": 0.4971817648915543, + "flos": 18744460439040.0, + "grad_norm": 1.633002968224224, + "learning_rate": 2.1153176070175293e-06, + "loss": 0.7428, + "num_input_tokens_seen": 179639920, + "step": 5513 + }, + { + "epoch": 0.4972719484150246, + "flos": 27597497520000.0, + "grad_norm": 1.5767798403445523, + "learning_rate": 2.114734363401352e-06, + "loss": 0.6415, + "num_input_tokens_seen": 179672375, + "step": 5514 + }, + { + "epoch": 0.4973621319384948, + "flos": 15319062624480.0, + "grad_norm": 1.9321403204229126, + "learning_rate": 2.1141511099954056e-06, + "loss": 0.7507, + "num_input_tokens_seen": 179698145, + "step": 5515 + }, + { + "epoch": 0.4974523154619651, + "flos": 25661578387200.0, + "grad_norm": 1.4188781848421732, + "learning_rate": 2.1135678468494576e-06, + "loss": 0.7348, + "num_input_tokens_seen": 179727845, + "step": 5516 + }, + { + "epoch": 0.49754249898543534, + "flos": 28507370736000.0, + "grad_norm": 1.9992178937426421, + "learning_rate": 2.112984574013275e-06, + "loss": 0.7912, + "num_input_tokens_seen": 179753800, + "step": 5517 + }, + { + "epoch": 0.49763268250890563, + "flos": 20054900888640.0, + "grad_norm": 1.5680182859220622, + "learning_rate": 2.112401291536625e-06, + "loss": 0.7127, + "num_input_tokens_seen": 179783330, + "step": 5518 + }, + { + "epoch": 0.49772286603237587, + "flos": 20711087527200.0, + "grad_norm": 1.9706548527985925, + "learning_rate": 2.111817999469278e-06, + "loss": 0.7881, + "num_input_tokens_seen": 179811860, + "step": 5519 + }, + { + "epoch": 0.49781304955584615, + "flos": 17979089014560.0, + "grad_norm": 1.636675704353588, + "learning_rate": 2.1112346978610016e-06, + "loss": 0.7037, + "num_input_tokens_seen": 179840180, + "step": 5520 + }, + { + "epoch": 0.4979032330793164, + "flos": 42715889112000.0, + "grad_norm": 1.779914957972469, + "learning_rate": 2.1106513867615678e-06, + "loss": 0.7077, + "num_input_tokens_seen": 179871550, + "step": 5521 + }, + { + "epoch": 0.4979934166027867, + "flos": 23258472343680.0, + "grad_norm": 1.6872793689183383, + "learning_rate": 2.110068066220748e-06, + "loss": 0.756, + "num_input_tokens_seen": 179901465, + "step": 5522 + }, + { + "epoch": 0.4980836001262569, + "flos": 24757392144480.0, + "grad_norm": 1.331966991496699, + "learning_rate": 2.109484736288313e-06, + "loss": 0.8197, + "num_input_tokens_seen": 179933000, + "step": 5523 + }, + { + "epoch": 0.4981737836497272, + "flos": 22933594208640.0, + "grad_norm": 1.8930522376025423, + "learning_rate": 2.108901397014037e-06, + "loss": 0.7689, + "num_input_tokens_seen": 179961825, + "step": 5524 + }, + { + "epoch": 0.49826396717319743, + "flos": 21622298854560.0, + "grad_norm": 2.1037830972589595, + "learning_rate": 2.1083180484476934e-06, + "loss": 0.7568, + "num_input_tokens_seen": 179990355, + "step": 5525 + }, + { + "epoch": 0.4983541506966677, + "flos": 14591201221440.0, + "grad_norm": 1.7411371323480977, + "learning_rate": 2.1077346906390567e-06, + "loss": 0.7734, + "num_input_tokens_seen": 180016555, + "step": 5526 + }, + { + "epoch": 0.498444334220138, + "flos": 16958321203680.0, + "grad_norm": 1.6928632068697878, + "learning_rate": 2.107151323637902e-06, + "loss": 0.7709, + "num_input_tokens_seen": 180043195, + "step": 5527 + }, + { + "epoch": 0.49853451774360824, + "flos": 20311635386400.0, + "grad_norm": 2.1231105482560175, + "learning_rate": 2.106567947494006e-06, + "loss": 0.7423, + "num_input_tokens_seen": 180071610, + "step": 5528 + }, + { + "epoch": 0.49862470126707853, + "flos": 26066494482720.0, + "grad_norm": 1.9479521341885802, + "learning_rate": 2.1059845622571447e-06, + "loss": 0.8196, + "num_input_tokens_seen": 180099190, + "step": 5529 + }, + { + "epoch": 0.49871488479054876, + "flos": 21730703075520.0, + "grad_norm": 1.6727335499091598, + "learning_rate": 2.1054011679770956e-06, + "loss": 0.7938, + "num_input_tokens_seen": 180127090, + "step": 5530 + }, + { + "epoch": 0.49880506831401905, + "flos": 29232741765120.0, + "grad_norm": 2.134852401491392, + "learning_rate": 2.104817764703638e-06, + "loss": 0.763, + "num_input_tokens_seen": 180157035, + "step": 5531 + }, + { + "epoch": 0.4988952518374893, + "flos": 23771569641600.0, + "grad_norm": 1.8090836127270415, + "learning_rate": 2.1042343524865516e-06, + "loss": 0.7582, + "num_input_tokens_seen": 180182480, + "step": 5532 + }, + { + "epoch": 0.4989854353609596, + "flos": 23403308663040.0, + "grad_norm": 1.573938235902086, + "learning_rate": 2.103650931375615e-06, + "loss": 0.862, + "num_input_tokens_seen": 180209845, + "step": 5533 + }, + { + "epoch": 0.4990756188844298, + "flos": 20455728310560.0, + "grad_norm": 1.8365646914277074, + "learning_rate": 2.1030675014206094e-06, + "loss": 0.6921, + "num_input_tokens_seen": 180234645, + "step": 5534 + }, + { + "epoch": 0.4991658024079001, + "flos": 24537089745120.0, + "grad_norm": 2.385962933178082, + "learning_rate": 2.1024840626713166e-06, + "loss": 0.6837, + "num_input_tokens_seen": 180259375, + "step": 5535 + }, + { + "epoch": 0.4992559859313703, + "flos": 22132162383360.0, + "grad_norm": 1.6540881749258978, + "learning_rate": 2.1019006151775177e-06, + "loss": 0.853, + "num_input_tokens_seen": 180287940, + "step": 5536 + }, + { + "epoch": 0.4993461694548406, + "flos": 20201261168160.0, + "grad_norm": 1.7608530831602203, + "learning_rate": 2.101317158988997e-06, + "loss": 0.7425, + "num_input_tokens_seen": 180314920, + "step": 5537 + }, + { + "epoch": 0.49943635297831085, + "flos": 19763852968800.0, + "grad_norm": 2.334454528189262, + "learning_rate": 2.1007336941555374e-06, + "loss": 0.7272, + "num_input_tokens_seen": 180341060, + "step": 5538 + }, + { + "epoch": 0.49952653650178114, + "flos": 26722569612000.0, + "grad_norm": 1.6102648746992103, + "learning_rate": 2.1001502207269238e-06, + "loss": 0.7203, + "num_input_tokens_seen": 180372125, + "step": 5539 + }, + { + "epoch": 0.49961672002525137, + "flos": 24463444983360.0, + "grad_norm": 2.072966059059564, + "learning_rate": 2.0995667387529407e-06, + "loss": 0.7105, + "num_input_tokens_seen": 180400080, + "step": 5540 + }, + { + "epoch": 0.49970690354872166, + "flos": 20489149658400.0, + "grad_norm": 1.5984782224000371, + "learning_rate": 2.098983248283375e-06, + "loss": 0.7047, + "num_input_tokens_seen": 180427650, + "step": 5541 + }, + { + "epoch": 0.4997970870721919, + "flos": 21803753121120.0, + "grad_norm": 2.039702917794375, + "learning_rate": 2.098399749368012e-06, + "loss": 0.7549, + "num_input_tokens_seen": 180456270, + "step": 5542 + }, + { + "epoch": 0.4998872705956622, + "flos": 70552719948480.0, + "grad_norm": 0.6502996615930317, + "learning_rate": 2.09781624205664e-06, + "loss": 0.5693, + "num_input_tokens_seen": 180543765, + "step": 5543 + }, + { + "epoch": 0.4999774541191324, + "flos": 24862525426560.0, + "grad_norm": 1.5355675221648717, + "learning_rate": 2.0972327263990477e-06, + "loss": 0.8014, + "num_input_tokens_seen": 180572170, + "step": 5544 + }, + { + "epoch": 0.5000676376426026, + "flos": 20602162929600.0, + "grad_norm": 1.848819229552701, + "learning_rate": 2.0966492024450226e-06, + "loss": 0.6839, + "num_input_tokens_seen": 180600010, + "step": 5545 + }, + { + "epoch": 0.500157821166073, + "flos": 29200547019360.0, + "grad_norm": 2.6679281787384643, + "learning_rate": 2.0960656702443545e-06, + "loss": 0.9011, + "num_input_tokens_seen": 180629160, + "step": 5546 + }, + { + "epoch": 0.5002480046895432, + "flos": 21148681575360.0, + "grad_norm": 2.39170832802599, + "learning_rate": 2.0954821298468343e-06, + "loss": 0.7527, + "num_input_tokens_seen": 180656555, + "step": 5547 + }, + { + "epoch": 0.5003381882130135, + "flos": 24022803014880.0, + "grad_norm": 1.7854980542191816, + "learning_rate": 2.0948985813022513e-06, + "loss": 0.8142, + "num_input_tokens_seen": 180685155, + "step": 5548 + }, + { + "epoch": 0.5004283717364837, + "flos": 24500062930560.0, + "grad_norm": 2.2311709178940005, + "learning_rate": 2.094315024660399e-06, + "loss": 0.7899, + "num_input_tokens_seen": 180715330, + "step": 5549 + }, + { + "epoch": 0.500518555259954, + "flos": 23006198217120.0, + "grad_norm": 2.048952867985159, + "learning_rate": 2.0937314599710676e-06, + "loss": 0.7754, + "num_input_tokens_seen": 180743795, + "step": 5550 + }, + { + "epoch": 0.5006087387834243, + "flos": 23917186525920.0, + "grad_norm": 2.180379668780306, + "learning_rate": 2.0931478872840526e-06, + "loss": 0.746, + "num_input_tokens_seen": 180773975, + "step": 5551 + }, + { + "epoch": 0.5006989223068945, + "flos": 21986768517600.0, + "grad_norm": 5.0955300058629245, + "learning_rate": 2.092564306649145e-06, + "loss": 0.7331, + "num_input_tokens_seen": 180802100, + "step": 5552 + }, + { + "epoch": 0.5007891058303648, + "flos": 24281358830880.0, + "grad_norm": 1.626831050107688, + "learning_rate": 2.091980718116141e-06, + "loss": 0.7903, + "num_input_tokens_seen": 180833345, + "step": 5553 + }, + { + "epoch": 0.5008792893538351, + "flos": 36268151090400.0, + "grad_norm": 2.495105243213828, + "learning_rate": 2.091397121734835e-06, + "loss": 0.7942, + "num_input_tokens_seen": 180864410, + "step": 5554 + }, + { + "epoch": 0.5009694728773053, + "flos": 24936727734720.0, + "grad_norm": 2.3333916202712834, + "learning_rate": 2.090813517555022e-06, + "loss": 0.774, + "num_input_tokens_seen": 180894305, + "step": 5555 + }, + { + "epoch": 0.5010596564007755, + "flos": 34405839549600.0, + "grad_norm": 1.822062621098871, + "learning_rate": 2.0902299056265e-06, + "loss": 0.6946, + "num_input_tokens_seen": 180922350, + "step": 5556 + }, + { + "epoch": 0.5011498399242459, + "flos": 19249714917600.0, + "grad_norm": 2.299770580127923, + "learning_rate": 2.0896462859990643e-06, + "loss": 0.734, + "num_input_tokens_seen": 180949390, + "step": 5557 + }, + { + "epoch": 0.5012400234477161, + "flos": 20856778751040.0, + "grad_norm": 1.8420702507879254, + "learning_rate": 2.089062658722513e-06, + "loss": 0.8134, + "num_input_tokens_seen": 180976425, + "step": 5558 + }, + { + "epoch": 0.5013302069711864, + "flos": 25231455460800.0, + "grad_norm": 3.227140730430941, + "learning_rate": 2.0884790238466452e-06, + "loss": 0.8015, + "num_input_tokens_seen": 181006855, + "step": 5559 + }, + { + "epoch": 0.5014203904946566, + "flos": 26321407662240.0, + "grad_norm": 2.1771714650496223, + "learning_rate": 2.087895381421259e-06, + "loss": 0.7842, + "num_input_tokens_seen": 181033045, + "step": 5560 + }, + { + "epoch": 0.5015105740181269, + "flos": 21767618380800.0, + "grad_norm": 4.231531518673856, + "learning_rate": 2.087311731496154e-06, + "loss": 0.737, + "num_input_tokens_seen": 181058715, + "step": 5561 + }, + { + "epoch": 0.5016007575415972, + "flos": 21623079419520.0, + "grad_norm": 1.7297388608452258, + "learning_rate": 2.08672807412113e-06, + "loss": 0.7455, + "num_input_tokens_seen": 181088750, + "step": 5562 + }, + { + "epoch": 0.5016909410650674, + "flos": 15902199217440.0, + "grad_norm": 2.583035226254575, + "learning_rate": 2.08614440934599e-06, + "loss": 0.8454, + "num_input_tokens_seen": 181115770, + "step": 5563 + }, + { + "epoch": 0.5017811245885376, + "flos": 29085935448480.0, + "grad_norm": 2.40346147540401, + "learning_rate": 2.0855607372205337e-06, + "loss": 0.6643, + "num_input_tokens_seen": 181144445, + "step": 5564 + }, + { + "epoch": 0.501871308112008, + "flos": 26649816924480.0, + "grad_norm": 2.9545341238058005, + "learning_rate": 2.0849770577945623e-06, + "loss": 0.767, + "num_input_tokens_seen": 181171880, + "step": 5565 + }, + { + "epoch": 0.5019614916354782, + "flos": 42676148904960.0, + "grad_norm": 1.6402668879993851, + "learning_rate": 2.084393371117881e-06, + "loss": 0.7844, + "num_input_tokens_seen": 181205690, + "step": 5566 + }, + { + "epoch": 0.5020516751589484, + "flos": 33935641888320.0, + "grad_norm": 1.8387216438477243, + "learning_rate": 2.0838096772402902e-06, + "loss": 0.6839, + "num_input_tokens_seen": 181238445, + "step": 5567 + }, + { + "epoch": 0.5021418586824187, + "flos": 44461730593920.0, + "grad_norm": 2.525321516963352, + "learning_rate": 2.0832259762115973e-06, + "loss": 0.6382, + "num_input_tokens_seen": 181271920, + "step": 5568 + }, + { + "epoch": 0.502232042205889, + "flos": 26172594178560.0, + "grad_norm": 4.58615730447737, + "learning_rate": 2.082642268081605e-06, + "loss": 0.7821, + "num_input_tokens_seen": 181302520, + "step": 5569 + }, + { + "epoch": 0.5023222257293593, + "flos": 21876059771520.0, + "grad_norm": 2.661949595122465, + "learning_rate": 2.082058552900118e-06, + "loss": 0.7089, + "num_input_tokens_seen": 181326805, + "step": 5570 + }, + { + "epoch": 0.5024124092528295, + "flos": 36194023121760.0, + "grad_norm": 1.7908530125069995, + "learning_rate": 2.081474830716944e-06, + "loss": 0.727, + "num_input_tokens_seen": 181358825, + "step": 5571 + }, + { + "epoch": 0.5025025927762997, + "flos": 20747705474400.0, + "grad_norm": 3.3666862946747043, + "learning_rate": 2.080891101581887e-06, + "loss": 0.7811, + "num_input_tokens_seen": 181387560, + "step": 5572 + }, + { + "epoch": 0.5025927762997701, + "flos": 10582815492960.0, + "grad_norm": 2.477828511496559, + "learning_rate": 2.080307365544755e-06, + "loss": 0.7945, + "num_input_tokens_seen": 181412465, + "step": 5573 + }, + { + "epoch": 0.5026829598232403, + "flos": 22276961532960.0, + "grad_norm": 2.050591741615988, + "learning_rate": 2.0797236226553567e-06, + "loss": 0.7921, + "num_input_tokens_seen": 181443230, + "step": 5574 + }, + { + "epoch": 0.5027731433467105, + "flos": 17141745467520.0, + "grad_norm": 4.088994882656449, + "learning_rate": 2.079139872963499e-06, + "loss": 0.6905, + "num_input_tokens_seen": 181470370, + "step": 5575 + }, + { + "epoch": 0.5028633268701809, + "flos": 19655151389760.0, + "grad_norm": 2.628509536897114, + "learning_rate": 2.078556116518991e-06, + "loss": 0.6547, + "num_input_tokens_seen": 181497195, + "step": 5576 + }, + { + "epoch": 0.5029535103936511, + "flos": 22021974013920.0, + "grad_norm": 1.9455931434265836, + "learning_rate": 2.077972353371642e-06, + "loss": 0.7246, + "num_input_tokens_seen": 181524855, + "step": 5577 + }, + { + "epoch": 0.5030436939171213, + "flos": 21072174742080.0, + "grad_norm": 4.149380621592417, + "learning_rate": 2.077388583571262e-06, + "loss": 0.6996, + "num_input_tokens_seen": 181554785, + "step": 5578 + }, + { + "epoch": 0.5031338774405916, + "flos": 64279708466880.0, + "grad_norm": 0.6489187545912602, + "learning_rate": 2.0768048071676608e-06, + "loss": 0.513, + "num_input_tokens_seen": 181650110, + "step": 5579 + }, + { + "epoch": 0.5032240609640619, + "flos": 23698965633120.0, + "grad_norm": 2.466745359437095, + "learning_rate": 2.0762210242106505e-06, + "loss": 0.7176, + "num_input_tokens_seen": 181679940, + "step": 5580 + }, + { + "epoch": 0.5033142444875321, + "flos": 27269051088000.0, + "grad_norm": 2.5164770732815276, + "learning_rate": 2.0756372347500424e-06, + "loss": 0.7683, + "num_input_tokens_seen": 181708885, + "step": 5581 + }, + { + "epoch": 0.5034044280110024, + "flos": 23043708238560.0, + "grad_norm": 2.6514675395373435, + "learning_rate": 2.0750534388356473e-06, + "loss": 0.8573, + "num_input_tokens_seen": 181732855, + "step": 5582 + }, + { + "epoch": 0.5034946115344726, + "flos": 32551631016480.0, + "grad_norm": 2.331020116849327, + "learning_rate": 2.07446963651728e-06, + "loss": 0.698, + "num_input_tokens_seen": 181763780, + "step": 5583 + }, + { + "epoch": 0.503584795057943, + "flos": 19982408389440.0, + "grad_norm": 2.1747584241683513, + "learning_rate": 2.0738858278447516e-06, + "loss": 0.6915, + "num_input_tokens_seen": 181790755, + "step": 5584 + }, + { + "epoch": 0.5036749785814132, + "flos": 25411088409120.0, + "grad_norm": 1.8261432829061823, + "learning_rate": 2.073302012867878e-06, + "loss": 0.8438, + "num_input_tokens_seen": 181817915, + "step": 5585 + }, + { + "epoch": 0.5037651621048834, + "flos": 20966706932160.0, + "grad_norm": 1.8816935906569014, + "learning_rate": 2.0727181916364725e-06, + "loss": 0.7873, + "num_input_tokens_seen": 181846675, + "step": 5586 + }, + { + "epoch": 0.5038553456283537, + "flos": 22350866483040.0, + "grad_norm": 1.970746012467702, + "learning_rate": 2.0721343642003493e-06, + "loss": 0.7858, + "num_input_tokens_seen": 181874040, + "step": 5587 + }, + { + "epoch": 0.503945529151824, + "flos": 19218077718240.0, + "grad_norm": 3.0748975845277307, + "learning_rate": 2.0715505306093247e-06, + "loss": 0.7813, + "num_input_tokens_seen": 181901380, + "step": 5588 + }, + { + "epoch": 0.5040357126752942, + "flos": 15751898943360.0, + "grad_norm": 2.747507092481489, + "learning_rate": 2.070966690913214e-06, + "loss": 0.7868, + "num_input_tokens_seen": 181925075, + "step": 5589 + }, + { + "epoch": 0.5041258961987645, + "flos": 22788200342880.0, + "grad_norm": 4.642276765368926, + "learning_rate": 2.0703828451618346e-06, + "loss": 0.638, + "num_input_tokens_seen": 181954645, + "step": 5590 + }, + { + "epoch": 0.5042160797222347, + "flos": 25084500465120.0, + "grad_norm": 2.993714582835011, + "learning_rate": 2.069798993405002e-06, + "loss": 0.7481, + "num_input_tokens_seen": 181986990, + "step": 5591 + }, + { + "epoch": 0.504306263245705, + "flos": 27633372072000.0, + "grad_norm": 1.9650331561370709, + "learning_rate": 2.0692151356925345e-06, + "loss": 0.7919, + "num_input_tokens_seen": 182017390, + "step": 5592 + }, + { + "epoch": 0.5043964467691753, + "flos": 25520421874080.0, + "grad_norm": 2.043201440575554, + "learning_rate": 2.068631272074251e-06, + "loss": 0.7762, + "num_input_tokens_seen": 182045495, + "step": 5593 + }, + { + "epoch": 0.5044866302926455, + "flos": 22386443676960.0, + "grad_norm": 2.38113494354788, + "learning_rate": 2.0680474025999676e-06, + "loss": 0.8793, + "num_input_tokens_seen": 182075315, + "step": 5594 + }, + { + "epoch": 0.5045768138161157, + "flos": 70612165861920.0, + "grad_norm": 0.8633369669475702, + "learning_rate": 2.0674635273195055e-06, + "loss": 0.6311, + "num_input_tokens_seen": 182151240, + "step": 5595 + }, + { + "epoch": 0.5046669973395861, + "flos": 23219884399200.0, + "grad_norm": 2.493145510421085, + "learning_rate": 2.066879646282682e-06, + "loss": 0.7513, + "num_input_tokens_seen": 182180670, + "step": 5596 + }, + { + "epoch": 0.5047571808630563, + "flos": 21622038666240.0, + "grad_norm": 2.0523166685201732, + "learning_rate": 2.0662957595393194e-06, + "loss": 0.7254, + "num_input_tokens_seen": 182208440, + "step": 5597 + }, + { + "epoch": 0.5048473643865266, + "flos": 21585680907360.0, + "grad_norm": 2.012424263802617, + "learning_rate": 2.0657118671392373e-06, + "loss": 0.778, + "num_input_tokens_seen": 182237340, + "step": 5598 + }, + { + "epoch": 0.5049375479099969, + "flos": 24864644102880.0, + "grad_norm": 5.282204626358362, + "learning_rate": 2.0651279691322558e-06, + "loss": 0.8104, + "num_input_tokens_seen": 182265050, + "step": 5599 + }, + { + "epoch": 0.5050277314334671, + "flos": 28723993329120.0, + "grad_norm": 1.8025527151708625, + "learning_rate": 2.0645440655681973e-06, + "loss": 0.707, + "num_input_tokens_seen": 182294545, + "step": 5600 + }, + { + "epoch": 0.5051179149569374, + "flos": 20747073588480.0, + "grad_norm": 4.154041822286303, + "learning_rate": 2.0639601564968826e-06, + "loss": 0.8175, + "num_input_tokens_seen": 182321200, + "step": 5601 + }, + { + "epoch": 0.5052080984804076, + "flos": 71970226507680.0, + "grad_norm": 0.6935036737431346, + "learning_rate": 2.0633762419681355e-06, + "loss": 0.6653, + "num_input_tokens_seen": 182415895, + "step": 5602 + }, + { + "epoch": 0.5052982820038779, + "flos": 38674528072800.0, + "grad_norm": 1.8886256089012279, + "learning_rate": 2.062792322031777e-06, + "loss": 0.775, + "num_input_tokens_seen": 182448420, + "step": 5603 + }, + { + "epoch": 0.5053884655273482, + "flos": 21257977870560.0, + "grad_norm": 2.8198144531080898, + "learning_rate": 2.062208396737632e-06, + "loss": 0.7126, + "num_input_tokens_seen": 182476620, + "step": 5604 + }, + { + "epoch": 0.5054786490508184, + "flos": 25739125973760.0, + "grad_norm": 1.7892473949857401, + "learning_rate": 2.0616244661355235e-06, + "loss": 0.7793, + "num_input_tokens_seen": 182507135, + "step": 5605 + }, + { + "epoch": 0.5055688325742886, + "flos": 23297803683360.0, + "grad_norm": 8.182129008236297, + "learning_rate": 2.0610405302752752e-06, + "loss": 0.7882, + "num_input_tokens_seen": 182535905, + "step": 5606 + }, + { + "epoch": 0.505659016097759, + "flos": 31162973924640.0, + "grad_norm": 1.9255543046978905, + "learning_rate": 2.060456589206713e-06, + "loss": 0.745, + "num_input_tokens_seen": 182565595, + "step": 5607 + }, + { + "epoch": 0.5057491996212292, + "flos": 26097425456640.0, + "grad_norm": 2.484259836658629, + "learning_rate": 2.0598726429796614e-06, + "loss": 0.7055, + "num_input_tokens_seen": 182594160, + "step": 5608 + }, + { + "epoch": 0.5058393831446995, + "flos": 29454679633920.0, + "grad_norm": 1.9589673361983329, + "learning_rate": 2.059288691643945e-06, + "loss": 0.717, + "num_input_tokens_seen": 182623220, + "step": 5609 + }, + { + "epoch": 0.5059295666681697, + "flos": 20855589318720.0, + "grad_norm": 2.024844354369942, + "learning_rate": 2.0587047352493913e-06, + "loss": 0.8325, + "num_input_tokens_seen": 182648970, + "step": 5610 + }, + { + "epoch": 0.50601975019164, + "flos": 23188730406720.0, + "grad_norm": 2.0114757798510143, + "learning_rate": 2.0581207738458248e-06, + "loss": 0.7324, + "num_input_tokens_seen": 182677860, + "step": 5611 + }, + { + "epoch": 0.5061099337151103, + "flos": 18306494693280.0, + "grad_norm": 2.4835784402851204, + "learning_rate": 2.0575368074830743e-06, + "loss": 0.7112, + "num_input_tokens_seen": 182703395, + "step": 5612 + }, + { + "epoch": 0.5062001172385805, + "flos": 30767090080800.0, + "grad_norm": 8.655708378672488, + "learning_rate": 2.0569528362109667e-06, + "loss": 0.6648, + "num_input_tokens_seen": 182734525, + "step": 5613 + }, + { + "epoch": 0.5062903007620507, + "flos": 34007242313280.0, + "grad_norm": 1.5903406514820488, + "learning_rate": 2.056368860079327e-06, + "loss": 0.7505, + "num_input_tokens_seen": 182766060, + "step": 5614 + }, + { + "epoch": 0.5063804842855211, + "flos": 21694642674720.0, + "grad_norm": 3.8493544495596566, + "learning_rate": 2.0557848791379874e-06, + "loss": 0.6952, + "num_input_tokens_seen": 182791465, + "step": 5615 + }, + { + "epoch": 0.5064706678089913, + "flos": 20092299400800.0, + "grad_norm": 1.856124729848818, + "learning_rate": 2.0552008934367734e-06, + "loss": 0.6556, + "num_input_tokens_seen": 182820075, + "step": 5616 + }, + { + "epoch": 0.5065608513324615, + "flos": 20018914827360.0, + "grad_norm": 2.2702272172046603, + "learning_rate": 2.0546169030255154e-06, + "loss": 0.8258, + "num_input_tokens_seen": 182844365, + "step": 5617 + }, + { + "epoch": 0.5066510348559318, + "flos": 23516024576160.0, + "grad_norm": 2.410844355308671, + "learning_rate": 2.054032907954041e-06, + "loss": 0.8272, + "num_input_tokens_seen": 182870210, + "step": 5618 + }, + { + "epoch": 0.5067412183794021, + "flos": 22205175259200.0, + "grad_norm": 1.9104560351570685, + "learning_rate": 2.053448908272182e-06, + "loss": 0.7184, + "num_input_tokens_seen": 182901155, + "step": 5619 + }, + { + "epoch": 0.5068314019028723, + "flos": 18598509026880.0, + "grad_norm": 2.32918949241122, + "learning_rate": 2.0528649040297673e-06, + "loss": 0.7887, + "num_input_tokens_seen": 182929120, + "step": 5620 + }, + { + "epoch": 0.5069215854263426, + "flos": 17542089682560.0, + "grad_norm": 1.8183382649737974, + "learning_rate": 2.0522808952766266e-06, + "loss": 0.7595, + "num_input_tokens_seen": 182955785, + "step": 5621 + }, + { + "epoch": 0.5070117689498129, + "flos": 24203997093120.0, + "grad_norm": 1.9061064744292124, + "learning_rate": 2.0516968820625925e-06, + "loss": 0.8002, + "num_input_tokens_seen": 182985100, + "step": 5622 + }, + { + "epoch": 0.5071019524732832, + "flos": 15938222448480.0, + "grad_norm": 2.1142441676450576, + "learning_rate": 2.051112864437495e-06, + "loss": 0.7462, + "num_input_tokens_seen": 183010050, + "step": 5623 + }, + { + "epoch": 0.5071921359967534, + "flos": 22059521205120.0, + "grad_norm": 2.9425471348904924, + "learning_rate": 2.050528842451166e-06, + "loss": 0.7231, + "num_input_tokens_seen": 183033555, + "step": 5624 + }, + { + "epoch": 0.5072823195202236, + "flos": 21439469306880.0, + "grad_norm": 2.1242857461951856, + "learning_rate": 2.049944816153438e-06, + "loss": 0.6357, + "num_input_tokens_seen": 183065455, + "step": 5625 + }, + { + "epoch": 0.507372503043694, + "flos": 24318831682560.0, + "grad_norm": 1.7869802959916288, + "learning_rate": 2.049360785594142e-06, + "loss": 0.7934, + "num_input_tokens_seen": 183093990, + "step": 5626 + }, + { + "epoch": 0.5074626865671642, + "flos": 67344390764160.0, + "grad_norm": 0.6163201817133498, + "learning_rate": 2.048776750823113e-06, + "loss": 0.5098, + "num_input_tokens_seen": 183193180, + "step": 5627 + }, + { + "epoch": 0.5075528700906344, + "flos": 65519403396000.0, + "grad_norm": 0.6398150976750043, + "learning_rate": 2.0481927118901817e-06, + "loss": 0.5627, + "num_input_tokens_seen": 183291965, + "step": 5628 + }, + { + "epoch": 0.5076430536141047, + "flos": 33608607907200.0, + "grad_norm": 2.6825490192296706, + "learning_rate": 2.0476086688451824e-06, + "loss": 0.6962, + "num_input_tokens_seen": 183320610, + "step": 5629 + }, + { + "epoch": 0.507733237137575, + "flos": 27812075776320.0, + "grad_norm": 2.6653616413994103, + "learning_rate": 2.04702462173795e-06, + "loss": 0.7436, + "num_input_tokens_seen": 183349725, + "step": 5630 + }, + { + "epoch": 0.5078234206610452, + "flos": 21147975349920.0, + "grad_norm": 2.686583961684189, + "learning_rate": 2.0464405706183167e-06, + "loss": 0.7318, + "num_input_tokens_seen": 183375705, + "step": 5631 + }, + { + "epoch": 0.5079136041845155, + "flos": 21104109299520.0, + "grad_norm": 1.8842990436080436, + "learning_rate": 2.045856515536118e-06, + "loss": 0.7698, + "num_input_tokens_seen": 183402885, + "step": 5632 + }, + { + "epoch": 0.5080037877079857, + "flos": 19472656369920.0, + "grad_norm": 2.7037432189658848, + "learning_rate": 2.045272456541188e-06, + "loss": 0.7693, + "num_input_tokens_seen": 183431145, + "step": 5633 + }, + { + "epoch": 0.508093971231456, + "flos": 23735137543200.0, + "grad_norm": 2.0541644572877384, + "learning_rate": 2.0446883936833635e-06, + "loss": 0.8257, + "num_input_tokens_seen": 183460040, + "step": 5634 + }, + { + "epoch": 0.5081841547549263, + "flos": 24354780574080.0, + "grad_norm": 2.4011754614457415, + "learning_rate": 2.0441043270124782e-06, + "loss": 0.7243, + "num_input_tokens_seen": 183487665, + "step": 5635 + }, + { + "epoch": 0.5082743382783965, + "flos": 29819297976000.0, + "grad_norm": 2.298509392563767, + "learning_rate": 2.0435202565783683e-06, + "loss": 0.7817, + "num_input_tokens_seen": 183518175, + "step": 5636 + }, + { + "epoch": 0.5083645218018668, + "flos": 25702545196320.0, + "grad_norm": 3.5952673321091235, + "learning_rate": 2.042936182430871e-06, + "loss": 0.7161, + "num_input_tokens_seen": 183546990, + "step": 5637 + }, + { + "epoch": 0.5084547053253371, + "flos": 24755905354080.0, + "grad_norm": 2.259284044811896, + "learning_rate": 2.0423521046198206e-06, + "loss": 0.7605, + "num_input_tokens_seen": 183575025, + "step": 5638 + }, + { + "epoch": 0.5085448888488073, + "flos": 41514930806400.0, + "grad_norm": 2.508725389360827, + "learning_rate": 2.041768023195056e-06, + "loss": 0.672, + "num_input_tokens_seen": 183609335, + "step": 5639 + }, + { + "epoch": 0.5086350723722776, + "flos": 21003176200320.0, + "grad_norm": 1.7580608277283583, + "learning_rate": 2.0411839382064126e-06, + "loss": 0.8227, + "num_input_tokens_seen": 183637370, + "step": 5640 + }, + { + "epoch": 0.5087252558957478, + "flos": 14259149322720.0, + "grad_norm": 2.0976171507645622, + "learning_rate": 2.040599849703729e-06, + "loss": 0.6489, + "num_input_tokens_seen": 183664290, + "step": 5641 + }, + { + "epoch": 0.5088154394192181, + "flos": 29346015224640.0, + "grad_norm": 2.3144814533801674, + "learning_rate": 2.040015757736843e-06, + "loss": 0.692, + "num_input_tokens_seen": 183693165, + "step": 5642 + }, + { + "epoch": 0.5089056229426884, + "flos": 23546918380320.0, + "grad_norm": 2.0679075848130624, + "learning_rate": 2.039431662355591e-06, + "loss": 0.6783, + "num_input_tokens_seen": 183720110, + "step": 5643 + }, + { + "epoch": 0.5089958064661586, + "flos": 22964636691840.0, + "grad_norm": 1.9484458109442087, + "learning_rate": 2.0388475636098126e-06, + "loss": 0.7693, + "num_input_tokens_seen": 183747735, + "step": 5644 + }, + { + "epoch": 0.5090859899896288, + "flos": 22421909361600.0, + "grad_norm": 1.9382436002175216, + "learning_rate": 2.038263461549346e-06, + "loss": 0.7854, + "num_input_tokens_seen": 183776810, + "step": 5645 + }, + { + "epoch": 0.5091761735130992, + "flos": 14002228976160.0, + "grad_norm": 2.931474227300098, + "learning_rate": 2.0376793562240297e-06, + "loss": 0.7547, + "num_input_tokens_seen": 183800370, + "step": 5646 + }, + { + "epoch": 0.5092663570365694, + "flos": 21366865298400.0, + "grad_norm": 2.8572835308593745, + "learning_rate": 2.037095247683703e-06, + "loss": 0.7662, + "num_input_tokens_seen": 183829470, + "step": 5647 + }, + { + "epoch": 0.5093565405600397, + "flos": 25338038363520.0, + "grad_norm": 2.6036191034035507, + "learning_rate": 2.0365111359782046e-06, + "loss": 0.7113, + "num_input_tokens_seen": 183859965, + "step": 5648 + }, + { + "epoch": 0.50944672408351, + "flos": 28544992266720.0, + "grad_norm": 2.1999399957276666, + "learning_rate": 2.0359270211573757e-06, + "loss": 0.7843, + "num_input_tokens_seen": 183892125, + "step": 5649 + }, + { + "epoch": 0.5095369076069802, + "flos": 15318839605920.0, + "grad_norm": 2.57002975624759, + "learning_rate": 2.0353429032710545e-06, + "loss": 0.817, + "num_input_tokens_seen": 183917305, + "step": 5650 + }, + { + "epoch": 0.5096270911304505, + "flos": 22861064539680.0, + "grad_norm": 2.6813490807070686, + "learning_rate": 2.0347587823690825e-06, + "loss": 0.7289, + "num_input_tokens_seen": 183947585, + "step": 5651 + }, + { + "epoch": 0.5097172746539207, + "flos": 28136433534720.0, + "grad_norm": 2.134903438269484, + "learning_rate": 2.034174658501299e-06, + "loss": 0.7525, + "num_input_tokens_seen": 183977760, + "step": 5652 + }, + { + "epoch": 0.509807458177391, + "flos": 31967639519040.0, + "grad_norm": 1.7201916977047695, + "learning_rate": 2.0335905317175453e-06, + "loss": 0.7872, + "num_input_tokens_seen": 184009240, + "step": 5653 + }, + { + "epoch": 0.5098976417008613, + "flos": 36194655007680.0, + "grad_norm": 2.3234295699311964, + "learning_rate": 2.033006402067663e-06, + "loss": 0.7528, + "num_input_tokens_seen": 184038605, + "step": 5654 + }, + { + "epoch": 0.5099878252243315, + "flos": 21549025790400.0, + "grad_norm": 1.7642743506322098, + "learning_rate": 2.0324222696014912e-06, + "loss": 0.6572, + "num_input_tokens_seen": 184066745, + "step": 5655 + }, + { + "epoch": 0.5100780087478017, + "flos": 22606262869440.0, + "grad_norm": 2.5454235036307207, + "learning_rate": 2.0318381343688733e-06, + "loss": 0.7917, + "num_input_tokens_seen": 184093980, + "step": 5656 + }, + { + "epoch": 0.5101681922712721, + "flos": 29928148234080.0, + "grad_norm": 2.3863794862955694, + "learning_rate": 2.0312539964196505e-06, + "loss": 0.7369, + "num_input_tokens_seen": 184123430, + "step": 5657 + }, + { + "epoch": 0.5102583757947423, + "flos": 22934003076000.0, + "grad_norm": 2.666677226601991, + "learning_rate": 2.030669855803664e-06, + "loss": 0.7312, + "num_input_tokens_seen": 184154415, + "step": 5658 + }, + { + "epoch": 0.5103485593182125, + "flos": 22056844982400.0, + "grad_norm": 4.901846884058687, + "learning_rate": 2.0300857125707563e-06, + "loss": 0.8063, + "num_input_tokens_seen": 184182980, + "step": 5659 + }, + { + "epoch": 0.5104387428416828, + "flos": 19726974833280.0, + "grad_norm": 2.684458497495197, + "learning_rate": 2.0295015667707697e-06, + "loss": 0.7816, + "num_input_tokens_seen": 184208790, + "step": 5660 + }, + { + "epoch": 0.5105289263651531, + "flos": 29162553791040.0, + "grad_norm": 2.2741832231100885, + "learning_rate": 2.0289174184535472e-06, + "loss": 0.7614, + "num_input_tokens_seen": 184238085, + "step": 5661 + }, + { + "epoch": 0.5106191098886234, + "flos": 26613421995840.0, + "grad_norm": 2.301066408474761, + "learning_rate": 2.02833326766893e-06, + "loss": 0.7258, + "num_input_tokens_seen": 184266625, + "step": 5662 + }, + { + "epoch": 0.5107092934120936, + "flos": 20050180329120.0, + "grad_norm": 2.077142468844269, + "learning_rate": 2.027749114466763e-06, + "loss": 0.8183, + "num_input_tokens_seen": 184294025, + "step": 5663 + }, + { + "epoch": 0.5107994769355638, + "flos": 22749203531040.0, + "grad_norm": 1.9129836798498696, + "learning_rate": 2.027164958896889e-06, + "loss": 0.7996, + "num_input_tokens_seen": 184324740, + "step": 5664 + }, + { + "epoch": 0.5108896604590342, + "flos": 17790014947200.0, + "grad_norm": 2.3311125768497893, + "learning_rate": 2.02658080100915e-06, + "loss": 0.8726, + "num_input_tokens_seen": 184347965, + "step": 5665 + }, + { + "epoch": 0.5109798439825044, + "flos": 27995462870400.0, + "grad_norm": 1.7580322128313062, + "learning_rate": 2.0259966408533915e-06, + "loss": 0.8263, + "num_input_tokens_seen": 184377000, + "step": 5666 + }, + { + "epoch": 0.5110700275059746, + "flos": 23771235113760.0, + "grad_norm": 2.204264677449038, + "learning_rate": 2.025412478479455e-06, + "loss": 0.8443, + "num_input_tokens_seen": 184404685, + "step": 5667 + }, + { + "epoch": 0.5111602110294449, + "flos": 37143078998400.0, + "grad_norm": 2.5044874924103517, + "learning_rate": 2.0248283139371862e-06, + "loss": 0.7966, + "num_input_tokens_seen": 184433650, + "step": 5668 + }, + { + "epoch": 0.5112503945529152, + "flos": 18853682394720.0, + "grad_norm": 1.996080730716776, + "learning_rate": 2.024244147276429e-06, + "loss": 0.7233, + "num_input_tokens_seen": 184462160, + "step": 5669 + }, + { + "epoch": 0.5113405780763854, + "flos": 26322411245760.0, + "grad_norm": 2.2062890889538624, + "learning_rate": 2.023659978547027e-06, + "loss": 0.8215, + "num_input_tokens_seen": 184492980, + "step": 5670 + }, + { + "epoch": 0.5114307615998557, + "flos": 24171765177600.0, + "grad_norm": 2.0534491342654144, + "learning_rate": 2.023075807798826e-06, + "loss": 0.7539, + "num_input_tokens_seen": 184523755, + "step": 5671 + }, + { + "epoch": 0.511520945123326, + "flos": 27268679390400.0, + "grad_norm": 1.845164193081288, + "learning_rate": 2.0224916350816696e-06, + "loss": 0.6804, + "num_input_tokens_seen": 184557005, + "step": 5672 + }, + { + "epoch": 0.5116111286467963, + "flos": 19108298216160.0, + "grad_norm": 1.702919623914739, + "learning_rate": 2.0219074604454026e-06, + "loss": 0.7785, + "num_input_tokens_seen": 184582065, + "step": 5673 + }, + { + "epoch": 0.5117013121702665, + "flos": 23801199673920.0, + "grad_norm": 2.59776757941063, + "learning_rate": 2.02132328393987e-06, + "loss": 0.7553, + "num_input_tokens_seen": 184607075, + "step": 5674 + }, + { + "epoch": 0.5117914956937367, + "flos": 19977836508960.0, + "grad_norm": 2.143290287473485, + "learning_rate": 2.0207391056149174e-06, + "loss": 0.7072, + "num_input_tokens_seen": 184635940, + "step": 5675 + }, + { + "epoch": 0.5118816792172071, + "flos": 31786222422240.0, + "grad_norm": 2.1687676684624533, + "learning_rate": 2.020154925520391e-06, + "loss": 0.7243, + "num_input_tokens_seen": 184666375, + "step": 5676 + }, + { + "epoch": 0.5119718627406773, + "flos": 28359746684640.0, + "grad_norm": 1.782296994006431, + "learning_rate": 2.0195707437061332e-06, + "loss": 0.8529, + "num_input_tokens_seen": 184694335, + "step": 5677 + }, + { + "epoch": 0.5120620462641475, + "flos": 18451554031200.0, + "grad_norm": 3.2483166957042258, + "learning_rate": 2.0189865602219934e-06, + "loss": 0.756, + "num_input_tokens_seen": 184721070, + "step": 5678 + }, + { + "epoch": 0.5121522297876178, + "flos": 19363397244480.0, + "grad_norm": 2.3698854085525762, + "learning_rate": 2.0184023751178154e-06, + "loss": 0.829, + "num_input_tokens_seen": 184748115, + "step": 5679 + }, + { + "epoch": 0.5122424133110881, + "flos": 24464597245920.0, + "grad_norm": 1.6847559199152058, + "learning_rate": 2.017818188443444e-06, + "loss": 0.7067, + "num_input_tokens_seen": 184779905, + "step": 5680 + }, + { + "epoch": 0.5123325968345583, + "flos": 16557939818880.0, + "grad_norm": 2.468784090603479, + "learning_rate": 2.017234000248728e-06, + "loss": 0.7681, + "num_input_tokens_seen": 184806825, + "step": 5681 + }, + { + "epoch": 0.5124227803580286, + "flos": 20419816588800.0, + "grad_norm": 2.0956826471215266, + "learning_rate": 2.0166498105835108e-06, + "loss": 0.6948, + "num_input_tokens_seen": 184836410, + "step": 5682 + }, + { + "epoch": 0.5125129638814988, + "flos": 31673766697440.0, + "grad_norm": 1.5940508838941514, + "learning_rate": 2.0160656194976407e-06, + "loss": 0.6924, + "num_input_tokens_seen": 184870635, + "step": 5683 + }, + { + "epoch": 0.5126031474049692, + "flos": 24900927522240.0, + "grad_norm": 1.945799801011359, + "learning_rate": 2.0154814270409634e-06, + "loss": 0.7072, + "num_input_tokens_seen": 184900185, + "step": 5684 + }, + { + "epoch": 0.5126933309284394, + "flos": 21804161988480.0, + "grad_norm": 2.0579431073820857, + "learning_rate": 2.0148972332633247e-06, + "loss": 0.7167, + "num_input_tokens_seen": 184925925, + "step": 5685 + }, + { + "epoch": 0.5127835144519096, + "flos": 23881088955360.0, + "grad_norm": 2.9719521369109154, + "learning_rate": 2.0143130382145733e-06, + "loss": 0.8144, + "num_input_tokens_seen": 184955615, + "step": 5686 + }, + { + "epoch": 0.5128736979753798, + "flos": 23006086707840.0, + "grad_norm": 3.4248404695432586, + "learning_rate": 2.0137288419445533e-06, + "loss": 0.7294, + "num_input_tokens_seen": 184983425, + "step": 5687 + }, + { + "epoch": 0.5129638814988502, + "flos": 19836568486560.0, + "grad_norm": 2.377629280305883, + "learning_rate": 2.0131446445031134e-06, + "loss": 0.7953, + "num_input_tokens_seen": 185011805, + "step": 5688 + }, + { + "epoch": 0.5130540650223204, + "flos": 26430815466720.0, + "grad_norm": 2.5403978797585407, + "learning_rate": 2.0125604459400994e-06, + "loss": 0.838, + "num_input_tokens_seen": 185039550, + "step": 5689 + }, + { + "epoch": 0.5131442485457907, + "flos": 21548468244000.0, + "grad_norm": 2.919828136559736, + "learning_rate": 2.0119762463053596e-06, + "loss": 0.8131, + "num_input_tokens_seen": 185067390, + "step": 5690 + }, + { + "epoch": 0.5132344320692609, + "flos": 19399643494080.0, + "grad_norm": 3.253261836652535, + "learning_rate": 2.0113920456487406e-06, + "loss": 0.7443, + "num_input_tokens_seen": 185094310, + "step": 5691 + }, + { + "epoch": 0.5133246155927312, + "flos": 20855366300160.0, + "grad_norm": 4.307070645339511, + "learning_rate": 2.010807844020088e-06, + "loss": 0.7493, + "num_input_tokens_seen": 185122070, + "step": 5692 + }, + { + "epoch": 0.5134147991162015, + "flos": 22459233534240.0, + "grad_norm": 2.0546109145363354, + "learning_rate": 2.0102236414692524e-06, + "loss": 0.8079, + "num_input_tokens_seen": 185151120, + "step": 5693 + }, + { + "epoch": 0.5135049826396717, + "flos": 22785821478240.0, + "grad_norm": 1.7136386816819966, + "learning_rate": 2.0096394380460777e-06, + "loss": 0.7463, + "num_input_tokens_seen": 185180375, + "step": 5694 + }, + { + "epoch": 0.513595166163142, + "flos": 27011796213600.0, + "grad_norm": 1.8369834355462527, + "learning_rate": 2.0090552338004136e-06, + "loss": 0.7867, + "num_input_tokens_seen": 185213025, + "step": 5695 + }, + { + "epoch": 0.5136853496866123, + "flos": 32511927979200.0, + "grad_norm": 1.858759417532464, + "learning_rate": 2.0084710287821077e-06, + "loss": 0.7522, + "num_input_tokens_seen": 185242200, + "step": 5696 + }, + { + "epoch": 0.5137755332100825, + "flos": 25301531925600.0, + "grad_norm": 2.041601945538924, + "learning_rate": 2.007886823041006e-06, + "loss": 0.7931, + "num_input_tokens_seen": 185270795, + "step": 5697 + }, + { + "epoch": 0.5138657167335527, + "flos": 26503977021600.0, + "grad_norm": 2.024297465211359, + "learning_rate": 2.0073026166269577e-06, + "loss": 0.7736, + "num_input_tokens_seen": 185299905, + "step": 5698 + }, + { + "epoch": 0.5139559002570231, + "flos": 23840493843840.0, + "grad_norm": 1.9819071284649, + "learning_rate": 2.0067184095898093e-06, + "loss": 0.8168, + "num_input_tokens_seen": 185330805, + "step": 5699 + }, + { + "epoch": 0.5140460837804933, + "flos": 22824558101760.0, + "grad_norm": 1.9412826406629538, + "learning_rate": 2.0061342019794094e-06, + "loss": 0.7699, + "num_input_tokens_seen": 185359400, + "step": 5700 + }, + { + "epoch": 0.5141362673039636, + "flos": 29163966241920.0, + "grad_norm": 2.998912773214669, + "learning_rate": 2.0055499938456058e-06, + "loss": 0.8351, + "num_input_tokens_seen": 185386880, + "step": 5701 + }, + { + "epoch": 0.5142264508274338, + "flos": 23189473801920.0, + "grad_norm": 2.582619650507173, + "learning_rate": 2.0049657852382464e-06, + "loss": 0.7266, + "num_input_tokens_seen": 185414790, + "step": 5702 + }, + { + "epoch": 0.5143166343509041, + "flos": 40055676873120.0, + "grad_norm": 2.6727767225832686, + "learning_rate": 2.0043815762071782e-06, + "loss": 0.6874, + "num_input_tokens_seen": 185440635, + "step": 5703 + }, + { + "epoch": 0.5144068178743744, + "flos": 24279017136000.0, + "grad_norm": 3.447795362537991, + "learning_rate": 2.0037973668022492e-06, + "loss": 0.7249, + "num_input_tokens_seen": 185470105, + "step": 5704 + }, + { + "epoch": 0.5144970013978446, + "flos": 28362199888800.0, + "grad_norm": 1.9561503058919918, + "learning_rate": 2.003213157073309e-06, + "loss": 0.637, + "num_input_tokens_seen": 185500555, + "step": 5705 + }, + { + "epoch": 0.5145871849213148, + "flos": 28326325336800.0, + "grad_norm": 2.108413703301323, + "learning_rate": 2.002628947070204e-06, + "loss": 0.7517, + "num_input_tokens_seen": 185531425, + "step": 5706 + }, + { + "epoch": 0.5146773684447852, + "flos": 27705976080480.0, + "grad_norm": 3.650542540269579, + "learning_rate": 2.002044736842783e-06, + "loss": 0.8168, + "num_input_tokens_seen": 185560345, + "step": 5707 + }, + { + "epoch": 0.5147675519682554, + "flos": 27740512521120.0, + "grad_norm": 2.2592246315542712, + "learning_rate": 2.001460526440894e-06, + "loss": 0.7377, + "num_input_tokens_seen": 185589670, + "step": 5708 + }, + { + "epoch": 0.5148577354917256, + "flos": 69764743576320.0, + "grad_norm": 0.6253240222241093, + "learning_rate": 2.0008763159143843e-06, + "loss": 0.5806, + "num_input_tokens_seen": 185679085, + "step": 5709 + }, + { + "epoch": 0.5149479190151959, + "flos": 25338075533280.0, + "grad_norm": 1.7009324826179328, + "learning_rate": 2.000292105313103e-06, + "loss": 0.7859, + "num_input_tokens_seen": 185706510, + "step": 5710 + }, + { + "epoch": 0.5150381025386662, + "flos": 19582361532480.0, + "grad_norm": 3.240947440281768, + "learning_rate": 1.999707894686897e-06, + "loss": 0.7952, + "num_input_tokens_seen": 185734120, + "step": 5711 + }, + { + "epoch": 0.5151282860621365, + "flos": 19981553484960.0, + "grad_norm": 1.677289519657076, + "learning_rate": 1.9991236840856155e-06, + "loss": 0.7433, + "num_input_tokens_seen": 185762205, + "step": 5712 + }, + { + "epoch": 0.5152184695856067, + "flos": 22641171007680.0, + "grad_norm": 2.7721507351393684, + "learning_rate": 1.9985394735591065e-06, + "loss": 0.7275, + "num_input_tokens_seen": 185793540, + "step": 5713 + }, + { + "epoch": 0.5153086531090769, + "flos": 20237990624640.0, + "grad_norm": 2.1617834987945486, + "learning_rate": 1.997955263157217e-06, + "loss": 0.7997, + "num_input_tokens_seen": 185819920, + "step": 5714 + }, + { + "epoch": 0.5153988366325473, + "flos": 23225199674880.0, + "grad_norm": 2.9661717512803034, + "learning_rate": 1.997371052929796e-06, + "loss": 0.7884, + "num_input_tokens_seen": 185846040, + "step": 5715 + }, + { + "epoch": 0.5154890201560175, + "flos": 26212334385600.0, + "grad_norm": 1.9914734402168133, + "learning_rate": 1.996786842926691e-06, + "loss": 0.7085, + "num_input_tokens_seen": 185878280, + "step": 5716 + }, + { + "epoch": 0.5155792036794877, + "flos": 25447706356320.0, + "grad_norm": 1.9013579083781025, + "learning_rate": 1.9962026331977506e-06, + "loss": 0.7664, + "num_input_tokens_seen": 185906900, + "step": 5717 + }, + { + "epoch": 0.5156693872029581, + "flos": 21330879237120.0, + "grad_norm": 2.0189583591136695, + "learning_rate": 1.9956184237928224e-06, + "loss": 0.7708, + "num_input_tokens_seen": 185936220, + "step": 5718 + }, + { + "epoch": 0.5157595707264283, + "flos": 16920253635840.0, + "grad_norm": 2.7745633271456525, + "learning_rate": 1.995034214761754e-06, + "loss": 0.7441, + "num_input_tokens_seen": 185962935, + "step": 5719 + }, + { + "epoch": 0.5158497542498985, + "flos": 21038827733760.0, + "grad_norm": 1.4698120497874685, + "learning_rate": 1.9944500061543945e-06, + "loss": 0.7293, + "num_input_tokens_seen": 185992110, + "step": 5720 + }, + { + "epoch": 0.5159399377733688, + "flos": 28398074440800.0, + "grad_norm": 3.256872686496192, + "learning_rate": 1.99386579802059e-06, + "loss": 0.7785, + "num_input_tokens_seen": 186020465, + "step": 5721 + }, + { + "epoch": 0.5160301212968391, + "flos": 23769227946720.0, + "grad_norm": 1.767988665203509, + "learning_rate": 1.993281590410191e-06, + "loss": 0.7342, + "num_input_tokens_seen": 186048365, + "step": 5722 + }, + { + "epoch": 0.5161203048203094, + "flos": 27990890989920.0, + "grad_norm": 1.712292417767506, + "learning_rate": 1.992697383373043e-06, + "loss": 0.7079, + "num_input_tokens_seen": 186077715, + "step": 5723 + }, + { + "epoch": 0.5162104883437796, + "flos": 63301208406720.0, + "grad_norm": 0.7750939250758673, + "learning_rate": 1.9921131769589937e-06, + "loss": 0.6209, + "num_input_tokens_seen": 186159385, + "step": 5724 + }, + { + "epoch": 0.5163006718672498, + "flos": 22313096273280.0, + "grad_norm": 1.7577442071152147, + "learning_rate": 1.991528971217893e-06, + "loss": 0.7559, + "num_input_tokens_seen": 186188000, + "step": 5725 + }, + { + "epoch": 0.5163908553907202, + "flos": 23553311579040.0, + "grad_norm": 1.6694836493254404, + "learning_rate": 1.9909447661995858e-06, + "loss": 0.7494, + "num_input_tokens_seen": 186216195, + "step": 5726 + }, + { + "epoch": 0.5164810389141904, + "flos": 20200443433440.0, + "grad_norm": 2.633556136431607, + "learning_rate": 1.990360561953922e-06, + "loss": 0.7715, + "num_input_tokens_seen": 186241935, + "step": 5727 + }, + { + "epoch": 0.5165712224376606, + "flos": 27666198703680.0, + "grad_norm": 7.444281555721773, + "learning_rate": 1.9897763585307483e-06, + "loss": 0.7761, + "num_input_tokens_seen": 186272950, + "step": 5728 + }, + { + "epoch": 0.5166614059611309, + "flos": 26428808299680.0, + "grad_norm": 8.726322946018163, + "learning_rate": 1.989192155979912e-06, + "loss": 0.7994, + "num_input_tokens_seen": 186302115, + "step": 5729 + }, + { + "epoch": 0.5167515894846012, + "flos": 28256286041760.0, + "grad_norm": 1.4330689511691095, + "learning_rate": 1.98860795435126e-06, + "loss": 0.7918, + "num_input_tokens_seen": 186336470, + "step": 5730 + }, + { + "epoch": 0.5168417730080714, + "flos": 26904432745920.0, + "grad_norm": 1.670219814201917, + "learning_rate": 1.9880237536946406e-06, + "loss": 0.7623, + "num_input_tokens_seen": 186367655, + "step": 5731 + }, + { + "epoch": 0.5169319565315417, + "flos": 14517816648000.0, + "grad_norm": 2.378705998152561, + "learning_rate": 1.987439554059901e-06, + "loss": 0.7233, + "num_input_tokens_seen": 186393460, + "step": 5732 + }, + { + "epoch": 0.5170221400550119, + "flos": 18634792446240.0, + "grad_norm": 2.0572250419805926, + "learning_rate": 1.9868553554968864e-06, + "loss": 0.7286, + "num_input_tokens_seen": 186420035, + "step": 5733 + }, + { + "epoch": 0.5171123235784822, + "flos": 30474778389120.0, + "grad_norm": 2.487500701597489, + "learning_rate": 1.986271158055447e-06, + "loss": 0.7052, + "num_input_tokens_seen": 186449390, + "step": 5734 + }, + { + "epoch": 0.5172025071019525, + "flos": 21473076503520.0, + "grad_norm": 2.116040922270385, + "learning_rate": 1.9856869617854273e-06, + "loss": 0.7488, + "num_input_tokens_seen": 186476915, + "step": 5735 + }, + { + "epoch": 0.5172926906254227, + "flos": 24092582121600.0, + "grad_norm": 2.7492623640342453, + "learning_rate": 1.9851027667366746e-06, + "loss": 0.7351, + "num_input_tokens_seen": 186504345, + "step": 5736 + }, + { + "epoch": 0.517382874148893, + "flos": 33710953457280.0, + "grad_norm": 2.2160160252304966, + "learning_rate": 1.984518572959037e-06, + "loss": 0.8522, + "num_input_tokens_seen": 186533150, + "step": 5737 + }, + { + "epoch": 0.5174730576723633, + "flos": 22601988347040.0, + "grad_norm": 1.91519797763399, + "learning_rate": 1.9839343805023587e-06, + "loss": 0.6815, + "num_input_tokens_seen": 186560800, + "step": 5738 + }, + { + "epoch": 0.5175632411958335, + "flos": 19181385431520.0, + "grad_norm": 5.4610567748396885, + "learning_rate": 1.9833501894164886e-06, + "loss": 0.8459, + "num_input_tokens_seen": 186587630, + "step": 5739 + }, + { + "epoch": 0.5176534247193038, + "flos": 19647494419200.0, + "grad_norm": 2.0909610221656023, + "learning_rate": 1.982765999751273e-06, + "loss": 0.7917, + "num_input_tokens_seen": 186614215, + "step": 5740 + }, + { + "epoch": 0.5177436082427741, + "flos": 21876505808640.0, + "grad_norm": 2.165665676511084, + "learning_rate": 1.9821818115565553e-06, + "loss": 0.7866, + "num_input_tokens_seen": 186639805, + "step": 5741 + }, + { + "epoch": 0.5178337917662443, + "flos": 26431558861920.0, + "grad_norm": 1.9777561247577324, + "learning_rate": 1.9815976248821853e-06, + "loss": 0.6644, + "num_input_tokens_seen": 186668245, + "step": 5742 + }, + { + "epoch": 0.5179239752897146, + "flos": 28325879299680.0, + "grad_norm": 2.845318037066448, + "learning_rate": 1.981013439778007e-06, + "loss": 0.7671, + "num_input_tokens_seen": 186697915, + "step": 5743 + }, + { + "epoch": 0.5180141588131848, + "flos": 19764075987360.0, + "grad_norm": 2.200098759428543, + "learning_rate": 1.9804292562938666e-06, + "loss": 0.7618, + "num_input_tokens_seen": 186725390, + "step": 5744 + }, + { + "epoch": 0.5181043423366551, + "flos": 23225645712000.0, + "grad_norm": 2.17812058654951, + "learning_rate": 1.97984507447961e-06, + "loss": 0.7913, + "num_input_tokens_seen": 186754705, + "step": 5745 + }, + { + "epoch": 0.5181945258601254, + "flos": 23363345437440.0, + "grad_norm": 2.3032889567883434, + "learning_rate": 1.9792608943850824e-06, + "loss": 0.6959, + "num_input_tokens_seen": 186782745, + "step": 5746 + }, + { + "epoch": 0.5182847093835956, + "flos": 24609991111680.0, + "grad_norm": 1.8316336975161904, + "learning_rate": 1.9786767160601305e-06, + "loss": 0.7358, + "num_input_tokens_seen": 186811825, + "step": 5747 + }, + { + "epoch": 0.5183748929070658, + "flos": 18744311760000.0, + "grad_norm": 1.7400617417529485, + "learning_rate": 1.9780925395545977e-06, + "loss": 0.6869, + "num_input_tokens_seen": 186838885, + "step": 5748 + }, + { + "epoch": 0.5184650764305362, + "flos": 68368689702240.0, + "grad_norm": 0.6391297078875978, + "learning_rate": 1.9775083649183306e-06, + "loss": 0.5664, + "num_input_tokens_seen": 186938240, + "step": 5749 + }, + { + "epoch": 0.5185552599540064, + "flos": 18926360742720.0, + "grad_norm": 2.2226595752872202, + "learning_rate": 1.976924192201174e-06, + "loss": 0.7957, + "num_input_tokens_seen": 186962980, + "step": 5750 + }, + { + "epoch": 0.5186454434774767, + "flos": 27665789836320.0, + "grad_norm": 1.8768945023891872, + "learning_rate": 1.9763400214529723e-06, + "loss": 0.6853, + "num_input_tokens_seen": 186992880, + "step": 5751 + }, + { + "epoch": 0.5187356270009469, + "flos": 24245335599840.0, + "grad_norm": 2.0173749793738307, + "learning_rate": 1.9757558527235713e-06, + "loss": 0.6272, + "num_input_tokens_seen": 187023420, + "step": 5752 + }, + { + "epoch": 0.5188258105244172, + "flos": 23662161837120.0, + "grad_norm": 2.401343557668916, + "learning_rate": 1.9751716860628136e-06, + "loss": 0.7427, + "num_input_tokens_seen": 187051580, + "step": 5753 + }, + { + "epoch": 0.5189159940478875, + "flos": 29341183155840.0, + "grad_norm": 1.9835554217631668, + "learning_rate": 1.974587521520545e-06, + "loss": 0.7815, + "num_input_tokens_seen": 187082360, + "step": 5754 + }, + { + "epoch": 0.5190061775713577, + "flos": 31236023970240.0, + "grad_norm": 2.914737899277459, + "learning_rate": 1.9740033591466088e-06, + "loss": 0.7895, + "num_input_tokens_seen": 187112605, + "step": 5755 + }, + { + "epoch": 0.5190963610948279, + "flos": 24645196608000.0, + "grad_norm": 2.1855173117485007, + "learning_rate": 1.97341919899085e-06, + "loss": 0.8671, + "num_input_tokens_seen": 187142385, + "step": 5756 + }, + { + "epoch": 0.5191865446182983, + "flos": 25664068761120.0, + "grad_norm": 1.821327810412216, + "learning_rate": 1.9728350411031114e-06, + "loss": 0.7817, + "num_input_tokens_seen": 187171080, + "step": 5757 + }, + { + "epoch": 0.5192767281417685, + "flos": 22423693510080.0, + "grad_norm": 3.423111226331168, + "learning_rate": 1.9722508855332367e-06, + "loss": 0.8317, + "num_input_tokens_seen": 187199310, + "step": 5758 + }, + { + "epoch": 0.5193669116652387, + "flos": 33607418474880.0, + "grad_norm": 1.5911712727695624, + "learning_rate": 1.97166673233107e-06, + "loss": 0.6892, + "num_input_tokens_seen": 187230240, + "step": 5759 + }, + { + "epoch": 0.519457095188709, + "flos": 30729468550080.0, + "grad_norm": 1.7123396192755462, + "learning_rate": 1.971082581546453e-06, + "loss": 0.7967, + "num_input_tokens_seen": 187261290, + "step": 5760 + }, + { + "epoch": 0.5195472787121793, + "flos": 19142611638240.0, + "grad_norm": 2.5350747117388672, + "learning_rate": 1.9704984332292306e-06, + "loss": 0.7831, + "num_input_tokens_seen": 187288870, + "step": 5761 + }, + { + "epoch": 0.5196374622356495, + "flos": 18598471857120.0, + "grad_norm": 2.269648064711021, + "learning_rate": 1.9699142874292444e-06, + "loss": 0.8465, + "num_input_tokens_seen": 187314065, + "step": 5762 + }, + { + "epoch": 0.5197276457591198, + "flos": 22933557038880.0, + "grad_norm": 1.9890892320681022, + "learning_rate": 1.969330144196336e-06, + "loss": 0.8451, + "num_input_tokens_seen": 187342340, + "step": 5763 + }, + { + "epoch": 0.51981782928259, + "flos": 22424139547200.0, + "grad_norm": 2.8317052614105163, + "learning_rate": 1.9687460035803497e-06, + "loss": 0.7028, + "num_input_tokens_seen": 187370290, + "step": 5764 + }, + { + "epoch": 0.5199080128060604, + "flos": 17286433107840.0, + "grad_norm": 2.2686086016852665, + "learning_rate": 1.9681618656311265e-06, + "loss": 0.7435, + "num_input_tokens_seen": 187396490, + "step": 5765 + }, + { + "epoch": 0.5199981963295306, + "flos": 15319657340640.0, + "grad_norm": 2.830886420005867, + "learning_rate": 1.9675777303985086e-06, + "loss": 0.669, + "num_input_tokens_seen": 187423965, + "step": 5766 + }, + { + "epoch": 0.5200883798530008, + "flos": 19797348656160.0, + "grad_norm": 2.07235593895591, + "learning_rate": 1.9669935979323376e-06, + "loss": 0.7877, + "num_input_tokens_seen": 187451870, + "step": 5767 + }, + { + "epoch": 0.5201785633764712, + "flos": 24973717379520.0, + "grad_norm": 2.650807358455504, + "learning_rate": 1.9664094682824545e-06, + "loss": 0.8139, + "num_input_tokens_seen": 187480400, + "step": 5768 + }, + { + "epoch": 0.5202687468999414, + "flos": 21039496789440.0, + "grad_norm": 1.6663689903818792, + "learning_rate": 1.965825341498701e-06, + "loss": 0.7205, + "num_input_tokens_seen": 187508500, + "step": 5769 + }, + { + "epoch": 0.5203589304234116, + "flos": 21510177657600.0, + "grad_norm": 3.144455125728419, + "learning_rate": 1.9652412176309177e-06, + "loss": 0.8308, + "num_input_tokens_seen": 187536400, + "step": 5770 + }, + { + "epoch": 0.5204491139468819, + "flos": 32839184978880.0, + "grad_norm": 1.8599958264909087, + "learning_rate": 1.9646570967289453e-06, + "loss": 0.6221, + "num_input_tokens_seen": 187568730, + "step": 5771 + }, + { + "epoch": 0.5205392974703522, + "flos": 21471664052640.0, + "grad_norm": 4.193960188531632, + "learning_rate": 1.9640729788426246e-06, + "loss": 0.7944, + "num_input_tokens_seen": 187596645, + "step": 5772 + }, + { + "epoch": 0.5206294809938224, + "flos": 22130489744160.0, + "grad_norm": 1.966565752134327, + "learning_rate": 1.963488864021795e-06, + "loss": 0.7855, + "num_input_tokens_seen": 187627925, + "step": 5773 + }, + { + "epoch": 0.5207196645172927, + "flos": 29738553790080.0, + "grad_norm": 2.0802752620784646, + "learning_rate": 1.962904752316298e-06, + "loss": 0.6511, + "num_input_tokens_seen": 187658630, + "step": 5774 + }, + { + "epoch": 0.5208098480407629, + "flos": 13715604257760.0, + "grad_norm": 2.5284316774008224, + "learning_rate": 1.9623206437759706e-06, + "loss": 0.6864, + "num_input_tokens_seen": 187685645, + "step": 5775 + }, + { + "epoch": 0.5209000315642333, + "flos": 25046470067040.0, + "grad_norm": 1.8676582285735646, + "learning_rate": 1.9617365384506545e-06, + "loss": 0.7171, + "num_input_tokens_seen": 187715155, + "step": 5776 + }, + { + "epoch": 0.5209902150877035, + "flos": 29162479451520.0, + "grad_norm": 1.505138021584657, + "learning_rate": 1.9611524363901872e-06, + "loss": 0.7428, + "num_input_tokens_seen": 187745305, + "step": 5777 + }, + { + "epoch": 0.5210803986111737, + "flos": 20855291960640.0, + "grad_norm": 2.0431966563320656, + "learning_rate": 1.960568337644409e-06, + "loss": 0.7504, + "num_input_tokens_seen": 187773725, + "step": 5778 + }, + { + "epoch": 0.521170582134644, + "flos": 23842835538720.0, + "grad_norm": 1.960383667568472, + "learning_rate": 1.9599842422631576e-06, + "loss": 0.7383, + "num_input_tokens_seen": 187804905, + "step": 5779 + }, + { + "epoch": 0.5212607656581143, + "flos": 32078311095360.0, + "grad_norm": 2.128852605581298, + "learning_rate": 1.9594001502962703e-06, + "loss": 0.7466, + "num_input_tokens_seen": 187834735, + "step": 5780 + }, + { + "epoch": 0.5213509491815845, + "flos": 24755756675040.0, + "grad_norm": 3.0941160351547037, + "learning_rate": 1.9588160617935868e-06, + "loss": 0.8152, + "num_input_tokens_seen": 187864330, + "step": 5781 + }, + { + "epoch": 0.5214411327050548, + "flos": 26758741522080.0, + "grad_norm": 3.451830452999875, + "learning_rate": 1.958231976804944e-06, + "loss": 0.8441, + "num_input_tokens_seen": 187890590, + "step": 5782 + }, + { + "epoch": 0.521531316228525, + "flos": 21258015040320.0, + "grad_norm": 2.3607947517830485, + "learning_rate": 1.957647895380179e-06, + "loss": 0.6802, + "num_input_tokens_seen": 187918565, + "step": 5783 + }, + { + "epoch": 0.5216214997519953, + "flos": 48073117385760.0, + "grad_norm": 1.722026337828681, + "learning_rate": 1.9570638175691297e-06, + "loss": 0.6832, + "num_input_tokens_seen": 187956435, + "step": 5784 + }, + { + "epoch": 0.5217116832754656, + "flos": 24135927795360.0, + "grad_norm": 2.374851422932782, + "learning_rate": 1.956479743421632e-06, + "loss": 0.7987, + "num_input_tokens_seen": 187984535, + "step": 5785 + }, + { + "epoch": 0.5218018667989358, + "flos": 19072349324640.0, + "grad_norm": 2.738385643377179, + "learning_rate": 1.955895672987522e-06, + "loss": 0.689, + "num_input_tokens_seen": 188011020, + "step": 5786 + }, + { + "epoch": 0.521892050322406, + "flos": 27634784522880.0, + "grad_norm": 1.615330611363751, + "learning_rate": 1.9553116063166367e-06, + "loss": 0.661, + "num_input_tokens_seen": 188043770, + "step": 5787 + }, + { + "epoch": 0.5219822338458764, + "flos": 30296557891680.0, + "grad_norm": 1.5637946296745069, + "learning_rate": 1.954727543458812e-06, + "loss": 0.6614, + "num_input_tokens_seen": 188078530, + "step": 5788 + }, + { + "epoch": 0.5220724173693466, + "flos": 22856938696320.0, + "grad_norm": 2.2096829039296995, + "learning_rate": 1.954143484463883e-06, + "loss": 0.7839, + "num_input_tokens_seen": 188107995, + "step": 5789 + }, + { + "epoch": 0.5221626008928169, + "flos": 20927189743680.0, + "grad_norm": 2.5825414571914993, + "learning_rate": 1.9535594293816836e-06, + "loss": 0.7394, + "num_input_tokens_seen": 188135125, + "step": 5790 + }, + { + "epoch": 0.5222527844162872, + "flos": 14736446408160.0, + "grad_norm": 3.660825686471028, + "learning_rate": 1.952975378262051e-06, + "loss": 0.7016, + "num_input_tokens_seen": 188162510, + "step": 5791 + }, + { + "epoch": 0.5223429679397574, + "flos": 17213829099360.0, + "grad_norm": 2.8808868510342287, + "learning_rate": 1.952391331154817e-06, + "loss": 0.7067, + "num_input_tokens_seen": 188188695, + "step": 5792 + }, + { + "epoch": 0.5224331514632277, + "flos": 26025602013120.0, + "grad_norm": 1.7425629078744274, + "learning_rate": 1.9518072881098185e-06, + "loss": 0.8122, + "num_input_tokens_seen": 188216970, + "step": 5793 + }, + { + "epoch": 0.5225233349866979, + "flos": 27343699433280.0, + "grad_norm": 4.918837999964774, + "learning_rate": 1.9512232491768867e-06, + "loss": 0.7446, + "num_input_tokens_seen": 188251140, + "step": 5794 + }, + { + "epoch": 0.5226135185101682, + "flos": 28145131258560.0, + "grad_norm": 1.5368538126623335, + "learning_rate": 1.9506392144058573e-06, + "loss": 0.7457, + "num_input_tokens_seen": 188284345, + "step": 5795 + }, + { + "epoch": 0.5227037020336385, + "flos": 20018766148320.0, + "grad_norm": 1.7718583525021934, + "learning_rate": 1.9500551838465623e-06, + "loss": 0.7475, + "num_input_tokens_seen": 188310420, + "step": 5796 + }, + { + "epoch": 0.5227938855571087, + "flos": 32439249631200.0, + "grad_norm": 1.6057675865477687, + "learning_rate": 1.9494711575488337e-06, + "loss": 0.7673, + "num_input_tokens_seen": 188341290, + "step": 5797 + }, + { + "epoch": 0.5228840690805789, + "flos": 18962123785440.0, + "grad_norm": 2.3089559048307704, + "learning_rate": 1.948887135562505e-06, + "loss": 0.6978, + "num_input_tokens_seen": 188368170, + "step": 5798 + }, + { + "epoch": 0.5229742526040493, + "flos": 34446769188960.0, + "grad_norm": 1.7572850092875023, + "learning_rate": 1.9483031179374074e-06, + "loss": 0.8011, + "num_input_tokens_seen": 188400240, + "step": 5799 + }, + { + "epoch": 0.5230644361275195, + "flos": 23003782182720.0, + "grad_norm": 1.6699430765512664, + "learning_rate": 1.9477191047233736e-06, + "loss": 0.8115, + "num_input_tokens_seen": 188429125, + "step": 5800 + }, + { + "epoch": 0.5231546196509897, + "flos": 33352505295360.0, + "grad_norm": 1.9964982013078092, + "learning_rate": 1.9471350959702334e-06, + "loss": 0.6367, + "num_input_tokens_seen": 188460810, + "step": 5801 + }, + { + "epoch": 0.52324480317446, + "flos": 21257345984640.0, + "grad_norm": 2.405318240802779, + "learning_rate": 1.9465510917278184e-06, + "loss": 0.76, + "num_input_tokens_seen": 188489630, + "step": 5802 + }, + { + "epoch": 0.5233349866979303, + "flos": 21877137694560.0, + "grad_norm": 1.8893089442894315, + "learning_rate": 1.9459670920459593e-06, + "loss": 0.7502, + "num_input_tokens_seen": 188517065, + "step": 5803 + }, + { + "epoch": 0.5234251702214006, + "flos": 50396743015200.0, + "grad_norm": 2.1712551192869496, + "learning_rate": 1.945383096974485e-06, + "loss": 0.6702, + "num_input_tokens_seen": 188548590, + "step": 5804 + }, + { + "epoch": 0.5235153537448708, + "flos": 36013014892320.0, + "grad_norm": 2.674669288085662, + "learning_rate": 1.944799106563227e-06, + "loss": 0.6746, + "num_input_tokens_seen": 188580005, + "step": 5805 + }, + { + "epoch": 0.523605537268341, + "flos": 26977334112480.0, + "grad_norm": 2.193236105521471, + "learning_rate": 1.9442151208620133e-06, + "loss": 0.7914, + "num_input_tokens_seen": 188606660, + "step": 5806 + }, + { + "epoch": 0.5236957207918114, + "flos": 22715187467040.0, + "grad_norm": 2.05863405365551, + "learning_rate": 1.943631139920672e-06, + "loss": 0.8026, + "num_input_tokens_seen": 188633975, + "step": 5807 + }, + { + "epoch": 0.5237859043152816, + "flos": 23189027764800.0, + "grad_norm": 2.142303922176007, + "learning_rate": 1.943047163789034e-06, + "loss": 0.7028, + "num_input_tokens_seen": 188662120, + "step": 5808 + }, + { + "epoch": 0.5238760878387518, + "flos": 30437900253600.0, + "grad_norm": 1.564427826828072, + "learning_rate": 1.942463192516925e-06, + "loss": 0.7104, + "num_input_tokens_seen": 188695885, + "step": 5809 + }, + { + "epoch": 0.5239662713622221, + "flos": 22712474074560.0, + "grad_norm": 1.8683923640639644, + "learning_rate": 1.9418792261541746e-06, + "loss": 0.7713, + "num_input_tokens_seen": 188726495, + "step": 5810 + }, + { + "epoch": 0.5240564548856924, + "flos": 17760124726560.0, + "grad_norm": 2.0094456543076666, + "learning_rate": 1.9412952647506094e-06, + "loss": 0.7079, + "num_input_tokens_seen": 188754070, + "step": 5811 + }, + { + "epoch": 0.5241466384091626, + "flos": 23698928463360.0, + "grad_norm": 2.117842681971405, + "learning_rate": 1.9407113083560552e-06, + "loss": 0.7285, + "num_input_tokens_seen": 188784965, + "step": 5812 + }, + { + "epoch": 0.5242368219326329, + "flos": 24317679420000.0, + "grad_norm": 2.3268864556624087, + "learning_rate": 1.940127357020339e-06, + "loss": 0.7345, + "num_input_tokens_seen": 188812945, + "step": 5813 + }, + { + "epoch": 0.5243270054561032, + "flos": 15647471886720.0, + "grad_norm": 1.8725637502984738, + "learning_rate": 1.939543410793287e-06, + "loss": 0.76, + "num_input_tokens_seen": 188840230, + "step": 5814 + }, + { + "epoch": 0.5244171889795735, + "flos": 19799839030080.0, + "grad_norm": 1.8400194606581164, + "learning_rate": 1.9389594697247246e-06, + "loss": 0.7888, + "num_input_tokens_seen": 188868210, + "step": 5815 + }, + { + "epoch": 0.5245073725030437, + "flos": 23005640670720.0, + "grad_norm": 2.2443427915288368, + "learning_rate": 1.9383755338644763e-06, + "loss": 0.8278, + "num_input_tokens_seen": 188895785, + "step": 5816 + }, + { + "epoch": 0.5245975560265139, + "flos": 18196938209760.0, + "grad_norm": 2.1522545770079593, + "learning_rate": 1.937791603262368e-06, + "loss": 0.7809, + "num_input_tokens_seen": 188921830, + "step": 5817 + }, + { + "epoch": 0.5246877395499843, + "flos": 23439443403360.0, + "grad_norm": 1.8124593532871613, + "learning_rate": 1.9372076779682235e-06, + "loss": 0.7037, + "num_input_tokens_seen": 188949075, + "step": 5818 + }, + { + "epoch": 0.5247779230734545, + "flos": 22198187344320.0, + "grad_norm": 2.4293522741512454, + "learning_rate": 1.9366237580318648e-06, + "loss": 0.7043, + "num_input_tokens_seen": 188977620, + "step": 5819 + }, + { + "epoch": 0.5248681065969247, + "flos": 19837757918880.0, + "grad_norm": 1.5566025637157923, + "learning_rate": 1.9360398435031176e-06, + "loss": 0.7579, + "num_input_tokens_seen": 189005385, + "step": 5820 + }, + { + "epoch": 0.524958290120395, + "flos": 24318385645440.0, + "grad_norm": 3.1518359147461794, + "learning_rate": 1.9354559344318025e-06, + "loss": 0.7602, + "num_input_tokens_seen": 189034875, + "step": 5821 + }, + { + "epoch": 0.5250484736438653, + "flos": 19253915100480.0, + "grad_norm": 2.044658946392226, + "learning_rate": 1.934872030867744e-06, + "loss": 0.805, + "num_input_tokens_seen": 189060820, + "step": 5822 + }, + { + "epoch": 0.5251386571673355, + "flos": 24245893146240.0, + "grad_norm": 3.027686157551329, + "learning_rate": 1.934288132860763e-06, + "loss": 0.7637, + "num_input_tokens_seen": 189090120, + "step": 5823 + }, + { + "epoch": 0.5252288406908058, + "flos": 25338744588960.0, + "grad_norm": 1.656441947322774, + "learning_rate": 1.93370424046068e-06, + "loss": 0.7504, + "num_input_tokens_seen": 189118475, + "step": 5824 + }, + { + "epoch": 0.525319024214276, + "flos": 20019732562080.0, + "grad_norm": 1.7966516107106478, + "learning_rate": 1.9331203537173177e-06, + "loss": 0.8133, + "num_input_tokens_seen": 189147845, + "step": 5825 + }, + { + "epoch": 0.5254092077377464, + "flos": 29928185403840.0, + "grad_norm": 1.4587372719962075, + "learning_rate": 1.9325364726804947e-06, + "loss": 0.8635, + "num_input_tokens_seen": 189179670, + "step": 5826 + }, + { + "epoch": 0.5254993912612166, + "flos": 17684509967520.0, + "grad_norm": 2.2818653896592274, + "learning_rate": 1.9319525974000327e-06, + "loss": 0.6802, + "num_input_tokens_seen": 189206735, + "step": 5827 + }, + { + "epoch": 0.5255895747846868, + "flos": 13927729309920.0, + "grad_norm": 3.246268551156171, + "learning_rate": 1.93136872792575e-06, + "loss": 0.8161, + "num_input_tokens_seen": 189231250, + "step": 5828 + }, + { + "epoch": 0.525679758308157, + "flos": 22678532350080.0, + "grad_norm": 2.2825641765952343, + "learning_rate": 1.9307848643074653e-06, + "loss": 0.7052, + "num_input_tokens_seen": 189261090, + "step": 5829 + }, + { + "epoch": 0.5257699418316274, + "flos": 21805797457920.0, + "grad_norm": 1.7566076600059009, + "learning_rate": 1.9302010065949985e-06, + "loss": 0.7845, + "num_input_tokens_seen": 189290970, + "step": 5830 + }, + { + "epoch": 0.5258601253550976, + "flos": 26576618199840.0, + "grad_norm": 1.7403168492837386, + "learning_rate": 1.9296171548381657e-06, + "loss": 0.7121, + "num_input_tokens_seen": 189320125, + "step": 5831 + }, + { + "epoch": 0.5259503088785679, + "flos": 24023769428640.0, + "grad_norm": 2.195870868089408, + "learning_rate": 1.9290333090867862e-06, + "loss": 0.7389, + "num_input_tokens_seen": 189348875, + "step": 5832 + }, + { + "epoch": 0.5260404924020381, + "flos": 18452446105440.0, + "grad_norm": 3.7700832456914477, + "learning_rate": 1.928449469390676e-06, + "loss": 0.7543, + "num_input_tokens_seen": 189373740, + "step": 5833 + }, + { + "epoch": 0.5261306759255084, + "flos": 20857447806720.0, + "grad_norm": 1.8997800498942172, + "learning_rate": 1.927865635799651e-06, + "loss": 0.7235, + "num_input_tokens_seen": 189401800, + "step": 5834 + }, + { + "epoch": 0.5262208594489787, + "flos": 26136013401120.0, + "grad_norm": 1.7530506655459417, + "learning_rate": 1.927281808363528e-06, + "loss": 0.6774, + "num_input_tokens_seen": 189430985, + "step": 5835 + }, + { + "epoch": 0.5263110429724489, + "flos": 28543691325120.0, + "grad_norm": 1.9684665415320692, + "learning_rate": 1.9266979871321216e-06, + "loss": 0.8274, + "num_input_tokens_seen": 189460745, + "step": 5836 + }, + { + "epoch": 0.5264012264959192, + "flos": 23183935507680.0, + "grad_norm": 2.0720948806012403, + "learning_rate": 1.9261141721552482e-06, + "loss": 0.7024, + "num_input_tokens_seen": 189488320, + "step": 5837 + }, + { + "epoch": 0.5264914100193895, + "flos": 18489398580480.0, + "grad_norm": 3.0552398884640812, + "learning_rate": 1.9255303634827204e-06, + "loss": 0.8187, + "num_input_tokens_seen": 189515905, + "step": 5838 + }, + { + "epoch": 0.5265815935428597, + "flos": 19798203560640.0, + "grad_norm": 1.9997700320591432, + "learning_rate": 1.924946561164352e-06, + "loss": 0.8186, + "num_input_tokens_seen": 189542705, + "step": 5839 + }, + { + "epoch": 0.52667177706633, + "flos": 27228976353120.0, + "grad_norm": 1.9115569862733963, + "learning_rate": 1.9243627652499582e-06, + "loss": 0.7025, + "num_input_tokens_seen": 189573370, + "step": 5840 + }, + { + "epoch": 0.5267619605898003, + "flos": 21401104380960.0, + "grad_norm": 1.8135877830898224, + "learning_rate": 1.9237789757893493e-06, + "loss": 0.8154, + "num_input_tokens_seen": 189601630, + "step": 5841 + }, + { + "epoch": 0.5268521441132705, + "flos": 14700051479520.0, + "grad_norm": 2.196187778171194, + "learning_rate": 1.9231951928323395e-06, + "loss": 0.7984, + "num_input_tokens_seen": 189626370, + "step": 5842 + }, + { + "epoch": 0.5269423276367408, + "flos": 25374470461920.0, + "grad_norm": 2.0414168406115807, + "learning_rate": 1.922611416428738e-06, + "loss": 0.7049, + "num_input_tokens_seen": 189654325, + "step": 5843 + }, + { + "epoch": 0.527032511160211, + "flos": 21360174741600.0, + "grad_norm": 2.1269442695594467, + "learning_rate": 1.922027646628358e-06, + "loss": 0.7775, + "num_input_tokens_seen": 189681730, + "step": 5844 + }, + { + "epoch": 0.5271226946836813, + "flos": 58737420459840.0, + "grad_norm": 0.6592752345032615, + "learning_rate": 1.9214438834810092e-06, + "loss": 0.6344, + "num_input_tokens_seen": 189768575, + "step": 5845 + }, + { + "epoch": 0.5272128782071516, + "flos": 21292700160000.0, + "grad_norm": 1.6172396886393352, + "learning_rate": 1.9208601270365008e-06, + "loss": 0.6813, + "num_input_tokens_seen": 189797560, + "step": 5846 + }, + { + "epoch": 0.5273030617306218, + "flos": 20638780876800.0, + "grad_norm": 1.7220304913939257, + "learning_rate": 1.9202763773446435e-06, + "loss": 0.8323, + "num_input_tokens_seen": 189826020, + "step": 5847 + }, + { + "epoch": 0.527393245254092, + "flos": 25156472587680.0, + "grad_norm": 2.0895008982418894, + "learning_rate": 1.9196926344552444e-06, + "loss": 0.6505, + "num_input_tokens_seen": 189853750, + "step": 5848 + }, + { + "epoch": 0.5274834287775624, + "flos": 25259264174880.0, + "grad_norm": 4.202629201923756, + "learning_rate": 1.919108898418113e-06, + "loss": 0.7786, + "num_input_tokens_seen": 189883910, + "step": 5849 + }, + { + "epoch": 0.5275736123010326, + "flos": 32953164663840.0, + "grad_norm": 1.4914630509396185, + "learning_rate": 1.918525169283057e-06, + "loss": 0.6482, + "num_input_tokens_seen": 189919365, + "step": 5850 + }, + { + "epoch": 0.5276637958245028, + "flos": 24496903500960.0, + "grad_norm": 3.018652152614051, + "learning_rate": 1.9179414470998817e-06, + "loss": 0.7762, + "num_input_tokens_seen": 189950805, + "step": 5851 + }, + { + "epoch": 0.5277539793479731, + "flos": 33640988501760.0, + "grad_norm": 2.874090465473059, + "learning_rate": 1.917357731918395e-06, + "loss": 0.711, + "num_input_tokens_seen": 189981365, + "step": 5852 + }, + { + "epoch": 0.5278441628714434, + "flos": 21141433472160.0, + "grad_norm": 1.6945000033138002, + "learning_rate": 1.9167740237884025e-06, + "loss": 0.8812, + "num_input_tokens_seen": 190007175, + "step": 5853 + }, + { + "epoch": 0.5279343463949137, + "flos": 21726354213600.0, + "grad_norm": 1.8383089373927592, + "learning_rate": 1.916190322759709e-06, + "loss": 0.7923, + "num_input_tokens_seen": 190034975, + "step": 5854 + }, + { + "epoch": 0.5280245299183839, + "flos": 23043410880480.0, + "grad_norm": 2.379395984169344, + "learning_rate": 1.91560662888212e-06, + "loss": 0.8519, + "num_input_tokens_seen": 190063040, + "step": 5855 + }, + { + "epoch": 0.5281147134418541, + "flos": 23071888650240.0, + "grad_norm": 1.8330502797299884, + "learning_rate": 1.915022942205438e-06, + "loss": 0.8004, + "num_input_tokens_seen": 190089630, + "step": 5856 + }, + { + "epoch": 0.5282048969653245, + "flos": 13897950598560.0, + "grad_norm": 2.2646181171159143, + "learning_rate": 1.914439262779468e-06, + "loss": 0.745, + "num_input_tokens_seen": 190114820, + "step": 5857 + }, + { + "epoch": 0.5282950804887947, + "flos": 21286827337920.0, + "grad_norm": 2.2727384246205777, + "learning_rate": 1.9138555906540103e-06, + "loss": 0.6859, + "num_input_tokens_seen": 190141440, + "step": 5858 + }, + { + "epoch": 0.5283852640122649, + "flos": 24682297762080.0, + "grad_norm": 1.6695924668252253, + "learning_rate": 1.91327192587887e-06, + "loss": 0.6607, + "num_input_tokens_seen": 190171425, + "step": 5859 + }, + { + "epoch": 0.5284754475357353, + "flos": 27159866302080.0, + "grad_norm": 2.1823933531408244, + "learning_rate": 1.912688268503846e-06, + "loss": 0.7278, + "num_input_tokens_seen": 190200040, + "step": 5860 + }, + { + "epoch": 0.5285656310592055, + "flos": 16740174650400.0, + "grad_norm": 2.4068119334217473, + "learning_rate": 1.912104618578741e-06, + "loss": 0.7146, + "num_input_tokens_seen": 190226500, + "step": 5861 + }, + { + "epoch": 0.5286558145826757, + "flos": 30327897732960.0, + "grad_norm": 1.6872957600515144, + "learning_rate": 1.9115209761533554e-06, + "loss": 0.7327, + "num_input_tokens_seen": 190260250, + "step": 5862 + }, + { + "epoch": 0.528745998106146, + "flos": 24537275593920.0, + "grad_norm": 2.4201927002723185, + "learning_rate": 1.9109373412774863e-06, + "loss": 0.6665, + "num_input_tokens_seen": 190288440, + "step": 5863 + }, + { + "epoch": 0.5288361816296163, + "flos": 24134032137600.0, + "grad_norm": 1.9225164444000595, + "learning_rate": 1.910353714000936e-06, + "loss": 0.7569, + "num_input_tokens_seen": 190318040, + "step": 5864 + }, + { + "epoch": 0.5289263651530866, + "flos": 28504025457600.0, + "grad_norm": 3.0949760088506872, + "learning_rate": 1.9097700943734997e-06, + "loss": 0.7103, + "num_input_tokens_seen": 190349635, + "step": 5865 + }, + { + "epoch": 0.5290165486765568, + "flos": 14518076836320.0, + "grad_norm": 2.11407681366557, + "learning_rate": 1.909186482444977e-06, + "loss": 0.6531, + "num_input_tokens_seen": 190376635, + "step": 5866 + }, + { + "epoch": 0.529106732200027, + "flos": 22678829708160.0, + "grad_norm": 2.0973282030062648, + "learning_rate": 1.9086028782651652e-06, + "loss": 0.6568, + "num_input_tokens_seen": 190405415, + "step": 5867 + }, + { + "epoch": 0.5291969157234974, + "flos": 19871513794560.0, + "grad_norm": 1.7687485334302666, + "learning_rate": 1.908019281883859e-06, + "loss": 0.7477, + "num_input_tokens_seen": 190432355, + "step": 5868 + }, + { + "epoch": 0.5292870992469676, + "flos": 25229002256640.0, + "grad_norm": 1.8492895411747692, + "learning_rate": 1.9074356933508545e-06, + "loss": 0.7437, + "num_input_tokens_seen": 190461165, + "step": 5869 + }, + { + "epoch": 0.5293772827704378, + "flos": 18335418500160.0, + "grad_norm": 2.8270919346145598, + "learning_rate": 1.9068521127159477e-06, + "loss": 0.7805, + "num_input_tokens_seen": 190484880, + "step": 5870 + }, + { + "epoch": 0.5294674662939081, + "flos": 24318385645440.0, + "grad_norm": 1.7413564846891054, + "learning_rate": 1.9062685400289322e-06, + "loss": 0.6985, + "num_input_tokens_seen": 190514035, + "step": 5871 + }, + { + "epoch": 0.5295576498173784, + "flos": 20311338028320.0, + "grad_norm": 3.1975694068134253, + "learning_rate": 1.9056849753396018e-06, + "loss": 0.7251, + "num_input_tokens_seen": 190542400, + "step": 5872 + }, + { + "epoch": 0.5296478333408486, + "flos": 23006941612320.0, + "grad_norm": 2.1378323792926106, + "learning_rate": 1.9051014186977485e-06, + "loss": 0.7838, + "num_input_tokens_seen": 190571505, + "step": 5873 + }, + { + "epoch": 0.5297380168643189, + "flos": 23438179631520.0, + "grad_norm": 7.058996631332396, + "learning_rate": 1.9045178701531664e-06, + "loss": 0.7501, + "num_input_tokens_seen": 190600755, + "step": 5874 + }, + { + "epoch": 0.5298282003877891, + "flos": 22495739972160.0, + "grad_norm": 1.5284508792570057, + "learning_rate": 1.903934329755645e-06, + "loss": 0.7455, + "num_input_tokens_seen": 190631375, + "step": 5875 + }, + { + "epoch": 0.5299183839112594, + "flos": 26977148263680.0, + "grad_norm": 1.945527765703855, + "learning_rate": 1.9033507975549775e-06, + "loss": 0.7873, + "num_input_tokens_seen": 190660420, + "step": 5876 + }, + { + "epoch": 0.5300085674347297, + "flos": 19689278963040.0, + "grad_norm": 1.6342961629667874, + "learning_rate": 1.9027672736009525e-06, + "loss": 0.7424, + "num_input_tokens_seen": 190687300, + "step": 5877 + }, + { + "epoch": 0.5300987509581999, + "flos": 30365110396320.0, + "grad_norm": 1.657227785360664, + "learning_rate": 1.9021837579433593e-06, + "loss": 0.6985, + "num_input_tokens_seen": 190717905, + "step": 5878 + }, + { + "epoch": 0.5301889344816701, + "flos": 35722227160800.0, + "grad_norm": 2.0048761530271766, + "learning_rate": 1.901600250631988e-06, + "loss": 0.7244, + "num_input_tokens_seen": 190749420, + "step": 5879 + }, + { + "epoch": 0.5302791180051405, + "flos": 23695285826880.0, + "grad_norm": 2.0658934449065876, + "learning_rate": 1.901016751716625e-06, + "loss": 0.7382, + "num_input_tokens_seen": 190778015, + "step": 5880 + }, + { + "epoch": 0.5303693015286107, + "flos": 23515764387840.0, + "grad_norm": 1.5561095208798448, + "learning_rate": 1.9004332612470593e-06, + "loss": 0.715, + "num_input_tokens_seen": 190808170, + "step": 5881 + }, + { + "epoch": 0.530459485052081, + "flos": 23697999219360.0, + "grad_norm": 1.3130033072976894, + "learning_rate": 1.8998497792730763e-06, + "loss": 0.7734, + "num_input_tokens_seen": 190838515, + "step": 5882 + }, + { + "epoch": 0.5305496685755512, + "flos": 25229076596160.0, + "grad_norm": 1.7110936341831258, + "learning_rate": 1.8992663058444629e-06, + "loss": 0.8023, + "num_input_tokens_seen": 190868365, + "step": 5883 + }, + { + "epoch": 0.5306398520990215, + "flos": 33061011338400.0, + "grad_norm": 2.210142821793336, + "learning_rate": 1.8986828410110032e-06, + "loss": 0.7888, + "num_input_tokens_seen": 190899650, + "step": 5884 + }, + { + "epoch": 0.5307300356224918, + "flos": 23953841642880.0, + "grad_norm": 2.1150278710948025, + "learning_rate": 1.8980993848224823e-06, + "loss": 0.7406, + "num_input_tokens_seen": 190931765, + "step": 5885 + }, + { + "epoch": 0.530820219145962, + "flos": 24500546137440.0, + "grad_norm": 1.7289506095508087, + "learning_rate": 1.8975159373286843e-06, + "loss": 0.709, + "num_input_tokens_seen": 190959490, + "step": 5886 + }, + { + "epoch": 0.5309104026694323, + "flos": 28358297064000.0, + "grad_norm": 2.241520160959639, + "learning_rate": 1.8969324985793904e-06, + "loss": 0.7878, + "num_input_tokens_seen": 190990035, + "step": 5887 + }, + { + "epoch": 0.5310005861929026, + "flos": 22751470886400.0, + "grad_norm": 2.358433011198441, + "learning_rate": 1.8963490686243851e-06, + "loss": 0.7456, + "num_input_tokens_seen": 191019465, + "step": 5888 + }, + { + "epoch": 0.5310907697163728, + "flos": 24129869124480.0, + "grad_norm": 2.3232409560211114, + "learning_rate": 1.8957656475134486e-06, + "loss": 0.6893, + "num_input_tokens_seen": 191048090, + "step": 5889 + }, + { + "epoch": 0.531180953239843, + "flos": 22605816832320.0, + "grad_norm": 2.781393754651489, + "learning_rate": 1.895182235296361e-06, + "loss": 0.7288, + "num_input_tokens_seen": 191077765, + "step": 5890 + }, + { + "epoch": 0.5312711367633134, + "flos": 30835754094720.0, + "grad_norm": 1.8347390666531431, + "learning_rate": 1.8945988320229042e-06, + "loss": 0.6722, + "num_input_tokens_seen": 191106765, + "step": 5891 + }, + { + "epoch": 0.5313613202867836, + "flos": 23479964175360.0, + "grad_norm": 1.820789913064239, + "learning_rate": 1.8940154377428553e-06, + "loss": 0.83, + "num_input_tokens_seen": 191135685, + "step": 5892 + }, + { + "epoch": 0.5314515038102539, + "flos": 24969331347840.0, + "grad_norm": 1.8298474535913962, + "learning_rate": 1.8934320525059944e-06, + "loss": 0.779, + "num_input_tokens_seen": 191164570, + "step": 5893 + }, + { + "epoch": 0.5315416873337241, + "flos": 22204803561600.0, + "grad_norm": 2.651805615224287, + "learning_rate": 1.8928486763620984e-06, + "loss": 0.6349, + "num_input_tokens_seen": 191191240, + "step": 5894 + }, + { + "epoch": 0.5316318708571944, + "flos": 18525310302240.0, + "grad_norm": 2.014907354302758, + "learning_rate": 1.892265309360943e-06, + "loss": 0.7638, + "num_input_tokens_seen": 191217395, + "step": 5895 + }, + { + "epoch": 0.5317220543806647, + "flos": 28651723848480.0, + "grad_norm": 1.8136892034855945, + "learning_rate": 1.8916819515523067e-06, + "loss": 0.736, + "num_input_tokens_seen": 191246515, + "step": 5896 + }, + { + "epoch": 0.5318122379041349, + "flos": 20675175805440.0, + "grad_norm": 1.9203393274703808, + "learning_rate": 1.891098602985963e-06, + "loss": 0.6676, + "num_input_tokens_seen": 191276280, + "step": 5897 + }, + { + "epoch": 0.5319024214276051, + "flos": 27487048962240.0, + "grad_norm": 1.8977422500791818, + "learning_rate": 1.8905152637116872e-06, + "loss": 0.7733, + "num_input_tokens_seen": 191306605, + "step": 5898 + }, + { + "epoch": 0.5319926049510755, + "flos": 18524492567520.0, + "grad_norm": 2.9506116394632547, + "learning_rate": 1.8899319337792527e-06, + "loss": 0.7908, + "num_input_tokens_seen": 191332155, + "step": 5899 + }, + { + "epoch": 0.5320827884745457, + "flos": 27998287772160.0, + "grad_norm": 1.467126395321147, + "learning_rate": 1.8893486132384325e-06, + "loss": 0.7859, + "num_input_tokens_seen": 191362685, + "step": 5900 + }, + { + "epoch": 0.5321729719980159, + "flos": 20783988893760.0, + "grad_norm": 2.420857664982897, + "learning_rate": 1.888765302138999e-06, + "loss": 0.7183, + "num_input_tokens_seen": 191390425, + "step": 5901 + }, + { + "epoch": 0.5322631555214862, + "flos": 32332926916800.0, + "grad_norm": 2.5920252753249886, + "learning_rate": 1.8881820005307224e-06, + "loss": 0.693, + "num_input_tokens_seen": 191418585, + "step": 5902 + }, + { + "epoch": 0.5323533390449565, + "flos": 24276192234240.0, + "grad_norm": 1.7626132766358655, + "learning_rate": 1.8875987084633748e-06, + "loss": 0.6816, + "num_input_tokens_seen": 191446470, + "step": 5903 + }, + { + "epoch": 0.5324435225684268, + "flos": 22496334688320.0, + "grad_norm": 3.490711693980636, + "learning_rate": 1.8870154259867246e-06, + "loss": 0.7811, + "num_input_tokens_seen": 191472705, + "step": 5904 + }, + { + "epoch": 0.532533706091897, + "flos": 21294707327040.0, + "grad_norm": 2.323949051283751, + "learning_rate": 1.886432153150542e-06, + "loss": 0.6813, + "num_input_tokens_seen": 191502695, + "step": 5905 + }, + { + "epoch": 0.5326238896153672, + "flos": 15683309268960.0, + "grad_norm": 1.7548548250933238, + "learning_rate": 1.8858488900045944e-06, + "loss": 0.7274, + "num_input_tokens_seen": 191528205, + "step": 5906 + }, + { + "epoch": 0.5327140731388376, + "flos": 24854236570080.0, + "grad_norm": 2.6845404719376447, + "learning_rate": 1.885265636598648e-06, + "loss": 0.7074, + "num_input_tokens_seen": 191548370, + "step": 5907 + }, + { + "epoch": 0.5328042566623078, + "flos": 28618042312320.0, + "grad_norm": 1.7248548114877031, + "learning_rate": 1.884682392982471e-06, + "loss": 0.7827, + "num_input_tokens_seen": 191581335, + "step": 5908 + }, + { + "epoch": 0.532894440185778, + "flos": 20055086737440.0, + "grad_norm": 3.133154186917168, + "learning_rate": 1.8840991592058274e-06, + "loss": 0.8006, + "num_input_tokens_seen": 191608040, + "step": 5909 + }, + { + "epoch": 0.5329846237092484, + "flos": 22969989137280.0, + "grad_norm": 2.123900634021809, + "learning_rate": 1.8835159353184828e-06, + "loss": 0.7665, + "num_input_tokens_seen": 191637940, + "step": 5910 + }, + { + "epoch": 0.5330748072327186, + "flos": 24247900313280.0, + "grad_norm": 2.3383648868569877, + "learning_rate": 1.8829327213702013e-06, + "loss": 0.8234, + "num_input_tokens_seen": 191668125, + "step": 5911 + }, + { + "epoch": 0.5331649907561888, + "flos": 24354594725280.0, + "grad_norm": 1.755967591953359, + "learning_rate": 1.8823495174107452e-06, + "loss": 0.8189, + "num_input_tokens_seen": 191697680, + "step": 5912 + }, + { + "epoch": 0.5332551742796591, + "flos": 20420188286400.0, + "grad_norm": 3.200518333655494, + "learning_rate": 1.8817663234898773e-06, + "loss": 0.7194, + "num_input_tokens_seen": 191726410, + "step": 5913 + }, + { + "epoch": 0.5333453578031294, + "flos": 24573893541120.0, + "grad_norm": 2.4801540385368415, + "learning_rate": 1.881183139657358e-06, + "loss": 0.7563, + "num_input_tokens_seen": 191755920, + "step": 5914 + }, + { + "epoch": 0.5334355413265996, + "flos": 37324198737120.0, + "grad_norm": 1.6716792624858892, + "learning_rate": 1.8805999659629488e-06, + "loss": 0.7623, + "num_input_tokens_seen": 191784555, + "step": 5915 + }, + { + "epoch": 0.5335257248500699, + "flos": 64853255261760.0, + "grad_norm": 0.6220766363615965, + "learning_rate": 1.880016802456409e-06, + "loss": 0.5355, + "num_input_tokens_seen": 191871975, + "step": 5916 + }, + { + "epoch": 0.5336159083735401, + "flos": 28981359712800.0, + "grad_norm": 3.040450144618257, + "learning_rate": 1.8794336491874964e-06, + "loss": 0.6864, + "num_input_tokens_seen": 191898560, + "step": 5917 + }, + { + "epoch": 0.5337060918970105, + "flos": 30475670463360.0, + "grad_norm": 2.242351778505254, + "learning_rate": 1.8788505062059708e-06, + "loss": 0.7067, + "num_input_tokens_seen": 191927275, + "step": 5918 + }, + { + "epoch": 0.5337962754204807, + "flos": 24755645165760.0, + "grad_norm": 1.5683022912683273, + "learning_rate": 1.8782673735615869e-06, + "loss": 0.732, + "num_input_tokens_seen": 191956930, + "step": 5919 + }, + { + "epoch": 0.5338864589439509, + "flos": 22163502224640.0, + "grad_norm": 2.132229344498162, + "learning_rate": 1.8776842513041026e-06, + "loss": 0.7068, + "num_input_tokens_seen": 191984755, + "step": 5920 + }, + { + "epoch": 0.5339766424674212, + "flos": 27633111883680.0, + "grad_norm": 2.2430758427770994, + "learning_rate": 1.8771011394832727e-06, + "loss": 0.7176, + "num_input_tokens_seen": 192013685, + "step": 5921 + }, + { + "epoch": 0.5340668259908915, + "flos": 22747530891840.0, + "grad_norm": 1.8553584979579618, + "learning_rate": 1.8765180381488501e-06, + "loss": 0.7303, + "num_input_tokens_seen": 192042830, + "step": 5922 + }, + { + "epoch": 0.5341570095143617, + "flos": 59011327386720.0, + "grad_norm": 0.7386453931429958, + "learning_rate": 1.8759349473505905e-06, + "loss": 0.6664, + "num_input_tokens_seen": 192133790, + "step": 5923 + }, + { + "epoch": 0.534247193037832, + "flos": 20929903136160.0, + "grad_norm": 2.3906271940732373, + "learning_rate": 1.8753518671382447e-06, + "loss": 0.6691, + "num_input_tokens_seen": 192163395, + "step": 5924 + }, + { + "epoch": 0.5343373765613022, + "flos": 30801440672640.0, + "grad_norm": 1.9837128142613591, + "learning_rate": 1.8747687975615649e-06, + "loss": 0.7289, + "num_input_tokens_seen": 192194835, + "step": 5925 + }, + { + "epoch": 0.5344275600847725, + "flos": 20893656886560.0, + "grad_norm": 2.0322072047897732, + "learning_rate": 1.874185738670302e-06, + "loss": 0.6413, + "num_input_tokens_seen": 192225120, + "step": 5926 + }, + { + "epoch": 0.5345177436082428, + "flos": 26504051361120.0, + "grad_norm": 1.8433192768331277, + "learning_rate": 1.8736026905142057e-06, + "loss": 0.7342, + "num_input_tokens_seen": 192254965, + "step": 5927 + }, + { + "epoch": 0.534607927131713, + "flos": 21800259163680.0, + "grad_norm": 1.870400764339379, + "learning_rate": 1.873019653143025e-06, + "loss": 0.7231, + "num_input_tokens_seen": 192281275, + "step": 5928 + }, + { + "epoch": 0.5346981106551832, + "flos": 21586052604960.0, + "grad_norm": 2.29717154113183, + "learning_rate": 1.8724366266065069e-06, + "loss": 0.7606, + "num_input_tokens_seen": 192311590, + "step": 5929 + }, + { + "epoch": 0.5347882941786536, + "flos": 16194585248640.0, + "grad_norm": 2.497879750475041, + "learning_rate": 1.8718536109543998e-06, + "loss": 0.823, + "num_input_tokens_seen": 192337100, + "step": 5930 + }, + { + "epoch": 0.5348784777021238, + "flos": 18197904623520.0, + "grad_norm": 2.1335511657644433, + "learning_rate": 1.8712706062364485e-06, + "loss": 0.7619, + "num_input_tokens_seen": 192366460, + "step": 5931 + }, + { + "epoch": 0.534968661225594, + "flos": 25444658436000.0, + "grad_norm": 1.810497295602013, + "learning_rate": 1.8706876125024e-06, + "loss": 0.6339, + "num_input_tokens_seen": 192395415, + "step": 5932 + }, + { + "epoch": 0.5350588447490644, + "flos": 19760135992800.0, + "grad_norm": 2.680653567545485, + "learning_rate": 1.870104629801997e-06, + "loss": 0.7881, + "num_input_tokens_seen": 192421135, + "step": 5933 + }, + { + "epoch": 0.5351490282725346, + "flos": 24390915314400.0, + "grad_norm": 3.4321708291315858, + "learning_rate": 1.8695216581849823e-06, + "loss": 0.7565, + "num_input_tokens_seen": 192450530, + "step": 5934 + }, + { + "epoch": 0.5352392117960049, + "flos": 21001763749440.0, + "grad_norm": 1.7307554399315876, + "learning_rate": 1.8689386977011003e-06, + "loss": 0.7145, + "num_input_tokens_seen": 192479120, + "step": 5935 + }, + { + "epoch": 0.5353293953194751, + "flos": 64252866166560.0, + "grad_norm": 0.6689889935342173, + "learning_rate": 1.8683557484000903e-06, + "loss": 0.5747, + "num_input_tokens_seen": 192569830, + "step": 5936 + }, + { + "epoch": 0.5354195788429454, + "flos": 23079211092960.0, + "grad_norm": 3.3700471493283723, + "learning_rate": 1.8677728103316947e-06, + "loss": 0.7828, + "num_input_tokens_seen": 192598085, + "step": 5937 + }, + { + "epoch": 0.5355097623664157, + "flos": 25006469671680.0, + "grad_norm": 1.506673352907474, + "learning_rate": 1.8671898835456518e-06, + "loss": 0.6979, + "num_input_tokens_seen": 192629650, + "step": 5938 + }, + { + "epoch": 0.5355999458898859, + "flos": 17942545406880.0, + "grad_norm": 1.8710024498822042, + "learning_rate": 1.8666069680917003e-06, + "loss": 0.8483, + "num_input_tokens_seen": 192656225, + "step": 5939 + }, + { + "epoch": 0.5356901294133561, + "flos": 20668187890560.0, + "grad_norm": 2.3427713003780894, + "learning_rate": 1.8660240640195775e-06, + "loss": 0.7019, + "num_input_tokens_seen": 192682580, + "step": 5940 + }, + { + "epoch": 0.5357803129368265, + "flos": 21292254122880.0, + "grad_norm": 1.6670073478916656, + "learning_rate": 1.8654411713790203e-06, + "loss": 0.7824, + "num_input_tokens_seen": 192711310, + "step": 5941 + }, + { + "epoch": 0.5358704964602967, + "flos": 23079359772000.0, + "grad_norm": 2.306327624164735, + "learning_rate": 1.8648582902197648e-06, + "loss": 0.7675, + "num_input_tokens_seen": 192740730, + "step": 5942 + }, + { + "epoch": 0.535960679983767, + "flos": 33644779817280.0, + "grad_norm": 2.302402307382505, + "learning_rate": 1.8642754205915452e-06, + "loss": 0.7504, + "num_input_tokens_seen": 192770135, + "step": 5943 + }, + { + "epoch": 0.5360508635072372, + "flos": 21987140215200.0, + "grad_norm": 2.3922192611928, + "learning_rate": 1.8636925625440943e-06, + "loss": 0.6938, + "num_input_tokens_seen": 192799660, + "step": 5944 + }, + { + "epoch": 0.5361410470307075, + "flos": 23808187588800.0, + "grad_norm": 2.189932097858562, + "learning_rate": 1.863109716127146e-06, + "loss": 0.6556, + "num_input_tokens_seen": 192827435, + "step": 5945 + }, + { + "epoch": 0.5362312305541778, + "flos": 23006681424000.0, + "grad_norm": 1.7099693729491034, + "learning_rate": 1.8625268813904311e-06, + "loss": 0.7891, + "num_input_tokens_seen": 192855275, + "step": 5946 + }, + { + "epoch": 0.536321414077648, + "flos": 19175772797760.0, + "grad_norm": 1.7868442693137925, + "learning_rate": 1.8619440583836814e-06, + "loss": 0.774, + "num_input_tokens_seen": 192881490, + "step": 5947 + }, + { + "epoch": 0.5364115976011182, + "flos": 22606225699680.0, + "grad_norm": 2.5187139087296755, + "learning_rate": 1.8613612471566249e-06, + "loss": 0.7524, + "num_input_tokens_seen": 192909780, + "step": 5948 + }, + { + "epoch": 0.5365017811245886, + "flos": 14663916739200.0, + "grad_norm": 2.258474074854425, + "learning_rate": 1.8607784477589922e-06, + "loss": 0.6937, + "num_input_tokens_seen": 192936325, + "step": 5949 + }, + { + "epoch": 0.5365919646480588, + "flos": 23589818016960.0, + "grad_norm": 1.9335272267005221, + "learning_rate": 1.8601956602405103e-06, + "loss": 0.7636, + "num_input_tokens_seen": 192964160, + "step": 5950 + }, + { + "epoch": 0.536682148171529, + "flos": 25007733443520.0, + "grad_norm": 1.7470716089421894, + "learning_rate": 1.8596128846509043e-06, + "loss": 0.7895, + "num_input_tokens_seen": 192993450, + "step": 5951 + }, + { + "epoch": 0.5367723316949993, + "flos": 22970063476800.0, + "grad_norm": 1.8000305994399894, + "learning_rate": 1.859030121039902e-06, + "loss": 0.7037, + "num_input_tokens_seen": 193020640, + "step": 5952 + }, + { + "epoch": 0.5368625152184696, + "flos": 29198539852320.0, + "grad_norm": 1.722547739638014, + "learning_rate": 1.8584473694572268e-06, + "loss": 0.7007, + "num_input_tokens_seen": 193051210, + "step": 5953 + }, + { + "epoch": 0.5369526987419398, + "flos": 21840631256640.0, + "grad_norm": 2.050155689561266, + "learning_rate": 1.8578646299526026e-06, + "loss": 0.7692, + "num_input_tokens_seen": 193080070, + "step": 5954 + }, + { + "epoch": 0.5370428822654101, + "flos": 20453089257600.0, + "grad_norm": 1.8828544770306053, + "learning_rate": 1.8572819025757518e-06, + "loss": 0.7199, + "num_input_tokens_seen": 193106930, + "step": 5955 + }, + { + "epoch": 0.5371330657888804, + "flos": 25265583034080.0, + "grad_norm": 2.2684519001815975, + "learning_rate": 1.8566991873763959e-06, + "loss": 0.7612, + "num_input_tokens_seen": 193137240, + "step": 5956 + }, + { + "epoch": 0.5372232493123507, + "flos": 18926100554400.0, + "grad_norm": 1.892170046183243, + "learning_rate": 1.856116484404256e-06, + "loss": 0.697, + "num_input_tokens_seen": 193165200, + "step": 5957 + }, + { + "epoch": 0.5373134328358209, + "flos": 31604991174240.0, + "grad_norm": 2.1585789141868252, + "learning_rate": 1.8555337937090502e-06, + "loss": 0.7132, + "num_input_tokens_seen": 193192835, + "step": 5958 + }, + { + "epoch": 0.5374036163592911, + "flos": 13568760771360.0, + "grad_norm": 5.032567959292013, + "learning_rate": 1.8549511153404984e-06, + "loss": 0.724, + "num_input_tokens_seen": 193216285, + "step": 5959 + }, + { + "epoch": 0.5374937998827615, + "flos": 20416173952320.0, + "grad_norm": 1.7756888597573224, + "learning_rate": 1.854368449348317e-06, + "loss": 0.8552, + "num_input_tokens_seen": 193241420, + "step": 5960 + }, + { + "epoch": 0.5375839834062317, + "flos": 19800396576480.0, + "grad_norm": 3.2994251964383268, + "learning_rate": 1.853785795782222e-06, + "loss": 0.7055, + "num_input_tokens_seen": 193268240, + "step": 5961 + }, + { + "epoch": 0.5376741669297019, + "flos": 26795917015680.0, + "grad_norm": 2.630703678635737, + "learning_rate": 1.85320315469193e-06, + "loss": 0.6131, + "num_input_tokens_seen": 193297340, + "step": 5962 + }, + { + "epoch": 0.5377643504531722, + "flos": 69964268079840.0, + "grad_norm": 0.6486461717286051, + "learning_rate": 1.8526205261271534e-06, + "loss": 0.5809, + "num_input_tokens_seen": 193381980, + "step": 5963 + }, + { + "epoch": 0.5378545339766425, + "flos": 35575755372000.0, + "grad_norm": 1.8892457202692787, + "learning_rate": 1.852037910137607e-06, + "loss": 0.6564, + "num_input_tokens_seen": 193414715, + "step": 5964 + }, + { + "epoch": 0.5379447175001127, + "flos": 25083831409440.0, + "grad_norm": 2.430999990126666, + "learning_rate": 1.851455306773002e-06, + "loss": 0.8315, + "num_input_tokens_seen": 193446700, + "step": 5965 + }, + { + "epoch": 0.538034901023583, + "flos": 31390487257440.0, + "grad_norm": 1.5215617051262542, + "learning_rate": 1.8508727160830483e-06, + "loss": 0.7218, + "num_input_tokens_seen": 193483375, + "step": 5966 + }, + { + "epoch": 0.5381250845470532, + "flos": 25630684583040.0, + "grad_norm": 1.7920845860247243, + "learning_rate": 1.8502901381174575e-06, + "loss": 0.8144, + "num_input_tokens_seen": 193512495, + "step": 5967 + }, + { + "epoch": 0.5382152680705236, + "flos": 26426838302400.0, + "grad_norm": 2.22187209417526, + "learning_rate": 1.8497075729259372e-06, + "loss": 0.6511, + "num_input_tokens_seen": 193541895, + "step": 5968 + }, + { + "epoch": 0.5383054515939938, + "flos": 58208854808160.0, + "grad_norm": 0.7627836950380852, + "learning_rate": 1.8491250205581963e-06, + "loss": 0.6408, + "num_input_tokens_seen": 193624570, + "step": 5969 + }, + { + "epoch": 0.538395635117464, + "flos": 29674275807840.0, + "grad_norm": 1.674680619607392, + "learning_rate": 1.8485424810639393e-06, + "loss": 0.7539, + "num_input_tokens_seen": 193656905, + "step": 5970 + }, + { + "epoch": 0.5384858186409343, + "flos": 22387335751200.0, + "grad_norm": 2.192060675456501, + "learning_rate": 1.847959954492874e-06, + "loss": 0.8059, + "num_input_tokens_seen": 193684725, + "step": 5971 + }, + { + "epoch": 0.5385760021644046, + "flos": 22168111274880.0, + "grad_norm": 7.7118944619741105, + "learning_rate": 1.8473774408947035e-06, + "loss": 0.8412, + "num_input_tokens_seen": 193710185, + "step": 5972 + }, + { + "epoch": 0.5386661856878748, + "flos": 67188589365600.0, + "grad_norm": 0.6911959139615941, + "learning_rate": 1.8467949403191308e-06, + "loss": 0.5527, + "num_input_tokens_seen": 193793105, + "step": 5973 + }, + { + "epoch": 0.5387563692113451, + "flos": 34847447931840.0, + "grad_norm": 1.9584591466651773, + "learning_rate": 1.8462124528158592e-06, + "loss": 0.8035, + "num_input_tokens_seen": 193823820, + "step": 5974 + }, + { + "epoch": 0.5388465527348153, + "flos": 23149250388000.0, + "grad_norm": 1.72176076575877, + "learning_rate": 1.8456299784345881e-06, + "loss": 0.7385, + "num_input_tokens_seen": 193854720, + "step": 5975 + }, + { + "epoch": 0.5389367362582856, + "flos": 22934300434080.0, + "grad_norm": 1.9621020502815327, + "learning_rate": 1.8450475172250194e-06, + "loss": 0.7676, + "num_input_tokens_seen": 193882220, + "step": 5976 + }, + { + "epoch": 0.5390269197817559, + "flos": 26325905203200.0, + "grad_norm": 1.6473285868360332, + "learning_rate": 1.844465069236851e-06, + "loss": 0.7272, + "num_input_tokens_seen": 193913945, + "step": 5977 + }, + { + "epoch": 0.5391171033052261, + "flos": 28068252727680.0, + "grad_norm": 2.490698103268635, + "learning_rate": 1.8438826345197796e-06, + "loss": 0.6541, + "num_input_tokens_seen": 193946220, + "step": 5978 + }, + { + "epoch": 0.5392072868286965, + "flos": 21986173801440.0, + "grad_norm": 2.9616992946781697, + "learning_rate": 1.8433002131235036e-06, + "loss": 0.7387, + "num_input_tokens_seen": 193972545, + "step": 5979 + }, + { + "epoch": 0.5392974703521667, + "flos": 30580840915200.0, + "grad_norm": 1.6561406657281093, + "learning_rate": 1.8427178050977167e-06, + "loss": 0.7672, + "num_input_tokens_seen": 194003540, + "step": 5980 + }, + { + "epoch": 0.5393876538756369, + "flos": 19362579509760.0, + "grad_norm": 2.5248210881366644, + "learning_rate": 1.8421354104921143e-06, + "loss": 0.8255, + "num_input_tokens_seen": 194031560, + "step": 5981 + }, + { + "epoch": 0.5394778373991072, + "flos": 69787125505440.0, + "grad_norm": 0.6034870819946857, + "learning_rate": 1.8415530293563894e-06, + "loss": 0.5487, + "num_input_tokens_seen": 194124320, + "step": 5982 + }, + { + "epoch": 0.5395680209225775, + "flos": 41368310338560.0, + "grad_norm": 1.7479059769944267, + "learning_rate": 1.8409706617402333e-06, + "loss": 0.729, + "num_input_tokens_seen": 194156910, + "step": 5983 + }, + { + "epoch": 0.5396582044460477, + "flos": 34406768793600.0, + "grad_norm": 1.6662957335003061, + "learning_rate": 1.8403883076933378e-06, + "loss": 0.7702, + "num_input_tokens_seen": 194190275, + "step": 5984 + }, + { + "epoch": 0.539748387969518, + "flos": 24681777385440.0, + "grad_norm": 1.7893471970231911, + "learning_rate": 1.839805967265391e-06, + "loss": 0.7381, + "num_input_tokens_seen": 194222535, + "step": 5985 + }, + { + "epoch": 0.5398385714929882, + "flos": 20966149385760.0, + "grad_norm": 2.4374019896536314, + "learning_rate": 1.839223640506083e-06, + "loss": 0.7998, + "num_input_tokens_seen": 194251750, + "step": 5986 + }, + { + "epoch": 0.5399287550164585, + "flos": 25775223544320.0, + "grad_norm": 2.8594575463558853, + "learning_rate": 1.8386413274650998e-06, + "loss": 0.7412, + "num_input_tokens_seen": 194279770, + "step": 5987 + }, + { + "epoch": 0.5400189385399288, + "flos": 22563400402560.0, + "grad_norm": 1.8563552184419725, + "learning_rate": 1.8380590281921294e-06, + "loss": 0.7965, + "num_input_tokens_seen": 194308025, + "step": 5988 + }, + { + "epoch": 0.540109122063399, + "flos": 34152933537120.0, + "grad_norm": 1.7112562326856957, + "learning_rate": 1.8374767427368552e-06, + "loss": 0.6552, + "num_input_tokens_seen": 194339895, + "step": 5989 + }, + { + "epoch": 0.5401993055868692, + "flos": 17614247653920.0, + "grad_norm": 2.8522292938296667, + "learning_rate": 1.8368944711489608e-06, + "loss": 0.7748, + "num_input_tokens_seen": 194367055, + "step": 5990 + }, + { + "epoch": 0.5402894891103396, + "flos": 20929754457120.0, + "grad_norm": 2.1134407702472435, + "learning_rate": 1.8363122134781304e-06, + "loss": 0.7755, + "num_input_tokens_seen": 194394465, + "step": 5991 + }, + { + "epoch": 0.5403796726338098, + "flos": 14809013246880.0, + "grad_norm": 2.5042822402290854, + "learning_rate": 1.835729969774044e-06, + "loss": 0.7737, + "num_input_tokens_seen": 194421230, + "step": 5992 + }, + { + "epoch": 0.54046985615728, + "flos": 21622633382400.0, + "grad_norm": 1.6836502525955874, + "learning_rate": 1.8351477400863823e-06, + "loss": 0.7557, + "num_input_tokens_seen": 194448955, + "step": 5993 + }, + { + "epoch": 0.5405600396807503, + "flos": 20054863718880.0, + "grad_norm": 2.0767765115717465, + "learning_rate": 1.8345655244648249e-06, + "loss": 0.7291, + "num_input_tokens_seen": 194478075, + "step": 5994 + }, + { + "epoch": 0.5406502232042206, + "flos": 46867512860160.0, + "grad_norm": 1.4309300094111856, + "learning_rate": 1.8339833229590486e-06, + "loss": 0.739, + "num_input_tokens_seen": 194510025, + "step": 5995 + }, + { + "epoch": 0.5407404067276909, + "flos": 23368363355040.0, + "grad_norm": 2.5777041929894082, + "learning_rate": 1.833401135618731e-06, + "loss": 0.8178, + "num_input_tokens_seen": 194537835, + "step": 5996 + }, + { + "epoch": 0.5408305902511611, + "flos": 16264959071520.0, + "grad_norm": 1.8979259790833434, + "learning_rate": 1.8328189624935466e-06, + "loss": 0.6752, + "num_input_tokens_seen": 194565275, + "step": 5997 + }, + { + "epoch": 0.5409207737746313, + "flos": 26794913432160.0, + "grad_norm": 2.39025517015284, + "learning_rate": 1.832236803633171e-06, + "loss": 0.7848, + "num_input_tokens_seen": 194596925, + "step": 5998 + }, + { + "epoch": 0.5410109572981017, + "flos": 25228556219520.0, + "grad_norm": 2.1055192534382394, + "learning_rate": 1.831654659087276e-06, + "loss": 0.7144, + "num_input_tokens_seen": 194624705, + "step": 5999 + }, + { + "epoch": 0.5411011408215719, + "flos": 20961577505280.0, + "grad_norm": 2.2012450943068473, + "learning_rate": 1.831072528905533e-06, + "loss": 0.763, + "num_input_tokens_seen": 194651895, + "step": 6000 + }, + { + "epoch": 0.5411913243450421, + "flos": 25628640246240.0, + "grad_norm": 1.6435361918940368, + "learning_rate": 1.8304904131376142e-06, + "loss": 0.7995, + "num_input_tokens_seen": 194681195, + "step": 6001 + }, + { + "epoch": 0.5412815078685124, + "flos": 20966446743840.0, + "grad_norm": 1.9801266466924605, + "learning_rate": 1.8299083118331874e-06, + "loss": 0.7664, + "num_input_tokens_seen": 194709105, + "step": 6002 + }, + { + "epoch": 0.5413716913919827, + "flos": 24026780179200.0, + "grad_norm": 2.1409788187360967, + "learning_rate": 1.8293262250419217e-06, + "loss": 0.8819, + "num_input_tokens_seen": 194735195, + "step": 6003 + }, + { + "epoch": 0.541461874915453, + "flos": 30110271556320.0, + "grad_norm": 2.0273001193899165, + "learning_rate": 1.828744152813484e-06, + "loss": 0.8616, + "num_input_tokens_seen": 194764470, + "step": 6004 + }, + { + "epoch": 0.5415520584389232, + "flos": 24536160501120.0, + "grad_norm": 1.8799420394485014, + "learning_rate": 1.8281620951975382e-06, + "loss": 0.7079, + "num_input_tokens_seen": 194793735, + "step": 6005 + }, + { + "epoch": 0.5416422419623935, + "flos": 20417846591520.0, + "grad_norm": 3.202295261005716, + "learning_rate": 1.827580052243751e-06, + "loss": 0.7095, + "num_input_tokens_seen": 194819605, + "step": 6006 + }, + { + "epoch": 0.5417324254858638, + "flos": 25702842554400.0, + "grad_norm": 2.4069078215067705, + "learning_rate": 1.826998024001784e-06, + "loss": 0.7804, + "num_input_tokens_seen": 194853410, + "step": 6007 + }, + { + "epoch": 0.541822609009334, + "flos": 18847995421440.0, + "grad_norm": 1.8585922773563024, + "learning_rate": 1.8264160105212995e-06, + "loss": 0.7766, + "num_input_tokens_seen": 194880640, + "step": 6008 + }, + { + "epoch": 0.5419127925328042, + "flos": 19873781149920.0, + "grad_norm": 2.838232535698366, + "learning_rate": 1.8258340118519582e-06, + "loss": 0.7166, + "num_input_tokens_seen": 194908395, + "step": 6009 + }, + { + "epoch": 0.5420029760562746, + "flos": 25149596182080.0, + "grad_norm": 2.1828332439903275, + "learning_rate": 1.82525202804342e-06, + "loss": 0.7342, + "num_input_tokens_seen": 194939175, + "step": 6010 + }, + { + "epoch": 0.5420931595797448, + "flos": 23006346896160.0, + "grad_norm": 2.2049343173738998, + "learning_rate": 1.8246700591453415e-06, + "loss": 0.7267, + "num_input_tokens_seen": 194965035, + "step": 6011 + }, + { + "epoch": 0.542183343103215, + "flos": 23737442068320.0, + "grad_norm": 2.3644273233132425, + "learning_rate": 1.8240881052073801e-06, + "loss": 0.7518, + "num_input_tokens_seen": 194998590, + "step": 6012 + }, + { + "epoch": 0.5422735266266853, + "flos": 21804161988480.0, + "grad_norm": 2.561987383672187, + "learning_rate": 1.8235061662791923e-06, + "loss": 0.717, + "num_input_tokens_seen": 195027705, + "step": 6013 + }, + { + "epoch": 0.5423637101501556, + "flos": 23545134231840.0, + "grad_norm": 2.571873709584465, + "learning_rate": 1.8229242424104309e-06, + "loss": 0.8194, + "num_input_tokens_seen": 195053665, + "step": 6014 + }, + { + "epoch": 0.5424538936736258, + "flos": 59909566467840.0, + "grad_norm": 0.613904478401158, + "learning_rate": 1.8223423336507503e-06, + "loss": 0.5592, + "num_input_tokens_seen": 195151325, + "step": 6015 + }, + { + "epoch": 0.5425440771970961, + "flos": 25261420020960.0, + "grad_norm": 2.6438207305753396, + "learning_rate": 1.8217604400498012e-06, + "loss": 0.7362, + "num_input_tokens_seen": 195179280, + "step": 6016 + }, + { + "epoch": 0.5426342607205663, + "flos": 23114193570720.0, + "grad_norm": 1.6351773599208825, + "learning_rate": 1.8211785616572333e-06, + "loss": 0.7961, + "num_input_tokens_seen": 195210115, + "step": 6017 + }, + { + "epoch": 0.5427244442440367, + "flos": 21075594360000.0, + "grad_norm": 1.993000178909868, + "learning_rate": 1.8205966985226975e-06, + "loss": 0.8158, + "num_input_tokens_seen": 195236225, + "step": 6018 + }, + { + "epoch": 0.5428146277675069, + "flos": 24099755885280.0, + "grad_norm": 2.303110805348685, + "learning_rate": 1.8200148506958397e-06, + "loss": 0.7605, + "num_input_tokens_seen": 195267200, + "step": 6019 + }, + { + "epoch": 0.5429048112909771, + "flos": 67157584052160.0, + "grad_norm": 0.6500258468110677, + "learning_rate": 1.819433018226308e-06, + "loss": 0.6053, + "num_input_tokens_seen": 195353450, + "step": 6020 + }, + { + "epoch": 0.5429949948144474, + "flos": 68585163616320.0, + "grad_norm": 0.6004625643417267, + "learning_rate": 1.8188512011637471e-06, + "loss": 0.5497, + "num_input_tokens_seen": 195449550, + "step": 6021 + }, + { + "epoch": 0.5430851783379177, + "flos": 28798827523200.0, + "grad_norm": 2.1418671728101897, + "learning_rate": 1.8182693995578e-06, + "loss": 0.7251, + "num_input_tokens_seen": 195477675, + "step": 6022 + }, + { + "epoch": 0.5431753618613879, + "flos": 23006012368320.0, + "grad_norm": 2.288397995252153, + "learning_rate": 1.8176876134581098e-06, + "loss": 0.6291, + "num_input_tokens_seen": 195509890, + "step": 6023 + }, + { + "epoch": 0.5432655453848582, + "flos": 25301866453440.0, + "grad_norm": 4.2415188706785525, + "learning_rate": 1.8171058429143176e-06, + "loss": 0.6967, + "num_input_tokens_seen": 195539145, + "step": 6024 + }, + { + "epoch": 0.5433557289083284, + "flos": 26606359741440.0, + "grad_norm": 2.4111912512852767, + "learning_rate": 1.8165240879760637e-06, + "loss": 0.7863, + "num_input_tokens_seen": 195568045, + "step": 6025 + }, + { + "epoch": 0.5434459124317987, + "flos": 28690051604640.0, + "grad_norm": 2.1042540377867858, + "learning_rate": 1.8159423486929862e-06, + "loss": 0.7785, + "num_input_tokens_seen": 195597845, + "step": 6026 + }, + { + "epoch": 0.543536095955269, + "flos": 20857224788160.0, + "grad_norm": 4.820500449005001, + "learning_rate": 1.815360625114722e-06, + "loss": 0.7502, + "num_input_tokens_seen": 195625660, + "step": 6027 + }, + { + "epoch": 0.5436262794787392, + "flos": 23371002408000.0, + "grad_norm": 1.7655451258594108, + "learning_rate": 1.814778917290908e-06, + "loss": 0.7366, + "num_input_tokens_seen": 195654850, + "step": 6028 + }, + { + "epoch": 0.5437164630022095, + "flos": 14663507871840.0, + "grad_norm": 16.44800953982778, + "learning_rate": 1.8141972252711773e-06, + "loss": 0.7652, + "num_input_tokens_seen": 195680435, + "step": 6029 + }, + { + "epoch": 0.5438066465256798, + "flos": 18888887891040.0, + "grad_norm": 2.408758663999281, + "learning_rate": 1.8136155491051645e-06, + "loss": 0.7475, + "num_input_tokens_seen": 195709115, + "step": 6030 + }, + { + "epoch": 0.54389683004915, + "flos": 58569904853280.0, + "grad_norm": 0.6662497458537978, + "learning_rate": 1.8130338888424998e-06, + "loss": 0.5933, + "num_input_tokens_seen": 195798145, + "step": 6031 + }, + { + "epoch": 0.5439870135726202, + "flos": 18453486858720.0, + "grad_norm": 2.032046587342397, + "learning_rate": 1.812452244532816e-06, + "loss": 0.7667, + "num_input_tokens_seen": 195825770, + "step": 6032 + }, + { + "epoch": 0.5440771970960906, + "flos": 28726335024000.0, + "grad_norm": 1.906061194403184, + "learning_rate": 1.8118706162257405e-06, + "loss": 0.6867, + "num_input_tokens_seen": 195855395, + "step": 6033 + }, + { + "epoch": 0.5441673806195608, + "flos": 22494253181760.0, + "grad_norm": 2.2323305606939696, + "learning_rate": 1.8112890039709002e-06, + "loss": 0.8308, + "num_input_tokens_seen": 195883610, + "step": 6034 + }, + { + "epoch": 0.5442575641430311, + "flos": 25557114160800.0, + "grad_norm": 3.6102221266864056, + "learning_rate": 1.8107074078179238e-06, + "loss": 0.7552, + "num_input_tokens_seen": 195912870, + "step": 6035 + }, + { + "epoch": 0.5443477476665013, + "flos": 34443275231520.0, + "grad_norm": 1.7714569863554255, + "learning_rate": 1.8101258278164348e-06, + "loss": 0.7022, + "num_input_tokens_seen": 195944340, + "step": 6036 + }, + { + "epoch": 0.5444379311899716, + "flos": 25301829283680.0, + "grad_norm": 1.8123117546455993, + "learning_rate": 1.8095442640160575e-06, + "loss": 0.8586, + "num_input_tokens_seen": 195972235, + "step": 6037 + }, + { + "epoch": 0.5445281147134419, + "flos": 20964774104640.0, + "grad_norm": 3.3139948800691457, + "learning_rate": 1.8089627164664132e-06, + "loss": 0.7134, + "num_input_tokens_seen": 195998100, + "step": 6038 + }, + { + "epoch": 0.5446182982369121, + "flos": 23623796911200.0, + "grad_norm": 1.5466665262326362, + "learning_rate": 1.8083811852171233e-06, + "loss": 0.6688, + "num_input_tokens_seen": 196028170, + "step": 6039 + }, + { + "epoch": 0.5447084817603823, + "flos": 20493535690080.0, + "grad_norm": 1.8194273405104284, + "learning_rate": 1.8077996703178078e-06, + "loss": 0.75, + "num_input_tokens_seen": 196058535, + "step": 6040 + }, + { + "epoch": 0.5447986652838527, + "flos": 21440138362560.0, + "grad_norm": 2.3590817300788656, + "learning_rate": 1.8072181718180833e-06, + "loss": 0.6931, + "num_input_tokens_seen": 196087960, + "step": 6041 + }, + { + "epoch": 0.5448888488073229, + "flos": 30256780514880.0, + "grad_norm": 2.6360731450642927, + "learning_rate": 1.806636689767568e-06, + "loss": 0.8342, + "num_input_tokens_seen": 196118610, + "step": 6042 + }, + { + "epoch": 0.5449790323307931, + "flos": 28580234932800.0, + "grad_norm": 1.8078526076184427, + "learning_rate": 1.8060552242158769e-06, + "loss": 0.6892, + "num_input_tokens_seen": 196149430, + "step": 6043 + }, + { + "epoch": 0.5450692158542634, + "flos": 26140325093280.0, + "grad_norm": 1.8320710720785296, + "learning_rate": 1.8054737752126224e-06, + "loss": 0.7215, + "num_input_tokens_seen": 196180640, + "step": 6044 + }, + { + "epoch": 0.5451593993777337, + "flos": 20674469580000.0, + "grad_norm": 2.289236537319289, + "learning_rate": 1.804892342807419e-06, + "loss": 0.7352, + "num_input_tokens_seen": 196209780, + "step": 6045 + }, + { + "epoch": 0.545249582901204, + "flos": 22891103439360.0, + "grad_norm": 3.8090935168895705, + "learning_rate": 1.8043109270498756e-06, + "loss": 0.8173, + "num_input_tokens_seen": 196235270, + "step": 6046 + }, + { + "epoch": 0.5453397664246742, + "flos": 24203848414080.0, + "grad_norm": 2.018056186933507, + "learning_rate": 1.803729527989604e-06, + "loss": 0.7598, + "num_input_tokens_seen": 196266940, + "step": 6047 + }, + { + "epoch": 0.5454299499481444, + "flos": 22132757099520.0, + "grad_norm": 1.837764723516276, + "learning_rate": 1.8031481456762112e-06, + "loss": 0.7405, + "num_input_tokens_seen": 196294550, + "step": 6048 + }, + { + "epoch": 0.5455201334716148, + "flos": 20674989956640.0, + "grad_norm": 1.9365585870289295, + "learning_rate": 1.8025667801593033e-06, + "loss": 0.6788, + "num_input_tokens_seen": 196323905, + "step": 6049 + }, + { + "epoch": 0.545610316995085, + "flos": 27408757980480.0, + "grad_norm": 2.8502063609585937, + "learning_rate": 1.8019854314884871e-06, + "loss": 0.6866, + "num_input_tokens_seen": 196353455, + "step": 6050 + }, + { + "epoch": 0.5457005005185552, + "flos": 28943886861120.0, + "grad_norm": 1.6751734202554367, + "learning_rate": 1.8014040997133652e-06, + "loss": 0.803, + "num_input_tokens_seen": 196385680, + "step": 6051 + }, + { + "epoch": 0.5457906840420256, + "flos": 68828293981920.0, + "grad_norm": 0.7584141967613621, + "learning_rate": 1.8008227848835414e-06, + "loss": 0.6137, + "num_input_tokens_seen": 196474990, + "step": 6052 + }, + { + "epoch": 0.5458808675654958, + "flos": 18452446105440.0, + "grad_norm": 2.5092962041243676, + "learning_rate": 1.8002414870486144e-06, + "loss": 0.7053, + "num_input_tokens_seen": 196502320, + "step": 6053 + }, + { + "epoch": 0.545971051088966, + "flos": 25663957251840.0, + "grad_norm": 1.8981623823085332, + "learning_rate": 1.7996602062581864e-06, + "loss": 0.7019, + "num_input_tokens_seen": 196534070, + "step": 6054 + }, + { + "epoch": 0.5460612346124363, + "flos": 19800062048640.0, + "grad_norm": 2.0567493609224203, + "learning_rate": 1.7990789425618544e-06, + "loss": 0.7972, + "num_input_tokens_seen": 196561130, + "step": 6055 + }, + { + "epoch": 0.5461514181359066, + "flos": 53718085470720.0, + "grad_norm": 1.5398695325700962, + "learning_rate": 1.7984976960092137e-06, + "loss": 0.7322, + "num_input_tokens_seen": 196595535, + "step": 6056 + }, + { + "epoch": 0.5462416016593769, + "flos": 29815209302400.0, + "grad_norm": 2.2142994260583158, + "learning_rate": 1.7979164666498617e-06, + "loss": 0.7153, + "num_input_tokens_seen": 196624955, + "step": 6057 + }, + { + "epoch": 0.5463317851828471, + "flos": 64840017093600.0, + "grad_norm": 0.6168461350525752, + "learning_rate": 1.7973352545333901e-06, + "loss": 0.5527, + "num_input_tokens_seen": 196714710, + "step": 6058 + }, + { + "epoch": 0.5464219687063173, + "flos": 34008097217760.0, + "grad_norm": 1.9133716815778408, + "learning_rate": 1.796754059709393e-06, + "loss": 0.667, + "num_input_tokens_seen": 196742230, + "step": 6059 + }, + { + "epoch": 0.5465121522297877, + "flos": 27232730498880.0, + "grad_norm": 2.7598757711731063, + "learning_rate": 1.7961728822274603e-06, + "loss": 0.8121, + "num_input_tokens_seen": 196770930, + "step": 6060 + }, + { + "epoch": 0.5466023357532579, + "flos": 25630164206400.0, + "grad_norm": 3.591707926525732, + "learning_rate": 1.7955917221371802e-06, + "loss": 0.7031, + "num_input_tokens_seen": 196799180, + "step": 6061 + }, + { + "epoch": 0.5466925192767281, + "flos": 24901224880320.0, + "grad_norm": 3.824753253432958, + "learning_rate": 1.7950105794881422e-06, + "loss": 0.6645, + "num_input_tokens_seen": 196827525, + "step": 6062 + }, + { + "epoch": 0.5467827028001984, + "flos": 29272965179040.0, + "grad_norm": 1.7481229115576982, + "learning_rate": 1.7944294543299317e-06, + "loss": 0.7315, + "num_input_tokens_seen": 196859825, + "step": 6063 + }, + { + "epoch": 0.5468728863236687, + "flos": 41878619904480.0, + "grad_norm": 1.8480815760826357, + "learning_rate": 1.7938483467121333e-06, + "loss": 0.8017, + "num_input_tokens_seen": 196891305, + "step": 6064 + }, + { + "epoch": 0.5469630698471389, + "flos": 17760347745120.0, + "grad_norm": 1.9871472195560476, + "learning_rate": 1.7932672566843313e-06, + "loss": 0.7315, + "num_input_tokens_seen": 196918535, + "step": 6065 + }, + { + "epoch": 0.5470532533706092, + "flos": 22897868335680.0, + "grad_norm": 2.29556683506902, + "learning_rate": 1.7926861842961065e-06, + "loss": 0.672, + "num_input_tokens_seen": 196947370, + "step": 6066 + }, + { + "epoch": 0.5471434368940794, + "flos": 26175679268640.0, + "grad_norm": 1.7272751662132164, + "learning_rate": 1.7921051295970399e-06, + "loss": 0.6691, + "num_input_tokens_seen": 196975910, + "step": 6067 + }, + { + "epoch": 0.5472336204175497, + "flos": 34986411429120.0, + "grad_norm": 1.793136090217405, + "learning_rate": 1.7915240926367092e-06, + "loss": 0.7165, + "num_input_tokens_seen": 197006610, + "step": 6068 + }, + { + "epoch": 0.54732380394102, + "flos": 68431449457920.0, + "grad_norm": 1.727335070894919, + "learning_rate": 1.7909430734646932e-06, + "loss": 0.7925, + "num_input_tokens_seen": 197040270, + "step": 6069 + }, + { + "epoch": 0.5474139874644902, + "flos": 63313214239200.0, + "grad_norm": 0.6800941916752974, + "learning_rate": 1.790362072130567e-06, + "loss": 0.5855, + "num_input_tokens_seen": 197133065, + "step": 6070 + }, + { + "epoch": 0.5475041709879604, + "flos": 24245893146240.0, + "grad_norm": 2.8215029415241695, + "learning_rate": 1.7897810886839037e-06, + "loss": 0.7324, + "num_input_tokens_seen": 197161675, + "step": 6071 + }, + { + "epoch": 0.5475943545114308, + "flos": 31893623059680.0, + "grad_norm": 3.44924709010335, + "learning_rate": 1.7892001231742782e-06, + "loss": 0.7575, + "num_input_tokens_seen": 197192245, + "step": 6072 + }, + { + "epoch": 0.547684538034901, + "flos": 26468288318400.0, + "grad_norm": 1.6642973111822734, + "learning_rate": 1.7886191756512594e-06, + "loss": 0.7574, + "num_input_tokens_seen": 197219660, + "step": 6073 + }, + { + "epoch": 0.5477747215583713, + "flos": 23915699735520.0, + "grad_norm": 2.1302201722304983, + "learning_rate": 1.7880382461644192e-06, + "loss": 0.828, + "num_input_tokens_seen": 197248955, + "step": 6074 + }, + { + "epoch": 0.5478649050818416, + "flos": 26977854489120.0, + "grad_norm": 2.182280061701567, + "learning_rate": 1.7874573347633235e-06, + "loss": 0.7268, + "num_input_tokens_seen": 197275155, + "step": 6075 + }, + { + "epoch": 0.5479550886053118, + "flos": 26431075655040.0, + "grad_norm": 1.9547169415568837, + "learning_rate": 1.7868764414975408e-06, + "loss": 0.8185, + "num_input_tokens_seen": 197305610, + "step": 6076 + }, + { + "epoch": 0.5480452721287821, + "flos": 23363419776960.0, + "grad_norm": 1.963768727524669, + "learning_rate": 1.7862955664166353e-06, + "loss": 0.6955, + "num_input_tokens_seen": 197332105, + "step": 6077 + }, + { + "epoch": 0.5481354556522523, + "flos": 22640873649600.0, + "grad_norm": 1.9120381140361657, + "learning_rate": 1.78571470957017e-06, + "loss": 0.8325, + "num_input_tokens_seen": 197361755, + "step": 6078 + }, + { + "epoch": 0.5482256391757226, + "flos": 19727235021600.0, + "grad_norm": 3.51640480717797, + "learning_rate": 1.7851338710077074e-06, + "loss": 0.816, + "num_input_tokens_seen": 197389230, + "step": 6079 + }, + { + "epoch": 0.5483158226991929, + "flos": 25405215587040.0, + "grad_norm": 2.843725518760523, + "learning_rate": 1.7845530507788076e-06, + "loss": 0.8554, + "num_input_tokens_seen": 197415395, + "step": 6080 + }, + { + "epoch": 0.5484060062226631, + "flos": 23443606416480.0, + "grad_norm": 1.861612336274679, + "learning_rate": 1.7839722489330298e-06, + "loss": 0.7198, + "num_input_tokens_seen": 197443795, + "step": 6081 + }, + { + "epoch": 0.5484961897461333, + "flos": 23078950904640.0, + "grad_norm": 1.8614876549675416, + "learning_rate": 1.7833914655199308e-06, + "loss": 0.7181, + "num_input_tokens_seen": 197473890, + "step": 6082 + }, + { + "epoch": 0.5485863732696037, + "flos": 25082716316640.0, + "grad_norm": 2.5000119209005334, + "learning_rate": 1.7828107005890658e-06, + "loss": 0.7937, + "num_input_tokens_seen": 197501725, + "step": 6083 + }, + { + "epoch": 0.5486765567930739, + "flos": 10983457066080.0, + "grad_norm": 2.892862639514471, + "learning_rate": 1.7822299541899898e-06, + "loss": 0.6671, + "num_input_tokens_seen": 197527125, + "step": 6084 + }, + { + "epoch": 0.5487667403165442, + "flos": 26466764358240.0, + "grad_norm": 2.5245537429622558, + "learning_rate": 1.7816492263722545e-06, + "loss": 0.7478, + "num_input_tokens_seen": 197559530, + "step": 6085 + }, + { + "epoch": 0.5488569238400144, + "flos": 23808113249280.0, + "grad_norm": 1.9987566182383028, + "learning_rate": 1.781068517185412e-06, + "loss": 0.7726, + "num_input_tokens_seen": 197587395, + "step": 6086 + }, + { + "epoch": 0.5489471073634847, + "flos": 20747036418720.0, + "grad_norm": 1.9461261319315146, + "learning_rate": 1.7804878266790104e-06, + "loss": 0.8142, + "num_input_tokens_seen": 197616465, + "step": 6087 + }, + { + "epoch": 0.549037290886955, + "flos": 25483729587360.0, + "grad_norm": 1.587958565510126, + "learning_rate": 1.779907154902597e-06, + "loss": 0.684, + "num_input_tokens_seen": 197646340, + "step": 6088 + }, + { + "epoch": 0.5491274744104252, + "flos": 32147012279040.0, + "grad_norm": 1.598701713256338, + "learning_rate": 1.7793265019057198e-06, + "loss": 0.6775, + "num_input_tokens_seen": 197678980, + "step": 6089 + }, + { + "epoch": 0.5492176579338954, + "flos": 70643988910080.0, + "grad_norm": 0.7379359492391215, + "learning_rate": 1.7787458677379212e-06, + "loss": 0.5946, + "num_input_tokens_seen": 197758945, + "step": 6090 + }, + { + "epoch": 0.5493078414573658, + "flos": 25884965876640.0, + "grad_norm": 1.695575254268106, + "learning_rate": 1.7781652524487463e-06, + "loss": 0.807, + "num_input_tokens_seen": 197789050, + "step": 6091 + }, + { + "epoch": 0.549398024980836, + "flos": 24063212277600.0, + "grad_norm": 1.9517408425394924, + "learning_rate": 1.777584656087735e-06, + "loss": 0.6964, + "num_input_tokens_seen": 197819295, + "step": 6092 + }, + { + "epoch": 0.5494882085043062, + "flos": 14260190076000.0, + "grad_norm": 2.1087298501617355, + "learning_rate": 1.777004078704427e-06, + "loss": 0.6682, + "num_input_tokens_seen": 197845450, + "step": 6093 + }, + { + "epoch": 0.5495783920277765, + "flos": 22966643858880.0, + "grad_norm": 2.227862169512287, + "learning_rate": 1.7764235203483603e-06, + "loss": 0.773, + "num_input_tokens_seen": 197872565, + "step": 6094 + }, + { + "epoch": 0.5496685755512468, + "flos": 22824446592480.0, + "grad_norm": 2.384221995232929, + "learning_rate": 1.775842981069072e-06, + "loss": 0.7211, + "num_input_tokens_seen": 197902980, + "step": 6095 + }, + { + "epoch": 0.549758759074717, + "flos": 26650411640640.0, + "grad_norm": 1.9138205239066803, + "learning_rate": 1.7752624609160966e-06, + "loss": 0.7602, + "num_input_tokens_seen": 197933210, + "step": 6096 + }, + { + "epoch": 0.5498489425981873, + "flos": 28398185950080.0, + "grad_norm": 2.3007381415740795, + "learning_rate": 1.7746819599389665e-06, + "loss": 0.7839, + "num_input_tokens_seen": 197962235, + "step": 6097 + }, + { + "epoch": 0.5499391261216575, + "flos": 26320478418240.0, + "grad_norm": 2.448651066992208, + "learning_rate": 1.774101478187215e-06, + "loss": 0.7535, + "num_input_tokens_seen": 197991480, + "step": 6098 + }, + { + "epoch": 0.5500293096451279, + "flos": 25482279966720.0, + "grad_norm": 2.0866768773146993, + "learning_rate": 1.773521015710371e-06, + "loss": 0.8213, + "num_input_tokens_seen": 198017780, + "step": 6099 + }, + { + "epoch": 0.5501194931685981, + "flos": 19108335385920.0, + "grad_norm": 2.7546859178630507, + "learning_rate": 1.7729405725579614e-06, + "loss": 0.8258, + "num_input_tokens_seen": 198046200, + "step": 6100 + }, + { + "epoch": 0.5502096766920683, + "flos": 25594438333440.0, + "grad_norm": 1.67292153946423, + "learning_rate": 1.7723601487795151e-06, + "loss": 0.723, + "num_input_tokens_seen": 198076775, + "step": 6101 + }, + { + "epoch": 0.5502998602155387, + "flos": 29344528434240.0, + "grad_norm": 2.152831532890762, + "learning_rate": 1.7717797444245557e-06, + "loss": 0.6971, + "num_input_tokens_seen": 198106960, + "step": 6102 + }, + { + "epoch": 0.5503900437390089, + "flos": 29892013493760.0, + "grad_norm": 1.645851819220275, + "learning_rate": 1.7711993595426076e-06, + "loss": 0.6866, + "num_input_tokens_seen": 198141085, + "step": 6103 + }, + { + "epoch": 0.5504802272624791, + "flos": 18561519382080.0, + "grad_norm": 2.0761649848486328, + "learning_rate": 1.7706189941831915e-06, + "loss": 0.8597, + "num_input_tokens_seen": 198166530, + "step": 6104 + }, + { + "epoch": 0.5505704107859494, + "flos": 41623260687840.0, + "grad_norm": 1.7952519923704944, + "learning_rate": 1.770038648395827e-06, + "loss": 0.7109, + "num_input_tokens_seen": 198200490, + "step": 6105 + }, + { + "epoch": 0.5506605943094197, + "flos": 62075154779520.0, + "grad_norm": 0.6078578206403219, + "learning_rate": 1.7694583222300336e-06, + "loss": 0.5935, + "num_input_tokens_seen": 198299740, + "step": 6106 + }, + { + "epoch": 0.55075077783289, + "flos": 21325155094080.0, + "grad_norm": 1.7223826772877373, + "learning_rate": 1.7688780157353272e-06, + "loss": 0.7292, + "num_input_tokens_seen": 198326685, + "step": 6107 + }, + { + "epoch": 0.5508409613563602, + "flos": 33498568216800.0, + "grad_norm": 1.9826394248318844, + "learning_rate": 1.768297728961223e-06, + "loss": 0.6931, + "num_input_tokens_seen": 198359795, + "step": 6108 + }, + { + "epoch": 0.5509311448798304, + "flos": 27593185827840.0, + "grad_norm": 2.075730744344352, + "learning_rate": 1.7677174619572342e-06, + "loss": 0.756, + "num_input_tokens_seen": 198387610, + "step": 6109 + }, + { + "epoch": 0.5510213284033008, + "flos": 20419630740000.0, + "grad_norm": 2.0359595062635747, + "learning_rate": 1.7671372147728717e-06, + "loss": 0.7661, + "num_input_tokens_seen": 198415190, + "step": 6110 + }, + { + "epoch": 0.551111511926771, + "flos": 18087530405280.0, + "grad_norm": 2.570414810695299, + "learning_rate": 1.7665569874576471e-06, + "loss": 0.8155, + "num_input_tokens_seen": 198440845, + "step": 6111 + }, + { + "epoch": 0.5512016954502412, + "flos": 25556705293440.0, + "grad_norm": 8.942325596440442, + "learning_rate": 1.7659767800610664e-06, + "loss": 0.7521, + "num_input_tokens_seen": 198470005, + "step": 6112 + }, + { + "epoch": 0.5512918789737115, + "flos": 22241904715680.0, + "grad_norm": 2.228710436852268, + "learning_rate": 1.7653965926326379e-06, + "loss": 0.6564, + "num_input_tokens_seen": 198499840, + "step": 6113 + }, + { + "epoch": 0.5513820624971818, + "flos": 21182994997440.0, + "grad_norm": 1.503139627429112, + "learning_rate": 1.764816425221866e-06, + "loss": 0.7069, + "num_input_tokens_seen": 198528900, + "step": 6114 + }, + { + "epoch": 0.551472246020652, + "flos": 29600259348480.0, + "grad_norm": 3.07072667247642, + "learning_rate": 1.7642362778782524e-06, + "loss": 0.6669, + "num_input_tokens_seen": 198562590, + "step": 6115 + }, + { + "epoch": 0.5515624295441223, + "flos": 20128322631840.0, + "grad_norm": 15.347760144564356, + "learning_rate": 1.7636561506513005e-06, + "loss": 0.7218, + "num_input_tokens_seen": 198591120, + "step": 6116 + }, + { + "epoch": 0.5516526130675925, + "flos": 48689972684640.0, + "grad_norm": 1.517541911946273, + "learning_rate": 1.7630760435905083e-06, + "loss": 0.7246, + "num_input_tokens_seen": 198625350, + "step": 6117 + }, + { + "epoch": 0.5517427965910628, + "flos": 16740137480640.0, + "grad_norm": 2.1915617154983646, + "learning_rate": 1.762495956745375e-06, + "loss": 0.7093, + "num_input_tokens_seen": 198652265, + "step": 6118 + }, + { + "epoch": 0.5518329801145331, + "flos": 24132247989120.0, + "grad_norm": 3.2433862147736265, + "learning_rate": 1.7619158901653962e-06, + "loss": 0.792, + "num_input_tokens_seen": 198678795, + "step": 6119 + }, + { + "epoch": 0.5519231636380033, + "flos": 24136299492960.0, + "grad_norm": 2.3470466317354086, + "learning_rate": 1.761335843900066e-06, + "loss": 0.749, + "num_input_tokens_seen": 198706805, + "step": 6120 + }, + { + "epoch": 0.5520133471614735, + "flos": 36194617837920.0, + "grad_norm": 2.001189712483637, + "learning_rate": 1.7607558179988785e-06, + "loss": 0.7334, + "num_input_tokens_seen": 198739215, + "step": 6121 + }, + { + "epoch": 0.5521035306849439, + "flos": 47485520421600.0, + "grad_norm": 1.6937077393837356, + "learning_rate": 1.760175812511323e-06, + "loss": 0.6771, + "num_input_tokens_seen": 198771985, + "step": 6122 + }, + { + "epoch": 0.5521937142084141, + "flos": 27269943162240.0, + "grad_norm": 2.1217048003712047, + "learning_rate": 1.75959582748689e-06, + "loss": 0.7073, + "num_input_tokens_seen": 198799345, + "step": 6123 + }, + { + "epoch": 0.5522838977318844, + "flos": 21039348110400.0, + "grad_norm": 2.0123131327621726, + "learning_rate": 1.7590158629750657e-06, + "loss": 0.7812, + "num_input_tokens_seen": 198825565, + "step": 6124 + }, + { + "epoch": 0.5523740812553547, + "flos": 19217520171840.0, + "grad_norm": 3.1318982091076832, + "learning_rate": 1.7584359190253376e-06, + "loss": 0.7986, + "num_input_tokens_seen": 198855235, + "step": 6125 + }, + { + "epoch": 0.5524642647788249, + "flos": 22824149234400.0, + "grad_norm": 1.974167469658399, + "learning_rate": 1.7578559956871892e-06, + "loss": 0.8112, + "num_input_tokens_seen": 198884170, + "step": 6126 + }, + { + "epoch": 0.5525544483022952, + "flos": 19982668577760.0, + "grad_norm": 1.5507338970651399, + "learning_rate": 1.7572760930101012e-06, + "loss": 0.7461, + "num_input_tokens_seen": 198913290, + "step": 6127 + }, + { + "epoch": 0.5526446318257654, + "flos": 64464545181600.0, + "grad_norm": 0.7672721893741328, + "learning_rate": 1.7566962110435563e-06, + "loss": 0.6334, + "num_input_tokens_seen": 199009470, + "step": 6128 + }, + { + "epoch": 0.5527348153492357, + "flos": 26103521297280.0, + "grad_norm": 1.9294249939071608, + "learning_rate": 1.7561163498370313e-06, + "loss": 0.7644, + "num_input_tokens_seen": 199042090, + "step": 6129 + }, + { + "epoch": 0.552824998872706, + "flos": 27452884219200.0, + "grad_norm": 1.7555736610562747, + "learning_rate": 1.755536509440005e-06, + "loss": 0.7139, + "num_input_tokens_seen": 199076435, + "step": 6130 + }, + { + "epoch": 0.5529151823961762, + "flos": 22601430800640.0, + "grad_norm": 3.2045145137825433, + "learning_rate": 1.7549566899019519e-06, + "loss": 0.7483, + "num_input_tokens_seen": 199104050, + "step": 6131 + }, + { + "epoch": 0.5530053659196464, + "flos": 24281730528480.0, + "grad_norm": 2.053999637005708, + "learning_rate": 1.754376891272344e-06, + "loss": 0.7665, + "num_input_tokens_seen": 199132555, + "step": 6132 + }, + { + "epoch": 0.5530955494431168, + "flos": 21622670552160.0, + "grad_norm": 1.5865957463199631, + "learning_rate": 1.753797113600655e-06, + "loss": 0.6979, + "num_input_tokens_seen": 199161600, + "step": 6133 + }, + { + "epoch": 0.553185732966587, + "flos": 24713340245280.0, + "grad_norm": 2.0476014689919206, + "learning_rate": 1.7532173569363535e-06, + "loss": 0.8844, + "num_input_tokens_seen": 199191125, + "step": 6134 + }, + { + "epoch": 0.5532759164900573, + "flos": 39109371558720.0, + "grad_norm": 1.9389411119801778, + "learning_rate": 1.7526376213289077e-06, + "loss": 0.5656, + "num_input_tokens_seen": 199222220, + "step": 6135 + }, + { + "epoch": 0.5533661000135275, + "flos": 22933817227200.0, + "grad_norm": 1.6162281231384261, + "learning_rate": 1.7520579068277844e-06, + "loss": 0.7319, + "num_input_tokens_seen": 199252045, + "step": 6136 + }, + { + "epoch": 0.5534562835369978, + "flos": 24682669459680.0, + "grad_norm": 1.8098606755709201, + "learning_rate": 1.7514782134824472e-06, + "loss": 0.7755, + "num_input_tokens_seen": 199282280, + "step": 6137 + }, + { + "epoch": 0.5535464670604681, + "flos": 20667779023200.0, + "grad_norm": 2.3473085980696133, + "learning_rate": 1.7508985413423599e-06, + "loss": 0.7422, + "num_input_tokens_seen": 199308975, + "step": 6138 + }, + { + "epoch": 0.5536366505839383, + "flos": 22496929404480.0, + "grad_norm": 1.8460137645749677, + "learning_rate": 1.7503188904569814e-06, + "loss": 0.8218, + "num_input_tokens_seen": 199337450, + "step": 6139 + }, + { + "epoch": 0.5537268341074085, + "flos": 21834461076480.0, + "grad_norm": 6.484711199398136, + "learning_rate": 1.7497392608757728e-06, + "loss": 0.7149, + "num_input_tokens_seen": 199367235, + "step": 6140 + }, + { + "epoch": 0.5538170176308789, + "flos": 25010149477920.0, + "grad_norm": 1.6521384830898767, + "learning_rate": 1.7491596526481897e-06, + "loss": 0.7657, + "num_input_tokens_seen": 199398660, + "step": 6141 + }, + { + "epoch": 0.5539072011543491, + "flos": 20784286251840.0, + "grad_norm": 2.3999564365565855, + "learning_rate": 1.7485800658236888e-06, + "loss": 0.6774, + "num_input_tokens_seen": 199426920, + "step": 6142 + }, + { + "epoch": 0.5539973846778193, + "flos": 22569013036320.0, + "grad_norm": 1.9265815975235785, + "learning_rate": 1.7480005004517232e-06, + "loss": 0.8178, + "num_input_tokens_seen": 199455255, + "step": 6143 + }, + { + "epoch": 0.5540875682012896, + "flos": 23842686859680.0, + "grad_norm": 2.5482058178284284, + "learning_rate": 1.7474209565817435e-06, + "loss": 0.8093, + "num_input_tokens_seen": 199483955, + "step": 6144 + }, + { + "epoch": 0.5541777517247599, + "flos": 31163494301280.0, + "grad_norm": 1.698007577283867, + "learning_rate": 1.7468414342632014e-06, + "loss": 0.7473, + "num_input_tokens_seen": 199515590, + "step": 6145 + }, + { + "epoch": 0.5542679352482301, + "flos": 22532952635520.0, + "grad_norm": 2.5249660338581688, + "learning_rate": 1.746261933545543e-06, + "loss": 0.8038, + "num_input_tokens_seen": 199544115, + "step": 6146 + }, + { + "epoch": 0.5543581187717004, + "flos": 23079768639360.0, + "grad_norm": 1.71782062719434, + "learning_rate": 1.7456824544782165e-06, + "loss": 0.7496, + "num_input_tokens_seen": 199571885, + "step": 6147 + }, + { + "epoch": 0.5544483022951707, + "flos": 24719287406880.0, + "grad_norm": 1.74928868687652, + "learning_rate": 1.7451029971106653e-06, + "loss": 0.754, + "num_input_tokens_seen": 199599795, + "step": 6148 + }, + { + "epoch": 0.554538485818641, + "flos": 21724904592960.0, + "grad_norm": 1.6587501944431862, + "learning_rate": 1.7445235614923313e-06, + "loss": 0.7782, + "num_input_tokens_seen": 199628205, + "step": 6149 + }, + { + "epoch": 0.5546286693421112, + "flos": 21513448596480.0, + "grad_norm": 2.108928152720861, + "learning_rate": 1.7439441476726556e-06, + "loss": 0.7383, + "num_input_tokens_seen": 199656490, + "step": 6150 + }, + { + "epoch": 0.5547188528655814, + "flos": 22824558101760.0, + "grad_norm": 3.554096415914763, + "learning_rate": 1.7433647557010776e-06, + "loss": 0.7573, + "num_input_tokens_seen": 199685155, + "step": 6151 + }, + { + "epoch": 0.5548090363890518, + "flos": 35612038791360.0, + "grad_norm": 1.9666926821372703, + "learning_rate": 1.7427853856270338e-06, + "loss": 0.7143, + "num_input_tokens_seen": 199714545, + "step": 6152 + }, + { + "epoch": 0.554899219912522, + "flos": 26576692539360.0, + "grad_norm": 2.3920166353266024, + "learning_rate": 1.7422060374999593e-06, + "loss": 0.7107, + "num_input_tokens_seen": 199744810, + "step": 6153 + }, + { + "epoch": 0.5549894034359922, + "flos": 21002878842240.0, + "grad_norm": 1.627883407035607, + "learning_rate": 1.7416267113692862e-06, + "loss": 0.7096, + "num_input_tokens_seen": 199774775, + "step": 6154 + }, + { + "epoch": 0.5550795869594625, + "flos": 23079173923200.0, + "grad_norm": 1.8445433510008145, + "learning_rate": 1.7410474072844475e-06, + "loss": 0.7784, + "num_input_tokens_seen": 199802785, + "step": 6155 + }, + { + "epoch": 0.5551697704829328, + "flos": 34154717685600.0, + "grad_norm": 1.8774211510407686, + "learning_rate": 1.740468125294871e-06, + "loss": 0.7327, + "num_input_tokens_seen": 199831905, + "step": 6156 + }, + { + "epoch": 0.555259954006403, + "flos": 21002321295840.0, + "grad_norm": 2.0487640050813187, + "learning_rate": 1.739888865449986e-06, + "loss": 0.6637, + "num_input_tokens_seen": 199855875, + "step": 6157 + }, + { + "epoch": 0.5553501375298733, + "flos": 19216776776640.0, + "grad_norm": 2.7378859295820415, + "learning_rate": 1.7393096277992174e-06, + "loss": 0.7748, + "num_input_tokens_seen": 199881720, + "step": 6158 + }, + { + "epoch": 0.5554403210533435, + "flos": 28976304625440.0, + "grad_norm": 1.893550658356982, + "learning_rate": 1.738730412391988e-06, + "loss": 0.7619, + "num_input_tokens_seen": 199911720, + "step": 6159 + }, + { + "epoch": 0.5555305045768139, + "flos": 22017179114880.0, + "grad_norm": 3.9641643073541597, + "learning_rate": 1.738151219277721e-06, + "loss": 0.801, + "num_input_tokens_seen": 199938545, + "step": 6160 + }, + { + "epoch": 0.5556206881002841, + "flos": 17177359831200.0, + "grad_norm": 1.977773182514772, + "learning_rate": 1.7375720485058349e-06, + "loss": 0.7052, + "num_input_tokens_seen": 199964435, + "step": 6161 + }, + { + "epoch": 0.5557108716237543, + "flos": 18197384246880.0, + "grad_norm": 2.2267780400402986, + "learning_rate": 1.7369929001257498e-06, + "loss": 0.7684, + "num_input_tokens_seen": 199992935, + "step": 6162 + }, + { + "epoch": 0.5558010551472246, + "flos": 25010223817440.0, + "grad_norm": 3.7700841310458277, + "learning_rate": 1.73641377418688e-06, + "loss": 0.6663, + "num_input_tokens_seen": 200021210, + "step": 6163 + }, + { + "epoch": 0.5558912386706949, + "flos": 21440398550880.0, + "grad_norm": 1.762188901567519, + "learning_rate": 1.7358346707386408e-06, + "loss": 0.6846, + "num_input_tokens_seen": 200048995, + "step": 6164 + }, + { + "epoch": 0.5559814221941651, + "flos": 17432161501440.0, + "grad_norm": 2.205845336977478, + "learning_rate": 1.7352555898304439e-06, + "loss": 0.799, + "num_input_tokens_seen": 200076185, + "step": 6165 + }, + { + "epoch": 0.5560716057176354, + "flos": 20085683183520.0, + "grad_norm": 2.500536479608336, + "learning_rate": 1.7346765315116996e-06, + "loss": 0.6675, + "num_input_tokens_seen": 200103840, + "step": 6166 + }, + { + "epoch": 0.5561617892411056, + "flos": 25698679541280.0, + "grad_norm": 2.0717411510103165, + "learning_rate": 1.734097495831817e-06, + "loss": 0.7114, + "num_input_tokens_seen": 200131530, + "step": 6167 + }, + { + "epoch": 0.5562519727645759, + "flos": 17505248716800.0, + "grad_norm": 2.5364985740232004, + "learning_rate": 1.7335184828402015e-06, + "loss": 0.7639, + "num_input_tokens_seen": 200158925, + "step": 6168 + }, + { + "epoch": 0.5563421562880462, + "flos": 22642583458560.0, + "grad_norm": 1.6415634377776611, + "learning_rate": 1.7329394925862595e-06, + "loss": 0.7276, + "num_input_tokens_seen": 200185555, + "step": 6169 + }, + { + "epoch": 0.5564323398115164, + "flos": 18816506901120.0, + "grad_norm": 2.2732758874244583, + "learning_rate": 1.7323605251193922e-06, + "loss": 0.7964, + "num_input_tokens_seen": 200211475, + "step": 6170 + }, + { + "epoch": 0.5565225233349868, + "flos": 14371344859200.0, + "grad_norm": 2.993974037786639, + "learning_rate": 1.7317815804890001e-06, + "loss": 0.77, + "num_input_tokens_seen": 200235680, + "step": 6171 + }, + { + "epoch": 0.556612706858457, + "flos": 25192309969920.0, + "grad_norm": 1.980744953946888, + "learning_rate": 1.731202658744483e-06, + "loss": 0.707, + "num_input_tokens_seen": 200264080, + "step": 6172 + }, + { + "epoch": 0.5567028903819272, + "flos": 24937508299680.0, + "grad_norm": 1.4720700663186408, + "learning_rate": 1.7306237599352365e-06, + "loss": 0.8092, + "num_input_tokens_seen": 200295155, + "step": 6173 + }, + { + "epoch": 0.5567930739053975, + "flos": 24061242280320.0, + "grad_norm": 1.6676656192962043, + "learning_rate": 1.730044884110657e-06, + "loss": 0.736, + "num_input_tokens_seen": 200324150, + "step": 6174 + }, + { + "epoch": 0.5568832574288678, + "flos": 31780646958240.0, + "grad_norm": 1.982808373268829, + "learning_rate": 1.7294660313201366e-06, + "loss": 0.7057, + "num_input_tokens_seen": 200352035, + "step": 6175 + }, + { + "epoch": 0.556973440952338, + "flos": 25301123058240.0, + "grad_norm": 14.560768908644128, + "learning_rate": 1.7288872016130652e-06, + "loss": 0.728, + "num_input_tokens_seen": 200374095, + "step": 6176 + }, + { + "epoch": 0.5570636244758083, + "flos": 32114706024000.0, + "grad_norm": 1.7917665261525348, + "learning_rate": 1.7283083950388334e-06, + "loss": 0.662, + "num_input_tokens_seen": 200406460, + "step": 6177 + }, + { + "epoch": 0.5571538079992785, + "flos": 43804428862560.0, + "grad_norm": 1.9136693433911096, + "learning_rate": 1.727729611646827e-06, + "loss": 0.6537, + "num_input_tokens_seen": 200439470, + "step": 6178 + }, + { + "epoch": 0.5572439915227488, + "flos": 22824781120320.0, + "grad_norm": 2.2021987807784873, + "learning_rate": 1.7271508514864318e-06, + "loss": 0.8022, + "num_input_tokens_seen": 200467750, + "step": 6179 + }, + { + "epoch": 0.5573341750462191, + "flos": 15647248868160.0, + "grad_norm": 2.4786576996132013, + "learning_rate": 1.7265721146070302e-06, + "loss": 0.7499, + "num_input_tokens_seen": 200492215, + "step": 6180 + }, + { + "epoch": 0.5574243585696893, + "flos": 20675175805440.0, + "grad_norm": 2.1830344260031143, + "learning_rate": 1.7259934010580035e-06, + "loss": 0.7485, + "num_input_tokens_seen": 200518870, + "step": 6181 + }, + { + "epoch": 0.5575145420931595, + "flos": 28868903988000.0, + "grad_norm": 3.1859929130939943, + "learning_rate": 1.725414710888731e-06, + "loss": 0.6942, + "num_input_tokens_seen": 200548485, + "step": 6182 + }, + { + "epoch": 0.5576047256166299, + "flos": 19836865844640.0, + "grad_norm": 2.058761097658654, + "learning_rate": 1.7248360441485885e-06, + "loss": 0.7243, + "num_input_tokens_seen": 200578805, + "step": 6183 + }, + { + "epoch": 0.5576949091401001, + "flos": 30325741886880.0, + "grad_norm": 2.11166359409188, + "learning_rate": 1.7242574008869528e-06, + "loss": 0.8209, + "num_input_tokens_seen": 200606930, + "step": 6184 + }, + { + "epoch": 0.5577850926635703, + "flos": 71174332976640.0, + "grad_norm": 0.672340453625399, + "learning_rate": 1.7236787811531951e-06, + "loss": 0.6291, + "num_input_tokens_seen": 200710520, + "step": 6185 + }, + { + "epoch": 0.5578752761870406, + "flos": 27047670765600.0, + "grad_norm": 1.9107644452691088, + "learning_rate": 1.7231001849966887e-06, + "loss": 0.7179, + "num_input_tokens_seen": 200741660, + "step": 6186 + }, + { + "epoch": 0.5579654597105109, + "flos": 69778279102560.0, + "grad_norm": 0.6516995475682013, + "learning_rate": 1.722521612466801e-06, + "loss": 0.6068, + "num_input_tokens_seen": 200839520, + "step": 6187 + }, + { + "epoch": 0.5580556432339812, + "flos": 18707805322080.0, + "grad_norm": 2.575567468131625, + "learning_rate": 1.7219430636128984e-06, + "loss": 0.6925, + "num_input_tokens_seen": 200866835, + "step": 6188 + }, + { + "epoch": 0.5581458267574514, + "flos": 20310111426240.0, + "grad_norm": 1.807158854622148, + "learning_rate": 1.7213645384843479e-06, + "loss": 0.6933, + "num_input_tokens_seen": 200894505, + "step": 6189 + }, + { + "epoch": 0.5582360102809216, + "flos": 35972196762240.0, + "grad_norm": 2.2225002777133764, + "learning_rate": 1.7207860371305108e-06, + "loss": 0.7396, + "num_input_tokens_seen": 200926070, + "step": 6190 + }, + { + "epoch": 0.558326193804392, + "flos": 39323094910560.0, + "grad_norm": 2.3079500543489107, + "learning_rate": 1.7202075596007487e-06, + "loss": 0.721, + "num_input_tokens_seen": 200957920, + "step": 6191 + }, + { + "epoch": 0.5584163773278622, + "flos": 48582349028640.0, + "grad_norm": 2.4136035075447, + "learning_rate": 1.7196291059444206e-06, + "loss": 0.7069, + "num_input_tokens_seen": 200990665, + "step": 6192 + }, + { + "epoch": 0.5585065608513324, + "flos": 30401802683040.0, + "grad_norm": 2.040953595984763, + "learning_rate": 1.7190506762108828e-06, + "loss": 0.7316, + "num_input_tokens_seen": 201019830, + "step": 6193 + }, + { + "epoch": 0.5585967443748028, + "flos": 20782910970720.0, + "grad_norm": 2.1694018876572634, + "learning_rate": 1.7184722704494907e-06, + "loss": 0.7467, + "num_input_tokens_seen": 201049905, + "step": 6194 + }, + { + "epoch": 0.558686927898273, + "flos": 25735520507040.0, + "grad_norm": 1.7534413560557385, + "learning_rate": 1.717893888709596e-06, + "loss": 0.7617, + "num_input_tokens_seen": 201077530, + "step": 6195 + }, + { + "epoch": 0.5587771114217432, + "flos": 20128917348000.0, + "grad_norm": 1.7756484445637433, + "learning_rate": 1.7173155310405515e-06, + "loss": 0.7767, + "num_input_tokens_seen": 201103635, + "step": 6196 + }, + { + "epoch": 0.5588672949452135, + "flos": 25116323513280.0, + "grad_norm": 1.6599113014377447, + "learning_rate": 1.7167371974917043e-06, + "loss": 0.8218, + "num_input_tokens_seen": 201133970, + "step": 6197 + }, + { + "epoch": 0.5589574784686838, + "flos": 18125263445280.0, + "grad_norm": 2.4674870604511083, + "learning_rate": 1.7161588881124003e-06, + "loss": 0.7631, + "num_input_tokens_seen": 201159615, + "step": 6198 + }, + { + "epoch": 0.559047661992154, + "flos": 29309285768160.0, + "grad_norm": 4.412544245943842, + "learning_rate": 1.7155806029519861e-06, + "loss": 0.7329, + "num_input_tokens_seen": 201188570, + "step": 6199 + }, + { + "epoch": 0.5591378455156243, + "flos": 19982259710400.0, + "grad_norm": 2.313202107011949, + "learning_rate": 1.7150023420598023e-06, + "loss": 0.7894, + "num_input_tokens_seen": 201215215, + "step": 6200 + }, + { + "epoch": 0.5592280290390945, + "flos": 32698028465760.0, + "grad_norm": 1.8305216020873991, + "learning_rate": 1.714424105485191e-06, + "loss": 0.8001, + "num_input_tokens_seen": 201247190, + "step": 6201 + }, + { + "epoch": 0.5593182125625649, + "flos": 23990385250560.0, + "grad_norm": 2.164277544595208, + "learning_rate": 1.7138458932774896e-06, + "loss": 0.7119, + "num_input_tokens_seen": 201276160, + "step": 6202 + }, + { + "epoch": 0.5594083960860351, + "flos": 60014625410400.0, + "grad_norm": 0.6501824902238165, + "learning_rate": 1.7132677054860335e-06, + "loss": 0.6061, + "num_input_tokens_seen": 201362540, + "step": 6203 + }, + { + "epoch": 0.5594985796095053, + "flos": 19397561987520.0, + "grad_norm": 2.651813257394552, + "learning_rate": 1.7126895421601586e-06, + "loss": 0.771, + "num_input_tokens_seen": 201390040, + "step": 6204 + }, + { + "epoch": 0.5595887631329756, + "flos": 29598326520960.0, + "grad_norm": 2.1809778256834815, + "learning_rate": 1.712111403349196e-06, + "loss": 0.6921, + "num_input_tokens_seen": 201422630, + "step": 6205 + }, + { + "epoch": 0.5596789466564459, + "flos": 24828658041600.0, + "grad_norm": 1.5763682038700106, + "learning_rate": 1.7115332891024757e-06, + "loss": 0.78, + "num_input_tokens_seen": 201451575, + "step": 6206 + }, + { + "epoch": 0.5597691301799161, + "flos": 58924598869440.0, + "grad_norm": 0.7135097793545142, + "learning_rate": 1.7109551994693257e-06, + "loss": 0.6563, + "num_input_tokens_seen": 201537395, + "step": 6207 + }, + { + "epoch": 0.5598593137033864, + "flos": 22096882547520.0, + "grad_norm": 1.8500832229659359, + "learning_rate": 1.7103771344990725e-06, + "loss": 0.6889, + "num_input_tokens_seen": 201566050, + "step": 6208 + }, + { + "epoch": 0.5599494972268566, + "flos": 40020285528000.0, + "grad_norm": 1.956439631231924, + "learning_rate": 1.709799094241039e-06, + "loss": 0.6632, + "num_input_tokens_seen": 201597410, + "step": 6209 + }, + { + "epoch": 0.560039680750327, + "flos": 26321519171520.0, + "grad_norm": 1.9883121270740667, + "learning_rate": 1.709221078744546e-06, + "loss": 0.781, + "num_input_tokens_seen": 201627210, + "step": 6210 + }, + { + "epoch": 0.5601298642737972, + "flos": 26937482396160.0, + "grad_norm": 1.6582808369645277, + "learning_rate": 1.7086430880589148e-06, + "loss": 0.7564, + "num_input_tokens_seen": 201658175, + "step": 6211 + }, + { + "epoch": 0.5602200477972674, + "flos": 18161546864640.0, + "grad_norm": 3.4051030178982837, + "learning_rate": 1.7080651222334612e-06, + "loss": 0.7541, + "num_input_tokens_seen": 201685555, + "step": 6212 + }, + { + "epoch": 0.5603102313207377, + "flos": 25520273195040.0, + "grad_norm": 2.125297749920672, + "learning_rate": 1.7074871813175018e-06, + "loss": 0.8065, + "num_input_tokens_seen": 201709805, + "step": 6213 + }, + { + "epoch": 0.560400414844208, + "flos": 25044983276640.0, + "grad_norm": 1.7964455796324739, + "learning_rate": 1.706909265360349e-06, + "loss": 0.6618, + "num_input_tokens_seen": 201741250, + "step": 6214 + }, + { + "epoch": 0.5604905983676782, + "flos": 68480699389920.0, + "grad_norm": 0.6983821488535672, + "learning_rate": 1.7063313744113128e-06, + "loss": 0.5955, + "num_input_tokens_seen": 201828300, + "step": 6215 + }, + { + "epoch": 0.5605807818911485, + "flos": 69190198931520.0, + "grad_norm": 0.6359456757715847, + "learning_rate": 1.7057535085197042e-06, + "loss": 0.5695, + "num_input_tokens_seen": 201920860, + "step": 6216 + }, + { + "epoch": 0.5606709654146187, + "flos": 19472619200160.0, + "grad_norm": 13.768486219209239, + "learning_rate": 1.705175667734828e-06, + "loss": 0.6691, + "num_input_tokens_seen": 201948450, + "step": 6217 + }, + { + "epoch": 0.560761148938089, + "flos": 27086964935520.0, + "grad_norm": 2.3860690517170218, + "learning_rate": 1.7045978521059894e-06, + "loss": 0.7155, + "num_input_tokens_seen": 201976215, + "step": 6218 + }, + { + "epoch": 0.5608513324615593, + "flos": 20821387405920.0, + "grad_norm": 1.575388055255893, + "learning_rate": 1.7040200616824914e-06, + "loss": 0.7473, + "num_input_tokens_seen": 202004825, + "step": 6219 + }, + { + "epoch": 0.5609415159850295, + "flos": 25339413644640.0, + "grad_norm": 1.9423570357923867, + "learning_rate": 1.7034422965136333e-06, + "loss": 0.6864, + "num_input_tokens_seen": 202033860, + "step": 6220 + }, + { + "epoch": 0.5610316995084998, + "flos": 23772127188000.0, + "grad_norm": 2.1737487175242856, + "learning_rate": 1.7028645566487137e-06, + "loss": 0.7868, + "num_input_tokens_seen": 202062115, + "step": 6221 + }, + { + "epoch": 0.5611218830319701, + "flos": 31309259864640.0, + "grad_norm": 1.8136450997848594, + "learning_rate": 1.7022868421370284e-06, + "loss": 0.7224, + "num_input_tokens_seen": 202090465, + "step": 6222 + }, + { + "epoch": 0.5612120665554403, + "flos": 21691334566080.0, + "grad_norm": 1.87720772785829, + "learning_rate": 1.701709153027872e-06, + "loss": 0.7929, + "num_input_tokens_seen": 202118450, + "step": 6223 + }, + { + "epoch": 0.5613022500789105, + "flos": 23261483094240.0, + "grad_norm": 1.8743609292983114, + "learning_rate": 1.7011314893705353e-06, + "loss": 0.7146, + "num_input_tokens_seen": 202147380, + "step": 6224 + }, + { + "epoch": 0.5613924336023809, + "flos": 20893917074880.0, + "grad_norm": 1.7351832525564534, + "learning_rate": 1.700553851214307e-06, + "loss": 0.7469, + "num_input_tokens_seen": 202176295, + "step": 6225 + }, + { + "epoch": 0.5614826171258511, + "flos": 43766844501600.0, + "grad_norm": 2.5783252985047618, + "learning_rate": 1.699976238608476e-06, + "loss": 0.6155, + "num_input_tokens_seen": 202206755, + "step": 6226 + }, + { + "epoch": 0.5615728006493214, + "flos": 25228779238080.0, + "grad_norm": 16.533550023125326, + "learning_rate": 1.699398651602326e-06, + "loss": 0.6979, + "num_input_tokens_seen": 202235045, + "step": 6227 + }, + { + "epoch": 0.5616629841727916, + "flos": 26395424121600.0, + "grad_norm": 1.9095939323413595, + "learning_rate": 1.6988210902451413e-06, + "loss": 0.7891, + "num_input_tokens_seen": 202264535, + "step": 6228 + }, + { + "epoch": 0.5617531676962619, + "flos": 20528815525920.0, + "grad_norm": 1.9348103712742826, + "learning_rate": 1.6982435545862011e-06, + "loss": 0.7271, + "num_input_tokens_seen": 202292165, + "step": 6229 + }, + { + "epoch": 0.5618433512197322, + "flos": 33019970189760.0, + "grad_norm": 2.1169138376093195, + "learning_rate": 1.6976660446747853e-06, + "loss": 0.7532, + "num_input_tokens_seen": 202322140, + "step": 6230 + }, + { + "epoch": 0.5619335347432024, + "flos": 24063063598560.0, + "grad_norm": 2.8330364258723932, + "learning_rate": 1.6970885605601696e-06, + "loss": 0.6754, + "num_input_tokens_seen": 202350580, + "step": 6231 + }, + { + "epoch": 0.5620237182666726, + "flos": 23297506325280.0, + "grad_norm": 2.1388089906816377, + "learning_rate": 1.6965111022916282e-06, + "loss": 0.8583, + "num_input_tokens_seen": 202377025, + "step": 6232 + }, + { + "epoch": 0.562113901790143, + "flos": 26066605992000.0, + "grad_norm": 1.8483260958741705, + "learning_rate": 1.6959336699184323e-06, + "loss": 0.7209, + "num_input_tokens_seen": 202405875, + "step": 6233 + }, + { + "epoch": 0.5622040853136132, + "flos": 22022197032480.0, + "grad_norm": 1.875593536529989, + "learning_rate": 1.6953562634898529e-06, + "loss": 0.7413, + "num_input_tokens_seen": 202435410, + "step": 6234 + }, + { + "epoch": 0.5622942688370834, + "flos": 17682354121440.0, + "grad_norm": 2.277175190414509, + "learning_rate": 1.6947788830551569e-06, + "loss": 0.7104, + "num_input_tokens_seen": 202459200, + "step": 6235 + }, + { + "epoch": 0.5623844523605537, + "flos": 59359776883200.0, + "grad_norm": 0.7664439822988806, + "learning_rate": 1.6942015286636093e-06, + "loss": 0.6292, + "num_input_tokens_seen": 202543200, + "step": 6236 + }, + { + "epoch": 0.562474635884024, + "flos": 22205138089440.0, + "grad_norm": 1.8845753628060893, + "learning_rate": 1.6936242003644735e-06, + "loss": 0.7315, + "num_input_tokens_seen": 202569910, + "step": 6237 + }, + { + "epoch": 0.5625648194074943, + "flos": 17650977110400.0, + "grad_norm": 3.199091609529128, + "learning_rate": 1.6930468982070106e-06, + "loss": 0.8164, + "num_input_tokens_seen": 202594580, + "step": 6238 + }, + { + "epoch": 0.5626550029309645, + "flos": 66994826174880.0, + "grad_norm": 0.5914672588289887, + "learning_rate": 1.692469622240478e-06, + "loss": 0.5178, + "num_input_tokens_seen": 202687580, + "step": 6239 + }, + { + "epoch": 0.5627451864544347, + "flos": 27591513188640.0, + "grad_norm": 1.6388984678099852, + "learning_rate": 1.6918923725141339e-06, + "loss": 0.7029, + "num_input_tokens_seen": 202716565, + "step": 6240 + }, + { + "epoch": 0.5628353699779051, + "flos": 30183172922880.0, + "grad_norm": 2.0342728643371246, + "learning_rate": 1.6913151490772312e-06, + "loss": 0.7967, + "num_input_tokens_seen": 202749115, + "step": 6241 + }, + { + "epoch": 0.5629255535013753, + "flos": 23987225820960.0, + "grad_norm": 2.2490301161090045, + "learning_rate": 1.6907379519790215e-06, + "loss": 0.7865, + "num_input_tokens_seen": 202779290, + "step": 6242 + }, + { + "epoch": 0.5630157370248455, + "flos": 38381026948800.0, + "grad_norm": 1.5852078417658895, + "learning_rate": 1.6901607812687558e-06, + "loss": 0.6546, + "num_input_tokens_seen": 202812325, + "step": 6243 + }, + { + "epoch": 0.5631059205483159, + "flos": 16994455944000.0, + "grad_norm": 2.276163570767952, + "learning_rate": 1.6895836369956794e-06, + "loss": 0.7016, + "num_input_tokens_seen": 202838160, + "step": 6244 + }, + { + "epoch": 0.5631961040717861, + "flos": 46570103177760.0, + "grad_norm": 0.6764788031841467, + "learning_rate": 1.6890065192090402e-06, + "loss": 0.5418, + "num_input_tokens_seen": 202919460, + "step": 6245 + }, + { + "epoch": 0.5632862875952563, + "flos": 20272973102400.0, + "grad_norm": 2.2477509596318446, + "learning_rate": 1.6884294279580793e-06, + "loss": 0.6461, + "num_input_tokens_seen": 202945375, + "step": 6246 + }, + { + "epoch": 0.5633764711187266, + "flos": 24646869247200.0, + "grad_norm": 2.3688280319204287, + "learning_rate": 1.6878523632920371e-06, + "loss": 0.7729, + "num_input_tokens_seen": 202973020, + "step": 6247 + }, + { + "epoch": 0.5634666546421969, + "flos": 17687483548320.0, + "grad_norm": 1.900818701083893, + "learning_rate": 1.6872753252601525e-06, + "loss": 0.8113, + "num_input_tokens_seen": 202999860, + "step": 6248 + }, + { + "epoch": 0.5635568381656672, + "flos": 19400163870720.0, + "grad_norm": 1.713256727217945, + "learning_rate": 1.6866983139116616e-06, + "loss": 0.6885, + "num_input_tokens_seen": 203029515, + "step": 6249 + }, + { + "epoch": 0.5636470216891374, + "flos": 20420076777120.0, + "grad_norm": 1.7354578984900904, + "learning_rate": 1.6861213292957981e-06, + "loss": 0.7162, + "num_input_tokens_seen": 203058315, + "step": 6250 + }, + { + "epoch": 0.5637372052126076, + "flos": 26358397307040.0, + "grad_norm": 2.0578624722170638, + "learning_rate": 1.685544371461793e-06, + "loss": 0.7791, + "num_input_tokens_seen": 203086955, + "step": 6251 + }, + { + "epoch": 0.563827388736078, + "flos": 18671224544640.0, + "grad_norm": 1.7355272744029502, + "learning_rate": 1.6849674404588767e-06, + "loss": 0.7696, + "num_input_tokens_seen": 203114645, + "step": 6252 + }, + { + "epoch": 0.5639175722595482, + "flos": 17572128582240.0, + "grad_norm": 2.170624297383835, + "learning_rate": 1.6843905363362758e-06, + "loss": 0.777, + "num_input_tokens_seen": 203139075, + "step": 6253 + }, + { + "epoch": 0.5640077557830184, + "flos": 24572518260000.0, + "grad_norm": 1.8289092900227697, + "learning_rate": 1.6838136591432136e-06, + "loss": 0.6862, + "num_input_tokens_seen": 203171345, + "step": 6254 + }, + { + "epoch": 0.5640979393064887, + "flos": 22168743160800.0, + "grad_norm": 2.9054331298023244, + "learning_rate": 1.6832368089289139e-06, + "loss": 0.772, + "num_input_tokens_seen": 203199575, + "step": 6255 + }, + { + "epoch": 0.564188122829959, + "flos": 23914473133440.0, + "grad_norm": 2.212283422354924, + "learning_rate": 1.682659985742596e-06, + "loss": 0.7636, + "num_input_tokens_seen": 203228270, + "step": 6256 + }, + { + "epoch": 0.5642783063534292, + "flos": 25374842159520.0, + "grad_norm": 2.4837951943250665, + "learning_rate": 1.6820831896334782e-06, + "loss": 0.7872, + "num_input_tokens_seen": 203256635, + "step": 6257 + }, + { + "epoch": 0.5643684898768995, + "flos": 27663485311200.0, + "grad_norm": 2.3148078743159477, + "learning_rate": 1.681506420650776e-06, + "loss": 0.7673, + "num_input_tokens_seen": 203287195, + "step": 6258 + }, + { + "epoch": 0.5644586734003697, + "flos": 22966718198400.0, + "grad_norm": 1.5975648438810435, + "learning_rate": 1.680929678843701e-06, + "loss": 0.7338, + "num_input_tokens_seen": 203315570, + "step": 6259 + }, + { + "epoch": 0.56454885692384, + "flos": 24645716984640.0, + "grad_norm": 1.7926676046417003, + "learning_rate": 1.6803529642614662e-06, + "loss": 0.7935, + "num_input_tokens_seen": 203345910, + "step": 6260 + }, + { + "epoch": 0.5646390404473103, + "flos": 28180188075840.0, + "grad_norm": 2.347061665879263, + "learning_rate": 1.6797762769532785e-06, + "loss": 0.7518, + "num_input_tokens_seen": 203376575, + "step": 6261 + }, + { + "epoch": 0.5647292239707805, + "flos": 24754790261280.0, + "grad_norm": 1.4694750193536033, + "learning_rate": 1.679199616968345e-06, + "loss": 0.7813, + "num_input_tokens_seen": 203406830, + "step": 6262 + }, + { + "epoch": 0.5648194074942507, + "flos": 20014900493280.0, + "grad_norm": 1.6673264707439077, + "learning_rate": 1.6786229843558689e-06, + "loss": 0.774, + "num_input_tokens_seen": 203435235, + "step": 6263 + }, + { + "epoch": 0.5649095910177211, + "flos": 18047938877280.0, + "grad_norm": 4.513358527432045, + "learning_rate": 1.6780463791650514e-06, + "loss": 0.7296, + "num_input_tokens_seen": 203460205, + "step": 6264 + }, + { + "epoch": 0.5649997745411913, + "flos": 28143978996000.0, + "grad_norm": 3.4724072945333497, + "learning_rate": 1.6774698014450928e-06, + "loss": 0.6934, + "num_input_tokens_seen": 203490825, + "step": 6265 + }, + { + "epoch": 0.5650899580646616, + "flos": 62923692157920.0, + "grad_norm": 0.6547868858047161, + "learning_rate": 1.6768932512451883e-06, + "loss": 0.5783, + "num_input_tokens_seen": 203575045, + "step": 6266 + }, + { + "epoch": 0.5651801415881319, + "flos": 23225757221280.0, + "grad_norm": 2.0977301273798052, + "learning_rate": 1.676316728614534e-06, + "loss": 0.6979, + "num_input_tokens_seen": 203606640, + "step": 6267 + }, + { + "epoch": 0.5652703251116021, + "flos": 68771784479520.0, + "grad_norm": 0.5717468253574214, + "learning_rate": 1.675740233602321e-06, + "loss": 0.5875, + "num_input_tokens_seen": 203704990, + "step": 6268 + }, + { + "epoch": 0.5653605086350724, + "flos": 23183377961280.0, + "grad_norm": 1.9665334424730718, + "learning_rate": 1.6751637662577385e-06, + "loss": 0.791, + "num_input_tokens_seen": 203733630, + "step": 6269 + }, + { + "epoch": 0.5654506921585426, + "flos": 58223542596960.0, + "grad_norm": 0.8107874501789454, + "learning_rate": 1.6745873266299753e-06, + "loss": 0.5314, + "num_input_tokens_seen": 203813550, + "step": 6270 + }, + { + "epoch": 0.565540875682013, + "flos": 18452669124000.0, + "grad_norm": 2.71002243384616, + "learning_rate": 1.6740109147682148e-06, + "loss": 0.708, + "num_input_tokens_seen": 203841740, + "step": 6271 + }, + { + "epoch": 0.5656310592054832, + "flos": 22423470491520.0, + "grad_norm": 2.5705249084085016, + "learning_rate": 1.6734345307216418e-06, + "loss": 0.7578, + "num_input_tokens_seen": 203871080, + "step": 6272 + }, + { + "epoch": 0.5657212427289534, + "flos": 18743865722880.0, + "grad_norm": 2.7797818541528123, + "learning_rate": 1.6728581745394346e-06, + "loss": 0.7188, + "num_input_tokens_seen": 203897750, + "step": 6273 + }, + { + "epoch": 0.5658114262524236, + "flos": 24608950358400.0, + "grad_norm": 1.6291664702406405, + "learning_rate": 1.672281846270772e-06, + "loss": 0.8206, + "num_input_tokens_seen": 203928845, + "step": 6274 + }, + { + "epoch": 0.565901609775894, + "flos": 22060301770080.0, + "grad_norm": 2.023188630930919, + "learning_rate": 1.6717055459648295e-06, + "loss": 0.7438, + "num_input_tokens_seen": 203958490, + "step": 6275 + }, + { + "epoch": 0.5659917932993642, + "flos": 26174118138720.0, + "grad_norm": 2.404876713437913, + "learning_rate": 1.6711292736707793e-06, + "loss": 0.6968, + "num_input_tokens_seen": 203986640, + "step": 6276 + }, + { + "epoch": 0.5660819768228345, + "flos": 18524901434880.0, + "grad_norm": 2.011995581680816, + "learning_rate": 1.6705530294377938e-06, + "loss": 0.8184, + "num_input_tokens_seen": 204014540, + "step": 6277 + }, + { + "epoch": 0.5661721603463047, + "flos": 26431298673600.0, + "grad_norm": 1.8925913523870195, + "learning_rate": 1.6699768133150395e-06, + "loss": 0.7352, + "num_input_tokens_seen": 204044795, + "step": 6278 + }, + { + "epoch": 0.566262343869775, + "flos": 21657132653280.0, + "grad_norm": 2.4867024585646833, + "learning_rate": 1.6694006253516837e-06, + "loss": 0.737, + "num_input_tokens_seen": 204071460, + "step": 6279 + }, + { + "epoch": 0.5663525273932453, + "flos": 21949890382080.0, + "grad_norm": 2.3916604785196705, + "learning_rate": 1.6688244655968896e-06, + "loss": 0.7409, + "num_input_tokens_seen": 204099495, + "step": 6280 + }, + { + "epoch": 0.5664427109167155, + "flos": 28908904383360.0, + "grad_norm": 3.1451993054915692, + "learning_rate": 1.6682483340998175e-06, + "loss": 0.6687, + "num_input_tokens_seen": 204131540, + "step": 6281 + }, + { + "epoch": 0.5665328944401857, + "flos": 22130489744160.0, + "grad_norm": 12.544157843238969, + "learning_rate": 1.6676722309096276e-06, + "loss": 0.7052, + "num_input_tokens_seen": 204158210, + "step": 6282 + }, + { + "epoch": 0.5666230779636561, + "flos": 65947890852960.0, + "grad_norm": 0.7431119281462893, + "learning_rate": 1.6670961560754744e-06, + "loss": 0.6462, + "num_input_tokens_seen": 204243985, + "step": 6283 + }, + { + "epoch": 0.5667132614871263, + "flos": 31969386497760.0, + "grad_norm": 1.849196777728931, + "learning_rate": 1.6665201096465138e-06, + "loss": 0.7413, + "num_input_tokens_seen": 204275950, + "step": 6284 + }, + { + "epoch": 0.5668034450105965, + "flos": 19982965935840.0, + "grad_norm": 2.06496386313974, + "learning_rate": 1.6659440916718961e-06, + "loss": 0.79, + "num_input_tokens_seen": 204304140, + "step": 6285 + }, + { + "epoch": 0.5668936285340668, + "flos": 20269888012320.0, + "grad_norm": 2.3078987120975465, + "learning_rate": 1.6653681022007696e-06, + "loss": 0.7597, + "num_input_tokens_seen": 204331660, + "step": 6286 + }, + { + "epoch": 0.5669838120575371, + "flos": 19800768274080.0, + "grad_norm": 3.7929932645561424, + "learning_rate": 1.6647921412822825e-06, + "loss": 0.6766, + "num_input_tokens_seen": 204359480, + "step": 6287 + }, + { + "epoch": 0.5670739955810074, + "flos": 22241681697120.0, + "grad_norm": 4.08244942161312, + "learning_rate": 1.6642162089655782e-06, + "loss": 0.7417, + "num_input_tokens_seen": 204387505, + "step": 6288 + }, + { + "epoch": 0.5671641791044776, + "flos": 20780606445600.0, + "grad_norm": 2.1740884822045743, + "learning_rate": 1.663640305299798e-06, + "loss": 0.7109, + "num_input_tokens_seen": 204415370, + "step": 6289 + }, + { + "epoch": 0.5672543626279479, + "flos": 64394245698240.0, + "grad_norm": 0.6616519033993423, + "learning_rate": 1.6630644303340824e-06, + "loss": 0.6098, + "num_input_tokens_seen": 204510470, + "step": 6290 + }, + { + "epoch": 0.5673445461514182, + "flos": 27268828069440.0, + "grad_norm": 2.9301054389391488, + "learning_rate": 1.662488584117567e-06, + "loss": 0.7538, + "num_input_tokens_seen": 204539820, + "step": 6291 + }, + { + "epoch": 0.5674347296748884, + "flos": 35283926887200.0, + "grad_norm": 1.8241892124047354, + "learning_rate": 1.6619127666993867e-06, + "loss": 0.8007, + "num_input_tokens_seen": 204572370, + "step": 6292 + }, + { + "epoch": 0.5675249131983586, + "flos": 20889085006080.0, + "grad_norm": 2.3502354727114394, + "learning_rate": 1.6613369781286727e-06, + "loss": 0.7466, + "num_input_tokens_seen": 204600150, + "step": 6293 + }, + { + "epoch": 0.567615096721829, + "flos": 32879594241600.0, + "grad_norm": 2.0259254507224016, + "learning_rate": 1.6607612184545562e-06, + "loss": 0.6826, + "num_input_tokens_seen": 204631265, + "step": 6294 + }, + { + "epoch": 0.5677052802452992, + "flos": 21764310272160.0, + "grad_norm": 2.0113241753138413, + "learning_rate": 1.6601854877261617e-06, + "loss": 0.8126, + "num_input_tokens_seen": 204659360, + "step": 6295 + }, + { + "epoch": 0.5677954637687694, + "flos": 19290310029120.0, + "grad_norm": 2.1826133642653875, + "learning_rate": 1.6596097859926163e-06, + "loss": 0.7431, + "num_input_tokens_seen": 204684120, + "step": 6296 + }, + { + "epoch": 0.5678856472922397, + "flos": 17978196940320.0, + "grad_norm": 2.380801693825478, + "learning_rate": 1.6590341133030407e-06, + "loss": 0.8289, + "num_input_tokens_seen": 204710310, + "step": 6297 + }, + { + "epoch": 0.56797583081571, + "flos": 26866104989760.0, + "grad_norm": 1.815026528345865, + "learning_rate": 1.658458469706554e-06, + "loss": 0.8213, + "num_input_tokens_seen": 204743710, + "step": 6298 + }, + { + "epoch": 0.5680660143391802, + "flos": 21512965389600.0, + "grad_norm": 1.7245034069714469, + "learning_rate": 1.6578828552522746e-06, + "loss": 0.6993, + "num_input_tokens_seen": 204772500, + "step": 6299 + }, + { + "epoch": 0.5681561978626505, + "flos": 15319880359200.0, + "grad_norm": 1.8882171809383441, + "learning_rate": 1.6573072699893156e-06, + "loss": 0.7923, + "num_input_tokens_seen": 204798100, + "step": 6300 + }, + { + "epoch": 0.5682463813861207, + "flos": 24937954336800.0, + "grad_norm": 3.3721412984853827, + "learning_rate": 1.6567317139667906e-06, + "loss": 0.714, + "num_input_tokens_seen": 204826915, + "step": 6301 + }, + { + "epoch": 0.5683365649095911, + "flos": 24638468881440.0, + "grad_norm": 2.2207732311177275, + "learning_rate": 1.6561561872338087e-06, + "loss": 0.7602, + "num_input_tokens_seen": 204853075, + "step": 6302 + }, + { + "epoch": 0.5684267484330613, + "flos": 57726359689920.0, + "grad_norm": 2.3116137894255613, + "learning_rate": 1.6555806898394764e-06, + "loss": 0.6585, + "num_input_tokens_seen": 204886640, + "step": 6303 + }, + { + "epoch": 0.5685169319565315, + "flos": 25812138849600.0, + "grad_norm": 2.25294471474453, + "learning_rate": 1.6550052218328987e-06, + "loss": 0.716, + "num_input_tokens_seen": 204916440, + "step": 6304 + }, + { + "epoch": 0.5686071154800018, + "flos": 24573521843520.0, + "grad_norm": 2.522908441691349, + "learning_rate": 1.6544297832631777e-06, + "loss": 0.716, + "num_input_tokens_seen": 204945825, + "step": 6305 + }, + { + "epoch": 0.5686972990034721, + "flos": 26539963082880.0, + "grad_norm": 2.820738276838551, + "learning_rate": 1.6538543741794135e-06, + "loss": 0.75, + "num_input_tokens_seen": 204971680, + "step": 6306 + }, + { + "epoch": 0.5687874825269423, + "flos": 31017654398400.0, + "grad_norm": 1.8194252438620058, + "learning_rate": 1.6532789946307028e-06, + "loss": 0.7138, + "num_input_tokens_seen": 205004650, + "step": 6307 + }, + { + "epoch": 0.5688776660504126, + "flos": 68390365405920.0, + "grad_norm": 0.6343910168284145, + "learning_rate": 1.6527036446661393e-06, + "loss": 0.6124, + "num_input_tokens_seen": 205099290, + "step": 6308 + }, + { + "epoch": 0.5689678495738828, + "flos": 18853347866880.0, + "grad_norm": 1.6396045962953973, + "learning_rate": 1.6521283243348165e-06, + "loss": 0.7771, + "num_input_tokens_seen": 205126530, + "step": 6309 + }, + { + "epoch": 0.5690580330973531, + "flos": 21181879904640.0, + "grad_norm": 2.861784873154739, + "learning_rate": 1.6515530336858227e-06, + "loss": 0.7356, + "num_input_tokens_seen": 205155170, + "step": 6310 + }, + { + "epoch": 0.5691482166208234, + "flos": 18889891474560.0, + "grad_norm": 3.194508346434464, + "learning_rate": 1.6509777727682457e-06, + "loss": 0.7918, + "num_input_tokens_seen": 205184080, + "step": 6311 + }, + { + "epoch": 0.5692384001442936, + "flos": 21767544041280.0, + "grad_norm": 1.6132954372862864, + "learning_rate": 1.65040254163117e-06, + "loss": 0.7833, + "num_input_tokens_seen": 205212315, + "step": 6312 + }, + { + "epoch": 0.569328583667764, + "flos": 16121051996160.0, + "grad_norm": 4.393892314662657, + "learning_rate": 1.649827340323676e-06, + "loss": 0.6526, + "num_input_tokens_seen": 205238755, + "step": 6313 + }, + { + "epoch": 0.5694187671912342, + "flos": 26504162870400.0, + "grad_norm": 1.9236125999594877, + "learning_rate": 1.6492521688948454e-06, + "loss": 0.7108, + "num_input_tokens_seen": 205266870, + "step": 6314 + }, + { + "epoch": 0.5695089507147044, + "flos": 22239265662720.0, + "grad_norm": 2.212260036063366, + "learning_rate": 1.6486770273937526e-06, + "loss": 0.7641, + "num_input_tokens_seen": 205297635, + "step": 6315 + }, + { + "epoch": 0.5695991342381747, + "flos": 25555627370400.0, + "grad_norm": 1.5879571391646485, + "learning_rate": 1.6481019158694738e-06, + "loss": 0.7214, + "num_input_tokens_seen": 205329095, + "step": 6316 + }, + { + "epoch": 0.569689317761645, + "flos": 66933862034880.0, + "grad_norm": 0.7581461142597574, + "learning_rate": 1.6475268343710792e-06, + "loss": 0.6513, + "num_input_tokens_seen": 205420330, + "step": 6317 + }, + { + "epoch": 0.5697795012851152, + "flos": 19035285340320.0, + "grad_norm": 2.049962983146107, + "learning_rate": 1.6469517829476396e-06, + "loss": 0.6754, + "num_input_tokens_seen": 205447085, + "step": 6318 + }, + { + "epoch": 0.5698696848085855, + "flos": 20966409574080.0, + "grad_norm": 2.866975615361447, + "learning_rate": 1.64637676164822e-06, + "loss": 0.7259, + "num_input_tokens_seen": 205474855, + "step": 6319 + }, + { + "epoch": 0.5699598683320557, + "flos": 24974906811840.0, + "grad_norm": 1.6219430192564854, + "learning_rate": 1.6458017705218848e-06, + "loss": 0.7804, + "num_input_tokens_seen": 205507115, + "step": 6320 + }, + { + "epoch": 0.570050051855526, + "flos": 22015543645440.0, + "grad_norm": 1.7318679285197607, + "learning_rate": 1.645226809617696e-06, + "loss": 0.7689, + "num_input_tokens_seen": 205537075, + "step": 6321 + }, + { + "epoch": 0.5701402353789963, + "flos": 23292265389120.0, + "grad_norm": 2.0682011040351416, + "learning_rate": 1.6446518789847112e-06, + "loss": 0.7784, + "num_input_tokens_seen": 205563860, + "step": 6322 + }, + { + "epoch": 0.5702304189024665, + "flos": 17541048929280.0, + "grad_norm": 3.231472463715642, + "learning_rate": 1.6440769786719883e-06, + "loss": 0.7642, + "num_input_tokens_seen": 205588750, + "step": 6323 + }, + { + "epoch": 0.5703206024259367, + "flos": 26283934810560.0, + "grad_norm": 1.5993672281950726, + "learning_rate": 1.6435021087285803e-06, + "loss": 0.7845, + "num_input_tokens_seen": 205618285, + "step": 6324 + }, + { + "epoch": 0.5704107859494071, + "flos": 22417486160160.0, + "grad_norm": 2.025952635494796, + "learning_rate": 1.642927269203537e-06, + "loss": 0.784, + "num_input_tokens_seen": 205646150, + "step": 6325 + }, + { + "epoch": 0.5705009694728773, + "flos": 25003012884000.0, + "grad_norm": 2.736444703585967, + "learning_rate": 1.642352460145909e-06, + "loss": 0.7955, + "num_input_tokens_seen": 205674240, + "step": 6326 + }, + { + "epoch": 0.5705911529963476, + "flos": 28908755704320.0, + "grad_norm": 1.7775159815245531, + "learning_rate": 1.6417776816047402e-06, + "loss": 0.7524, + "num_input_tokens_seen": 205703900, + "step": 6327 + }, + { + "epoch": 0.5706813365198178, + "flos": 23006830103040.0, + "grad_norm": 2.3272883812963094, + "learning_rate": 1.6412029336290755e-06, + "loss": 0.7317, + "num_input_tokens_seen": 205730875, + "step": 6328 + }, + { + "epoch": 0.5707715200432881, + "flos": 20456025668640.0, + "grad_norm": 2.8066391808831965, + "learning_rate": 1.6406282162679551e-06, + "loss": 0.7431, + "num_input_tokens_seen": 205756680, + "step": 6329 + }, + { + "epoch": 0.5708617035667584, + "flos": 30765751969440.0, + "grad_norm": 2.9638132910565003, + "learning_rate": 1.6400535295704162e-06, + "loss": 0.7179, + "num_input_tokens_seen": 205788695, + "step": 6330 + }, + { + "epoch": 0.5709518870902286, + "flos": 22860098125920.0, + "grad_norm": 1.7199625073272173, + "learning_rate": 1.6394788735854955e-06, + "loss": 0.6836, + "num_input_tokens_seen": 205818635, + "step": 6331 + }, + { + "epoch": 0.5710420706136988, + "flos": 23547847624320.0, + "grad_norm": 2.507409844798265, + "learning_rate": 1.6389042483622246e-06, + "loss": 0.8124, + "num_input_tokens_seen": 205847555, + "step": 6332 + }, + { + "epoch": 0.5711322541371692, + "flos": 36558269766240.0, + "grad_norm": 4.062998814403559, + "learning_rate": 1.638329653949635e-06, + "loss": 0.7415, + "num_input_tokens_seen": 205878665, + "step": 6333 + }, + { + "epoch": 0.5712224376606394, + "flos": 18157272342240.0, + "grad_norm": 1.7441016659545117, + "learning_rate": 1.637755090396753e-06, + "loss": 0.7094, + "num_input_tokens_seen": 205905390, + "step": 6334 + }, + { + "epoch": 0.5713126211841096, + "flos": 69712031123040.0, + "grad_norm": 0.6493957058025898, + "learning_rate": 1.6371805577526039e-06, + "loss": 0.6101, + "num_input_tokens_seen": 206005600, + "step": 6335 + }, + { + "epoch": 0.5714028047075799, + "flos": 23335796911680.0, + "grad_norm": 2.065329604733875, + "learning_rate": 1.636606056066211e-06, + "loss": 0.6951, + "num_input_tokens_seen": 206033570, + "step": 6336 + }, + { + "epoch": 0.5714929882310502, + "flos": 23001700676160.0, + "grad_norm": 2.4108047972949613, + "learning_rate": 1.636031585386592e-06, + "loss": 0.6781, + "num_input_tokens_seen": 206060920, + "step": 6337 + }, + { + "epoch": 0.5715831717545204, + "flos": 23516767971360.0, + "grad_norm": 1.617155212964607, + "learning_rate": 1.635457145762766e-06, + "loss": 0.8042, + "num_input_tokens_seen": 206090440, + "step": 6338 + }, + { + "epoch": 0.5716733552779907, + "flos": 17823807992640.0, + "grad_norm": 2.4723689920337084, + "learning_rate": 1.6348827372437456e-06, + "loss": 0.7735, + "num_input_tokens_seen": 206113265, + "step": 6339 + }, + { + "epoch": 0.571763538801461, + "flos": 26463716437920.0, + "grad_norm": 1.8168566073709385, + "learning_rate": 1.634308359878544e-06, + "loss": 0.7475, + "num_input_tokens_seen": 206144485, + "step": 6340 + }, + { + "epoch": 0.5718537223249313, + "flos": 41222470435680.0, + "grad_norm": 1.872204413110768, + "learning_rate": 1.6337340137161695e-06, + "loss": 0.8243, + "num_input_tokens_seen": 206174345, + "step": 6341 + }, + { + "epoch": 0.5719439058484015, + "flos": 25769499401280.0, + "grad_norm": 2.1063821541731085, + "learning_rate": 1.6331596988056277e-06, + "loss": 0.7467, + "num_input_tokens_seen": 206201865, + "step": 6342 + }, + { + "epoch": 0.5720340893718717, + "flos": 24969814554720.0, + "grad_norm": 1.907961155553987, + "learning_rate": 1.632585415195924e-06, + "loss": 0.7758, + "num_input_tokens_seen": 206228955, + "step": 6343 + }, + { + "epoch": 0.5721242728953421, + "flos": 25483209210720.0, + "grad_norm": 2.0332677356373745, + "learning_rate": 1.6320111629360583e-06, + "loss": 0.7735, + "num_input_tokens_seen": 206261020, + "step": 6344 + }, + { + "epoch": 0.5722144564188123, + "flos": 21402702680640.0, + "grad_norm": 2.090623143399131, + "learning_rate": 1.631436942075029e-06, + "loss": 0.7222, + "num_input_tokens_seen": 206284380, + "step": 6345 + }, + { + "epoch": 0.5723046399422825, + "flos": 19144581635520.0, + "grad_norm": 2.601966162172009, + "learning_rate": 1.630862752661833e-06, + "loss": 0.7205, + "num_input_tokens_seen": 206311340, + "step": 6346 + }, + { + "epoch": 0.5723948234657528, + "flos": 46215675083520.0, + "grad_norm": 1.7995523160943656, + "learning_rate": 1.6302885947454612e-06, + "loss": 0.6721, + "num_input_tokens_seen": 206348290, + "step": 6347 + }, + { + "epoch": 0.5724850069892231, + "flos": 33061345866240.0, + "grad_norm": 2.1059562953750453, + "learning_rate": 1.6297144683749057e-06, + "loss": 0.7595, + "num_input_tokens_seen": 206376495, + "step": 6348 + }, + { + "epoch": 0.5725751905126933, + "flos": 19436112762240.0, + "grad_norm": 2.042348618051049, + "learning_rate": 1.629140373599153e-06, + "loss": 0.7557, + "num_input_tokens_seen": 206402270, + "step": 6349 + }, + { + "epoch": 0.5726653740361636, + "flos": 27995871737760.0, + "grad_norm": 2.4136807531837747, + "learning_rate": 1.628566310467189e-06, + "loss": 0.7614, + "num_input_tokens_seen": 206432625, + "step": 6350 + }, + { + "epoch": 0.5727555575596338, + "flos": 16375519138560.0, + "grad_norm": 2.141665908674806, + "learning_rate": 1.6279922790279957e-06, + "loss": 0.7102, + "num_input_tokens_seen": 206457685, + "step": 6351 + }, + { + "epoch": 0.5728457410831042, + "flos": 22861027369920.0, + "grad_norm": 2.2081453735206353, + "learning_rate": 1.6274182793305512e-06, + "loss": 0.7617, + "num_input_tokens_seen": 206485725, + "step": 6352 + }, + { + "epoch": 0.5729359246065744, + "flos": 15282593356320.0, + "grad_norm": 3.081153580277021, + "learning_rate": 1.626844311423835e-06, + "loss": 0.77, + "num_input_tokens_seen": 206512355, + "step": 6353 + }, + { + "epoch": 0.5730261081300446, + "flos": 24245038241760.0, + "grad_norm": 2.2319433678993215, + "learning_rate": 1.6262703753568181e-06, + "loss": 0.8258, + "num_input_tokens_seen": 206539385, + "step": 6354 + }, + { + "epoch": 0.5731162916535149, + "flos": 13607088527520.0, + "grad_norm": 2.1422917914514956, + "learning_rate": 1.6256964711784747e-06, + "loss": 0.8216, + "num_input_tokens_seen": 206565190, + "step": 6355 + }, + { + "epoch": 0.5732064751769852, + "flos": 25410084825600.0, + "grad_norm": 2.3671243489217155, + "learning_rate": 1.6251225989377723e-06, + "loss": 0.7622, + "num_input_tokens_seen": 206595045, + "step": 6356 + }, + { + "epoch": 0.5732966587004554, + "flos": 24682074743520.0, + "grad_norm": 1.9078186959830732, + "learning_rate": 1.624548758683676e-06, + "loss": 0.749, + "num_input_tokens_seen": 206621875, + "step": 6357 + }, + { + "epoch": 0.5733868422239257, + "flos": 20232972707040.0, + "grad_norm": 1.9826675638625666, + "learning_rate": 1.6239749504651505e-06, + "loss": 0.8423, + "num_input_tokens_seen": 206647675, + "step": 6358 + }, + { + "epoch": 0.5734770257473959, + "flos": 23151517743360.0, + "grad_norm": 1.8096838152266101, + "learning_rate": 1.6234011743311552e-06, + "loss": 0.7278, + "num_input_tokens_seen": 206677975, + "step": 6359 + }, + { + "epoch": 0.5735672092708662, + "flos": 20995890927360.0, + "grad_norm": 2.085323629168072, + "learning_rate": 1.6228274303306483e-06, + "loss": 0.6835, + "num_input_tokens_seen": 206706520, + "step": 6360 + }, + { + "epoch": 0.5736573927943365, + "flos": 24646014342720.0, + "grad_norm": 6.085021355245915, + "learning_rate": 1.6222537185125847e-06, + "loss": 0.8343, + "num_input_tokens_seen": 206735485, + "step": 6361 + }, + { + "epoch": 0.5737475763178067, + "flos": 22201086585600.0, + "grad_norm": 1.8608924379826381, + "learning_rate": 1.6216800389259172e-06, + "loss": 0.7687, + "num_input_tokens_seen": 206764290, + "step": 6362 + }, + { + "epoch": 0.573837759841277, + "flos": 26721974895840.0, + "grad_norm": 1.5119918852492464, + "learning_rate": 1.6211063916195949e-06, + "loss": 0.7534, + "num_input_tokens_seen": 206795650, + "step": 6363 + }, + { + "epoch": 0.5739279433647473, + "flos": 26066382973440.0, + "grad_norm": 2.0416915139483396, + "learning_rate": 1.6205327766425633e-06, + "loss": 0.7594, + "num_input_tokens_seen": 206822670, + "step": 6364 + }, + { + "epoch": 0.5740181268882175, + "flos": 30761365937760.0, + "grad_norm": 1.8222992496311627, + "learning_rate": 1.6199591940437689e-06, + "loss": 0.7615, + "num_input_tokens_seen": 206853560, + "step": 6365 + }, + { + "epoch": 0.5741083104116878, + "flos": 19181348261760.0, + "grad_norm": 1.7236877951349094, + "learning_rate": 1.6193856438721505e-06, + "loss": 0.7104, + "num_input_tokens_seen": 206882290, + "step": 6366 + }, + { + "epoch": 0.5741984939351581, + "flos": 21476050084320.0, + "grad_norm": 2.374060595593097, + "learning_rate": 1.6188121261766483e-06, + "loss": 0.7587, + "num_input_tokens_seen": 206910570, + "step": 6367 + }, + { + "epoch": 0.5742886774586283, + "flos": 31598077598880.0, + "grad_norm": 1.736613504410046, + "learning_rate": 1.6182386410061976e-06, + "loss": 0.6841, + "num_input_tokens_seen": 206940785, + "step": 6368 + }, + { + "epoch": 0.5743788609820986, + "flos": 23188581727680.0, + "grad_norm": 2.870327012840964, + "learning_rate": 1.61766518840973e-06, + "loss": 0.6595, + "num_input_tokens_seen": 206970395, + "step": 6369 + }, + { + "epoch": 0.5744690445055688, + "flos": 23552902711680.0, + "grad_norm": 2.1208691557111607, + "learning_rate": 1.6170917684361779e-06, + "loss": 0.6237, + "num_input_tokens_seen": 206997210, + "step": 6370 + }, + { + "epoch": 0.5745592280290391, + "flos": 23803467029280.0, + "grad_norm": 1.8474675217226935, + "learning_rate": 1.6165183811344662e-06, + "loss": 0.8283, + "num_input_tokens_seen": 207024805, + "step": 6371 + }, + { + "epoch": 0.5746494115525094, + "flos": 24500137270080.0, + "grad_norm": 3.2681098500796297, + "learning_rate": 1.6159450265535218e-06, + "loss": 0.7472, + "num_input_tokens_seen": 207053830, + "step": 6372 + }, + { + "epoch": 0.5747395950759796, + "flos": 21507687283680.0, + "grad_norm": 2.1766682157564516, + "learning_rate": 1.6153717047422652e-06, + "loss": 0.8353, + "num_input_tokens_seen": 207082910, + "step": 6373 + }, + { + "epoch": 0.5748297785994498, + "flos": 27305669035200.0, + "grad_norm": 2.7212434055597168, + "learning_rate": 1.6147984157496155e-06, + "loss": 0.6374, + "num_input_tokens_seen": 207114490, + "step": 6374 + }, + { + "epoch": 0.5749199621229202, + "flos": 24245372769600.0, + "grad_norm": 1.8555461024172373, + "learning_rate": 1.6142251596244886e-06, + "loss": 0.6776, + "num_input_tokens_seen": 207142820, + "step": 6375 + }, + { + "epoch": 0.5750101456463904, + "flos": 18052139060160.0, + "grad_norm": 1.610363702861302, + "learning_rate": 1.6136519364157983e-06, + "loss": 0.8053, + "num_input_tokens_seen": 207170355, + "step": 6376 + }, + { + "epoch": 0.5751003291698606, + "flos": 14444692262880.0, + "grad_norm": 2.117585655020277, + "learning_rate": 1.6130787461724555e-06, + "loss": 0.6936, + "num_input_tokens_seen": 207197795, + "step": 6377 + }, + { + "epoch": 0.5751905126933309, + "flos": 22530796789440.0, + "grad_norm": 1.647106627764304, + "learning_rate": 1.6125055889433679e-06, + "loss": 0.8289, + "num_input_tokens_seen": 207227075, + "step": 6378 + }, + { + "epoch": 0.5752806962168012, + "flos": 21181731225600.0, + "grad_norm": 1.6821448514982646, + "learning_rate": 1.6119324647774386e-06, + "loss": 0.7762, + "num_input_tokens_seen": 207255100, + "step": 6379 + }, + { + "epoch": 0.5753708797402715, + "flos": 20231820444480.0, + "grad_norm": 1.9464991982921611, + "learning_rate": 1.6113593737235724e-06, + "loss": 0.7596, + "num_input_tokens_seen": 207280285, + "step": 6380 + }, + { + "epoch": 0.5754610632637417, + "flos": 22714964448480.0, + "grad_norm": 2.9738972262881602, + "learning_rate": 1.6107863158306665e-06, + "loss": 0.6989, + "num_input_tokens_seen": 207308845, + "step": 6381 + }, + { + "epoch": 0.5755512467872119, + "flos": 15683234929440.0, + "grad_norm": 2.4530603679719736, + "learning_rate": 1.610213291147619e-06, + "loss": 0.6469, + "num_input_tokens_seen": 207334510, + "step": 6382 + }, + { + "epoch": 0.5756414303106823, + "flos": 21294298459680.0, + "grad_norm": 1.6967410911145193, + "learning_rate": 1.609640299723322e-06, + "loss": 0.7713, + "num_input_tokens_seen": 207363795, + "step": 6383 + }, + { + "epoch": 0.5757316138341525, + "flos": 22787605626720.0, + "grad_norm": 1.6292293969034486, + "learning_rate": 1.609067341606668e-06, + "loss": 0.7367, + "num_input_tokens_seen": 207392350, + "step": 6384 + }, + { + "epoch": 0.5758217973576227, + "flos": 23371076747520.0, + "grad_norm": 2.9216785058845964, + "learning_rate": 1.6084944168465438e-06, + "loss": 0.7444, + "num_input_tokens_seen": 207420415, + "step": 6385 + }, + { + "epoch": 0.5759119808810931, + "flos": 22781695634880.0, + "grad_norm": 1.8560594120505414, + "learning_rate": 1.6079215254918339e-06, + "loss": 0.7724, + "num_input_tokens_seen": 207446825, + "step": 6386 + }, + { + "epoch": 0.5760021644045633, + "flos": 30147001012800.0, + "grad_norm": 2.974215966782747, + "learning_rate": 1.6073486675914222e-06, + "loss": 0.6616, + "num_input_tokens_seen": 207474460, + "step": 6387 + }, + { + "epoch": 0.5760923479280335, + "flos": 16623332893920.0, + "grad_norm": 4.460646103729787, + "learning_rate": 1.606775843194187e-06, + "loss": 0.7798, + "num_input_tokens_seen": 207497105, + "step": 6388 + }, + { + "epoch": 0.5761825314515038, + "flos": 12580968271200.0, + "grad_norm": 2.5111739307842926, + "learning_rate": 1.6062030523490053e-06, + "loss": 0.7422, + "num_input_tokens_seen": 207518475, + "step": 6389 + }, + { + "epoch": 0.5762727149749741, + "flos": 18598620536160.0, + "grad_norm": 3.1695660986711123, + "learning_rate": 1.60563029510475e-06, + "loss": 0.6747, + "num_input_tokens_seen": 207543685, + "step": 6390 + }, + { + "epoch": 0.5763628984984444, + "flos": 67565064861120.0, + "grad_norm": 0.6500387757913156, + "learning_rate": 1.6050575715102927e-06, + "loss": 0.5684, + "num_input_tokens_seen": 207645705, + "step": 6391 + }, + { + "epoch": 0.5764530820219146, + "flos": 31892061929760.0, + "grad_norm": 2.0498526078248047, + "learning_rate": 1.6044848816145014e-06, + "loss": 0.8092, + "num_input_tokens_seen": 207673795, + "step": 6392 + }, + { + "epoch": 0.5765432655453848, + "flos": 17104569973920.0, + "grad_norm": 2.564240213636386, + "learning_rate": 1.60391222546624e-06, + "loss": 0.777, + "num_input_tokens_seen": 207700805, + "step": 6393 + }, + { + "epoch": 0.5766334490688552, + "flos": 24245186920800.0, + "grad_norm": 1.7254000932305442, + "learning_rate": 1.6033396031143725e-06, + "loss": 0.7449, + "num_input_tokens_seen": 207729245, + "step": 6394 + }, + { + "epoch": 0.5767236325923254, + "flos": 23837854790880.0, + "grad_norm": 2.1035721823510323, + "learning_rate": 1.602767014607757e-06, + "loss": 0.7524, + "num_input_tokens_seen": 207756835, + "step": 6395 + }, + { + "epoch": 0.5768138161157956, + "flos": 39402761173440.0, + "grad_norm": 1.6935626021314527, + "learning_rate": 1.6021944599952493e-06, + "loss": 0.7641, + "num_input_tokens_seen": 207789430, + "step": 6396 + }, + { + "epoch": 0.5769039996392659, + "flos": 20379332986560.0, + "grad_norm": 2.12097639739246, + "learning_rate": 1.6016219393257048e-06, + "loss": 0.7647, + "num_input_tokens_seen": 207815195, + "step": 6397 + }, + { + "epoch": 0.5769941831627362, + "flos": 19763964478080.0, + "grad_norm": 1.823072771926928, + "learning_rate": 1.6010494526479722e-06, + "loss": 0.702, + "num_input_tokens_seen": 207843000, + "step": 6398 + }, + { + "epoch": 0.5770843666862064, + "flos": 23334458800320.0, + "grad_norm": 2.957789375549298, + "learning_rate": 1.6004770000109006e-06, + "loss": 0.6958, + "num_input_tokens_seen": 207870535, + "step": 6399 + }, + { + "epoch": 0.5771745502096767, + "flos": 28214687346720.0, + "grad_norm": 2.124157233573328, + "learning_rate": 1.5999045814633348e-06, + "loss": 0.7173, + "num_input_tokens_seen": 207900725, + "step": 6400 + }, + { + "epoch": 0.5772647337331469, + "flos": 25958015922240.0, + "grad_norm": 1.876654403684853, + "learning_rate": 1.5993321970541151e-06, + "loss": 0.7598, + "num_input_tokens_seen": 207929940, + "step": 6401 + }, + { + "epoch": 0.5773549172566173, + "flos": 20565396303360.0, + "grad_norm": 2.0415053660714406, + "learning_rate": 1.5987598468320825e-06, + "loss": 0.7555, + "num_input_tokens_seen": 207956210, + "step": 6402 + }, + { + "epoch": 0.5774451007800875, + "flos": 39471982733760.0, + "grad_norm": 1.8284993236771028, + "learning_rate": 1.5981875308460717e-06, + "loss": 0.7359, + "num_input_tokens_seen": 207989250, + "step": 6403 + }, + { + "epoch": 0.5775352843035577, + "flos": 22824632441280.0, + "grad_norm": 2.197835212632468, + "learning_rate": 1.5976152491449169e-06, + "loss": 0.7363, + "num_input_tokens_seen": 208016700, + "step": 6404 + }, + { + "epoch": 0.577625467827028, + "flos": 20346394845600.0, + "grad_norm": 3.361850553850266, + "learning_rate": 1.5970430017774468e-06, + "loss": 0.8112, + "num_input_tokens_seen": 208043050, + "step": 6405 + }, + { + "epoch": 0.5777156513504983, + "flos": 38198754947520.0, + "grad_norm": 2.436806213237734, + "learning_rate": 1.5964707887924904e-06, + "loss": 0.7637, + "num_input_tokens_seen": 208078210, + "step": 6406 + }, + { + "epoch": 0.5778058348739685, + "flos": 36848871648960.0, + "grad_norm": 1.571742986212909, + "learning_rate": 1.5958986102388714e-06, + "loss": 0.6364, + "num_input_tokens_seen": 208109240, + "step": 6407 + }, + { + "epoch": 0.5778960183974388, + "flos": 37178693362080.0, + "grad_norm": 1.765318177157184, + "learning_rate": 1.5953264661654104e-06, + "loss": 0.7057, + "num_input_tokens_seen": 208138595, + "step": 6408 + }, + { + "epoch": 0.5779862019209091, + "flos": 25226697731520.0, + "grad_norm": 1.819258683633046, + "learning_rate": 1.5947543566209276e-06, + "loss": 0.7856, + "num_input_tokens_seen": 208168150, + "step": 6409 + }, + { + "epoch": 0.5780763854443793, + "flos": 23626138606080.0, + "grad_norm": 2.1037604314372063, + "learning_rate": 1.5941822816542367e-06, + "loss": 0.7336, + "num_input_tokens_seen": 208197340, + "step": 6410 + }, + { + "epoch": 0.5781665689678496, + "flos": 30766272346080.0, + "grad_norm": 2.6590581075309494, + "learning_rate": 1.5936102413141519e-06, + "loss": 0.7071, + "num_input_tokens_seen": 208229700, + "step": 6411 + }, + { + "epoch": 0.5782567524913198, + "flos": 23735137543200.0, + "grad_norm": 2.5177366504226266, + "learning_rate": 1.5930382356494823e-06, + "loss": 0.6714, + "num_input_tokens_seen": 208257135, + "step": 6412 + }, + { + "epoch": 0.5783469360147901, + "flos": 23735583580320.0, + "grad_norm": 2.0960025342387247, + "learning_rate": 1.5924662647090335e-06, + "loss": 0.799, + "num_input_tokens_seen": 208285420, + "step": 6413 + }, + { + "epoch": 0.5784371195382604, + "flos": 22531056977760.0, + "grad_norm": 2.1003636908085688, + "learning_rate": 1.5918943285416108e-06, + "loss": 0.7712, + "num_input_tokens_seen": 208312720, + "step": 6414 + }, + { + "epoch": 0.5785273030617306, + "flos": 40675134055200.0, + "grad_norm": 1.748872666327467, + "learning_rate": 1.5913224271960139e-06, + "loss": 0.6859, + "num_input_tokens_seen": 208348435, + "step": 6415 + }, + { + "epoch": 0.5786174865852008, + "flos": 36117479118720.0, + "grad_norm": 2.4171089776076187, + "learning_rate": 1.590750560721041e-06, + "loss": 0.7633, + "num_input_tokens_seen": 208379435, + "step": 6416 + }, + { + "epoch": 0.5787076701086712, + "flos": 52043026679040.0, + "grad_norm": 1.4538436261218357, + "learning_rate": 1.5901787291654874e-06, + "loss": 0.7145, + "num_input_tokens_seen": 208412495, + "step": 6417 + }, + { + "epoch": 0.5787978536321414, + "flos": 46357760840640.0, + "grad_norm": 1.885444036745255, + "learning_rate": 1.5896069325781435e-06, + "loss": 0.6213, + "num_input_tokens_seen": 208447970, + "step": 6418 + }, + { + "epoch": 0.5788880371556117, + "flos": 24246822390240.0, + "grad_norm": 1.802253700391188, + "learning_rate": 1.5890351710077998e-06, + "loss": 0.7873, + "num_input_tokens_seen": 208476930, + "step": 6419 + }, + { + "epoch": 0.5789782206790819, + "flos": 22892590229760.0, + "grad_norm": 1.7909489975977708, + "learning_rate": 1.5884634445032406e-06, + "loss": 0.8082, + "num_input_tokens_seen": 208506485, + "step": 6420 + }, + { + "epoch": 0.5790684042025522, + "flos": 69683293164960.0, + "grad_norm": 0.6053217063286418, + "learning_rate": 1.5878917531132501e-06, + "loss": 0.5974, + "num_input_tokens_seen": 208612895, + "step": 6421 + }, + { + "epoch": 0.5791585877260225, + "flos": 23079285432480.0, + "grad_norm": 2.0807847519676312, + "learning_rate": 1.5873200968866077e-06, + "loss": 0.6238, + "num_input_tokens_seen": 208640285, + "step": 6422 + }, + { + "epoch": 0.5792487712494927, + "flos": 25337517986880.0, + "grad_norm": 2.9541659664983597, + "learning_rate": 1.586748475872089e-06, + "loss": 0.7223, + "num_input_tokens_seen": 208666520, + "step": 6423 + }, + { + "epoch": 0.5793389547729629, + "flos": 27815309545440.0, + "grad_norm": 2.6958837815939107, + "learning_rate": 1.58617689011847e-06, + "loss": 0.6168, + "num_input_tokens_seen": 208696500, + "step": 6424 + }, + { + "epoch": 0.5794291382964333, + "flos": 23953507115040.0, + "grad_norm": 2.17270562185453, + "learning_rate": 1.5856053396745198e-06, + "loss": 0.7845, + "num_input_tokens_seen": 208726755, + "step": 6425 + }, + { + "epoch": 0.5795193218199035, + "flos": 20343607113600.0, + "grad_norm": 2.196546757405975, + "learning_rate": 1.5850338245890078e-06, + "loss": 0.8147, + "num_input_tokens_seen": 208752600, + "step": 6426 + }, + { + "epoch": 0.5796095053433737, + "flos": 32806804384320.0, + "grad_norm": 2.700449683571193, + "learning_rate": 1.5844623449106974e-06, + "loss": 0.733, + "num_input_tokens_seen": 208780575, + "step": 6427 + }, + { + "epoch": 0.579699688866844, + "flos": 29379994118880.0, + "grad_norm": 1.9199914774109514, + "learning_rate": 1.583890900688351e-06, + "loss": 0.7908, + "num_input_tokens_seen": 208811490, + "step": 6428 + }, + { + "epoch": 0.5797898723903143, + "flos": 13643371946880.0, + "grad_norm": 2.8812261419073906, + "learning_rate": 1.583319491970728e-06, + "loss": 0.8298, + "num_input_tokens_seen": 208836070, + "step": 6429 + }, + { + "epoch": 0.5798800559137846, + "flos": 24352513218720.0, + "grad_norm": 2.092974163189388, + "learning_rate": 1.5827481188065828e-06, + "loss": 0.7651, + "num_input_tokens_seen": 208867075, + "step": 6430 + }, + { + "epoch": 0.5799702394372548, + "flos": 18741932895360.0, + "grad_norm": 2.194642531499026, + "learning_rate": 1.5821767812446689e-06, + "loss": 0.718, + "num_input_tokens_seen": 208895590, + "step": 6431 + }, + { + "epoch": 0.5800604229607251, + "flos": 20310520293600.0, + "grad_norm": 1.99571944159339, + "learning_rate": 1.581605479333736e-06, + "loss": 0.8648, + "num_input_tokens_seen": 208922705, + "step": 6432 + }, + { + "epoch": 0.5801506064841954, + "flos": 22459568062080.0, + "grad_norm": 5.775419687321197, + "learning_rate": 1.5810342131225308e-06, + "loss": 0.7569, + "num_input_tokens_seen": 208950925, + "step": 6433 + }, + { + "epoch": 0.5802407900076656, + "flos": 18705872494560.0, + "grad_norm": 3.1326846431983015, + "learning_rate": 1.580462982659797e-06, + "loss": 0.79, + "num_input_tokens_seen": 208977440, + "step": 6434 + }, + { + "epoch": 0.5803309735311358, + "flos": 11093422416960.0, + "grad_norm": 2.458100830549014, + "learning_rate": 1.5798917879942736e-06, + "loss": 0.8235, + "num_input_tokens_seen": 209003420, + "step": 6435 + }, + { + "epoch": 0.5804211570546062, + "flos": 24281098642560.0, + "grad_norm": 1.841099595960051, + "learning_rate": 1.5793206291747006e-06, + "loss": 0.6912, + "num_input_tokens_seen": 209033920, + "step": 6436 + }, + { + "epoch": 0.5805113405780764, + "flos": 22935452696640.0, + "grad_norm": 1.4672178027646448, + "learning_rate": 1.57874950624981e-06, + "loss": 0.6869, + "num_input_tokens_seen": 209063120, + "step": 6437 + }, + { + "epoch": 0.5806015241015466, + "flos": 22642917986400.0, + "grad_norm": 1.683277711169649, + "learning_rate": 1.5781784192683351e-06, + "loss": 0.7474, + "num_input_tokens_seen": 209094175, + "step": 6438 + }, + { + "epoch": 0.5806917076250169, + "flos": 57033332085600.0, + "grad_norm": 1.40156601615133, + "learning_rate": 1.5776073682790033e-06, + "loss": 0.7096, + "num_input_tokens_seen": 209130850, + "step": 6439 + }, + { + "epoch": 0.5807818911484872, + "flos": 33207668976000.0, + "grad_norm": 2.0368585477845196, + "learning_rate": 1.5770363533305393e-06, + "loss": 0.7752, + "num_input_tokens_seen": 209162175, + "step": 6440 + }, + { + "epoch": 0.5808720746719575, + "flos": 35611964451840.0, + "grad_norm": 2.0350615905889025, + "learning_rate": 1.5764653744716665e-06, + "loss": 0.7512, + "num_input_tokens_seen": 209190180, + "step": 6441 + }, + { + "epoch": 0.5809622581954277, + "flos": 22387707448800.0, + "grad_norm": 2.140287832172841, + "learning_rate": 1.575894431751103e-06, + "loss": 0.7786, + "num_input_tokens_seen": 209215165, + "step": 6442 + }, + { + "epoch": 0.5810524417188979, + "flos": 21512482182720.0, + "grad_norm": 1.9476360248077114, + "learning_rate": 1.575323525217565e-06, + "loss": 0.7779, + "num_input_tokens_seen": 209240980, + "step": 6443 + }, + { + "epoch": 0.5811426252423683, + "flos": 21216118987200.0, + "grad_norm": 2.1748714496726738, + "learning_rate": 1.574752654919766e-06, + "loss": 0.8325, + "num_input_tokens_seen": 209269740, + "step": 6444 + }, + { + "epoch": 0.5812328087658385, + "flos": 15975100584000.0, + "grad_norm": 2.209233790178565, + "learning_rate": 1.5741818209064146e-06, + "loss": 0.7212, + "num_input_tokens_seen": 209295850, + "step": 6445 + }, + { + "epoch": 0.5813229922893087, + "flos": 36661953427680.0, + "grad_norm": 2.517741479893521, + "learning_rate": 1.5736110232262183e-06, + "loss": 0.7601, + "num_input_tokens_seen": 209326075, + "step": 6446 + }, + { + "epoch": 0.581413175812779, + "flos": 21180839151360.0, + "grad_norm": 1.9892183804618426, + "learning_rate": 1.5730402619278804e-06, + "loss": 0.7939, + "num_input_tokens_seen": 209353870, + "step": 6447 + }, + { + "epoch": 0.5815033593362493, + "flos": 65962833096480.0, + "grad_norm": 0.6673342049893426, + "learning_rate": 1.5724695370601024e-06, + "loss": 0.5554, + "num_input_tokens_seen": 209446630, + "step": 6448 + }, + { + "epoch": 0.5815935428597195, + "flos": 21184630466880.0, + "grad_norm": 1.975154390215762, + "learning_rate": 1.5718988486715798e-06, + "loss": 0.6715, + "num_input_tokens_seen": 209476485, + "step": 6449 + }, + { + "epoch": 0.5816837263831898, + "flos": 25193164874400.0, + "grad_norm": 1.8795505616495123, + "learning_rate": 1.5713281968110087e-06, + "loss": 0.732, + "num_input_tokens_seen": 209505510, + "step": 6450 + }, + { + "epoch": 0.58177390990666, + "flos": 25041229130880.0, + "grad_norm": 1.9413017997045205, + "learning_rate": 1.5707575815270796e-06, + "loss": 0.7328, + "num_input_tokens_seen": 209534260, + "step": 6451 + }, + { + "epoch": 0.5818640934301303, + "flos": 20820792689760.0, + "grad_norm": 2.1429148212118085, + "learning_rate": 1.57018700286848e-06, + "loss": 0.7516, + "num_input_tokens_seen": 209562570, + "step": 6452 + }, + { + "epoch": 0.5819542769536006, + "flos": 14809013246880.0, + "grad_norm": 2.299196691792868, + "learning_rate": 1.5696164608838956e-06, + "loss": 0.7657, + "num_input_tokens_seen": 209587430, + "step": 6453 + }, + { + "epoch": 0.5820444604770708, + "flos": 26831494209600.0, + "grad_norm": 1.8173576898493633, + "learning_rate": 1.5690459556220073e-06, + "loss": 0.7705, + "num_input_tokens_seen": 209616865, + "step": 6454 + }, + { + "epoch": 0.582134644000541, + "flos": 22533249993600.0, + "grad_norm": 1.9529671567078242, + "learning_rate": 1.5684754871314949e-06, + "loss": 0.8095, + "num_input_tokens_seen": 209644995, + "step": 6455 + }, + { + "epoch": 0.5822248275240114, + "flos": 18634086220800.0, + "grad_norm": 1.9312154303854763, + "learning_rate": 1.5679050554610335e-06, + "loss": 0.7519, + "num_input_tokens_seen": 209670190, + "step": 6456 + }, + { + "epoch": 0.5823150110474816, + "flos": 25301346076800.0, + "grad_norm": 2.3273050797378465, + "learning_rate": 1.567334660659295e-06, + "loss": 0.7539, + "num_input_tokens_seen": 209700590, + "step": 6457 + }, + { + "epoch": 0.5824051945709519, + "flos": 37761086559840.0, + "grad_norm": 1.772150253095518, + "learning_rate": 1.5667643027749488e-06, + "loss": 0.7675, + "num_input_tokens_seen": 209732115, + "step": 6458 + }, + { + "epoch": 0.5824953780944222, + "flos": 28543765664640.0, + "grad_norm": 2.3933999903790704, + "learning_rate": 1.5661939818566614e-06, + "loss": 0.816, + "num_input_tokens_seen": 209760895, + "step": 6459 + }, + { + "epoch": 0.5825855616178924, + "flos": 27993901740480.0, + "grad_norm": 2.0138061360278496, + "learning_rate": 1.5656236979530956e-06, + "loss": 0.6158, + "num_input_tokens_seen": 209790770, + "step": 6460 + }, + { + "epoch": 0.5826757451413627, + "flos": 24493781241120.0, + "grad_norm": 2.0937164076202706, + "learning_rate": 1.5650534511129106e-06, + "loss": 0.6906, + "num_input_tokens_seen": 209819930, + "step": 6461 + }, + { + "epoch": 0.5827659286648329, + "flos": 20711422055040.0, + "grad_norm": 2.2537054703005874, + "learning_rate": 1.5644832413847635e-06, + "loss": 0.7334, + "num_input_tokens_seen": 209847645, + "step": 6462 + }, + { + "epoch": 0.5828561121883032, + "flos": 21766280269440.0, + "grad_norm": 2.4329839098453583, + "learning_rate": 1.5639130688173082e-06, + "loss": 0.7912, + "num_input_tokens_seen": 209876155, + "step": 6463 + }, + { + "epoch": 0.5829462957117735, + "flos": 23586175380480.0, + "grad_norm": 1.78724631530066, + "learning_rate": 1.5633429334591932e-06, + "loss": 0.7289, + "num_input_tokens_seen": 209904415, + "step": 6464 + }, + { + "epoch": 0.5830364792352437, + "flos": 18342889621920.0, + "grad_norm": 2.0599338867170625, + "learning_rate": 1.562772835359068e-06, + "loss": 0.7854, + "num_input_tokens_seen": 209929770, + "step": 6465 + }, + { + "epoch": 0.5831266627587139, + "flos": 19072237815360.0, + "grad_norm": 3.4266915759308096, + "learning_rate": 1.5622027745655753e-06, + "loss": 0.7208, + "num_input_tokens_seen": 209955960, + "step": 6466 + }, + { + "epoch": 0.5832168462821843, + "flos": 19690951602240.0, + "grad_norm": 1.8301291280537972, + "learning_rate": 1.561632751127355e-06, + "loss": 0.7064, + "num_input_tokens_seen": 209984200, + "step": 6467 + }, + { + "epoch": 0.5833070298056545, + "flos": 21002507144640.0, + "grad_norm": 2.2979700501351967, + "learning_rate": 1.561062765093046e-06, + "loss": 0.7921, + "num_input_tokens_seen": 210010535, + "step": 6468 + }, + { + "epoch": 0.5833972133291248, + "flos": 19943746105440.0, + "grad_norm": 40.550037509134164, + "learning_rate": 1.5604928165112817e-06, + "loss": 0.7404, + "num_input_tokens_seen": 210038940, + "step": 6469 + }, + { + "epoch": 0.583487396852595, + "flos": 26431187164320.0, + "grad_norm": 2.0733318839522976, + "learning_rate": 1.5599229054306945e-06, + "loss": 0.7422, + "num_input_tokens_seen": 210070180, + "step": 6470 + }, + { + "epoch": 0.5835775803760653, + "flos": 22569867940800.0, + "grad_norm": 2.099270389604601, + "learning_rate": 1.5593530318999111e-06, + "loss": 0.7451, + "num_input_tokens_seen": 210097300, + "step": 6471 + }, + { + "epoch": 0.5836677638995356, + "flos": 22787977324320.0, + "grad_norm": 1.745724769196908, + "learning_rate": 1.5587831959675572e-06, + "loss": 0.7757, + "num_input_tokens_seen": 210124575, + "step": 6472 + }, + { + "epoch": 0.5837579474230058, + "flos": 37392714072000.0, + "grad_norm": 2.574038509374267, + "learning_rate": 1.5582133976822534e-06, + "loss": 0.7346, + "num_input_tokens_seen": 210157190, + "step": 6473 + }, + { + "epoch": 0.583848130946476, + "flos": 59909869559520.0, + "grad_norm": 1.7490993634500687, + "learning_rate": 1.5576436370926185e-06, + "loss": 0.7072, + "num_input_tokens_seen": 210190525, + "step": 6474 + }, + { + "epoch": 0.5839383144699464, + "flos": 21512519352480.0, + "grad_norm": 1.6945491074025212, + "learning_rate": 1.5570739142472679e-06, + "loss": 0.7323, + "num_input_tokens_seen": 210219920, + "step": 6475 + }, + { + "epoch": 0.5840284979934166, + "flos": 24901856766240.0, + "grad_norm": 1.855964289297009, + "learning_rate": 1.5565042291948127e-06, + "loss": 0.8157, + "num_input_tokens_seen": 210250475, + "step": 6476 + }, + { + "epoch": 0.5841186815168868, + "flos": 19691955185760.0, + "grad_norm": 2.1871182789853805, + "learning_rate": 1.5559345819838624e-06, + "loss": 0.7851, + "num_input_tokens_seen": 210278515, + "step": 6477 + }, + { + "epoch": 0.5842088650403571, + "flos": 23696177901120.0, + "grad_norm": 1.6460889384412578, + "learning_rate": 1.5553649726630226e-06, + "loss": 0.7333, + "num_input_tokens_seen": 210308610, + "step": 6478 + }, + { + "epoch": 0.5842990485638274, + "flos": 20163974165280.0, + "grad_norm": 2.2064244879433543, + "learning_rate": 1.5547954012808942e-06, + "loss": 0.6826, + "num_input_tokens_seen": 210338550, + "step": 6479 + }, + { + "epoch": 0.5843892320872976, + "flos": 23763801161760.0, + "grad_norm": 2.1522588972753436, + "learning_rate": 1.5542258678860776e-06, + "loss": 0.7276, + "num_input_tokens_seen": 210367215, + "step": 6480 + }, + { + "epoch": 0.5844794156107679, + "flos": 24020349810720.0, + "grad_norm": 2.2981441390978556, + "learning_rate": 1.553656372527167e-06, + "loss": 0.8729, + "num_input_tokens_seen": 210394805, + "step": 6481 + }, + { + "epoch": 0.5845695991342382, + "flos": 22314360045120.0, + "grad_norm": 3.0372878828669903, + "learning_rate": 1.5530869152527568e-06, + "loss": 0.8458, + "num_input_tokens_seen": 210422320, + "step": 6482 + }, + { + "epoch": 0.5846597826577085, + "flos": 25738085220480.0, + "grad_norm": 1.7620960178431346, + "learning_rate": 1.5525174961114353e-06, + "loss": 0.768, + "num_input_tokens_seen": 210452570, + "step": 6483 + }, + { + "epoch": 0.5847499661811787, + "flos": 18044667938400.0, + "grad_norm": 1.8365690403150516, + "learning_rate": 1.5519481151517875e-06, + "loss": 0.7832, + "num_input_tokens_seen": 210478145, + "step": 6484 + }, + { + "epoch": 0.5848401497046489, + "flos": 27197041795680.0, + "grad_norm": 1.944156778340129, + "learning_rate": 1.551378772422398e-06, + "loss": 0.7532, + "num_input_tokens_seen": 210508615, + "step": 6485 + }, + { + "epoch": 0.5849303332281193, + "flos": 28759161655680.0, + "grad_norm": 1.9844374759480998, + "learning_rate": 1.5508094679718447e-06, + "loss": 0.7645, + "num_input_tokens_seen": 210538095, + "step": 6486 + }, + { + "epoch": 0.5850205167515895, + "flos": 27959402469600.0, + "grad_norm": 2.475968154203253, + "learning_rate": 1.5502402018487048e-06, + "loss": 0.6533, + "num_input_tokens_seen": 210570505, + "step": 6487 + }, + { + "epoch": 0.5851107002750597, + "flos": 28289149843200.0, + "grad_norm": 2.208835856523624, + "learning_rate": 1.54967097410155e-06, + "loss": 0.7552, + "num_input_tokens_seen": 210599870, + "step": 6488 + }, + { + "epoch": 0.58520088379853, + "flos": 31640568368160.0, + "grad_norm": 1.7811559685766887, + "learning_rate": 1.5491017847789519e-06, + "loss": 0.6701, + "num_input_tokens_seen": 210630380, + "step": 6489 + }, + { + "epoch": 0.5852910673220003, + "flos": 27079902681120.0, + "grad_norm": 2.3265946241918902, + "learning_rate": 1.5485326339294755e-06, + "loss": 0.6937, + "num_input_tokens_seen": 210659365, + "step": 6490 + }, + { + "epoch": 0.5853812508454705, + "flos": 21001540730880.0, + "grad_norm": 2.0917022712179802, + "learning_rate": 1.5479635216016832e-06, + "loss": 0.7191, + "num_input_tokens_seen": 210688270, + "step": 6491 + }, + { + "epoch": 0.5854714343689408, + "flos": 29382633171840.0, + "grad_norm": 1.9437411906051942, + "learning_rate": 1.547394447844137e-06, + "loss": 0.6669, + "num_input_tokens_seen": 210716785, + "step": 6492 + }, + { + "epoch": 0.585561617892411, + "flos": 54047200958400.0, + "grad_norm": 1.721048448185462, + "learning_rate": 1.546825412705391e-06, + "loss": 0.6674, + "num_input_tokens_seen": 210751405, + "step": 6493 + }, + { + "epoch": 0.5856518014158814, + "flos": 21545160135360.0, + "grad_norm": 1.993795845198227, + "learning_rate": 1.5462564162340007e-06, + "loss": 0.7787, + "num_input_tokens_seen": 210780255, + "step": 6494 + }, + { + "epoch": 0.5857419849393516, + "flos": 14546294417760.0, + "grad_norm": 2.0744934474853536, + "learning_rate": 1.5456874584785144e-06, + "loss": 0.7546, + "num_input_tokens_seen": 210804585, + "step": 6495 + }, + { + "epoch": 0.5858321684628218, + "flos": 34081890658560.0, + "grad_norm": 5.329051007210493, + "learning_rate": 1.5451185394874785e-06, + "loss": 0.7251, + "num_input_tokens_seen": 210833360, + "step": 6496 + }, + { + "epoch": 0.5859223519862921, + "flos": 22788534870720.0, + "grad_norm": 2.028656227826253, + "learning_rate": 1.5445496593094381e-06, + "loss": 0.7621, + "num_input_tokens_seen": 210863000, + "step": 6497 + }, + { + "epoch": 0.5860125355097624, + "flos": 19071643099200.0, + "grad_norm": 2.01847460495171, + "learning_rate": 1.5439808179929316e-06, + "loss": 0.7655, + "num_input_tokens_seen": 210888285, + "step": 6498 + }, + { + "epoch": 0.5861027190332326, + "flos": 19581841155840.0, + "grad_norm": 2.1418173038135255, + "learning_rate": 1.543412015586496e-06, + "loss": 0.7115, + "num_input_tokens_seen": 210917385, + "step": 6499 + }, + { + "epoch": 0.5861929025567029, + "flos": 31382904626400.0, + "grad_norm": 1.8387945791891944, + "learning_rate": 1.5428432521386655e-06, + "loss": 0.7156, + "num_input_tokens_seen": 210947885, + "step": 6500 + }, + { + "epoch": 0.5862830860801731, + "flos": 29125415467200.0, + "grad_norm": 2.247825419424502, + "learning_rate": 1.5422745276979688e-06, + "loss": 0.7861, + "num_input_tokens_seen": 210978045, + "step": 6501 + }, + { + "epoch": 0.5863732696036434, + "flos": 21433968182400.0, + "grad_norm": 1.7337041021211053, + "learning_rate": 1.5417058423129336e-06, + "loss": 0.7422, + "num_input_tokens_seen": 211004070, + "step": 6502 + }, + { + "epoch": 0.5864634531271137, + "flos": 26212631743680.0, + "grad_norm": 2.4356074688921434, + "learning_rate": 1.5411371960320822e-06, + "loss": 0.7856, + "num_input_tokens_seen": 211034495, + "step": 6503 + }, + { + "epoch": 0.5865536366505839, + "flos": 26213746836480.0, + "grad_norm": 1.6061905400889296, + "learning_rate": 1.5405685889039363e-06, + "loss": 0.7846, + "num_input_tokens_seen": 211064025, + "step": 6504 + }, + { + "epoch": 0.5866438201740543, + "flos": 30618573955200.0, + "grad_norm": 2.1695460724590436, + "learning_rate": 1.5400000209770118e-06, + "loss": 0.7482, + "num_input_tokens_seen": 211094890, + "step": 6505 + }, + { + "epoch": 0.5867340036975245, + "flos": 33091793633280.0, + "grad_norm": 2.153660532398794, + "learning_rate": 1.5394314922998208e-06, + "loss": 0.761, + "num_input_tokens_seen": 211123300, + "step": 6506 + }, + { + "epoch": 0.5868241872209947, + "flos": 24499133686560.0, + "grad_norm": 2.3329231946249416, + "learning_rate": 1.5388630029208756e-06, + "loss": 0.8002, + "num_input_tokens_seen": 211145540, + "step": 6507 + }, + { + "epoch": 0.586914370744465, + "flos": 35499768915360.0, + "grad_norm": 1.8435683080083376, + "learning_rate": 1.5382945528886806e-06, + "loss": 0.7742, + "num_input_tokens_seen": 211180120, + "step": 6508 + }, + { + "epoch": 0.5870045542679353, + "flos": 22344844981920.0, + "grad_norm": 2.2667882629578027, + "learning_rate": 1.5377261422517412e-06, + "loss": 0.737, + "num_input_tokens_seen": 211205795, + "step": 6509 + }, + { + "epoch": 0.5870947377914055, + "flos": 20092187891520.0, + "grad_norm": 1.8932564447913596, + "learning_rate": 1.5371577710585553e-06, + "loss": 0.7086, + "num_input_tokens_seen": 211233725, + "step": 6510 + }, + { + "epoch": 0.5871849213148758, + "flos": 70880571664320.0, + "grad_norm": 0.8420978901879137, + "learning_rate": 1.536589439357621e-06, + "loss": 0.5646, + "num_input_tokens_seen": 211332340, + "step": 6511 + }, + { + "epoch": 0.587275104838346, + "flos": 34078954247520.0, + "grad_norm": 1.8645580084020361, + "learning_rate": 1.5360211471974315e-06, + "loss": 0.754, + "num_input_tokens_seen": 211361720, + "step": 6512 + }, + { + "epoch": 0.5873652883618163, + "flos": 21329652635040.0, + "grad_norm": 2.4797448731523355, + "learning_rate": 1.5354528946264753e-06, + "loss": 0.8035, + "num_input_tokens_seen": 211389445, + "step": 6513 + }, + { + "epoch": 0.5874554718852866, + "flos": 37470707695680.0, + "grad_norm": 1.814972013883021, + "learning_rate": 1.5348846816932399e-06, + "loss": 0.6844, + "num_input_tokens_seen": 211419920, + "step": 6514 + }, + { + "epoch": 0.5875456554087568, + "flos": 68196230517600.0, + "grad_norm": 0.6300998284326201, + "learning_rate": 1.5343165084462077e-06, + "loss": 0.5735, + "num_input_tokens_seen": 211511895, + "step": 6515 + }, + { + "epoch": 0.587635838932227, + "flos": 21804682365120.0, + "grad_norm": 2.6616263339673925, + "learning_rate": 1.5337483749338595e-06, + "loss": 0.7021, + "num_input_tokens_seen": 211539590, + "step": 6516 + }, + { + "epoch": 0.5877260224556974, + "flos": 25367185188960.0, + "grad_norm": 2.1808522165035096, + "learning_rate": 1.5331802812046708e-06, + "loss": 0.7197, + "num_input_tokens_seen": 211566150, + "step": 6517 + }, + { + "epoch": 0.5878162059791676, + "flos": 38526755342400.0, + "grad_norm": 1.7698388242884548, + "learning_rate": 1.5326122273071133e-06, + "loss": 0.5997, + "num_input_tokens_seen": 211598505, + "step": 6518 + }, + { + "epoch": 0.5879063895026378, + "flos": 21433224787200.0, + "grad_norm": 1.8612611324266877, + "learning_rate": 1.532044213289659e-06, + "loss": 0.7875, + "num_input_tokens_seen": 211624295, + "step": 6519 + }, + { + "epoch": 0.5879965730261081, + "flos": 28544583399360.0, + "grad_norm": 1.8106070532154581, + "learning_rate": 1.5314762392007718e-06, + "loss": 0.6357, + "num_input_tokens_seen": 211655170, + "step": 6520 + }, + { + "epoch": 0.5880867565495784, + "flos": 24169014615360.0, + "grad_norm": 2.035991709551767, + "learning_rate": 1.530908305088916e-06, + "loss": 0.7992, + "num_input_tokens_seen": 211684815, + "step": 6521 + }, + { + "epoch": 0.5881769400730487, + "flos": 25410976899840.0, + "grad_norm": 2.0068707465427007, + "learning_rate": 1.5303404110025501e-06, + "loss": 0.7436, + "num_input_tokens_seen": 211714195, + "step": 6522 + }, + { + "epoch": 0.5882671235965189, + "flos": 23331708238080.0, + "grad_norm": 1.5714548994311115, + "learning_rate": 1.5297725569901293e-06, + "loss": 0.7067, + "num_input_tokens_seen": 211744165, + "step": 6523 + }, + { + "epoch": 0.5883573071199891, + "flos": 30395297975040.0, + "grad_norm": 1.8534268035348473, + "learning_rate": 1.5292047431001077e-06, + "loss": 0.6477, + "num_input_tokens_seen": 211770650, + "step": 6524 + }, + { + "epoch": 0.5884474906434595, + "flos": 62135344088160.0, + "grad_norm": 0.6711773244134581, + "learning_rate": 1.5286369693809321e-06, + "loss": 0.5882, + "num_input_tokens_seen": 211860470, + "step": 6525 + }, + { + "epoch": 0.5885376741669297, + "flos": 17650642582560.0, + "grad_norm": 3.184278917594472, + "learning_rate": 1.5280692358810506e-06, + "loss": 0.7592, + "num_input_tokens_seen": 211887805, + "step": 6526 + }, + { + "epoch": 0.5886278576903999, + "flos": 22169300707200.0, + "grad_norm": 2.0044195696327427, + "learning_rate": 1.527501542648904e-06, + "loss": 0.6991, + "num_input_tokens_seen": 211916430, + "step": 6527 + }, + { + "epoch": 0.5887180412138703, + "flos": 24858473922720.0, + "grad_norm": 1.9531401366600956, + "learning_rate": 1.5269338897329308e-06, + "loss": 0.6565, + "num_input_tokens_seen": 211943740, + "step": 6528 + }, + { + "epoch": 0.5888082247373405, + "flos": 28686743496000.0, + "grad_norm": 2.326262477900735, + "learning_rate": 1.5263662771815662e-06, + "loss": 0.7755, + "num_input_tokens_seen": 211972245, + "step": 6529 + }, + { + "epoch": 0.5888984082608107, + "flos": 30065996638560.0, + "grad_norm": 2.4110993900199733, + "learning_rate": 1.5257987050432429e-06, + "loss": 0.8242, + "num_input_tokens_seen": 212000705, + "step": 6530 + }, + { + "epoch": 0.588988591784281, + "flos": 24208903501440.0, + "grad_norm": 4.419914754675551, + "learning_rate": 1.5252311733663887e-06, + "loss": 0.6538, + "num_input_tokens_seen": 212030140, + "step": 6531 + }, + { + "epoch": 0.5890787753077513, + "flos": 16193841853440.0, + "grad_norm": 3.6019739673103746, + "learning_rate": 1.5246636821994281e-06, + "loss": 0.6921, + "num_input_tokens_seen": 212056590, + "step": 6532 + }, + { + "epoch": 0.5891689588312216, + "flos": 31820089807200.0, + "grad_norm": 1.9084776697448982, + "learning_rate": 1.524096231590784e-06, + "loss": 0.7964, + "num_input_tokens_seen": 212087880, + "step": 6533 + }, + { + "epoch": 0.5892591423546918, + "flos": 23188767576480.0, + "grad_norm": 1.885345717582783, + "learning_rate": 1.5235288215888736e-06, + "loss": 0.7939, + "num_input_tokens_seen": 212113960, + "step": 6534 + }, + { + "epoch": 0.589349325878162, + "flos": 24937136602080.0, + "grad_norm": 2.3567118993875953, + "learning_rate": 1.5229614522421102e-06, + "loss": 0.7543, + "num_input_tokens_seen": 212139170, + "step": 6535 + }, + { + "epoch": 0.5894395094016324, + "flos": 18197049719040.0, + "grad_norm": 2.0203113340535235, + "learning_rate": 1.5223941235989071e-06, + "loss": 0.5881, + "num_input_tokens_seen": 212167185, + "step": 6536 + }, + { + "epoch": 0.5895296929251026, + "flos": 24643040761920.0, + "grad_norm": 2.1303498859211114, + "learning_rate": 1.52182683570767e-06, + "loss": 0.7195, + "num_input_tokens_seen": 212197725, + "step": 6537 + }, + { + "epoch": 0.5896198764485728, + "flos": 15568065812160.0, + "grad_norm": 1.8342143311720693, + "learning_rate": 1.5212595886168046e-06, + "loss": 0.7718, + "num_input_tokens_seen": 212223375, + "step": 6538 + }, + { + "epoch": 0.5897100599720431, + "flos": 24203105018880.0, + "grad_norm": 1.7114714046388517, + "learning_rate": 1.520692382374711e-06, + "loss": 0.7629, + "num_input_tokens_seen": 212252080, + "step": 6539 + }, + { + "epoch": 0.5898002434955134, + "flos": 24676462109760.0, + "grad_norm": 2.375250150657451, + "learning_rate": 1.5201252170297854e-06, + "loss": 0.7744, + "num_input_tokens_seen": 212279505, + "step": 6540 + }, + { + "epoch": 0.5898904270189836, + "flos": 26394457707840.0, + "grad_norm": 2.2410355110742337, + "learning_rate": 1.5195580926304232e-06, + "loss": 0.6378, + "num_input_tokens_seen": 212309830, + "step": 6541 + }, + { + "epoch": 0.5899806105424539, + "flos": 17865481027200.0, + "grad_norm": 2.5322229846556867, + "learning_rate": 1.5189910092250131e-06, + "loss": 0.7684, + "num_input_tokens_seen": 212336605, + "step": 6542 + }, + { + "epoch": 0.5900707940659241, + "flos": 32187830409120.0, + "grad_norm": 2.462995558703341, + "learning_rate": 1.5184239668619427e-06, + "loss": 0.7592, + "num_input_tokens_seen": 212366515, + "step": 6543 + }, + { + "epoch": 0.5901609775893945, + "flos": 30213880878240.0, + "grad_norm": 1.781006445044669, + "learning_rate": 1.5178569655895946e-06, + "loss": 0.7495, + "num_input_tokens_seen": 212395725, + "step": 6544 + }, + { + "epoch": 0.5902511611128647, + "flos": 21658842462240.0, + "grad_norm": 1.5252531529412587, + "learning_rate": 1.5172900054563487e-06, + "loss": 0.7799, + "num_input_tokens_seen": 212425820, + "step": 6545 + }, + { + "epoch": 0.5903413446363349, + "flos": 22459791080640.0, + "grad_norm": 2.5550465929295725, + "learning_rate": 1.5167230865105814e-06, + "loss": 0.768, + "num_input_tokens_seen": 212453970, + "step": 6546 + }, + { + "epoch": 0.5904315281598052, + "flos": 31204200922080.0, + "grad_norm": 2.5136797474320685, + "learning_rate": 1.5161562088006644e-06, + "loss": 0.6667, + "num_input_tokens_seen": 212483150, + "step": 6547 + }, + { + "epoch": 0.5905217116832755, + "flos": 23116200737760.0, + "grad_norm": 2.2303046652170013, + "learning_rate": 1.5155893723749685e-06, + "loss": 0.774, + "num_input_tokens_seen": 212510615, + "step": 6548 + }, + { + "epoch": 0.5906118952067457, + "flos": 27560061838080.0, + "grad_norm": 1.8645722657058124, + "learning_rate": 1.5150225772818582e-06, + "loss": 0.7557, + "num_input_tokens_seen": 212541285, + "step": 6549 + }, + { + "epoch": 0.590702078730216, + "flos": 26359363720800.0, + "grad_norm": 1.971467277342364, + "learning_rate": 1.5144558235696949e-06, + "loss": 0.8008, + "num_input_tokens_seen": 212572485, + "step": 6550 + }, + { + "epoch": 0.5907922622536863, + "flos": 28762952971200.0, + "grad_norm": 1.7674882223122657, + "learning_rate": 1.5138891112868388e-06, + "loss": 0.6666, + "num_input_tokens_seen": 212604265, + "step": 6551 + }, + { + "epoch": 0.5908824457771565, + "flos": 71924687818080.0, + "grad_norm": 0.6940332066633197, + "learning_rate": 1.5133224404816433e-06, + "loss": 0.5781, + "num_input_tokens_seen": 212696925, + "step": 6552 + }, + { + "epoch": 0.5909726293006268, + "flos": 19466895057120.0, + "grad_norm": 1.8785262328268997, + "learning_rate": 1.5127558112024617e-06, + "loss": 0.6776, + "num_input_tokens_seen": 212722600, + "step": 6553 + }, + { + "epoch": 0.591062812824097, + "flos": 69248003641920.0, + "grad_norm": 0.6135807642328781, + "learning_rate": 1.5121892234976404e-06, + "loss": 0.5918, + "num_input_tokens_seen": 212822580, + "step": 6554 + }, + { + "epoch": 0.5911529963475673, + "flos": 20739788315520.0, + "grad_norm": 1.9393652889341733, + "learning_rate": 1.5116226774155243e-06, + "loss": 0.7359, + "num_input_tokens_seen": 212847770, + "step": 6555 + }, + { + "epoch": 0.5912431798710376, + "flos": 21366716619360.0, + "grad_norm": 3.515472001879135, + "learning_rate": 1.5110561730044547e-06, + "loss": 0.7208, + "num_input_tokens_seen": 212875880, + "step": 6556 + }, + { + "epoch": 0.5913333633945078, + "flos": 18379098701760.0, + "grad_norm": 1.8797030276246185, + "learning_rate": 1.510489710312768e-06, + "loss": 0.7508, + "num_input_tokens_seen": 212903885, + "step": 6557 + }, + { + "epoch": 0.591423546917978, + "flos": 22497003744000.0, + "grad_norm": 2.251102918938637, + "learning_rate": 1.5099232893887987e-06, + "loss": 0.7417, + "num_input_tokens_seen": 212935205, + "step": 6558 + }, + { + "epoch": 0.5915137304414484, + "flos": 31929757800000.0, + "grad_norm": 1.9542589481691068, + "learning_rate": 1.5093569102808758e-06, + "loss": 0.7195, + "num_input_tokens_seen": 212966575, + "step": 6559 + }, + { + "epoch": 0.5916039139649186, + "flos": 38854607058240.0, + "grad_norm": 2.222061909085257, + "learning_rate": 1.5087905730373275e-06, + "loss": 0.7534, + "num_input_tokens_seen": 212998070, + "step": 6560 + }, + { + "epoch": 0.5916940974883889, + "flos": 71139387668640.0, + "grad_norm": 0.617233033251035, + "learning_rate": 1.508224277706476e-06, + "loss": 0.59, + "num_input_tokens_seen": 213098825, + "step": 6561 + }, + { + "epoch": 0.5917842810118591, + "flos": 23553237239520.0, + "grad_norm": 4.990832889743916, + "learning_rate": 1.5076580243366399e-06, + "loss": 0.7036, + "num_input_tokens_seen": 213127245, + "step": 6562 + }, + { + "epoch": 0.5918744645353294, + "flos": 52588950608640.0, + "grad_norm": 2.5159749800525133, + "learning_rate": 1.507091812976137e-06, + "loss": 0.6564, + "num_input_tokens_seen": 213161175, + "step": 6563 + }, + { + "epoch": 0.5919646480587997, + "flos": 29053703532960.0, + "grad_norm": 1.9367807960804548, + "learning_rate": 1.5065256436732773e-06, + "loss": 0.6976, + "num_input_tokens_seen": 213192075, + "step": 6564 + }, + { + "epoch": 0.5920548315822699, + "flos": 15312520746720.0, + "grad_norm": 2.759559834252702, + "learning_rate": 1.5059595164763717e-06, + "loss": 0.6766, + "num_input_tokens_seen": 213217800, + "step": 6565 + }, + { + "epoch": 0.5921450151057401, + "flos": 56796960882720.0, + "grad_norm": 0.5774124238875973, + "learning_rate": 1.5053934314337243e-06, + "loss": 0.5849, + "num_input_tokens_seen": 213312885, + "step": 6566 + }, + { + "epoch": 0.5922351986292105, + "flos": 30363995303520.0, + "grad_norm": 1.6007075563207749, + "learning_rate": 1.5048273885936356e-06, + "loss": 0.737, + "num_input_tokens_seen": 213343285, + "step": 6567 + }, + { + "epoch": 0.5923253821526807, + "flos": 23479778326560.0, + "grad_norm": 1.5757099216792545, + "learning_rate": 1.5042613880044053e-06, + "loss": 0.6638, + "num_input_tokens_seen": 213376115, + "step": 6568 + }, + { + "epoch": 0.592415565676151, + "flos": 19655002710720.0, + "grad_norm": 1.7705467964673611, + "learning_rate": 1.5036954297143265e-06, + "loss": 0.7746, + "num_input_tokens_seen": 213401410, + "step": 6569 + }, + { + "epoch": 0.5925057491996212, + "flos": 39147253277760.0, + "grad_norm": 2.230277298762694, + "learning_rate": 1.50312951377169e-06, + "loss": 0.7012, + "num_input_tokens_seen": 213435305, + "step": 6570 + }, + { + "epoch": 0.5925959327230915, + "flos": 64183384417920.0, + "grad_norm": 0.5906190387485543, + "learning_rate": 1.502563640224784e-06, + "loss": 0.5874, + "num_input_tokens_seen": 213531180, + "step": 6571 + }, + { + "epoch": 0.5926861162465618, + "flos": 26064375806400.0, + "grad_norm": 4.421159774036513, + "learning_rate": 1.5019978091218903e-06, + "loss": 0.7442, + "num_input_tokens_seen": 213560215, + "step": 6572 + }, + { + "epoch": 0.592776299770032, + "flos": 23773353790080.0, + "grad_norm": 1.501200354626332, + "learning_rate": 1.50143202051129e-06, + "loss": 0.8071, + "num_input_tokens_seen": 213591030, + "step": 6573 + }, + { + "epoch": 0.5928664832935022, + "flos": 23438700008160.0, + "grad_norm": 1.8805510049286251, + "learning_rate": 1.500866274441258e-06, + "loss": 0.8002, + "num_input_tokens_seen": 213616360, + "step": 6574 + }, + { + "epoch": 0.5929566668169726, + "flos": 24968996820000.0, + "grad_norm": 2.469823579704036, + "learning_rate": 1.5003005709600682e-06, + "loss": 0.73, + "num_input_tokens_seen": 213645915, + "step": 6575 + }, + { + "epoch": 0.5930468503404428, + "flos": 20420113946880.0, + "grad_norm": 2.364075683234268, + "learning_rate": 1.4997349101159885e-06, + "loss": 0.8233, + "num_input_tokens_seen": 213673590, + "step": 6576 + }, + { + "epoch": 0.593137033863913, + "flos": 18889519776960.0, + "grad_norm": 2.3073945262935602, + "learning_rate": 1.4991692919572854e-06, + "loss": 0.7031, + "num_input_tokens_seen": 213700135, + "step": 6577 + }, + { + "epoch": 0.5932272173873834, + "flos": 23915067849600.0, + "grad_norm": 1.9661154906771534, + "learning_rate": 1.4986037165322199e-06, + "loss": 0.7633, + "num_input_tokens_seen": 213726980, + "step": 6578 + }, + { + "epoch": 0.5933174009108536, + "flos": 70298550164160.0, + "grad_norm": 0.6703705244581777, + "learning_rate": 1.498038183889049e-06, + "loss": 0.6144, + "num_input_tokens_seen": 213814760, + "step": 6579 + }, + { + "epoch": 0.5934075844343238, + "flos": 24573150145920.0, + "grad_norm": 1.8088135554833633, + "learning_rate": 1.4974726940760292e-06, + "loss": 0.6659, + "num_input_tokens_seen": 213844550, + "step": 6580 + }, + { + "epoch": 0.5934977679577941, + "flos": 21913681302240.0, + "grad_norm": 1.957350957952678, + "learning_rate": 1.496907247141409e-06, + "loss": 0.8173, + "num_input_tokens_seen": 213872075, + "step": 6581 + }, + { + "epoch": 0.5935879514812644, + "flos": 69961740536160.0, + "grad_norm": 0.7438027723807938, + "learning_rate": 1.4963418431334372e-06, + "loss": 0.656, + "num_input_tokens_seen": 213956755, + "step": 6582 + }, + { + "epoch": 0.5936781350047347, + "flos": 19763555610720.0, + "grad_norm": 2.175015710357402, + "learning_rate": 1.4957764821003566e-06, + "loss": 0.7285, + "num_input_tokens_seen": 213983905, + "step": 6583 + }, + { + "epoch": 0.5937683185282049, + "flos": 21911785644480.0, + "grad_norm": 2.0625154899246696, + "learning_rate": 1.4952111640904063e-06, + "loss": 0.7498, + "num_input_tokens_seen": 214011985, + "step": 6584 + }, + { + "epoch": 0.5938585020516751, + "flos": 18926881119360.0, + "grad_norm": 2.0287992511630555, + "learning_rate": 1.494645889151823e-06, + "loss": 0.73, + "num_input_tokens_seen": 214040685, + "step": 6585 + }, + { + "epoch": 0.5939486855751455, + "flos": 65990604640800.0, + "grad_norm": 0.652670863935218, + "learning_rate": 1.494080657332839e-06, + "loss": 0.5605, + "num_input_tokens_seen": 214128285, + "step": 6586 + }, + { + "epoch": 0.5940388690986157, + "flos": 23552419504800.0, + "grad_norm": 2.0406562744361207, + "learning_rate": 1.4935154686816832e-06, + "loss": 0.7046, + "num_input_tokens_seen": 214156535, + "step": 6587 + }, + { + "epoch": 0.5941290526220859, + "flos": 22532952635520.0, + "grad_norm": 5.1396694976589705, + "learning_rate": 1.4929503232465802e-06, + "loss": 0.8248, + "num_input_tokens_seen": 214184325, + "step": 6588 + }, + { + "epoch": 0.5942192361455562, + "flos": 19944675349440.0, + "grad_norm": 2.994879007586211, + "learning_rate": 1.492385221075751e-06, + "loss": 0.652, + "num_input_tokens_seen": 214213025, + "step": 6589 + }, + { + "epoch": 0.5943094196690265, + "flos": 19291610970720.0, + "grad_norm": 2.3681860098854126, + "learning_rate": 1.4918201622174142e-06, + "loss": 0.7796, + "num_input_tokens_seen": 214243220, + "step": 6590 + }, + { + "epoch": 0.5943996031924967, + "flos": 21182697639360.0, + "grad_norm": 1.903511519244655, + "learning_rate": 1.4912551467197827e-06, + "loss": 0.768, + "num_input_tokens_seen": 214271750, + "step": 6591 + }, + { + "epoch": 0.594489786715967, + "flos": 15829223511360.0, + "grad_norm": 3.573827703377735, + "learning_rate": 1.4906901746310678e-06, + "loss": 0.8306, + "num_input_tokens_seen": 214298535, + "step": 6592 + }, + { + "epoch": 0.5945799702394372, + "flos": 30181128586080.0, + "grad_norm": 1.7214447397785402, + "learning_rate": 1.4901252459994757e-06, + "loss": 0.7068, + "num_input_tokens_seen": 214332460, + "step": 6593 + }, + { + "epoch": 0.5946701537629075, + "flos": 20018989166880.0, + "grad_norm": 2.0035281056926286, + "learning_rate": 1.489560360873208e-06, + "loss": 0.7312, + "num_input_tokens_seen": 214358210, + "step": 6594 + }, + { + "epoch": 0.5947603372863778, + "flos": 23989121478720.0, + "grad_norm": 1.9412758244198713, + "learning_rate": 1.4889955193004659e-06, + "loss": 0.7211, + "num_input_tokens_seen": 214386815, + "step": 6595 + }, + { + "epoch": 0.594850520809848, + "flos": 58231534095360.0, + "grad_norm": 0.6360843751273196, + "learning_rate": 1.4884307213294428e-06, + "loss": 0.5746, + "num_input_tokens_seen": 214477700, + "step": 6596 + }, + { + "epoch": 0.5949407043333182, + "flos": 25519938667200.0, + "grad_norm": 3.1451788383543366, + "learning_rate": 1.4878659670083321e-06, + "loss": 0.7317, + "num_input_tokens_seen": 214504570, + "step": 6597 + }, + { + "epoch": 0.5950308878567886, + "flos": 29017940490240.0, + "grad_norm": 1.7697846018661603, + "learning_rate": 1.4873012563853208e-06, + "loss": 0.6674, + "num_input_tokens_seen": 214532920, + "step": 6598 + }, + { + "epoch": 0.5951210713802588, + "flos": 12914841488160.0, + "grad_norm": 2.609606247209033, + "learning_rate": 1.4867365895085935e-06, + "loss": 0.7537, + "num_input_tokens_seen": 214559360, + "step": 6599 + }, + { + "epoch": 0.5952112549037291, + "flos": 25590944376000.0, + "grad_norm": 1.5854921514292186, + "learning_rate": 1.4861719664263301e-06, + "loss": 0.7272, + "num_input_tokens_seen": 214589710, + "step": 6600 + }, + { + "epoch": 0.5953014384271994, + "flos": 23771866999680.0, + "grad_norm": 2.112405253435626, + "learning_rate": 1.485607387186708e-06, + "loss": 0.8516, + "num_input_tokens_seen": 214615420, + "step": 6601 + }, + { + "epoch": 0.5953916219506696, + "flos": 25041117621600.0, + "grad_norm": 2.349938119925814, + "learning_rate": 1.4850428518379001e-06, + "loss": 0.7423, + "num_input_tokens_seen": 214640850, + "step": 6602 + }, + { + "epoch": 0.5954818054741399, + "flos": 23803095331680.0, + "grad_norm": 1.9122444081772667, + "learning_rate": 1.4844783604280746e-06, + "loss": 0.7391, + "num_input_tokens_seen": 214671930, + "step": 6603 + }, + { + "epoch": 0.5955719889976101, + "flos": 33097926643680.0, + "grad_norm": 2.967974430971465, + "learning_rate": 1.483913913005399e-06, + "loss": 0.6886, + "num_input_tokens_seen": 214699645, + "step": 6604 + }, + { + "epoch": 0.5956621725210804, + "flos": 32041061262240.0, + "grad_norm": 2.0492487531087873, + "learning_rate": 1.483349509618034e-06, + "loss": 0.8072, + "num_input_tokens_seen": 214728005, + "step": 6605 + }, + { + "epoch": 0.5957523560445507, + "flos": 32806581365760.0, + "grad_norm": 2.4370567701299413, + "learning_rate": 1.4827851503141367e-06, + "loss": 0.6811, + "num_input_tokens_seen": 214757810, + "step": 6606 + }, + { + "epoch": 0.5958425395680209, + "flos": 23990719778400.0, + "grad_norm": 2.302984507195144, + "learning_rate": 1.482220835141863e-06, + "loss": 0.6729, + "num_input_tokens_seen": 214788145, + "step": 6607 + }, + { + "epoch": 0.5959327230914911, + "flos": 24755607996000.0, + "grad_norm": 2.1152959164783587, + "learning_rate": 1.481656564149362e-06, + "loss": 0.8228, + "num_input_tokens_seen": 214817360, + "step": 6608 + }, + { + "epoch": 0.5960229066149615, + "flos": 20632238999040.0, + "grad_norm": 1.9843338489020228, + "learning_rate": 1.4810923373847818e-06, + "loss": 0.7478, + "num_input_tokens_seen": 214844290, + "step": 6609 + }, + { + "epoch": 0.5961130901384317, + "flos": 20275017439200.0, + "grad_norm": 2.149616708522892, + "learning_rate": 1.4805281548962647e-06, + "loss": 0.7522, + "num_input_tokens_seen": 214873245, + "step": 6610 + }, + { + "epoch": 0.596203273661902, + "flos": 24388313431200.0, + "grad_norm": 2.4651951365581213, + "learning_rate": 1.4799640167319488e-06, + "loss": 0.6892, + "num_input_tokens_seen": 214902775, + "step": 6611 + }, + { + "epoch": 0.5962934571853722, + "flos": 34006647597120.0, + "grad_norm": 1.9194660130656993, + "learning_rate": 1.4793999229399714e-06, + "loss": 0.7312, + "num_input_tokens_seen": 214933550, + "step": 6612 + }, + { + "epoch": 0.5963836407088425, + "flos": 25261122662880.0, + "grad_norm": 3.006881292285413, + "learning_rate": 1.4788358735684626e-06, + "loss": 0.8112, + "num_input_tokens_seen": 214962570, + "step": 6613 + }, + { + "epoch": 0.5964738242323128, + "flos": 23188767576480.0, + "grad_norm": 1.8986618474571453, + "learning_rate": 1.4782718686655514e-06, + "loss": 0.8139, + "num_input_tokens_seen": 214992120, + "step": 6614 + }, + { + "epoch": 0.596564007755783, + "flos": 33207929164320.0, + "grad_norm": 2.2772157086713545, + "learning_rate": 1.4777079082793605e-06, + "loss": 0.7956, + "num_input_tokens_seen": 215021575, + "step": 6615 + }, + { + "epoch": 0.5966541912792532, + "flos": 52152137125440.0, + "grad_norm": 2.0349097514467815, + "learning_rate": 1.4771439924580108e-06, + "loss": 0.7181, + "num_input_tokens_seen": 215054545, + "step": 6616 + }, + { + "epoch": 0.5967443748027236, + "flos": 19761920141280.0, + "grad_norm": 2.496203400736142, + "learning_rate": 1.4765801212496189e-06, + "loss": 0.788, + "num_input_tokens_seen": 215082525, + "step": 6617 + }, + { + "epoch": 0.5968345583261938, + "flos": 23079545620800.0, + "grad_norm": 2.1655166092427014, + "learning_rate": 1.4760162947022968e-06, + "loss": 0.6634, + "num_input_tokens_seen": 215111760, + "step": 6618 + }, + { + "epoch": 0.596924741849664, + "flos": 26063446562400.0, + "grad_norm": 1.738382469402035, + "learning_rate": 1.475452512864154e-06, + "loss": 0.7369, + "num_input_tokens_seen": 215141550, + "step": 6619 + }, + { + "epoch": 0.5970149253731343, + "flos": 33823966728480.0, + "grad_norm": 1.6366011911931668, + "learning_rate": 1.4748887757832945e-06, + "loss": 0.7547, + "num_input_tokens_seen": 215173370, + "step": 6620 + }, + { + "epoch": 0.5971051088966046, + "flos": 38417421877440.0, + "grad_norm": 1.8703675583054848, + "learning_rate": 1.4743250835078209e-06, + "loss": 0.7805, + "num_input_tokens_seen": 215204535, + "step": 6621 + }, + { + "epoch": 0.5971952924200749, + "flos": 29121772830720.0, + "grad_norm": 1.6580809064021251, + "learning_rate": 1.4737614360858297e-06, + "loss": 0.7041, + "num_input_tokens_seen": 215237090, + "step": 6622 + }, + { + "epoch": 0.5972854759435451, + "flos": 19435815404160.0, + "grad_norm": 2.2641968533896644, + "learning_rate": 1.4731978335654133e-06, + "loss": 0.702, + "num_input_tokens_seen": 215263720, + "step": 6623 + }, + { + "epoch": 0.5973756594670154, + "flos": 30907391689440.0, + "grad_norm": 2.4981145424584583, + "learning_rate": 1.4726342759946638e-06, + "loss": 0.6852, + "num_input_tokens_seen": 215292635, + "step": 6624 + }, + { + "epoch": 0.5974658429904857, + "flos": 22387856127840.0, + "grad_norm": 2.266862412957563, + "learning_rate": 1.4720707634216653e-06, + "loss": 0.7428, + "num_input_tokens_seen": 215323090, + "step": 6625 + }, + { + "epoch": 0.5975560265139559, + "flos": 29672789017440.0, + "grad_norm": 2.3311664760650705, + "learning_rate": 1.4715072958945e-06, + "loss": 0.7528, + "num_input_tokens_seen": 215352625, + "step": 6626 + }, + { + "epoch": 0.5976462100374261, + "flos": 22600575896160.0, + "grad_norm": 2.101993105426083, + "learning_rate": 1.470943873461247e-06, + "loss": 0.7445, + "num_input_tokens_seen": 215381630, + "step": 6627 + }, + { + "epoch": 0.5977363935608965, + "flos": 20310148596000.0, + "grad_norm": 4.576874337556388, + "learning_rate": 1.470380496169979e-06, + "loss": 0.7061, + "num_input_tokens_seen": 215409295, + "step": 6628 + }, + { + "epoch": 0.5978265770843667, + "flos": 24318571494240.0, + "grad_norm": 1.8437342885527943, + "learning_rate": 1.4698171640687682e-06, + "loss": 0.7394, + "num_input_tokens_seen": 215437450, + "step": 6629 + }, + { + "epoch": 0.5979167606078369, + "flos": 24536569368480.0, + "grad_norm": 1.7213135761163298, + "learning_rate": 1.4692538772056792e-06, + "loss": 0.6806, + "num_input_tokens_seen": 215466600, + "step": 6630 + }, + { + "epoch": 0.5980069441313072, + "flos": 35648582399040.0, + "grad_norm": 1.566746929155344, + "learning_rate": 1.4686906356287772e-06, + "loss": 0.6582, + "num_input_tokens_seen": 215498645, + "step": 6631 + }, + { + "epoch": 0.5980971276547775, + "flos": 22060450449120.0, + "grad_norm": 1.6299507579503423, + "learning_rate": 1.4681274393861194e-06, + "loss": 0.7781, + "num_input_tokens_seen": 215527905, + "step": 6632 + }, + { + "epoch": 0.5981873111782477, + "flos": 18482038968000.0, + "grad_norm": 2.048380174569431, + "learning_rate": 1.4675642885257603e-06, + "loss": 0.8416, + "num_input_tokens_seen": 215553940, + "step": 6633 + }, + { + "epoch": 0.598277494701718, + "flos": 16558869062880.0, + "grad_norm": 2.1803540813674043, + "learning_rate": 1.4670011830957529e-06, + "loss": 0.8318, + "num_input_tokens_seen": 215580065, + "step": 6634 + }, + { + "epoch": 0.5983676782251882, + "flos": 22017587982240.0, + "grad_norm": 2.7497287096242755, + "learning_rate": 1.4664381231441427e-06, + "loss": 0.7097, + "num_input_tokens_seen": 215604470, + "step": 6635 + }, + { + "epoch": 0.5984578617486586, + "flos": 25447632016800.0, + "grad_norm": 2.0138042417546167, + "learning_rate": 1.4658751087189746e-06, + "loss": 0.7571, + "num_input_tokens_seen": 215633765, + "step": 6636 + }, + { + "epoch": 0.5985480452721288, + "flos": 25265062657440.0, + "grad_norm": 3.0001052202210277, + "learning_rate": 1.4653121398682874e-06, + "loss": 0.7956, + "num_input_tokens_seen": 215664115, + "step": 6637 + }, + { + "epoch": 0.598638228795599, + "flos": 25847604534240.0, + "grad_norm": 1.443666417728656, + "learning_rate": 1.4647492166401159e-06, + "loss": 0.6887, + "num_input_tokens_seen": 215697575, + "step": 6638 + }, + { + "epoch": 0.5987284123190693, + "flos": 19690988772000.0, + "grad_norm": 2.301229281106868, + "learning_rate": 1.4641863390824934e-06, + "loss": 0.7315, + "num_input_tokens_seen": 215725375, + "step": 6639 + }, + { + "epoch": 0.5988185958425396, + "flos": 29567135358720.0, + "grad_norm": 2.4487109471586015, + "learning_rate": 1.4636235072434465e-06, + "loss": 0.7355, + "num_input_tokens_seen": 215759495, + "step": 6640 + }, + { + "epoch": 0.5989087793660098, + "flos": 26826104594400.0, + "grad_norm": 1.5896513351487136, + "learning_rate": 1.4630607211709994e-06, + "loss": 0.7725, + "num_input_tokens_seen": 215788360, + "step": 6641 + }, + { + "epoch": 0.5989989628894801, + "flos": 22604032683840.0, + "grad_norm": 2.493288950737583, + "learning_rate": 1.4624979809131723e-06, + "loss": 0.7531, + "num_input_tokens_seen": 215817320, + "step": 6642 + }, + { + "epoch": 0.5990891464129503, + "flos": 19544293964640.0, + "grad_norm": 2.0262701172156548, + "learning_rate": 1.4619352865179814e-06, + "loss": 0.6557, + "num_input_tokens_seen": 215845675, + "step": 6643 + }, + { + "epoch": 0.5991793299364206, + "flos": 20310817651680.0, + "grad_norm": 2.6374339850241006, + "learning_rate": 1.4613726380334391e-06, + "loss": 0.7157, + "num_input_tokens_seen": 215873490, + "step": 6644 + }, + { + "epoch": 0.5992695134598909, + "flos": 34003153639680.0, + "grad_norm": 2.429876289515298, + "learning_rate": 1.4608100355075522e-06, + "loss": 0.6002, + "num_input_tokens_seen": 215903230, + "step": 6645 + }, + { + "epoch": 0.5993596969833611, + "flos": 66799619097120.0, + "grad_norm": 0.6215381591848617, + "learning_rate": 1.460247478988327e-06, + "loss": 0.5557, + "num_input_tokens_seen": 215995655, + "step": 6646 + }, + { + "epoch": 0.5994498805068315, + "flos": 23188470218400.0, + "grad_norm": 1.8049643167999927, + "learning_rate": 1.4596849685237623e-06, + "loss": 0.7609, + "num_input_tokens_seen": 216024715, + "step": 6647 + }, + { + "epoch": 0.5995400640303017, + "flos": 30944046806400.0, + "grad_norm": 1.5603764408700518, + "learning_rate": 1.459122504161856e-06, + "loss": 0.653, + "num_input_tokens_seen": 216056480, + "step": 6648 + }, + { + "epoch": 0.5996302475537719, + "flos": 28835928677280.0, + "grad_norm": 1.7983259893956218, + "learning_rate": 1.4585600859506001e-06, + "loss": 0.7172, + "num_input_tokens_seen": 216087450, + "step": 6649 + }, + { + "epoch": 0.5997204310772422, + "flos": 23917855581600.0, + "grad_norm": 1.6993695773129018, + "learning_rate": 1.4579977139379826e-06, + "loss": 0.803, + "num_input_tokens_seen": 216118355, + "step": 6650 + }, + { + "epoch": 0.5998106146007125, + "flos": 17140964902560.0, + "grad_norm": 2.3403701826870247, + "learning_rate": 1.4574353881719895e-06, + "loss": 0.8089, + "num_input_tokens_seen": 216142830, + "step": 6651 + }, + { + "epoch": 0.5999007981241827, + "flos": 23698519596000.0, + "grad_norm": 1.9409824576550616, + "learning_rate": 1.4568731087005998e-06, + "loss": 0.7487, + "num_input_tokens_seen": 216169365, + "step": 6652 + }, + { + "epoch": 0.599990981647653, + "flos": 43589739096960.0, + "grad_norm": 1.6960160865380665, + "learning_rate": 1.4563108755717916e-06, + "loss": 0.7718, + "num_input_tokens_seen": 216202810, + "step": 6653 + }, + { + "epoch": 0.6000811651711232, + "flos": 20926780876320.0, + "grad_norm": 3.4538705978848707, + "learning_rate": 1.455748688833538e-06, + "loss": 0.7716, + "num_input_tokens_seen": 216232310, + "step": 6654 + }, + { + "epoch": 0.6001713486945935, + "flos": 38161988321280.0, + "grad_norm": 2.657938970015839, + "learning_rate": 1.4551865485338065e-06, + "loss": 0.6603, + "num_input_tokens_seen": 216263550, + "step": 6655 + }, + { + "epoch": 0.6002615322180638, + "flos": 28799719597440.0, + "grad_norm": 1.8876982805807951, + "learning_rate": 1.4546244547205629e-06, + "loss": 0.8077, + "num_input_tokens_seen": 216293505, + "step": 6656 + }, + { + "epoch": 0.600351715741534, + "flos": 19363360074720.0, + "grad_norm": 1.98984644350095, + "learning_rate": 1.4540624074417678e-06, + "loss": 0.7716, + "num_input_tokens_seen": 216320925, + "step": 6657 + }, + { + "epoch": 0.6004418992650042, + "flos": 40637512524480.0, + "grad_norm": 9.545268971014066, + "learning_rate": 1.453500406745379e-06, + "loss": 0.8035, + "num_input_tokens_seen": 216355210, + "step": 6658 + }, + { + "epoch": 0.6005320827884746, + "flos": 18742936478880.0, + "grad_norm": 2.0171471104910887, + "learning_rate": 1.4529384526793486e-06, + "loss": 0.8228, + "num_input_tokens_seen": 216380975, + "step": 6659 + }, + { + "epoch": 0.6006222663119448, + "flos": 19833371887200.0, + "grad_norm": 1.7653141929711536, + "learning_rate": 1.4523765452916252e-06, + "loss": 0.7984, + "num_input_tokens_seen": 216407085, + "step": 6660 + }, + { + "epoch": 0.600712449835415, + "flos": 44173173048000.0, + "grad_norm": 3.2269260303305076, + "learning_rate": 1.4518146846301554e-06, + "loss": 0.7731, + "num_input_tokens_seen": 216439550, + "step": 6661 + }, + { + "epoch": 0.6008026333588853, + "flos": 18741077990880.0, + "grad_norm": 1.643510587829336, + "learning_rate": 1.4512528707428787e-06, + "loss": 0.7209, + "num_input_tokens_seen": 216466810, + "step": 6662 + }, + { + "epoch": 0.6008928168823556, + "flos": 28435398613440.0, + "grad_norm": 4.762768298824343, + "learning_rate": 1.4506911036777335e-06, + "loss": 0.7268, + "num_input_tokens_seen": 216497430, + "step": 6663 + }, + { + "epoch": 0.6009830004058259, + "flos": 53270784381600.0, + "grad_norm": 0.6735866611561642, + "learning_rate": 1.450129383482651e-06, + "loss": 0.5479, + "num_input_tokens_seen": 216591450, + "step": 6664 + }, + { + "epoch": 0.6010731839292961, + "flos": 23698816954080.0, + "grad_norm": 1.6333620889558769, + "learning_rate": 1.4495677102055629e-06, + "loss": 0.7006, + "num_input_tokens_seen": 216623950, + "step": 6665 + }, + { + "epoch": 0.6011633674527663, + "flos": 24096150418560.0, + "grad_norm": 1.9077634587963541, + "learning_rate": 1.4490060838943924e-06, + "loss": 0.7814, + "num_input_tokens_seen": 216653050, + "step": 6666 + }, + { + "epoch": 0.6012535509762367, + "flos": 22495628462880.0, + "grad_norm": 1.9434032339389704, + "learning_rate": 1.4484445045970609e-06, + "loss": 0.7483, + "num_input_tokens_seen": 216683815, + "step": 6667 + }, + { + "epoch": 0.6013437344997069, + "flos": 23224753637760.0, + "grad_norm": 2.6124716104671504, + "learning_rate": 1.447882972361485e-06, + "loss": 0.7426, + "num_input_tokens_seen": 216713475, + "step": 6668 + }, + { + "epoch": 0.6014339180231771, + "flos": 26977519961280.0, + "grad_norm": 5.086258885414986, + "learning_rate": 1.4473214872355785e-06, + "loss": 0.7327, + "num_input_tokens_seen": 216743830, + "step": 6669 + }, + { + "epoch": 0.6015241015466475, + "flos": 20741721143040.0, + "grad_norm": 2.1143057379455605, + "learning_rate": 1.4467600492672508e-06, + "loss": 0.7803, + "num_input_tokens_seen": 216765350, + "step": 6670 + }, + { + "epoch": 0.6016142850701177, + "flos": 18707545133760.0, + "grad_norm": 2.225216428538831, + "learning_rate": 1.4461986585044054e-06, + "loss": 0.726, + "num_input_tokens_seen": 216793285, + "step": 6671 + }, + { + "epoch": 0.601704468593588, + "flos": 23661306932640.0, + "grad_norm": 1.7319689720977822, + "learning_rate": 1.4456373149949446e-06, + "loss": 0.7208, + "num_input_tokens_seen": 216822340, + "step": 6672 + }, + { + "epoch": 0.6017946521170582, + "flos": 23334272951520.0, + "grad_norm": 2.0995809863633124, + "learning_rate": 1.4450760187867648e-06, + "loss": 0.7527, + "num_input_tokens_seen": 216854535, + "step": 6673 + }, + { + "epoch": 0.6018848356405285, + "flos": 24501066514080.0, + "grad_norm": 2.194119492412586, + "learning_rate": 1.4445147699277581e-06, + "loss": 0.739, + "num_input_tokens_seen": 216885420, + "step": 6674 + }, + { + "epoch": 0.6019750191639988, + "flos": 23043336540960.0, + "grad_norm": 1.9438953676479502, + "learning_rate": 1.4439535684658154e-06, + "loss": 0.7968, + "num_input_tokens_seen": 216915065, + "step": 6675 + }, + { + "epoch": 0.602065202687469, + "flos": 53463055048320.0, + "grad_norm": 0.7667477608030177, + "learning_rate": 1.44339241444882e-06, + "loss": 0.6173, + "num_input_tokens_seen": 216988430, + "step": 6676 + }, + { + "epoch": 0.6021553862109392, + "flos": 22750913340000.0, + "grad_norm": 2.54612280848666, + "learning_rate": 1.4428313079246518e-06, + "loss": 0.6976, + "num_input_tokens_seen": 217014720, + "step": 6677 + }, + { + "epoch": 0.6022455697344096, + "flos": 64027248491520.0, + "grad_norm": 0.6669247818003102, + "learning_rate": 1.4422702489411896e-06, + "loss": 0.5949, + "num_input_tokens_seen": 217100535, + "step": 6678 + }, + { + "epoch": 0.6023357532578798, + "flos": 20711682243360.0, + "grad_norm": 2.687212019284015, + "learning_rate": 1.4417092375463043e-06, + "loss": 0.6938, + "num_input_tokens_seen": 217127790, + "step": 6679 + }, + { + "epoch": 0.60242593678135, + "flos": 21982568334720.0, + "grad_norm": 2.001544237016298, + "learning_rate": 1.441148273787866e-06, + "loss": 0.7819, + "num_input_tokens_seen": 217156185, + "step": 6680 + }, + { + "epoch": 0.6025161203048203, + "flos": 67632502272960.0, + "grad_norm": 0.7833841927906458, + "learning_rate": 1.4405873577137383e-06, + "loss": 0.6013, + "num_input_tokens_seen": 217245550, + "step": 6681 + }, + { + "epoch": 0.6026063038282906, + "flos": 29673606752160.0, + "grad_norm": 2.359236757384171, + "learning_rate": 1.4400264893717816e-06, + "loss": 0.7565, + "num_input_tokens_seen": 217274505, + "step": 6682 + }, + { + "epoch": 0.6026964873517608, + "flos": 23844842705760.0, + "grad_norm": 2.1714153214996905, + "learning_rate": 1.4394656688098526e-06, + "loss": 0.7969, + "num_input_tokens_seen": 217303680, + "step": 6683 + }, + { + "epoch": 0.6027866708752311, + "flos": 19801028462400.0, + "grad_norm": 2.184482782458005, + "learning_rate": 1.4389048960758032e-06, + "loss": 0.7704, + "num_input_tokens_seen": 217332415, + "step": 6684 + }, + { + "epoch": 0.6028768543987013, + "flos": 24973791719040.0, + "grad_norm": 9.351834171335806, + "learning_rate": 1.4383441712174826e-06, + "loss": 0.7759, + "num_input_tokens_seen": 217359785, + "step": 6685 + }, + { + "epoch": 0.6029670379221717, + "flos": 19830472645920.0, + "grad_norm": 2.2357509617863665, + "learning_rate": 1.4377834942827333e-06, + "loss": 0.8135, + "num_input_tokens_seen": 217385910, + "step": 6686 + }, + { + "epoch": 0.6030572214456419, + "flos": 19618607782080.0, + "grad_norm": 2.076695689227658, + "learning_rate": 1.437222865319397e-06, + "loss": 0.6873, + "num_input_tokens_seen": 217413340, + "step": 6687 + }, + { + "epoch": 0.6031474049691121, + "flos": 21549806355360.0, + "grad_norm": 1.7297804176025748, + "learning_rate": 1.4366622843753092e-06, + "loss": 0.6725, + "num_input_tokens_seen": 217441965, + "step": 6688 + }, + { + "epoch": 0.6032375884925824, + "flos": 33680096822880.0, + "grad_norm": 2.1249184312311815, + "learning_rate": 1.4361017514983006e-06, + "loss": 0.7639, + "num_input_tokens_seen": 217474830, + "step": 6689 + }, + { + "epoch": 0.6033277720160527, + "flos": 19249826426880.0, + "grad_norm": 1.6493798766763341, + "learning_rate": 1.4355412667362006e-06, + "loss": 0.8038, + "num_input_tokens_seen": 217504600, + "step": 6690 + }, + { + "epoch": 0.6034179555395229, + "flos": 19764150326880.0, + "grad_norm": 1.778608471620906, + "learning_rate": 1.4349808301368311e-06, + "loss": 0.7046, + "num_input_tokens_seen": 217533130, + "step": 6691 + }, + { + "epoch": 0.6035081390629932, + "flos": 20710976017920.0, + "grad_norm": 2.2600738921787293, + "learning_rate": 1.4344204417480139e-06, + "loss": 0.7906, + "num_input_tokens_seen": 217558795, + "step": 6692 + }, + { + "epoch": 0.6035983225864634, + "flos": 22970026307040.0, + "grad_norm": 1.9577006339273295, + "learning_rate": 1.4338601016175628e-06, + "loss": 0.7003, + "num_input_tokens_seen": 217587960, + "step": 6693 + }, + { + "epoch": 0.6036885061099337, + "flos": 35757321147840.0, + "grad_norm": 1.8349827658028446, + "learning_rate": 1.433299809793289e-06, + "loss": 0.7465, + "num_input_tokens_seen": 217618820, + "step": 6694 + }, + { + "epoch": 0.603778689633404, + "flos": 22491093752160.0, + "grad_norm": 2.9819305155057183, + "learning_rate": 1.432739566323001e-06, + "loss": 0.62, + "num_input_tokens_seen": 217646600, + "step": 6695 + }, + { + "epoch": 0.6038688731568742, + "flos": 24756165542400.0, + "grad_norm": 2.26073596533296, + "learning_rate": 1.4321793712545004e-06, + "loss": 0.7158, + "num_input_tokens_seen": 217675795, + "step": 6696 + }, + { + "epoch": 0.6039590566803446, + "flos": 22452468637920.0, + "grad_norm": 1.9537451408067097, + "learning_rate": 1.4316192246355873e-06, + "loss": 0.7961, + "num_input_tokens_seen": 217702135, + "step": 6697 + }, + { + "epoch": 0.6040492402038148, + "flos": 64495253136960.0, + "grad_norm": 0.6029826406365123, + "learning_rate": 1.4310591265140555e-06, + "loss": 0.5784, + "num_input_tokens_seen": 217799715, + "step": 6698 + }, + { + "epoch": 0.604139423727285, + "flos": 24427050054720.0, + "grad_norm": 1.567717576448518, + "learning_rate": 1.4304990769376963e-06, + "loss": 0.7539, + "num_input_tokens_seen": 217831895, + "step": 6699 + }, + { + "epoch": 0.6042296072507553, + "flos": 30033950571840.0, + "grad_norm": 2.2990142824675948, + "learning_rate": 1.4299390759542962e-06, + "loss": 0.7312, + "num_input_tokens_seen": 217862145, + "step": 6700 + }, + { + "epoch": 0.6043197907742256, + "flos": 23807072496000.0, + "grad_norm": 1.8820247664286476, + "learning_rate": 1.4293791236116368e-06, + "loss": 0.7295, + "num_input_tokens_seen": 217889520, + "step": 6701 + }, + { + "epoch": 0.6044099742976958, + "flos": 23152186799040.0, + "grad_norm": 2.3514025053934526, + "learning_rate": 1.4288192199574978e-06, + "loss": 0.7094, + "num_input_tokens_seen": 217917645, + "step": 6702 + }, + { + "epoch": 0.6045001578211661, + "flos": 67416660244800.0, + "grad_norm": 0.640589247845677, + "learning_rate": 1.4282593650396524e-06, + "loss": 0.551, + "num_input_tokens_seen": 218014925, + "step": 6703 + }, + { + "epoch": 0.6045903413446363, + "flos": 23480224363680.0, + "grad_norm": 2.100804756007509, + "learning_rate": 1.4276995589058695e-06, + "loss": 0.8453, + "num_input_tokens_seen": 218043850, + "step": 6704 + }, + { + "epoch": 0.6046805248681066, + "flos": 66439684144800.0, + "grad_norm": 0.5956392848092142, + "learning_rate": 1.4271398016039168e-06, + "loss": 0.5478, + "num_input_tokens_seen": 218140795, + "step": 6705 + }, + { + "epoch": 0.6047707083915769, + "flos": 19581580967520.0, + "grad_norm": 2.214889498983256, + "learning_rate": 1.4265800931815542e-06, + "loss": 0.7893, + "num_input_tokens_seen": 218168385, + "step": 6706 + }, + { + "epoch": 0.6048608919150471, + "flos": 20711682243360.0, + "grad_norm": 1.6606991587007047, + "learning_rate": 1.4260204336865406e-06, + "loss": 0.7546, + "num_input_tokens_seen": 218195640, + "step": 6707 + }, + { + "epoch": 0.6049510754385173, + "flos": 23186463051360.0, + "grad_norm": 1.7497999894421528, + "learning_rate": 1.4254608231666281e-06, + "loss": 0.7439, + "num_input_tokens_seen": 218225395, + "step": 6708 + }, + { + "epoch": 0.6050412589619877, + "flos": 22310197032000.0, + "grad_norm": 1.7227128553713666, + "learning_rate": 1.4249012616695661e-06, + "loss": 0.7557, + "num_input_tokens_seen": 218252700, + "step": 6709 + }, + { + "epoch": 0.6051314424854579, + "flos": 47486338156320.0, + "grad_norm": 1.8202043754980244, + "learning_rate": 1.4243417492431e-06, + "loss": 0.7534, + "num_input_tokens_seen": 218285420, + "step": 6710 + }, + { + "epoch": 0.6052216260089281, + "flos": 16703816891520.0, + "grad_norm": 2.3011036049580063, + "learning_rate": 1.4237822859349696e-06, + "loss": 0.7248, + "num_input_tokens_seen": 218312445, + "step": 6711 + }, + { + "epoch": 0.6053118095323984, + "flos": 24568987132800.0, + "grad_norm": 1.4737302066680653, + "learning_rate": 1.423222871792912e-06, + "loss": 0.7175, + "num_input_tokens_seen": 218341610, + "step": 6712 + }, + { + "epoch": 0.6054019930558687, + "flos": 19614593448000.0, + "grad_norm": 1.6496931599922586, + "learning_rate": 1.4226635068646586e-06, + "loss": 0.8035, + "num_input_tokens_seen": 218369215, + "step": 6713 + }, + { + "epoch": 0.605492176579339, + "flos": 61724071963680.0, + "grad_norm": 0.6208487693309913, + "learning_rate": 1.4221041911979393e-06, + "loss": 0.5718, + "num_input_tokens_seen": 218460980, + "step": 6714 + }, + { + "epoch": 0.6055823601028092, + "flos": 18812789925120.0, + "grad_norm": 2.3999940872119545, + "learning_rate": 1.4215449248404765e-06, + "loss": 0.7628, + "num_input_tokens_seen": 218487470, + "step": 6715 + }, + { + "epoch": 0.6056725436262794, + "flos": 27153956310240.0, + "grad_norm": 1.6667147708944383, + "learning_rate": 1.4209857078399896e-06, + "loss": 0.7662, + "num_input_tokens_seen": 218520385, + "step": 6716 + }, + { + "epoch": 0.6057627271497498, + "flos": 13606456641600.0, + "grad_norm": 2.2198502942032636, + "learning_rate": 1.4204265402441955e-06, + "loss": 0.717, + "num_input_tokens_seen": 218546935, + "step": 6717 + }, + { + "epoch": 0.60585291067322, + "flos": 19800619595040.0, + "grad_norm": 1.6566790618790965, + "learning_rate": 1.419867422100804e-06, + "loss": 0.6251, + "num_input_tokens_seen": 218575980, + "step": 6718 + }, + { + "epoch": 0.6059430941966902, + "flos": 23552716862880.0, + "grad_norm": 2.1023187233956677, + "learning_rate": 1.4193083534575236e-06, + "loss": 0.732, + "num_input_tokens_seen": 218605525, + "step": 6719 + }, + { + "epoch": 0.6060332777201606, + "flos": 22277890776960.0, + "grad_norm": 2.4647893884531293, + "learning_rate": 1.4187493343620567e-06, + "loss": 0.7731, + "num_input_tokens_seen": 218633270, + "step": 6720 + }, + { + "epoch": 0.6061234612436308, + "flos": 21184630466880.0, + "grad_norm": 1.9701600247949838, + "learning_rate": 1.4181903648621006e-06, + "loss": 0.7797, + "num_input_tokens_seen": 218661065, + "step": 6721 + }, + { + "epoch": 0.606213644767101, + "flos": 23697999219360.0, + "grad_norm": 1.7695728827625177, + "learning_rate": 1.4176314450053512e-06, + "loss": 0.7789, + "num_input_tokens_seen": 218692070, + "step": 6722 + }, + { + "epoch": 0.6063038282905713, + "flos": 19181459771040.0, + "grad_norm": 2.1748547867279027, + "learning_rate": 1.4170725748394977e-06, + "loss": 0.7359, + "num_input_tokens_seen": 218720305, + "step": 6723 + }, + { + "epoch": 0.6063940118140416, + "flos": 23917632563040.0, + "grad_norm": 1.7159350839353666, + "learning_rate": 1.4165137544122266e-06, + "loss": 0.6582, + "num_input_tokens_seen": 218748800, + "step": 6724 + }, + { + "epoch": 0.6064841953375119, + "flos": 24675941733120.0, + "grad_norm": 2.002040180081287, + "learning_rate": 1.4159549837712194e-06, + "loss": 0.8168, + "num_input_tokens_seen": 218777870, + "step": 6725 + }, + { + "epoch": 0.6065743788609821, + "flos": 34880794940160.0, + "grad_norm": 1.5803946612986979, + "learning_rate": 1.415396262964153e-06, + "loss": 0.6657, + "num_input_tokens_seen": 218809165, + "step": 6726 + }, + { + "epoch": 0.6066645623844523, + "flos": 22491651298560.0, + "grad_norm": 1.9972547525236413, + "learning_rate": 1.4148375920387016e-06, + "loss": 0.8231, + "num_input_tokens_seen": 218837250, + "step": 6727 + }, + { + "epoch": 0.6067547459079227, + "flos": 67341268504320.0, + "grad_norm": 0.6408537014509322, + "learning_rate": 1.4142789710425325e-06, + "loss": 0.5751, + "num_input_tokens_seen": 218935630, + "step": 6728 + }, + { + "epoch": 0.6068449294313929, + "flos": 22022717409120.0, + "grad_norm": 2.110054754297052, + "learning_rate": 1.4137204000233118e-06, + "loss": 0.6849, + "num_input_tokens_seen": 218961615, + "step": 6729 + }, + { + "epoch": 0.6069351129548631, + "flos": 37069397066880.0, + "grad_norm": 2.095364076575553, + "learning_rate": 1.4131618790286987e-06, + "loss": 0.7432, + "num_input_tokens_seen": 218992570, + "step": 6730 + }, + { + "epoch": 0.6070252964783334, + "flos": 23990422420320.0, + "grad_norm": 2.761323676656836, + "learning_rate": 1.4126034081063506e-06, + "loss": 0.7331, + "num_input_tokens_seen": 219020830, + "step": 6731 + }, + { + "epoch": 0.6071154800018037, + "flos": 28143644468160.0, + "grad_norm": 1.9975601453642227, + "learning_rate": 1.4120449873039186e-06, + "loss": 0.7348, + "num_input_tokens_seen": 219049860, + "step": 6732 + }, + { + "epoch": 0.6072056635252739, + "flos": 26140325093280.0, + "grad_norm": 2.519059673800372, + "learning_rate": 1.4114866166690494e-06, + "loss": 0.7863, + "num_input_tokens_seen": 219077655, + "step": 6733 + }, + { + "epoch": 0.6072958470487442, + "flos": 46432520695200.0, + "grad_norm": 3.561700162276363, + "learning_rate": 1.4109282962493877e-06, + "loss": 0.7066, + "num_input_tokens_seen": 219113560, + "step": 6734 + }, + { + "epoch": 0.6073860305722144, + "flos": 21512816710560.0, + "grad_norm": 2.737282999574045, + "learning_rate": 1.4103700260925716e-06, + "loss": 0.7368, + "num_input_tokens_seen": 219142165, + "step": 6735 + }, + { + "epoch": 0.6074762140956848, + "flos": 27888582609600.0, + "grad_norm": 1.890570143226953, + "learning_rate": 1.4098118062462364e-06, + "loss": 0.7863, + "num_input_tokens_seen": 219171830, + "step": 6736 + }, + { + "epoch": 0.607566397619155, + "flos": 22528194906240.0, + "grad_norm": 1.7170855005045877, + "learning_rate": 1.4092536367580123e-06, + "loss": 0.7412, + "num_input_tokens_seen": 219199560, + "step": 6737 + }, + { + "epoch": 0.6076565811426252, + "flos": 24427161564000.0, + "grad_norm": 1.872595198570679, + "learning_rate": 1.4086955176755248e-06, + "loss": 0.66, + "num_input_tokens_seen": 219225500, + "step": 6738 + }, + { + "epoch": 0.6077467646660955, + "flos": 19764373345440.0, + "grad_norm": 2.3754753841665566, + "learning_rate": 1.4081374490463964e-06, + "loss": 0.6876, + "num_input_tokens_seen": 219254455, + "step": 6739 + }, + { + "epoch": 0.6078369481895658, + "flos": 15719852876640.0, + "grad_norm": 2.109790563268239, + "learning_rate": 1.4075794309182443e-06, + "loss": 0.7602, + "num_input_tokens_seen": 219282080, + "step": 6740 + }, + { + "epoch": 0.607927131713036, + "flos": 21074999643840.0, + "grad_norm": 2.6963605971416573, + "learning_rate": 1.407021463338682e-06, + "loss": 0.759, + "num_input_tokens_seen": 219309055, + "step": 6741 + }, + { + "epoch": 0.6080173152365063, + "flos": 70961836226880.0, + "grad_norm": 0.5938381581361608, + "learning_rate": 1.4064635463553177e-06, + "loss": 0.5204, + "num_input_tokens_seen": 219402485, + "step": 6742 + }, + { + "epoch": 0.6081074987599766, + "flos": 23297283306720.0, + "grad_norm": 2.942682280748659, + "learning_rate": 1.4059056800157563e-06, + "loss": 0.7766, + "num_input_tokens_seen": 219427825, + "step": 6743 + }, + { + "epoch": 0.6081976822834468, + "flos": 22605370795200.0, + "grad_norm": 1.781158913408302, + "learning_rate": 1.4053478643675982e-06, + "loss": 0.6766, + "num_input_tokens_seen": 219455130, + "step": 6744 + }, + { + "epoch": 0.6082878658069171, + "flos": 21148681575360.0, + "grad_norm": 1.7092570234519584, + "learning_rate": 1.4047900994584389e-06, + "loss": 0.6426, + "num_input_tokens_seen": 219482580, + "step": 6745 + }, + { + "epoch": 0.6083780493303873, + "flos": 26649408057120.0, + "grad_norm": 2.385124010904632, + "learning_rate": 1.404232385335871e-06, + "loss": 0.6563, + "num_input_tokens_seen": 219512435, + "step": 6746 + }, + { + "epoch": 0.6084682328538576, + "flos": 26353082031360.0, + "grad_norm": 2.0147097379943197, + "learning_rate": 1.4036747220474806e-06, + "loss": 0.7459, + "num_input_tokens_seen": 219540760, + "step": 6747 + }, + { + "epoch": 0.6085584163773279, + "flos": 24391361351520.0, + "grad_norm": 3.7714436636040656, + "learning_rate": 1.4031171096408506e-06, + "loss": 0.6961, + "num_input_tokens_seen": 219571715, + "step": 6748 + }, + { + "epoch": 0.6086485999007981, + "flos": 37215125460480.0, + "grad_norm": 6.105589022827677, + "learning_rate": 1.4025595481635607e-06, + "loss": 0.7685, + "num_input_tokens_seen": 219601185, + "step": 6749 + }, + { + "epoch": 0.6087387834242683, + "flos": 16336373647680.0, + "grad_norm": 2.0777470524647312, + "learning_rate": 1.4020020376631836e-06, + "loss": 0.7646, + "num_input_tokens_seen": 219626650, + "step": 6750 + }, + { + "epoch": 0.6088289669477387, + "flos": 25410679541760.0, + "grad_norm": 2.6009934178100043, + "learning_rate": 1.4014445781872908e-06, + "loss": 0.758, + "num_input_tokens_seen": 219654795, + "step": 6751 + }, + { + "epoch": 0.6089191504712089, + "flos": 26322299736480.0, + "grad_norm": 1.7499691415518352, + "learning_rate": 1.4008871697834465e-06, + "loss": 0.8067, + "num_input_tokens_seen": 219683745, + "step": 6752 + }, + { + "epoch": 0.6090093339946792, + "flos": 22387893297600.0, + "grad_norm": 1.6701965463070203, + "learning_rate": 1.400329812499213e-06, + "loss": 0.7817, + "num_input_tokens_seen": 219716275, + "step": 6753 + }, + { + "epoch": 0.6090995175181494, + "flos": 24204740488320.0, + "grad_norm": 1.5630776672640219, + "learning_rate": 1.3997725063821458e-06, + "loss": 0.7563, + "num_input_tokens_seen": 219745050, + "step": 6754 + }, + { + "epoch": 0.6091897010416197, + "flos": 28835296791360.0, + "grad_norm": 5.019145930006341, + "learning_rate": 1.3992152514797978e-06, + "loss": 0.7108, + "num_input_tokens_seen": 219775850, + "step": 6755 + }, + { + "epoch": 0.60927988456509, + "flos": 65146236009120.0, + "grad_norm": 0.6206992954066392, + "learning_rate": 1.398658047839718e-06, + "loss": 0.5953, + "num_input_tokens_seen": 219866185, + "step": 6756 + }, + { + "epoch": 0.6093700680885602, + "flos": 72383059762080.0, + "grad_norm": 0.6031258815304338, + "learning_rate": 1.3981008955094481e-06, + "loss": 0.5649, + "num_input_tokens_seen": 219970975, + "step": 6757 + }, + { + "epoch": 0.6094602516120304, + "flos": 40493159412000.0, + "grad_norm": 2.4700863275615874, + "learning_rate": 1.39754379453653e-06, + "loss": 0.7225, + "num_input_tokens_seen": 220004395, + "step": 6758 + }, + { + "epoch": 0.6095504351355008, + "flos": 22125211638240.0, + "grad_norm": 2.267333760923356, + "learning_rate": 1.3969867449684972e-06, + "loss": 0.7539, + "num_input_tokens_seen": 220032035, + "step": 6759 + }, + { + "epoch": 0.609640618658971, + "flos": 29852830833120.0, + "grad_norm": 3.2489983408915757, + "learning_rate": 1.396429746852879e-06, + "loss": 0.738, + "num_input_tokens_seen": 220060380, + "step": 6760 + }, + { + "epoch": 0.6097308021824412, + "flos": 23073189591840.0, + "grad_norm": 1.8083637659819933, + "learning_rate": 1.395872800237204e-06, + "loss": 0.7244, + "num_input_tokens_seen": 220087330, + "step": 6761 + }, + { + "epoch": 0.6098209857059115, + "flos": 60166040777280.0, + "grad_norm": 0.7562040188895203, + "learning_rate": 1.3953159051689918e-06, + "loss": 0.5972, + "num_input_tokens_seen": 220171350, + "step": 6762 + }, + { + "epoch": 0.6099111692293818, + "flos": 26319846532320.0, + "grad_norm": 1.8137553898884395, + "learning_rate": 1.3947590616957618e-06, + "loss": 0.7904, + "num_input_tokens_seen": 220200330, + "step": 6763 + }, + { + "epoch": 0.610001352752852, + "flos": 21585792416640.0, + "grad_norm": 2.0969405276297217, + "learning_rate": 1.3942022698650258e-06, + "loss": 0.7741, + "num_input_tokens_seen": 220228015, + "step": 6764 + }, + { + "epoch": 0.6100915362763223, + "flos": 26824543464480.0, + "grad_norm": 1.7275472741069322, + "learning_rate": 1.3936455297242917e-06, + "loss": 0.7797, + "num_input_tokens_seen": 220255785, + "step": 6765 + }, + { + "epoch": 0.6101817197997926, + "flos": 23953990321920.0, + "grad_norm": 1.7546351220320724, + "learning_rate": 1.3930888413210652e-06, + "loss": 0.7644, + "num_input_tokens_seen": 220284775, + "step": 6766 + }, + { + "epoch": 0.6102719033232629, + "flos": 22386964053600.0, + "grad_norm": 2.4961118981025514, + "learning_rate": 1.392532204702845e-06, + "loss": 0.7092, + "num_input_tokens_seen": 220313055, + "step": 6767 + }, + { + "epoch": 0.6103620868467331, + "flos": 16011867210240.0, + "grad_norm": 2.3273620379377045, + "learning_rate": 1.3919756199171266e-06, + "loss": 0.675, + "num_input_tokens_seen": 220339510, + "step": 6768 + }, + { + "epoch": 0.6104522703702033, + "flos": 22823926215840.0, + "grad_norm": 1.9335397423802205, + "learning_rate": 1.3914190870114009e-06, + "loss": 0.8195, + "num_input_tokens_seen": 220368105, + "step": 6769 + }, + { + "epoch": 0.6105424538936737, + "flos": 22715373315840.0, + "grad_norm": 1.9851881148518085, + "learning_rate": 1.3908626060331541e-06, + "loss": 0.7973, + "num_input_tokens_seen": 220397040, + "step": 6770 + }, + { + "epoch": 0.6106326374171439, + "flos": 15792940092000.0, + "grad_norm": 1.8897143606273814, + "learning_rate": 1.3903061770298693e-06, + "loss": 0.7784, + "num_input_tokens_seen": 220424615, + "step": 6771 + }, + { + "epoch": 0.6107228209406141, + "flos": 26825732896800.0, + "grad_norm": 2.082481578914871, + "learning_rate": 1.3897498000490223e-06, + "loss": 0.7206, + "num_input_tokens_seen": 220450560, + "step": 6772 + }, + { + "epoch": 0.6108130044640844, + "flos": 21805351420800.0, + "grad_norm": 2.0054703050049363, + "learning_rate": 1.3891934751380879e-06, + "loss": 0.8172, + "num_input_tokens_seen": 220479610, + "step": 6773 + }, + { + "epoch": 0.6109031879875547, + "flos": 20777558525280.0, + "grad_norm": 1.882612922189509, + "learning_rate": 1.3886372023445334e-06, + "loss": 0.766, + "num_input_tokens_seen": 220505630, + "step": 6774 + }, + { + "epoch": 0.610993371511025, + "flos": 20384127885600.0, + "grad_norm": 1.7303019630534313, + "learning_rate": 1.3880809817158246e-06, + "loss": 0.8164, + "num_input_tokens_seen": 220532110, + "step": 6775 + }, + { + "epoch": 0.6110835550344952, + "flos": 21403594754880.0, + "grad_norm": 2.3928513476010806, + "learning_rate": 1.3875248132994206e-06, + "loss": 0.7388, + "num_input_tokens_seen": 220558430, + "step": 6776 + }, + { + "epoch": 0.6111737385579654, + "flos": 18853310697120.0, + "grad_norm": 1.8664734564425867, + "learning_rate": 1.386968697142776e-06, + "loss": 0.6707, + "num_input_tokens_seen": 220587075, + "step": 6777 + }, + { + "epoch": 0.6112639220814358, + "flos": 22168929009600.0, + "grad_norm": 2.630340139781554, + "learning_rate": 1.386412633293343e-06, + "loss": 0.7923, + "num_input_tokens_seen": 220616810, + "step": 6778 + }, + { + "epoch": 0.611354105604906, + "flos": 16265219259840.0, + "grad_norm": 2.9152126139186807, + "learning_rate": 1.3858566217985672e-06, + "loss": 0.7201, + "num_input_tokens_seen": 220643250, + "step": 6779 + }, + { + "epoch": 0.6114442891283762, + "flos": 22673774620800.0, + "grad_norm": 2.1306905286206304, + "learning_rate": 1.3853006627058905e-06, + "loss": 0.8303, + "num_input_tokens_seen": 220670590, + "step": 6780 + }, + { + "epoch": 0.6115344726518465, + "flos": 25592877203520.0, + "grad_norm": 2.383841470644709, + "learning_rate": 1.3847447560627512e-06, + "loss": 0.63, + "num_input_tokens_seen": 220696200, + "step": 6781 + }, + { + "epoch": 0.6116246561753168, + "flos": 26060733169920.0, + "grad_norm": 1.9123346743464553, + "learning_rate": 1.3841889019165812e-06, + "loss": 0.7871, + "num_input_tokens_seen": 220725460, + "step": 6782 + }, + { + "epoch": 0.611714839698787, + "flos": 30474815558880.0, + "grad_norm": 2.8429465364037534, + "learning_rate": 1.3836331003148106e-06, + "loss": 0.7029, + "num_input_tokens_seen": 220755220, + "step": 6783 + }, + { + "epoch": 0.6118050232222573, + "flos": 28253200951680.0, + "grad_norm": 1.6643019669567738, + "learning_rate": 1.3830773513048612e-06, + "loss": 0.7836, + "num_input_tokens_seen": 220785245, + "step": 6784 + }, + { + "epoch": 0.6118952067457275, + "flos": 21366865298400.0, + "grad_norm": 1.751370574481642, + "learning_rate": 1.382521654934155e-06, + "loss": 0.763, + "num_input_tokens_seen": 220813030, + "step": 6785 + }, + { + "epoch": 0.6119853902691978, + "flos": 22350717804000.0, + "grad_norm": 1.9512340088994866, + "learning_rate": 1.3819660112501057e-06, + "loss": 0.6949, + "num_input_tokens_seen": 220840650, + "step": 6786 + }, + { + "epoch": 0.6120755737926681, + "flos": 24900667333920.0, + "grad_norm": 1.725811272736038, + "learning_rate": 1.3814104203001234e-06, + "loss": 0.7115, + "num_input_tokens_seen": 220868635, + "step": 6787 + }, + { + "epoch": 0.6121657573161383, + "flos": 21003510728160.0, + "grad_norm": 1.592490427775936, + "learning_rate": 1.3808548821316156e-06, + "loss": 0.7881, + "num_input_tokens_seen": 220898490, + "step": 6788 + }, + { + "epoch": 0.6122559408396087, + "flos": 25735260318720.0, + "grad_norm": 1.679915088937363, + "learning_rate": 1.3802993967919824e-06, + "loss": 0.6798, + "num_input_tokens_seen": 220928425, + "step": 6789 + }, + { + "epoch": 0.6123461243630789, + "flos": 25010223817440.0, + "grad_norm": 2.2235684793888715, + "learning_rate": 1.3797439643286227e-06, + "loss": 0.7072, + "num_input_tokens_seen": 220956770, + "step": 6790 + }, + { + "epoch": 0.6124363078865491, + "flos": 26030359742400.0, + "grad_norm": 1.5582382829342813, + "learning_rate": 1.3791885847889277e-06, + "loss": 0.7344, + "num_input_tokens_seen": 220986840, + "step": 6791 + }, + { + "epoch": 0.6125264914100194, + "flos": 17578335932160.0, + "grad_norm": 1.6512887170195818, + "learning_rate": 1.3786332582202853e-06, + "loss": 0.7848, + "num_input_tokens_seen": 221012035, + "step": 6792 + }, + { + "epoch": 0.6126166749334897, + "flos": 25666373286240.0, + "grad_norm": 1.838242273205243, + "learning_rate": 1.3780779846700799e-06, + "loss": 0.747, + "num_input_tokens_seen": 221041210, + "step": 6793 + }, + { + "epoch": 0.6127068584569599, + "flos": 29964543162720.0, + "grad_norm": 1.6106780619349754, + "learning_rate": 1.3775227641856899e-06, + "loss": 0.7572, + "num_input_tokens_seen": 221073255, + "step": 6794 + }, + { + "epoch": 0.6127970419804302, + "flos": 26649593905920.0, + "grad_norm": 1.8511391972694102, + "learning_rate": 1.37696759681449e-06, + "loss": 0.7458, + "num_input_tokens_seen": 221104765, + "step": 6795 + }, + { + "epoch": 0.6128872255039004, + "flos": 64968796076640.0, + "grad_norm": 0.6921737204529915, + "learning_rate": 1.37641248260385e-06, + "loss": 0.6061, + "num_input_tokens_seen": 221195910, + "step": 6796 + }, + { + "epoch": 0.6129774090273707, + "flos": 23151926610720.0, + "grad_norm": 1.7866152220882607, + "learning_rate": 1.375857421601136e-06, + "loss": 0.7972, + "num_input_tokens_seen": 221224585, + "step": 6797 + }, + { + "epoch": 0.613067592550841, + "flos": 27597348840960.0, + "grad_norm": 2.3275446840579925, + "learning_rate": 1.3753024138537082e-06, + "loss": 0.758, + "num_input_tokens_seen": 221254530, + "step": 6798 + }, + { + "epoch": 0.6131577760743112, + "flos": 21476607630720.0, + "grad_norm": 1.5309986764614332, + "learning_rate": 1.3747474594089221e-06, + "loss": 0.8134, + "num_input_tokens_seen": 221281130, + "step": 6799 + }, + { + "epoch": 0.6132479595977814, + "flos": 25411199918400.0, + "grad_norm": 2.5517845342195393, + "learning_rate": 1.374192558314131e-06, + "loss": 0.7283, + "num_input_tokens_seen": 221309240, + "step": 6800 + }, + { + "epoch": 0.6133381431212518, + "flos": 22707790684800.0, + "grad_norm": 2.0274128984775364, + "learning_rate": 1.373637710616681e-06, + "loss": 0.8092, + "num_input_tokens_seen": 221333510, + "step": 6801 + }, + { + "epoch": 0.613428326644722, + "flos": 15099652299360.0, + "grad_norm": 2.1899791293260185, + "learning_rate": 1.373082916363916e-06, + "loss": 0.8569, + "num_input_tokens_seen": 221356760, + "step": 6802 + }, + { + "epoch": 0.6135185101681923, + "flos": 27378458892480.0, + "grad_norm": 1.5694948411500065, + "learning_rate": 1.3725281756031732e-06, + "loss": 0.7905, + "num_input_tokens_seen": 221388135, + "step": 6803 + }, + { + "epoch": 0.6136086936916625, + "flos": 20856555732480.0, + "grad_norm": 1.707377453228122, + "learning_rate": 1.3719734883817858e-06, + "loss": 0.749, + "num_input_tokens_seen": 221416840, + "step": 6804 + }, + { + "epoch": 0.6136988772151328, + "flos": 24354483216000.0, + "grad_norm": 1.6765968454332423, + "learning_rate": 1.371418854747084e-06, + "loss": 0.8641, + "num_input_tokens_seen": 221445045, + "step": 6805 + }, + { + "epoch": 0.6137890607386031, + "flos": 27232656159360.0, + "grad_norm": 2.361787080927747, + "learning_rate": 1.3708642747463905e-06, + "loss": 0.784, + "num_input_tokens_seen": 221473385, + "step": 6806 + }, + { + "epoch": 0.6138792442620733, + "flos": 20454873406080.0, + "grad_norm": 1.7485442237224749, + "learning_rate": 1.370309748427027e-06, + "loss": 0.7785, + "num_input_tokens_seen": 221503280, + "step": 6807 + }, + { + "epoch": 0.6139694277855435, + "flos": 22496446197600.0, + "grad_norm": 1.9392948450999088, + "learning_rate": 1.3697552758363079e-06, + "loss": 0.8238, + "num_input_tokens_seen": 221531120, + "step": 6808 + }, + { + "epoch": 0.6140596113090139, + "flos": 21913644132480.0, + "grad_norm": 3.2997737633576127, + "learning_rate": 1.3692008570215432e-06, + "loss": 0.7536, + "num_input_tokens_seen": 221560715, + "step": 6809 + }, + { + "epoch": 0.6141497948324841, + "flos": 23618593144800.0, + "grad_norm": 1.7004050137002846, + "learning_rate": 1.3686464920300398e-06, + "loss": 0.7043, + "num_input_tokens_seen": 221589830, + "step": 6810 + }, + { + "epoch": 0.6142399783559543, + "flos": 28507965452160.0, + "grad_norm": 2.2481641697381085, + "learning_rate": 1.3680921809090985e-06, + "loss": 0.7257, + "num_input_tokens_seen": 221621060, + "step": 6811 + }, + { + "epoch": 0.6143301618794246, + "flos": 30875828829600.0, + "grad_norm": 1.969515212771875, + "learning_rate": 1.3675379237060175e-06, + "loss": 0.7864, + "num_input_tokens_seen": 221649290, + "step": 6812 + }, + { + "epoch": 0.6144203454028949, + "flos": 33206219355360.0, + "grad_norm": 1.67542353726926, + "learning_rate": 1.366983720468088e-06, + "loss": 0.7447, + "num_input_tokens_seen": 221680085, + "step": 6813 + }, + { + "epoch": 0.6145105289263652, + "flos": 24718581181440.0, + "grad_norm": 2.6525783519336374, + "learning_rate": 1.3664295712425972e-06, + "loss": 0.7562, + "num_input_tokens_seen": 221711265, + "step": 6814 + }, + { + "epoch": 0.6146007124498354, + "flos": 29997704322240.0, + "grad_norm": 2.0139663140624937, + "learning_rate": 1.3658754760768296e-06, + "loss": 0.6849, + "num_input_tokens_seen": 221739960, + "step": 6815 + }, + { + "epoch": 0.6146908959733057, + "flos": 20092448079840.0, + "grad_norm": 1.9529731386274096, + "learning_rate": 1.3653214350180621e-06, + "loss": 0.7865, + "num_input_tokens_seen": 221765960, + "step": 6816 + }, + { + "epoch": 0.614781079496776, + "flos": 20528778356160.0, + "grad_norm": 1.6468081263477183, + "learning_rate": 1.3647674481135703e-06, + "loss": 0.8173, + "num_input_tokens_seen": 221792505, + "step": 6817 + }, + { + "epoch": 0.6148712630202462, + "flos": 23298658587840.0, + "grad_norm": 1.4016903600984623, + "learning_rate": 1.3642135154106217e-06, + "loss": 0.6486, + "num_input_tokens_seen": 221821440, + "step": 6818 + }, + { + "epoch": 0.6149614465437164, + "flos": 24349688316960.0, + "grad_norm": 2.332293267282877, + "learning_rate": 1.363659636956482e-06, + "loss": 0.704, + "num_input_tokens_seen": 221850450, + "step": 6819 + }, + { + "epoch": 0.6150516300671868, + "flos": 36194692177440.0, + "grad_norm": 1.655981437933823, + "learning_rate": 1.3631058127984112e-06, + "loss": 0.7193, + "num_input_tokens_seen": 221881280, + "step": 6820 + }, + { + "epoch": 0.615141813590657, + "flos": 22275623421600.0, + "grad_norm": 3.0242691670469704, + "learning_rate": 1.3625520429836632e-06, + "loss": 0.7251, + "num_input_tokens_seen": 221909015, + "step": 6821 + }, + { + "epoch": 0.6152319971141272, + "flos": 27087002105280.0, + "grad_norm": 1.720170630374897, + "learning_rate": 1.361998327559491e-06, + "loss": 0.6544, + "num_input_tokens_seen": 221937540, + "step": 6822 + }, + { + "epoch": 0.6153221806375975, + "flos": 64653730758240.0, + "grad_norm": 0.6601174472939478, + "learning_rate": 1.3614446665731385e-06, + "loss": 0.5805, + "num_input_tokens_seen": 222031135, + "step": 6823 + }, + { + "epoch": 0.6154123641610678, + "flos": 22022977597440.0, + "grad_norm": 2.6899957822745346, + "learning_rate": 1.3608910600718484e-06, + "loss": 0.7572, + "num_input_tokens_seen": 222060030, + "step": 6824 + }, + { + "epoch": 0.615502547684538, + "flos": 21073103986080.0, + "grad_norm": 1.8069581118838995, + "learning_rate": 1.360337508102857e-06, + "loss": 0.7518, + "num_input_tokens_seen": 222089435, + "step": 6825 + }, + { + "epoch": 0.6155927312080083, + "flos": 19799913369600.0, + "grad_norm": 2.187913365136075, + "learning_rate": 1.3597840107133962e-06, + "loss": 0.8459, + "num_input_tokens_seen": 222114305, + "step": 6826 + }, + { + "epoch": 0.6156829147314785, + "flos": 28683584066400.0, + "grad_norm": 1.67016242899159, + "learning_rate": 1.3592305679506944e-06, + "loss": 0.7586, + "num_input_tokens_seen": 222144415, + "step": 6827 + }, + { + "epoch": 0.6157730982549489, + "flos": 23879118958080.0, + "grad_norm": 1.6267447275182116, + "learning_rate": 1.3586771798619726e-06, + "loss": 0.6587, + "num_input_tokens_seen": 222174560, + "step": 6828 + }, + { + "epoch": 0.6158632817784191, + "flos": 27087150784320.0, + "grad_norm": 1.6905566189346426, + "learning_rate": 1.358123846494451e-06, + "loss": 0.8134, + "num_input_tokens_seen": 222203535, + "step": 6829 + }, + { + "epoch": 0.6159534653018893, + "flos": 24017450569440.0, + "grad_norm": 2.454239124551835, + "learning_rate": 1.3575705678953422e-06, + "loss": 0.753, + "num_input_tokens_seen": 222225690, + "step": 6830 + }, + { + "epoch": 0.6160436488253596, + "flos": 20310483123840.0, + "grad_norm": 1.8577023844894447, + "learning_rate": 1.357017344111854e-06, + "loss": 0.788, + "num_input_tokens_seen": 222253080, + "step": 6831 + }, + { + "epoch": 0.6161338323488299, + "flos": 19872889075680.0, + "grad_norm": 1.754153703594862, + "learning_rate": 1.356464175191192e-06, + "loss": 0.7612, + "num_input_tokens_seen": 222281635, + "step": 6832 + }, + { + "epoch": 0.6162240158723001, + "flos": 28253015102880.0, + "grad_norm": 2.3677065440779668, + "learning_rate": 1.3559110611805542e-06, + "loss": 0.7767, + "num_input_tokens_seen": 222310170, + "step": 6833 + }, + { + "epoch": 0.6163141993957704, + "flos": 20015904076800.0, + "grad_norm": 2.1582350508455947, + "learning_rate": 1.3553580021271372e-06, + "loss": 0.79, + "num_input_tokens_seen": 222338115, + "step": 6834 + }, + { + "epoch": 0.6164043829192406, + "flos": 23589669337920.0, + "grad_norm": 1.7348063851755966, + "learning_rate": 1.3548049980781297e-06, + "loss": 0.6934, + "num_input_tokens_seen": 222367615, + "step": 6835 + }, + { + "epoch": 0.616494566442711, + "flos": 22605630983520.0, + "grad_norm": 1.8842269206658766, + "learning_rate": 1.3542520490807166e-06, + "loss": 0.7409, + "num_input_tokens_seen": 222394500, + "step": 6836 + }, + { + "epoch": 0.6165847499661812, + "flos": 15209060103840.0, + "grad_norm": 2.176481014569391, + "learning_rate": 1.3536991551820802e-06, + "loss": 0.7397, + "num_input_tokens_seen": 222420095, + "step": 6837 + }, + { + "epoch": 0.6166749334896514, + "flos": 37214344895520.0, + "grad_norm": 1.715029626039906, + "learning_rate": 1.3531463164293952e-06, + "loss": 0.7202, + "num_input_tokens_seen": 222455885, + "step": 6838 + }, + { + "epoch": 0.6167651170131218, + "flos": 20492234748480.0, + "grad_norm": 2.0030464573546576, + "learning_rate": 1.3525935328698332e-06, + "loss": 0.6603, + "num_input_tokens_seen": 222484030, + "step": 6839 + }, + { + "epoch": 0.616855300536592, + "flos": 23219810059680.0, + "grad_norm": 3.275597053442098, + "learning_rate": 1.3520408045505605e-06, + "loss": 0.6894, + "num_input_tokens_seen": 222512020, + "step": 6840 + }, + { + "epoch": 0.6169454840600622, + "flos": 24828546532320.0, + "grad_norm": 1.986453072098031, + "learning_rate": 1.3514881315187396e-06, + "loss": 0.7159, + "num_input_tokens_seen": 222541050, + "step": 6841 + }, + { + "epoch": 0.6170356675835325, + "flos": 24132508177440.0, + "grad_norm": 1.6641000376304602, + "learning_rate": 1.3509355138215273e-06, + "loss": 0.809, + "num_input_tokens_seen": 222568680, + "step": 6842 + }, + { + "epoch": 0.6171258511070028, + "flos": 21731632319520.0, + "grad_norm": 6.1543036932682, + "learning_rate": 1.350382951506075e-06, + "loss": 0.832, + "num_input_tokens_seen": 222596105, + "step": 6843 + }, + { + "epoch": 0.617216034630473, + "flos": 69451653988800.0, + "grad_norm": 0.6400769378192945, + "learning_rate": 1.349830444619532e-06, + "loss": 0.5669, + "num_input_tokens_seen": 222688050, + "step": 6844 + }, + { + "epoch": 0.6173062181539433, + "flos": 28617038728800.0, + "grad_norm": 3.127632100755813, + "learning_rate": 1.3492779932090397e-06, + "loss": 0.7647, + "num_input_tokens_seen": 222716850, + "step": 6845 + }, + { + "epoch": 0.6173964016774135, + "flos": 23261594603520.0, + "grad_norm": 1.863975859272868, + "learning_rate": 1.3487255973217377e-06, + "loss": 0.777, + "num_input_tokens_seen": 222747930, + "step": 6846 + }, + { + "epoch": 0.6174865852008838, + "flos": 23953730133600.0, + "grad_norm": 1.7026745961830347, + "learning_rate": 1.3481732570047592e-06, + "loss": 0.8216, + "num_input_tokens_seen": 222775960, + "step": 6847 + }, + { + "epoch": 0.6175767687243541, + "flos": 26940790504800.0, + "grad_norm": 2.9922152286370216, + "learning_rate": 1.3476209723052314e-06, + "loss": 0.6812, + "num_input_tokens_seen": 222804295, + "step": 6848 + }, + { + "epoch": 0.6176669522478243, + "flos": 23187875502240.0, + "grad_norm": 1.9338124777645103, + "learning_rate": 1.3470687432702806e-06, + "loss": 0.7753, + "num_input_tokens_seen": 222830945, + "step": 6849 + }, + { + "epoch": 0.6177571357712945, + "flos": 23152075289760.0, + "grad_norm": 2.4186508868115237, + "learning_rate": 1.346516569947024e-06, + "loss": 0.6814, + "num_input_tokens_seen": 222858835, + "step": 6850 + }, + { + "epoch": 0.6178473192947649, + "flos": 15283373921280.0, + "grad_norm": 2.0483545677919843, + "learning_rate": 1.3459644523825774e-06, + "loss": 0.7367, + "num_input_tokens_seen": 222887175, + "step": 6851 + }, + { + "epoch": 0.6179375028182351, + "flos": 16084099521120.0, + "grad_norm": 2.251370648291833, + "learning_rate": 1.34541239062405e-06, + "loss": 0.7893, + "num_input_tokens_seen": 222915485, + "step": 6852 + }, + { + "epoch": 0.6180276863417054, + "flos": 24318274136160.0, + "grad_norm": 1.543093055836341, + "learning_rate": 1.3448603847185464e-06, + "loss": 0.7504, + "num_input_tokens_seen": 222946700, + "step": 6853 + }, + { + "epoch": 0.6181178698651756, + "flos": 22819986221280.0, + "grad_norm": 1.7155610063913176, + "learning_rate": 1.344308434713168e-06, + "loss": 0.6906, + "num_input_tokens_seen": 222975075, + "step": 6854 + }, + { + "epoch": 0.6182080533886459, + "flos": 22860469823520.0, + "grad_norm": 2.2601983687177967, + "learning_rate": 1.3437565406550083e-06, + "loss": 0.6817, + "num_input_tokens_seen": 223006805, + "step": 6855 + }, + { + "epoch": 0.6182982369121162, + "flos": 13862001707040.0, + "grad_norm": 2.3066911683774993, + "learning_rate": 1.34320470259116e-06, + "loss": 0.7341, + "num_input_tokens_seen": 223033685, + "step": 6856 + }, + { + "epoch": 0.6183884204355864, + "flos": 27411657221760.0, + "grad_norm": 2.601264730372372, + "learning_rate": 1.3426529205687078e-06, + "loss": 0.6532, + "num_input_tokens_seen": 223063585, + "step": 6857 + }, + { + "epoch": 0.6184786039590566, + "flos": 24097823057760.0, + "grad_norm": 1.7787020345347562, + "learning_rate": 1.3421011946347323e-06, + "loss": 0.8061, + "num_input_tokens_seen": 223091145, + "step": 6858 + }, + { + "epoch": 0.618568787482527, + "flos": 19435518046080.0, + "grad_norm": 1.8964232131414325, + "learning_rate": 1.3415495248363113e-06, + "loss": 0.7118, + "num_input_tokens_seen": 223120735, + "step": 6859 + }, + { + "epoch": 0.6186589710059972, + "flos": 15975360772320.0, + "grad_norm": 2.638700698254632, + "learning_rate": 1.3409979112205148e-06, + "loss": 0.7733, + "num_input_tokens_seen": 223146820, + "step": 6860 + }, + { + "epoch": 0.6187491545294674, + "flos": 13279162472160.0, + "grad_norm": 2.034971496044988, + "learning_rate": 1.3404463538344107e-06, + "loss": 0.7795, + "num_input_tokens_seen": 223172565, + "step": 6861 + }, + { + "epoch": 0.6188393380529378, + "flos": 20928156157440.0, + "grad_norm": 2.171270821602733, + "learning_rate": 1.3398948527250602e-06, + "loss": 0.838, + "num_input_tokens_seen": 223194875, + "step": 6862 + }, + { + "epoch": 0.618929521576408, + "flos": 69727047706080.0, + "grad_norm": 1.458091125356416, + "learning_rate": 1.3393434079395207e-06, + "loss": 0.5909, + "num_input_tokens_seen": 223284280, + "step": 6863 + }, + { + "epoch": 0.6190197050998782, + "flos": 24931709817120.0, + "grad_norm": 1.7514468751256285, + "learning_rate": 1.3387920195248449e-06, + "loss": 0.7365, + "num_input_tokens_seen": 223314470, + "step": 6864 + }, + { + "epoch": 0.6191098886233485, + "flos": 28322459681760.0, + "grad_norm": 5.467173147387116, + "learning_rate": 1.3382406875280791e-06, + "loss": 0.7648, + "num_input_tokens_seen": 223345545, + "step": 6865 + }, + { + "epoch": 0.6192000721468188, + "flos": 17608932378240.0, + "grad_norm": 1.9576771292804986, + "learning_rate": 1.3376894119962672e-06, + "loss": 0.7741, + "num_input_tokens_seen": 223371965, + "step": 6866 + }, + { + "epoch": 0.6192902556702891, + "flos": 20966149385760.0, + "grad_norm": 2.262839979898642, + "learning_rate": 1.3371381929764464e-06, + "loss": 0.7773, + "num_input_tokens_seen": 223397085, + "step": 6867 + }, + { + "epoch": 0.6193804391937593, + "flos": 67807860698880.0, + "grad_norm": 0.6383853001104386, + "learning_rate": 1.3365870305156506e-06, + "loss": 0.5696, + "num_input_tokens_seen": 223486750, + "step": 6868 + }, + { + "epoch": 0.6194706227172295, + "flos": 22205992993920.0, + "grad_norm": 2.4023556592692614, + "learning_rate": 1.3360359246609073e-06, + "loss": 0.7535, + "num_input_tokens_seen": 223512795, + "step": 6869 + }, + { + "epoch": 0.6195608062406999, + "flos": 21184630466880.0, + "grad_norm": 1.7026204753042926, + "learning_rate": 1.3354848754592387e-06, + "loss": 0.7457, + "num_input_tokens_seen": 223541925, + "step": 6870 + }, + { + "epoch": 0.6196509897641701, + "flos": 29520853273920.0, + "grad_norm": 2.219655993674499, + "learning_rate": 1.334933882957666e-06, + "loss": 0.8385, + "num_input_tokens_seen": 223569980, + "step": 6871 + }, + { + "epoch": 0.6197411732876403, + "flos": 24317939608320.0, + "grad_norm": 1.948312980876226, + "learning_rate": 1.3343829472032004e-06, + "loss": 0.7141, + "num_input_tokens_seen": 223600295, + "step": 6872 + }, + { + "epoch": 0.6198313568111106, + "flos": 23152298308320.0, + "grad_norm": 2.637007995564415, + "learning_rate": 1.3338320682428527e-06, + "loss": 0.7354, + "num_input_tokens_seen": 223628995, + "step": 6873 + }, + { + "epoch": 0.6199215403345809, + "flos": 24677577202560.0, + "grad_norm": 2.091579621873437, + "learning_rate": 1.3332812461236263e-06, + "loss": 0.7129, + "num_input_tokens_seen": 223656880, + "step": 6874 + }, + { + "epoch": 0.6200117238580511, + "flos": 20893433868000.0, + "grad_norm": 1.5391394891217225, + "learning_rate": 1.3327304808925192e-06, + "loss": 0.7142, + "num_input_tokens_seen": 223686125, + "step": 6875 + }, + { + "epoch": 0.6201019073815214, + "flos": 30219716530560.0, + "grad_norm": 2.834246338738887, + "learning_rate": 1.332179772596528e-06, + "loss": 0.6125, + "num_input_tokens_seen": 223715500, + "step": 6876 + }, + { + "epoch": 0.6201920909049916, + "flos": 26973877324800.0, + "grad_norm": 1.838192078923896, + "learning_rate": 1.3316291212826402e-06, + "loss": 0.6751, + "num_input_tokens_seen": 223742125, + "step": 6877 + }, + { + "epoch": 0.620282274428462, + "flos": 19832405473440.0, + "grad_norm": 1.9752642925495167, + "learning_rate": 1.3310785269978413e-06, + "loss": 0.7993, + "num_input_tokens_seen": 223770450, + "step": 6878 + }, + { + "epoch": 0.6203724579519322, + "flos": 20310408784320.0, + "grad_norm": 6.405350189284211, + "learning_rate": 1.3305279897891111e-06, + "loss": 0.7902, + "num_input_tokens_seen": 223797500, + "step": 6879 + }, + { + "epoch": 0.6204626414754024, + "flos": 34955480455200.0, + "grad_norm": 2.7280762413535466, + "learning_rate": 1.329977509703424e-06, + "loss": 0.6145, + "num_input_tokens_seen": 223827655, + "step": 6880 + }, + { + "epoch": 0.6205528249988727, + "flos": 18962569822560.0, + "grad_norm": 2.1352724972571355, + "learning_rate": 1.32942708678775e-06, + "loss": 0.8371, + "num_input_tokens_seen": 223852665, + "step": 6881 + }, + { + "epoch": 0.620643008522343, + "flos": 25295250236160.0, + "grad_norm": 1.7455541813925461, + "learning_rate": 1.3288767210890548e-06, + "loss": 0.762, + "num_input_tokens_seen": 223882040, + "step": 6882 + }, + { + "epoch": 0.6207331920458132, + "flos": 23844656856960.0, + "grad_norm": 2.214882825072805, + "learning_rate": 1.3283264126542986e-06, + "loss": 0.715, + "num_input_tokens_seen": 223911150, + "step": 6883 + }, + { + "epoch": 0.6208233755692835, + "flos": 28284689472000.0, + "grad_norm": 1.7274539769138124, + "learning_rate": 1.3277761615304356e-06, + "loss": 0.7487, + "num_input_tokens_seen": 223941150, + "step": 6884 + }, + { + "epoch": 0.6209135590927538, + "flos": 24099198338880.0, + "grad_norm": 1.9859400907510085, + "learning_rate": 1.3272259677644178e-06, + "loss": 0.8494, + "num_input_tokens_seen": 223967865, + "step": 6885 + }, + { + "epoch": 0.621003742616224, + "flos": 24572964297120.0, + "grad_norm": 2.3409924370253656, + "learning_rate": 1.32667583140319e-06, + "loss": 0.6229, + "num_input_tokens_seen": 223998340, + "step": 6886 + }, + { + "epoch": 0.6210939261396943, + "flos": 13752036356160.0, + "grad_norm": 3.1346875902962523, + "learning_rate": 1.3261257524936924e-06, + "loss": 0.7892, + "num_input_tokens_seen": 224024125, + "step": 6887 + }, + { + "epoch": 0.6211841096631645, + "flos": 25590981545760.0, + "grad_norm": 1.7111445610828269, + "learning_rate": 1.3255757310828619e-06, + "loss": 0.7778, + "num_input_tokens_seen": 224054910, + "step": 6888 + }, + { + "epoch": 0.6212742931866349, + "flos": 13971112153440.0, + "grad_norm": 2.2641148236680526, + "learning_rate": 1.3250257672176282e-06, + "loss": 0.7676, + "num_input_tokens_seen": 224082005, + "step": 6889 + }, + { + "epoch": 0.6213644767101051, + "flos": 24825981818880.0, + "grad_norm": 1.982867111043189, + "learning_rate": 1.3244758609449183e-06, + "loss": 0.6564, + "num_input_tokens_seen": 224112300, + "step": 6890 + }, + { + "epoch": 0.6214546602335753, + "flos": 20447625302880.0, + "grad_norm": 1.9392902962826957, + "learning_rate": 1.323926012311653e-06, + "loss": 0.7691, + "num_input_tokens_seen": 224137325, + "step": 6891 + }, + { + "epoch": 0.6215448437570456, + "flos": 21913792811520.0, + "grad_norm": 3.8803155188722704, + "learning_rate": 1.3233762213647476e-06, + "loss": 0.7772, + "num_input_tokens_seen": 224165040, + "step": 6892 + }, + { + "epoch": 0.6216350272805159, + "flos": 27194885949600.0, + "grad_norm": 1.915212729368874, + "learning_rate": 1.3228264881511137e-06, + "loss": 0.725, + "num_input_tokens_seen": 224194335, + "step": 6893 + }, + { + "epoch": 0.6217252108039861, + "flos": 20783765875200.0, + "grad_norm": 1.722686974906696, + "learning_rate": 1.322276812717658e-06, + "loss": 0.8311, + "num_input_tokens_seen": 224223025, + "step": 6894 + }, + { + "epoch": 0.6218153943274564, + "flos": 27669878509920.0, + "grad_norm": 1.921485054960555, + "learning_rate": 1.3217271951112818e-06, + "loss": 0.6589, + "num_input_tokens_seen": 224256850, + "step": 6895 + }, + { + "epoch": 0.6219055778509266, + "flos": 21907102254720.0, + "grad_norm": 2.174932509504246, + "learning_rate": 1.321177635378881e-06, + "loss": 0.7322, + "num_input_tokens_seen": 224283900, + "step": 6896 + }, + { + "epoch": 0.6219957613743969, + "flos": 35539694971200.0, + "grad_norm": 2.4660110732214346, + "learning_rate": 1.3206281335673475e-06, + "loss": 0.642, + "num_input_tokens_seen": 224317700, + "step": 6897 + }, + { + "epoch": 0.6220859448978672, + "flos": 25264876808640.0, + "grad_norm": 2.255743537339739, + "learning_rate": 1.3200786897235677e-06, + "loss": 0.7252, + "num_input_tokens_seen": 224344865, + "step": 6898 + }, + { + "epoch": 0.6221761284213374, + "flos": 21767506871520.0, + "grad_norm": 1.8409487894489447, + "learning_rate": 1.3195293038944227e-06, + "loss": 0.7659, + "num_input_tokens_seen": 224372430, + "step": 6899 + }, + { + "epoch": 0.6222663119448076, + "flos": 34770495061440.0, + "grad_norm": 1.6539479217541149, + "learning_rate": 1.3189799761267902e-06, + "loss": 0.743, + "num_input_tokens_seen": 224406740, + "step": 6900 + }, + { + "epoch": 0.622356495468278, + "flos": 19837014523680.0, + "grad_norm": 3.511100332696952, + "learning_rate": 1.3184307064675412e-06, + "loss": 0.7457, + "num_input_tokens_seen": 224434025, + "step": 6901 + }, + { + "epoch": 0.6224466789917482, + "flos": 62763201636000.0, + "grad_norm": 0.6727708566041715, + "learning_rate": 1.3178814949635416e-06, + "loss": 0.6238, + "num_input_tokens_seen": 224524915, + "step": 6902 + }, + { + "epoch": 0.6225368625152184, + "flos": 27406787983200.0, + "grad_norm": 1.9639290411497214, + "learning_rate": 1.3173323416616549e-06, + "loss": 0.6982, + "num_input_tokens_seen": 224554220, + "step": 6903 + }, + { + "epoch": 0.6226270460386887, + "flos": 26539740064320.0, + "grad_norm": 2.4260435592187553, + "learning_rate": 1.3167832466087361e-06, + "loss": 0.804, + "num_input_tokens_seen": 224582590, + "step": 6904 + }, + { + "epoch": 0.622717229562159, + "flos": 26321147473920.0, + "grad_norm": 1.6772173794394323, + "learning_rate": 1.3162342098516388e-06, + "loss": 0.6988, + "num_input_tokens_seen": 224614030, + "step": 6905 + }, + { + "epoch": 0.6228074130856293, + "flos": 21913495453440.0, + "grad_norm": 1.7595107627511293, + "learning_rate": 1.3156852314372086e-06, + "loss": 0.6703, + "num_input_tokens_seen": 224643485, + "step": 6906 + }, + { + "epoch": 0.6228975966090995, + "flos": 68429548066560.0, + "grad_norm": 0.6769942714119163, + "learning_rate": 1.3151363114122882e-06, + "loss": 0.5595, + "num_input_tokens_seen": 224734525, + "step": 6907 + }, + { + "epoch": 0.6229877801325698, + "flos": 23953692963840.0, + "grad_norm": 1.722583518363457, + "learning_rate": 1.3145874498237133e-06, + "loss": 0.7633, + "num_input_tokens_seen": 224765535, + "step": 6908 + }, + { + "epoch": 0.6230779636560401, + "flos": 22750578812160.0, + "grad_norm": 1.6675493843280844, + "learning_rate": 1.3140386467183166e-06, + "loss": 0.7981, + "num_input_tokens_seen": 224793320, + "step": 6909 + }, + { + "epoch": 0.6231681471795103, + "flos": 24607798095840.0, + "grad_norm": 1.5380560978008246, + "learning_rate": 1.3134899021429258e-06, + "loss": 0.7909, + "num_input_tokens_seen": 224822165, + "step": 6910 + }, + { + "epoch": 0.6232583307029805, + "flos": 27268047504480.0, + "grad_norm": 2.150336075176635, + "learning_rate": 1.3129412161443604e-06, + "loss": 0.7334, + "num_input_tokens_seen": 224851335, + "step": 6911 + }, + { + "epoch": 0.6233485142264509, + "flos": 23698482426240.0, + "grad_norm": 1.7277524136027016, + "learning_rate": 1.3123925887694402e-06, + "loss": 0.7561, + "num_input_tokens_seen": 224880280, + "step": 6912 + }, + { + "epoch": 0.6234386977499211, + "flos": 24753117622080.0, + "grad_norm": 1.6631161860046444, + "learning_rate": 1.3118440200649756e-06, + "loss": 0.7605, + "num_input_tokens_seen": 224908900, + "step": 6913 + }, + { + "epoch": 0.6235288812733913, + "flos": 21439171948800.0, + "grad_norm": 1.6969455289848507, + "learning_rate": 1.3112955100777727e-06, + "loss": 0.7113, + "num_input_tokens_seen": 224939425, + "step": 6914 + }, + { + "epoch": 0.6236190647968616, + "flos": 24971264175360.0, + "grad_norm": 2.262044142011587, + "learning_rate": 1.3107470588546353e-06, + "loss": 0.7874, + "num_input_tokens_seen": 224969360, + "step": 6915 + }, + { + "epoch": 0.6237092483203319, + "flos": 24354483216000.0, + "grad_norm": 2.053895514842143, + "learning_rate": 1.3101986664423583e-06, + "loss": 0.7212, + "num_input_tokens_seen": 224997775, + "step": 6916 + }, + { + "epoch": 0.6237994318438022, + "flos": 30289272618720.0, + "grad_norm": 2.0171296174175657, + "learning_rate": 1.3096503328877358e-06, + "loss": 0.7252, + "num_input_tokens_seen": 225025580, + "step": 6917 + }, + { + "epoch": 0.6238896153672724, + "flos": 20598483123360.0, + "grad_norm": 1.7804113722791752, + "learning_rate": 1.309102058237553e-06, + "loss": 0.77, + "num_input_tokens_seen": 225054800, + "step": 6918 + }, + { + "epoch": 0.6239797988907426, + "flos": 25664886495840.0, + "grad_norm": 6.391780235876405, + "learning_rate": 1.3085538425385917e-06, + "loss": 0.801, + "num_input_tokens_seen": 225085815, + "step": 6919 + }, + { + "epoch": 0.624069982414213, + "flos": 19545780755040.0, + "grad_norm": 2.2641425182452686, + "learning_rate": 1.3080056858376298e-06, + "loss": 0.7837, + "num_input_tokens_seen": 225113960, + "step": 6920 + }, + { + "epoch": 0.6241601659376832, + "flos": 29892533870400.0, + "grad_norm": 2.4058106256604086, + "learning_rate": 1.3074575881814383e-06, + "loss": 0.7359, + "num_input_tokens_seen": 225140825, + "step": 6921 + }, + { + "epoch": 0.6242503494611534, + "flos": 29782865877600.0, + "grad_norm": 3.4364587767432, + "learning_rate": 1.3069095496167841e-06, + "loss": 0.6727, + "num_input_tokens_seen": 225170010, + "step": 6922 + }, + { + "epoch": 0.6243405329846237, + "flos": 22164171280320.0, + "grad_norm": 2.037916424094231, + "learning_rate": 1.3063615701904285e-06, + "loss": 0.6642, + "num_input_tokens_seen": 225197905, + "step": 6923 + }, + { + "epoch": 0.624430716508094, + "flos": 23407137148320.0, + "grad_norm": 2.44367794701387, + "learning_rate": 1.3058136499491283e-06, + "loss": 0.782, + "num_input_tokens_seen": 225225685, + "step": 6924 + }, + { + "epoch": 0.6245209000315642, + "flos": 20565544982400.0, + "grad_norm": 2.273775370552241, + "learning_rate": 1.3052657889396357e-06, + "loss": 0.7325, + "num_input_tokens_seen": 225254530, + "step": 6925 + }, + { + "epoch": 0.6246110835550345, + "flos": 21986917196640.0, + "grad_norm": 2.659642017196572, + "learning_rate": 1.304717987208696e-06, + "loss": 0.737, + "num_input_tokens_seen": 225280145, + "step": 6926 + }, + { + "epoch": 0.6247012670785047, + "flos": 27232953517440.0, + "grad_norm": 1.6605094982634825, + "learning_rate": 1.304170244803052e-06, + "loss": 0.7374, + "num_input_tokens_seen": 225310445, + "step": 6927 + }, + { + "epoch": 0.624791450601975, + "flos": 31204275261600.0, + "grad_norm": 2.0507524179520864, + "learning_rate": 1.3036225617694387e-06, + "loss": 0.7168, + "num_input_tokens_seen": 225342000, + "step": 6928 + }, + { + "epoch": 0.6248816341254453, + "flos": 20420113946880.0, + "grad_norm": 1.606755393112246, + "learning_rate": 1.3030749381545892e-06, + "loss": 0.7803, + "num_input_tokens_seen": 225369505, + "step": 6929 + }, + { + "epoch": 0.6249718176489155, + "flos": 15865767119040.0, + "grad_norm": 2.5240528785377667, + "learning_rate": 1.3025273740052285e-06, + "loss": 0.6865, + "num_input_tokens_seen": 225397155, + "step": 6930 + }, + { + "epoch": 0.6250620011723858, + "flos": 24681703045920.0, + "grad_norm": 1.534965134389343, + "learning_rate": 1.3019798693680774e-06, + "loss": 0.7942, + "num_input_tokens_seen": 225429335, + "step": 6931 + }, + { + "epoch": 0.6251521846958561, + "flos": 25300900039680.0, + "grad_norm": 1.9517390712638751, + "learning_rate": 1.3014324242898536e-06, + "loss": 0.8659, + "num_input_tokens_seen": 225458050, + "step": 6932 + }, + { + "epoch": 0.6252423682193263, + "flos": 21767841399360.0, + "grad_norm": 4.059896016034246, + "learning_rate": 1.3008850388172668e-06, + "loss": 0.8061, + "num_input_tokens_seen": 225485915, + "step": 6933 + }, + { + "epoch": 0.6253325517427966, + "flos": 28220931866400.0, + "grad_norm": 1.4504333506238922, + "learning_rate": 1.3003377129970233e-06, + "loss": 0.7056, + "num_input_tokens_seen": 225518585, + "step": 6934 + }, + { + "epoch": 0.6254227352662669, + "flos": 33972185496000.0, + "grad_norm": 1.4327280200609511, + "learning_rate": 1.2997904468758243e-06, + "loss": 0.668, + "num_input_tokens_seen": 225553685, + "step": 6935 + }, + { + "epoch": 0.6255129187897371, + "flos": 26432413766400.0, + "grad_norm": 1.8709345931053996, + "learning_rate": 1.299243240500365e-06, + "loss": 0.7639, + "num_input_tokens_seen": 225581025, + "step": 6936 + }, + { + "epoch": 0.6256031023132074, + "flos": 25957458375840.0, + "grad_norm": 1.8461847554883843, + "learning_rate": 1.2986960939173368e-06, + "loss": 0.7654, + "num_input_tokens_seen": 225608885, + "step": 6937 + }, + { + "epoch": 0.6256932858366776, + "flos": 15792717073440.0, + "grad_norm": 2.0770339970164096, + "learning_rate": 1.298149007173424e-06, + "loss": 0.7909, + "num_input_tokens_seen": 225633985, + "step": 6938 + }, + { + "epoch": 0.625783469360148, + "flos": 17651162959200.0, + "grad_norm": 2.1047747145766205, + "learning_rate": 1.2976019803153087e-06, + "loss": 0.7798, + "num_input_tokens_seen": 225660600, + "step": 6939 + }, + { + "epoch": 0.6258736528836182, + "flos": 30544929193440.0, + "grad_norm": 1.919343350893895, + "learning_rate": 1.2970550133896652e-06, + "loss": 0.694, + "num_input_tokens_seen": 225690795, + "step": 6940 + }, + { + "epoch": 0.6259638364070884, + "flos": 20711199036480.0, + "grad_norm": 2.1457467169702653, + "learning_rate": 1.2965081064431634e-06, + "loss": 0.8319, + "num_input_tokens_seen": 225717000, + "step": 6941 + }, + { + "epoch": 0.6260540199305586, + "flos": 19691397639360.0, + "grad_norm": 2.7542505360201415, + "learning_rate": 1.2959612595224698e-06, + "loss": 0.7704, + "num_input_tokens_seen": 225743030, + "step": 6942 + }, + { + "epoch": 0.626144203454029, + "flos": 15645501889440.0, + "grad_norm": 2.4015216692380323, + "learning_rate": 1.2954144726742424e-06, + "loss": 0.7615, + "num_input_tokens_seen": 225766075, + "step": 6943 + }, + { + "epoch": 0.6262343869774992, + "flos": 26540817987360.0, + "grad_norm": 1.7561225736282098, + "learning_rate": 1.2948677459451385e-06, + "loss": 0.7438, + "num_input_tokens_seen": 225797525, + "step": 6944 + }, + { + "epoch": 0.6263245705009695, + "flos": 23881609332000.0, + "grad_norm": 2.464922194887572, + "learning_rate": 1.2943210793818064e-06, + "loss": 0.7946, + "num_input_tokens_seen": 225824965, + "step": 6945 + }, + { + "epoch": 0.6264147540244397, + "flos": 32369730712800.0, + "grad_norm": 2.0610080293353694, + "learning_rate": 1.2937744730308899e-06, + "loss": 0.7282, + "num_input_tokens_seen": 225856305, + "step": 6946 + }, + { + "epoch": 0.62650493754791, + "flos": 26650002773280.0, + "grad_norm": 2.7612490760452846, + "learning_rate": 1.2932279269390305e-06, + "loss": 0.7444, + "num_input_tokens_seen": 225884515, + "step": 6947 + }, + { + "epoch": 0.6265951210713803, + "flos": 22278113795520.0, + "grad_norm": 1.9546138124428152, + "learning_rate": 1.292681441152861e-06, + "loss": 0.6522, + "num_input_tokens_seen": 225912110, + "step": 6948 + }, + { + "epoch": 0.6266853045948505, + "flos": 24463816680960.0, + "grad_norm": 2.1064582155811347, + "learning_rate": 1.292135015719011e-06, + "loss": 0.7635, + "num_input_tokens_seen": 225940985, + "step": 6949 + }, + { + "epoch": 0.6267754881183207, + "flos": 18889519776960.0, + "grad_norm": 1.931510897794035, + "learning_rate": 1.2915886506841046e-06, + "loss": 0.7062, + "num_input_tokens_seen": 225969400, + "step": 6950 + }, + { + "epoch": 0.6268656716417911, + "flos": 35575569523200.0, + "grad_norm": 1.6319921157403374, + "learning_rate": 1.2910423460947613e-06, + "loss": 0.6818, + "num_input_tokens_seen": 226001060, + "step": 6951 + }, + { + "epoch": 0.6269558551652613, + "flos": 24134106477120.0, + "grad_norm": 1.9627641172113939, + "learning_rate": 1.290496101997594e-06, + "loss": 0.718, + "num_input_tokens_seen": 226029105, + "step": 6952 + }, + { + "epoch": 0.6270460386887315, + "flos": 62796771662880.0, + "grad_norm": 0.7183115077704342, + "learning_rate": 1.2899499184392105e-06, + "loss": 0.5809, + "num_input_tokens_seen": 226119540, + "step": 6953 + }, + { + "epoch": 0.6271362222122018, + "flos": 24352327369920.0, + "grad_norm": 2.0072052627388057, + "learning_rate": 1.289403795466216e-06, + "loss": 0.7272, + "num_input_tokens_seen": 226148495, + "step": 6954 + }, + { + "epoch": 0.6272264057356721, + "flos": 17977936752000.0, + "grad_norm": 2.618709838850569, + "learning_rate": 1.288857733125207e-06, + "loss": 0.7637, + "num_input_tokens_seen": 226173755, + "step": 6955 + }, + { + "epoch": 0.6273165892591424, + "flos": 24864309575040.0, + "grad_norm": 2.015654807646604, + "learning_rate": 1.2883117314627785e-06, + "loss": 0.7424, + "num_input_tokens_seen": 226202820, + "step": 6956 + }, + { + "epoch": 0.6274067727826126, + "flos": 25556407935360.0, + "grad_norm": 1.7151823991142883, + "learning_rate": 1.2877657905255168e-06, + "loss": 0.688, + "num_input_tokens_seen": 226233255, + "step": 6957 + }, + { + "epoch": 0.6274969563060829, + "flos": 23110996971360.0, + "grad_norm": 1.8517033004341557, + "learning_rate": 1.2872199103600046e-06, + "loss": 0.8109, + "num_input_tokens_seen": 226261610, + "step": 6958 + }, + { + "epoch": 0.6275871398295532, + "flos": 24938103015840.0, + "grad_norm": 2.0102157755940087, + "learning_rate": 1.286674091012821e-06, + "loss": 0.8172, + "num_input_tokens_seen": 226291955, + "step": 6959 + }, + { + "epoch": 0.6276773233530234, + "flos": 19873818319680.0, + "grad_norm": 1.7921757824888036, + "learning_rate": 1.2861283325305356e-06, + "loss": 0.805, + "num_input_tokens_seen": 226317010, + "step": 6960 + }, + { + "epoch": 0.6277675068764936, + "flos": 21035668304160.0, + "grad_norm": 2.5337040625737806, + "learning_rate": 1.2855826349597185e-06, + "loss": 0.7464, + "num_input_tokens_seen": 226343745, + "step": 6961 + }, + { + "epoch": 0.627857690399964, + "flos": 19035545528640.0, + "grad_norm": 1.5593863935952594, + "learning_rate": 1.2850369983469302e-06, + "loss": 0.6046, + "num_input_tokens_seen": 226372575, + "step": 6962 + }, + { + "epoch": 0.6279478739234342, + "flos": 26314902954240.0, + "grad_norm": 1.8739635781989952, + "learning_rate": 1.2844914227387266e-06, + "loss": 0.775, + "num_input_tokens_seen": 226400760, + "step": 6963 + }, + { + "epoch": 0.6280380574469044, + "flos": 29382038455680.0, + "grad_norm": 1.5088047853903666, + "learning_rate": 1.2839459081816606e-06, + "loss": 0.7712, + "num_input_tokens_seen": 226432885, + "step": 6964 + }, + { + "epoch": 0.6281282409703747, + "flos": 21942939636960.0, + "grad_norm": 2.165705198534188, + "learning_rate": 1.283400454722278e-06, + "loss": 0.8168, + "num_input_tokens_seen": 226457660, + "step": 6965 + }, + { + "epoch": 0.628218424493845, + "flos": 21907325273280.0, + "grad_norm": 1.8520124830233111, + "learning_rate": 1.28285506240712e-06, + "loss": 0.8122, + "num_input_tokens_seen": 226485795, + "step": 6966 + }, + { + "epoch": 0.6283086080173153, + "flos": 25630610243520.0, + "grad_norm": 1.6016393875646484, + "learning_rate": 1.2823097312827225e-06, + "loss": 0.7773, + "num_input_tokens_seen": 226517045, + "step": 6967 + }, + { + "epoch": 0.6283987915407855, + "flos": 24171690838080.0, + "grad_norm": 1.7515874883007565, + "learning_rate": 1.2817644613956153e-06, + "loss": 0.7583, + "num_input_tokens_seen": 226545995, + "step": 6968 + }, + { + "epoch": 0.6284889750642557, + "flos": 22715075957760.0, + "grad_norm": 6.822885419080005, + "learning_rate": 1.2812192527923253e-06, + "loss": 0.8098, + "num_input_tokens_seen": 226572860, + "step": 6969 + }, + { + "epoch": 0.6285791585877261, + "flos": 27738170826240.0, + "grad_norm": 2.088073208625324, + "learning_rate": 1.2806741055193712e-06, + "loss": 0.7814, + "num_input_tokens_seen": 226601890, + "step": 6970 + }, + { + "epoch": 0.6286693421111963, + "flos": 13898099277600.0, + "grad_norm": 3.0577659607290517, + "learning_rate": 1.2801290196232695e-06, + "loss": 0.7059, + "num_input_tokens_seen": 226628605, + "step": 6971 + }, + { + "epoch": 0.6287595256346665, + "flos": 24318236966400.0, + "grad_norm": 2.184170505190895, + "learning_rate": 1.2795839951505282e-06, + "loss": 0.7408, + "num_input_tokens_seen": 226658180, + "step": 6972 + }, + { + "epoch": 0.6288497091581368, + "flos": 14882100462240.0, + "grad_norm": 2.715278426026909, + "learning_rate": 1.2790390321476538e-06, + "loss": 0.7692, + "num_input_tokens_seen": 226684695, + "step": 6973 + }, + { + "epoch": 0.6289398926816071, + "flos": 24499802742240.0, + "grad_norm": 2.201036694495932, + "learning_rate": 1.2784941306611446e-06, + "loss": 0.7226, + "num_input_tokens_seen": 226712750, + "step": 6974 + }, + { + "epoch": 0.6290300762050773, + "flos": 16886534929920.0, + "grad_norm": 2.2565591362596753, + "learning_rate": 1.2779492907374935e-06, + "loss": 0.8091, + "num_input_tokens_seen": 226739740, + "step": 6975 + }, + { + "epoch": 0.6291202597285476, + "flos": 20893285188960.0, + "grad_norm": 1.9189462443342202, + "learning_rate": 1.2774045124231911e-06, + "loss": 0.7359, + "num_input_tokens_seen": 226766265, + "step": 6976 + }, + { + "epoch": 0.6292104432520178, + "flos": 16339644586560.0, + "grad_norm": 1.5416232266191252, + "learning_rate": 1.2768597957647197e-06, + "loss": 0.6318, + "num_input_tokens_seen": 226792870, + "step": 6977 + }, + { + "epoch": 0.6293006267754881, + "flos": 20712053940960.0, + "grad_norm": 2.836304359772528, + "learning_rate": 1.2763151408085582e-06, + "loss": 0.7605, + "num_input_tokens_seen": 226819630, + "step": 6978 + }, + { + "epoch": 0.6293908102989584, + "flos": 26759150389440.0, + "grad_norm": 1.9369064775575937, + "learning_rate": 1.2757705476011788e-06, + "loss": 0.7176, + "num_input_tokens_seen": 226849570, + "step": 6979 + }, + { + "epoch": 0.6294809938224286, + "flos": 23480484552000.0, + "grad_norm": 2.308766831366089, + "learning_rate": 1.27522601618905e-06, + "loss": 0.695, + "num_input_tokens_seen": 226879025, + "step": 6980 + }, + { + "epoch": 0.629571177345899, + "flos": 64620160731360.0, + "grad_norm": 0.920095925357728, + "learning_rate": 1.2746815466186337e-06, + "loss": 0.6033, + "num_input_tokens_seen": 226976485, + "step": 6981 + }, + { + "epoch": 0.6296613608693692, + "flos": 29195566271520.0, + "grad_norm": 1.872614169135365, + "learning_rate": 1.274137138936387e-06, + "loss": 0.7516, + "num_input_tokens_seen": 227006190, + "step": 6982 + }, + { + "epoch": 0.6297515443928394, + "flos": 46609551760320.0, + "grad_norm": 6.446654689794787, + "learning_rate": 1.2735927931887625e-06, + "loss": 0.749, + "num_input_tokens_seen": 227041610, + "step": 6983 + }, + { + "epoch": 0.6298417279163097, + "flos": 66909621617760.0, + "grad_norm": 0.6837574789973699, + "learning_rate": 1.2730485094222061e-06, + "loss": 0.5852, + "num_input_tokens_seen": 227130895, + "step": 6984 + }, + { + "epoch": 0.62993191143978, + "flos": 22125843524160.0, + "grad_norm": 1.7887352390066793, + "learning_rate": 1.2725042876831586e-06, + "loss": 0.7868, + "num_input_tokens_seen": 227159620, + "step": 6985 + }, + { + "epoch": 0.6300220949632502, + "flos": 25514883579840.0, + "grad_norm": 2.418755768367385, + "learning_rate": 1.2719601280180573e-06, + "loss": 0.6467, + "num_input_tokens_seen": 227188910, + "step": 6986 + }, + { + "epoch": 0.6301122784867205, + "flos": 25659422541120.0, + "grad_norm": 2.2444186391175798, + "learning_rate": 1.2714160304733317e-06, + "loss": 0.8155, + "num_input_tokens_seen": 227216055, + "step": 6987 + }, + { + "epoch": 0.6302024620101907, + "flos": 24971821721760.0, + "grad_norm": 1.9278454965954315, + "learning_rate": 1.2708719950954082e-06, + "loss": 0.7622, + "num_input_tokens_seen": 227245510, + "step": 6988 + }, + { + "epoch": 0.630292645533661, + "flos": 18307275258240.0, + "grad_norm": 1.9002641594611056, + "learning_rate": 1.2703280219307065e-06, + "loss": 0.7668, + "num_input_tokens_seen": 227272845, + "step": 6989 + }, + { + "epoch": 0.6303828290571313, + "flos": 20780643615360.0, + "grad_norm": 1.9549705026064719, + "learning_rate": 1.2697841110256411e-06, + "loss": 0.6811, + "num_input_tokens_seen": 227299065, + "step": 6990 + }, + { + "epoch": 0.6304730125806015, + "flos": 21950522268000.0, + "grad_norm": 2.8844199725743884, + "learning_rate": 1.2692402624266221e-06, + "loss": 0.7946, + "num_input_tokens_seen": 227327740, + "step": 6991 + }, + { + "epoch": 0.6305631961040717, + "flos": 23079768639360.0, + "grad_norm": 2.603066224605866, + "learning_rate": 1.2686964761800529e-06, + "loss": 0.6583, + "num_input_tokens_seen": 227355865, + "step": 6992 + }, + { + "epoch": 0.6306533796275421, + "flos": 70546066561440.0, + "grad_norm": 0.6395500282123563, + "learning_rate": 1.268152752332333e-06, + "loss": 0.5833, + "num_input_tokens_seen": 227449640, + "step": 6993 + }, + { + "epoch": 0.6307435631510123, + "flos": 33863632596000.0, + "grad_norm": 5.738802746576269, + "learning_rate": 1.2676090909298549e-06, + "loss": 0.6649, + "num_input_tokens_seen": 227480610, + "step": 6994 + }, + { + "epoch": 0.6308337466744826, + "flos": 26467693602240.0, + "grad_norm": 3.0811040569043846, + "learning_rate": 1.2670654920190086e-06, + "loss": 0.795, + "num_input_tokens_seen": 227510345, + "step": 6995 + }, + { + "epoch": 0.6309239301979528, + "flos": 26065825427040.0, + "grad_norm": 1.6849540295541643, + "learning_rate": 1.2665219556461754e-06, + "loss": 0.7997, + "num_input_tokens_seen": 227541515, + "step": 6996 + }, + { + "epoch": 0.6310141137214231, + "flos": 21986880026880.0, + "grad_norm": 2.843926392838053, + "learning_rate": 1.2659784818577329e-06, + "loss": 0.848, + "num_input_tokens_seen": 227569365, + "step": 6997 + }, + { + "epoch": 0.6311042972448934, + "flos": 31712949358080.0, + "grad_norm": 1.5839935565875232, + "learning_rate": 1.2654350707000542e-06, + "loss": 0.6346, + "num_input_tokens_seen": 227601145, + "step": 6998 + }, + { + "epoch": 0.6311944807683636, + "flos": 23990236571520.0, + "grad_norm": 1.8128745251141853, + "learning_rate": 1.264891722219505e-06, + "loss": 0.7353, + "num_input_tokens_seen": 227629000, + "step": 6999 + }, + { + "epoch": 0.6312846642918338, + "flos": 23079545620800.0, + "grad_norm": 8.075643545672246, + "learning_rate": 1.2643484364624483e-06, + "loss": 0.7428, + "num_input_tokens_seen": 227657965, + "step": 7000 + }, + { + "epoch": 0.6313748478153042, + "flos": 42716000621280.0, + "grad_norm": 2.0018858363310534, + "learning_rate": 1.2638052134752393e-06, + "loss": 0.8004, + "num_input_tokens_seen": 227691325, + "step": 7001 + }, + { + "epoch": 0.6314650313387744, + "flos": 22203985826880.0, + "grad_norm": 1.8969233247486799, + "learning_rate": 1.2632620533042277e-06, + "loss": 0.8231, + "num_input_tokens_seen": 227718515, + "step": 7002 + }, + { + "epoch": 0.6315552148622446, + "flos": 21471069336480.0, + "grad_norm": 2.339135675808239, + "learning_rate": 1.2627189559957612e-06, + "loss": 0.7188, + "num_input_tokens_seen": 227745980, + "step": 7003 + }, + { + "epoch": 0.631645398385715, + "flos": 18707284945440.0, + "grad_norm": 2.955738828309971, + "learning_rate": 1.262175921596178e-06, + "loss": 0.8072, + "num_input_tokens_seen": 227771435, + "step": 7004 + }, + { + "epoch": 0.6317355819091852, + "flos": 20528109300480.0, + "grad_norm": 2.983291828854227, + "learning_rate": 1.2616329501518137e-06, + "loss": 0.6228, + "num_input_tokens_seen": 227798750, + "step": 7005 + }, + { + "epoch": 0.6318257654326555, + "flos": 27556456371360.0, + "grad_norm": 6.4075581029191735, + "learning_rate": 1.2610900417089978e-06, + "loss": 0.6714, + "num_input_tokens_seen": 227829925, + "step": 7006 + }, + { + "epoch": 0.6319159489561257, + "flos": 25703102742720.0, + "grad_norm": 1.8185827853490664, + "learning_rate": 1.2605471963140535e-06, + "loss": 0.7592, + "num_input_tokens_seen": 227860935, + "step": 7007 + }, + { + "epoch": 0.632006132479596, + "flos": 25444323908160.0, + "grad_norm": 1.9740046057062808, + "learning_rate": 1.2600044140133e-06, + "loss": 0.8232, + "num_input_tokens_seen": 227889130, + "step": 7008 + }, + { + "epoch": 0.6320963160030663, + "flos": 19326593448480.0, + "grad_norm": 1.7197030632844006, + "learning_rate": 1.2594616948530493e-06, + "loss": 0.7012, + "num_input_tokens_seen": 227917565, + "step": 7009 + }, + { + "epoch": 0.6321864995265365, + "flos": 17432756217600.0, + "grad_norm": 2.192182488411948, + "learning_rate": 1.258919038879611e-06, + "loss": 0.7719, + "num_input_tokens_seen": 227944315, + "step": 7010 + }, + { + "epoch": 0.6322766830500067, + "flos": 16266631710720.0, + "grad_norm": 2.0173852612320555, + "learning_rate": 1.2583764461392859e-06, + "loss": 0.7468, + "num_input_tokens_seen": 227972430, + "step": 7011 + }, + { + "epoch": 0.6323668665734771, + "flos": 30767238759840.0, + "grad_norm": 1.6661203760664909, + "learning_rate": 1.2578339166783724e-06, + "loss": 0.7113, + "num_input_tokens_seen": 228003960, + "step": 7012 + }, + { + "epoch": 0.6324570500969473, + "flos": 23553608937120.0, + "grad_norm": 3.7636788432002826, + "learning_rate": 1.2572914505431613e-06, + "loss": 0.6855, + "num_input_tokens_seen": 228033730, + "step": 7013 + }, + { + "epoch": 0.6325472336204175, + "flos": 35212437971520.0, + "grad_norm": 1.5633666877333452, + "learning_rate": 1.2567490477799383e-06, + "loss": 0.6477, + "num_input_tokens_seen": 228070940, + "step": 7014 + }, + { + "epoch": 0.6326374171438878, + "flos": 23915216528640.0, + "grad_norm": 2.2939865738692826, + "learning_rate": 1.2562067084349852e-06, + "loss": 0.8422, + "num_input_tokens_seen": 228100705, + "step": 7015 + }, + { + "epoch": 0.6327276006673581, + "flos": 22967870460960.0, + "grad_norm": 2.0372268774536937, + "learning_rate": 1.2556644325545764e-06, + "loss": 0.7315, + "num_input_tokens_seen": 228130795, + "step": 7016 + }, + { + "epoch": 0.6328177841908283, + "flos": 21695497579200.0, + "grad_norm": 1.8404040607172958, + "learning_rate": 1.255122220184983e-06, + "loss": 0.7303, + "num_input_tokens_seen": 228158830, + "step": 7017 + }, + { + "epoch": 0.6329079677142986, + "flos": 22350940822560.0, + "grad_norm": 2.2625281906479224, + "learning_rate": 1.2545800713724694e-06, + "loss": 0.6564, + "num_input_tokens_seen": 228185845, + "step": 7018 + }, + { + "epoch": 0.6329981512377688, + "flos": 36376741160160.0, + "grad_norm": 1.8490602993496734, + "learning_rate": 1.254037986163294e-06, + "loss": 0.6126, + "num_input_tokens_seen": 228217215, + "step": 7019 + }, + { + "epoch": 0.6330883347612392, + "flos": 53500830991680.0, + "grad_norm": 1.8059745921801054, + "learning_rate": 1.2534959646037104e-06, + "loss": 0.7042, + "num_input_tokens_seen": 228250005, + "step": 7020 + }, + { + "epoch": 0.6331785182847094, + "flos": 25739088804000.0, + "grad_norm": 1.9484519902893487, + "learning_rate": 1.2529540067399675e-06, + "loss": 0.7524, + "num_input_tokens_seen": 228279680, + "step": 7021 + }, + { + "epoch": 0.6332687018081796, + "flos": 19436633138880.0, + "grad_norm": 1.9836833679700403, + "learning_rate": 1.252412112618308e-06, + "loss": 0.68, + "num_input_tokens_seen": 228307030, + "step": 7022 + }, + { + "epoch": 0.6333588853316499, + "flos": 25994299341600.0, + "grad_norm": 1.9919464563887326, + "learning_rate": 1.2518702822849696e-06, + "loss": 0.6306, + "num_input_tokens_seen": 228334980, + "step": 7023 + }, + { + "epoch": 0.6334490688551202, + "flos": 17578633290240.0, + "grad_norm": 2.483500488680566, + "learning_rate": 1.2513285157861831e-06, + "loss": 0.6906, + "num_input_tokens_seen": 228362090, + "step": 7024 + }, + { + "epoch": 0.6335392523785904, + "flos": 25409564448960.0, + "grad_norm": 1.7044568934698165, + "learning_rate": 1.2507868131681764e-06, + "loss": 0.7418, + "num_input_tokens_seen": 228391845, + "step": 7025 + }, + { + "epoch": 0.6336294359020607, + "flos": 24098194755360.0, + "grad_norm": 1.9182830988020663, + "learning_rate": 1.250245174477169e-06, + "loss": 0.7312, + "num_input_tokens_seen": 228421115, + "step": 7026 + }, + { + "epoch": 0.6337196194255309, + "flos": 70621198113600.0, + "grad_norm": 0.6475089675326529, + "learning_rate": 1.2497035997593783e-06, + "loss": 0.5193, + "num_input_tokens_seen": 228507195, + "step": 7027 + }, + { + "epoch": 0.6338098029490012, + "flos": 30438532139520.0, + "grad_norm": 1.6365196817102072, + "learning_rate": 1.2491620890610135e-06, + "loss": 0.7452, + "num_input_tokens_seen": 228537300, + "step": 7028 + }, + { + "epoch": 0.6338999864724715, + "flos": 23366765055360.0, + "grad_norm": 1.720001320062175, + "learning_rate": 1.2486206424282788e-06, + "loss": 0.7309, + "num_input_tokens_seen": 228564240, + "step": 7029 + }, + { + "epoch": 0.6339901699959417, + "flos": 21840705596160.0, + "grad_norm": 1.8764215802365267, + "learning_rate": 1.2480792599073743e-06, + "loss": 0.8031, + "num_input_tokens_seen": 228592225, + "step": 7030 + }, + { + "epoch": 0.634080353519412, + "flos": 19980624240960.0, + "grad_norm": 1.73463032535649, + "learning_rate": 1.247537941544493e-06, + "loss": 0.7895, + "num_input_tokens_seen": 228621715, + "step": 7031 + }, + { + "epoch": 0.6341705370428823, + "flos": 22235288498400.0, + "grad_norm": 14.620170643615536, + "learning_rate": 1.2469966873858242e-06, + "loss": 0.7798, + "num_input_tokens_seen": 228650240, + "step": 7032 + }, + { + "epoch": 0.6342607205663525, + "flos": 23733873771360.0, + "grad_norm": 2.04175643983033, + "learning_rate": 1.2464554974775496e-06, + "loss": 0.6695, + "num_input_tokens_seen": 228675155, + "step": 7033 + }, + { + "epoch": 0.6343509040898228, + "flos": 17396101100640.0, + "grad_norm": 2.7092110800957667, + "learning_rate": 1.2459143718658474e-06, + "loss": 0.7464, + "num_input_tokens_seen": 228701805, + "step": 7034 + }, + { + "epoch": 0.6344410876132931, + "flos": 23698445256480.0, + "grad_norm": 2.266327847607622, + "learning_rate": 1.2453733105968886e-06, + "loss": 0.6825, + "num_input_tokens_seen": 228730775, + "step": 7035 + }, + { + "epoch": 0.6345312711367633, + "flos": 67556441476800.0, + "grad_norm": 0.627844441324307, + "learning_rate": 1.2448323137168399e-06, + "loss": 0.5845, + "num_input_tokens_seen": 228819535, + "step": 7036 + }, + { + "epoch": 0.6346214546602336, + "flos": 41113136970720.0, + "grad_norm": 1.7378370767662028, + "learning_rate": 1.2442913812718625e-06, + "loss": 0.6686, + "num_input_tokens_seen": 228852595, + "step": 7037 + }, + { + "epoch": 0.6347116381837038, + "flos": 34769974684800.0, + "grad_norm": 1.592318621356588, + "learning_rate": 1.2437505133081108e-06, + "loss": 0.7307, + "num_input_tokens_seen": 228882965, + "step": 7038 + }, + { + "epoch": 0.6348018217071741, + "flos": 20018543129760.0, + "grad_norm": 6.32139670523308, + "learning_rate": 1.2432097098717358e-06, + "loss": 0.7607, + "num_input_tokens_seen": 228908605, + "step": 7039 + }, + { + "epoch": 0.6348920052306444, + "flos": 21798735203520.0, + "grad_norm": 2.000266534211314, + "learning_rate": 1.2426689710088813e-06, + "loss": 0.7502, + "num_input_tokens_seen": 228936750, + "step": 7040 + }, + { + "epoch": 0.6349821887541146, + "flos": 17393610726720.0, + "grad_norm": 5.116971496216372, + "learning_rate": 1.2421282967656855e-06, + "loss": 0.8019, + "num_input_tokens_seen": 228963145, + "step": 7041 + }, + { + "epoch": 0.6350723722775848, + "flos": 21293480724960.0, + "grad_norm": 1.7956277250168993, + "learning_rate": 1.2415876871882827e-06, + "loss": 0.7738, + "num_input_tokens_seen": 228991915, + "step": 7042 + }, + { + "epoch": 0.6351625558010552, + "flos": 69728385817440.0, + "grad_norm": 0.674115658654854, + "learning_rate": 1.2410471423227998e-06, + "loss": 0.6487, + "num_input_tokens_seen": 229095590, + "step": 7043 + }, + { + "epoch": 0.6352527393245254, + "flos": 24937136602080.0, + "grad_norm": 1.7779567457887242, + "learning_rate": 1.24050666221536e-06, + "loss": 0.7386, + "num_input_tokens_seen": 229126765, + "step": 7044 + }, + { + "epoch": 0.6353429228479957, + "flos": 20565805170720.0, + "grad_norm": 1.982751016587942, + "learning_rate": 1.23996624691208e-06, + "loss": 0.8113, + "num_input_tokens_seen": 229154925, + "step": 7045 + }, + { + "epoch": 0.6354331063714659, + "flos": 27485487832320.0, + "grad_norm": 1.5776218329777842, + "learning_rate": 1.2394258964590693e-06, + "loss": 0.73, + "num_input_tokens_seen": 229186020, + "step": 7046 + }, + { + "epoch": 0.6355232898949362, + "flos": 24864160896000.0, + "grad_norm": 1.912453422727192, + "learning_rate": 1.238885610902436e-06, + "loss": 0.6951, + "num_input_tokens_seen": 229213365, + "step": 7047 + }, + { + "epoch": 0.6356134734184065, + "flos": 28723881819840.0, + "grad_norm": 2.624365775375343, + "learning_rate": 1.2383453902882787e-06, + "loss": 0.7576, + "num_input_tokens_seen": 229242345, + "step": 7048 + }, + { + "epoch": 0.6357036569418767, + "flos": 37282079665440.0, + "grad_norm": 1.5573757068152645, + "learning_rate": 1.2378052346626927e-06, + "loss": 0.8098, + "num_input_tokens_seen": 229276435, + "step": 7049 + }, + { + "epoch": 0.6357938404653469, + "flos": 14845296666240.0, + "grad_norm": 2.5456152298606587, + "learning_rate": 1.2372651440717665e-06, + "loss": 0.6739, + "num_input_tokens_seen": 229302800, + "step": 7050 + }, + { + "epoch": 0.6358840239888173, + "flos": 22023163446240.0, + "grad_norm": 1.8656941591156564, + "learning_rate": 1.236725118561584e-06, + "loss": 0.7795, + "num_input_tokens_seen": 229328955, + "step": 7051 + }, + { + "epoch": 0.6359742075122875, + "flos": 24498018593760.0, + "grad_norm": 1.933149128666788, + "learning_rate": 1.2361851581782232e-06, + "loss": 0.7454, + "num_input_tokens_seen": 229359215, + "step": 7052 + }, + { + "epoch": 0.6360643910357577, + "flos": 22095767454720.0, + "grad_norm": 1.7877297580550853, + "learning_rate": 1.2356452629677554e-06, + "loss": 0.8098, + "num_input_tokens_seen": 229386675, + "step": 7053 + }, + { + "epoch": 0.6361545745592281, + "flos": 34227953580000.0, + "grad_norm": 1.9366616311831293, + "learning_rate": 1.2351054329762494e-06, + "loss": 0.6778, + "num_input_tokens_seen": 229419910, + "step": 7054 + }, + { + "epoch": 0.6362447580826983, + "flos": 15356238118080.0, + "grad_norm": 1.848566262668998, + "learning_rate": 1.2345656682497648e-06, + "loss": 0.7687, + "num_input_tokens_seen": 229444965, + "step": 7055 + }, + { + "epoch": 0.6363349416061685, + "flos": 23189027764800.0, + "grad_norm": 1.9691978429628685, + "learning_rate": 1.2340259688343583e-06, + "loss": 0.7194, + "num_input_tokens_seen": 229476660, + "step": 7056 + }, + { + "epoch": 0.6364251251296388, + "flos": 24208383124800.0, + "grad_norm": 1.6460808274133465, + "learning_rate": 1.2334863347760803e-06, + "loss": 0.6468, + "num_input_tokens_seen": 229509835, + "step": 7057 + }, + { + "epoch": 0.6365153086531091, + "flos": 54270991581600.0, + "grad_norm": 0.6408276586277206, + "learning_rate": 1.2329467661209734e-06, + "loss": 0.5439, + "num_input_tokens_seen": 229600380, + "step": 7058 + }, + { + "epoch": 0.6366054921765794, + "flos": 27160684036800.0, + "grad_norm": 1.9201560047468298, + "learning_rate": 1.2324072629150788e-06, + "loss": 0.7889, + "num_input_tokens_seen": 229627965, + "step": 7059 + }, + { + "epoch": 0.6366956757000496, + "flos": 18196306323840.0, + "grad_norm": 3.053593198081984, + "learning_rate": 1.2318678252044287e-06, + "loss": 0.7971, + "num_input_tokens_seen": 229653010, + "step": 7060 + }, + { + "epoch": 0.6367858592235198, + "flos": 20491974560160.0, + "grad_norm": 2.616526141513754, + "learning_rate": 1.2313284530350512e-06, + "loss": 0.7994, + "num_input_tokens_seen": 229679875, + "step": 7061 + }, + { + "epoch": 0.6368760427469902, + "flos": 20378515251840.0, + "grad_norm": 1.9754004398439309, + "learning_rate": 1.230789146452969e-06, + "loss": 0.6822, + "num_input_tokens_seen": 229707335, + "step": 7062 + }, + { + "epoch": 0.6369662262704604, + "flos": 24463370643840.0, + "grad_norm": 2.0850778015404035, + "learning_rate": 1.2302499055041974e-06, + "loss": 0.7514, + "num_input_tokens_seen": 229735585, + "step": 7063 + }, + { + "epoch": 0.6370564097939306, + "flos": 64696110018240.0, + "grad_norm": 0.6621466207015198, + "learning_rate": 1.2297107302347488e-06, + "loss": 0.5814, + "num_input_tokens_seen": 229825195, + "step": 7064 + }, + { + "epoch": 0.6371465933174009, + "flos": 24569433169920.0, + "grad_norm": 1.8787900131340247, + "learning_rate": 1.2291716206906275e-06, + "loss": 0.8142, + "num_input_tokens_seen": 229853755, + "step": 7065 + }, + { + "epoch": 0.6372367768408712, + "flos": 31093826703840.0, + "grad_norm": 2.462204476840359, + "learning_rate": 1.2286325769178345e-06, + "loss": 0.812, + "num_input_tokens_seen": 229879935, + "step": 7066 + }, + { + "epoch": 0.6373269603643414, + "flos": 22678606689600.0, + "grad_norm": 1.7873914485595792, + "learning_rate": 1.2280935989623633e-06, + "loss": 0.7732, + "num_input_tokens_seen": 229907050, + "step": 7067 + }, + { + "epoch": 0.6374171438878117, + "flos": 24427273073280.0, + "grad_norm": 2.301151783353131, + "learning_rate": 1.2275546868702017e-06, + "loss": 0.6807, + "num_input_tokens_seen": 229934645, + "step": 7068 + }, + { + "epoch": 0.6375073274112819, + "flos": 17542052512800.0, + "grad_norm": 1.7426319325222104, + "learning_rate": 1.2270158406873341e-06, + "loss": 0.6835, + "num_input_tokens_seen": 229962705, + "step": 7069 + }, + { + "epoch": 0.6375975109347523, + "flos": 25810540549920.0, + "grad_norm": 1.895268688025735, + "learning_rate": 1.2264770604597363e-06, + "loss": 0.7871, + "num_input_tokens_seen": 229994060, + "step": 7070 + }, + { + "epoch": 0.6376876944582225, + "flos": 22019000433120.0, + "grad_norm": 1.7238461629864739, + "learning_rate": 1.2259383462333819e-06, + "loss": 0.845, + "num_input_tokens_seen": 230021920, + "step": 7071 + }, + { + "epoch": 0.6377778779816927, + "flos": 23509111000800.0, + "grad_norm": 1.8808994308753946, + "learning_rate": 1.2253996980542359e-06, + "loss": 0.7022, + "num_input_tokens_seen": 230049790, + "step": 7072 + }, + { + "epoch": 0.637868061505163, + "flos": 23115531682080.0, + "grad_norm": 1.7579078478914834, + "learning_rate": 1.2248611159682578e-06, + "loss": 0.7284, + "num_input_tokens_seen": 230076705, + "step": 7073 + }, + { + "epoch": 0.6379582450286333, + "flos": 63608016304800.0, + "grad_norm": 0.6889035679571833, + "learning_rate": 1.2243226000214044e-06, + "loss": 0.5624, + "num_input_tokens_seen": 230165935, + "step": 7074 + }, + { + "epoch": 0.6380484285521035, + "flos": 25666968002400.0, + "grad_norm": 2.097829346290114, + "learning_rate": 1.2237841502596232e-06, + "loss": 0.7499, + "num_input_tokens_seen": 230195730, + "step": 7075 + }, + { + "epoch": 0.6381386120755738, + "flos": 27487792357440.0, + "grad_norm": 2.0523787026102718, + "learning_rate": 1.2232457667288583e-06, + "loss": 0.7528, + "num_input_tokens_seen": 230226665, + "step": 7076 + }, + { + "epoch": 0.6382287955990441, + "flos": 23516470613280.0, + "grad_norm": 2.9668396274281377, + "learning_rate": 1.2227074494750476e-06, + "loss": 0.7499, + "num_input_tokens_seen": 230253840, + "step": 7077 + }, + { + "epoch": 0.6383189791225143, + "flos": 25006358162400.0, + "grad_norm": 2.2818259988255054, + "learning_rate": 1.2221691985441238e-06, + "loss": 0.7796, + "num_input_tokens_seen": 230282550, + "step": 7078 + }, + { + "epoch": 0.6384091626459846, + "flos": 23370816559200.0, + "grad_norm": 2.194738672814935, + "learning_rate": 1.2216310139820128e-06, + "loss": 0.6935, + "num_input_tokens_seen": 230311240, + "step": 7079 + }, + { + "epoch": 0.6384993461694548, + "flos": 26429254336800.0, + "grad_norm": 1.682680594854147, + "learning_rate": 1.2210928958346347e-06, + "loss": 0.7897, + "num_input_tokens_seen": 230343450, + "step": 7080 + }, + { + "epoch": 0.6385895296929252, + "flos": 25555924728480.0, + "grad_norm": 1.550822990105581, + "learning_rate": 1.2205548441479065e-06, + "loss": 0.6753, + "num_input_tokens_seen": 230374280, + "step": 7081 + }, + { + "epoch": 0.6386797132163954, + "flos": 24390134749440.0, + "grad_norm": 2.0856595086297043, + "learning_rate": 1.2200168589677357e-06, + "loss": 0.6835, + "num_input_tokens_seen": 230403590, + "step": 7082 + }, + { + "epoch": 0.6387698967398656, + "flos": 24861819201120.0, + "grad_norm": 1.6377204979522122, + "learning_rate": 1.2194789403400284e-06, + "loss": 0.7155, + "num_input_tokens_seen": 230432565, + "step": 7083 + }, + { + "epoch": 0.6388600802633358, + "flos": 29855655734880.0, + "grad_norm": 3.030979105062776, + "learning_rate": 1.2189410883106816e-06, + "loss": 0.7266, + "num_input_tokens_seen": 230461275, + "step": 7084 + }, + { + "epoch": 0.6389502637868062, + "flos": 20929271250240.0, + "grad_norm": 1.968641611173402, + "learning_rate": 1.2184033029255872e-06, + "loss": 0.8119, + "num_input_tokens_seen": 230492695, + "step": 7085 + }, + { + "epoch": 0.6390404473102764, + "flos": 69316593316320.0, + "grad_norm": 0.644458720432296, + "learning_rate": 1.2178655842306334e-06, + "loss": 0.58, + "num_input_tokens_seen": 230589815, + "step": 7086 + }, + { + "epoch": 0.6391306308337467, + "flos": 45156691025760.0, + "grad_norm": 2.1734035245213006, + "learning_rate": 1.2173279322716999e-06, + "loss": 0.7465, + "num_input_tokens_seen": 230620740, + "step": 7087 + }, + { + "epoch": 0.6392208143572169, + "flos": 20820123634080.0, + "grad_norm": 2.009389888536222, + "learning_rate": 1.216790347094663e-06, + "loss": 0.7484, + "num_input_tokens_seen": 230648515, + "step": 7088 + }, + { + "epoch": 0.6393109978806872, + "flos": 28252754914560.0, + "grad_norm": 1.8759348128447446, + "learning_rate": 1.2162528287453927e-06, + "loss": 0.7871, + "num_input_tokens_seen": 230677825, + "step": 7089 + }, + { + "epoch": 0.6394011814041575, + "flos": 21440101192800.0, + "grad_norm": 2.452752832257307, + "learning_rate": 1.215715377269752e-06, + "loss": 0.7511, + "num_input_tokens_seen": 230702650, + "step": 7090 + }, + { + "epoch": 0.6394913649276277, + "flos": 26136310759200.0, + "grad_norm": 2.1748964438504625, + "learning_rate": 1.2151779927136003e-06, + "loss": 0.7286, + "num_input_tokens_seen": 230732775, + "step": 7091 + }, + { + "epoch": 0.6395815484510979, + "flos": 22165583731200.0, + "grad_norm": 1.8295441686193856, + "learning_rate": 1.2146406751227893e-06, + "loss": 0.8182, + "num_input_tokens_seen": 230759430, + "step": 7092 + }, + { + "epoch": 0.6396717319745683, + "flos": 22605556644000.0, + "grad_norm": 1.8396329009771668, + "learning_rate": 1.214103424543167e-06, + "loss": 0.7745, + "num_input_tokens_seen": 230787025, + "step": 7093 + }, + { + "epoch": 0.6397619154980385, + "flos": 26468028130080.0, + "grad_norm": 3.3970477189246076, + "learning_rate": 1.2135662410205735e-06, + "loss": 0.6607, + "num_input_tokens_seen": 230819110, + "step": 7094 + }, + { + "epoch": 0.6398520990215087, + "flos": 15242146923840.0, + "grad_norm": 2.305632109947751, + "learning_rate": 1.2130291246008444e-06, + "loss": 0.6645, + "num_input_tokens_seen": 230843075, + "step": 7095 + }, + { + "epoch": 0.639942282544979, + "flos": 33497750482080.0, + "grad_norm": 2.631155245200656, + "learning_rate": 1.21249207532981e-06, + "loss": 0.6644, + "num_input_tokens_seen": 230875390, + "step": 7096 + }, + { + "epoch": 0.6400324660684493, + "flos": 21907882819680.0, + "grad_norm": 1.985267138245877, + "learning_rate": 1.2119550932532936e-06, + "loss": 0.6674, + "num_input_tokens_seen": 230904030, + "step": 7097 + }, + { + "epoch": 0.6401226495919196, + "flos": 22315103440320.0, + "grad_norm": 1.8687028681192248, + "learning_rate": 1.2114181784171144e-06, + "loss": 0.7219, + "num_input_tokens_seen": 230932695, + "step": 7098 + }, + { + "epoch": 0.6402128331153898, + "flos": 23697404503200.0, + "grad_norm": 1.927427999930418, + "learning_rate": 1.2108813308670837e-06, + "loss": 0.6491, + "num_input_tokens_seen": 230964590, + "step": 7099 + }, + { + "epoch": 0.6403030166388601, + "flos": 21986731347840.0, + "grad_norm": 2.3921358408168234, + "learning_rate": 1.2103445506490099e-06, + "loss": 0.7491, + "num_input_tokens_seen": 230992845, + "step": 7100 + }, + { + "epoch": 0.6403932001623304, + "flos": 71050726323840.0, + "grad_norm": 0.7203868802739352, + "learning_rate": 1.209807837808693e-06, + "loss": 0.5919, + "num_input_tokens_seen": 231078160, + "step": 7101 + }, + { + "epoch": 0.6404833836858006, + "flos": 33571432413600.0, + "grad_norm": 2.3298787929459626, + "learning_rate": 1.2092711923919282e-06, + "loss": 0.7853, + "num_input_tokens_seen": 231104890, + "step": 7102 + }, + { + "epoch": 0.6405735672092708, + "flos": 19690356886080.0, + "grad_norm": 1.60279188654945, + "learning_rate": 1.2087346144445053e-06, + "loss": 0.7062, + "num_input_tokens_seen": 231132885, + "step": 7103 + }, + { + "epoch": 0.6406637507327412, + "flos": 23880865936800.0, + "grad_norm": 2.1398101146626876, + "learning_rate": 1.2081981040122081e-06, + "loss": 0.755, + "num_input_tokens_seen": 231158950, + "step": 7104 + }, + { + "epoch": 0.6407539342562114, + "flos": 38637241069920.0, + "grad_norm": 1.692268697916281, + "learning_rate": 1.2076616611408148e-06, + "loss": 0.6959, + "num_input_tokens_seen": 231194430, + "step": 7105 + }, + { + "epoch": 0.6408441177796816, + "flos": 28616815710240.0, + "grad_norm": 2.0181712546294714, + "learning_rate": 1.2071252858760972e-06, + "loss": 0.7771, + "num_input_tokens_seen": 231224475, + "step": 7106 + }, + { + "epoch": 0.6409343013031519, + "flos": 20711793752640.0, + "grad_norm": 2.3615846707528845, + "learning_rate": 1.2065889782638218e-06, + "loss": 0.7311, + "num_input_tokens_seen": 231252115, + "step": 7107 + }, + { + "epoch": 0.6410244848266222, + "flos": 20274348383520.0, + "grad_norm": 1.7271438931598733, + "learning_rate": 1.2060527383497506e-06, + "loss": 0.7236, + "num_input_tokens_seen": 231280115, + "step": 7108 + }, + { + "epoch": 0.6411146683500925, + "flos": 21727655155200.0, + "grad_norm": 1.8462706969798577, + "learning_rate": 1.2055165661796363e-06, + "loss": 0.8578, + "num_input_tokens_seen": 231309250, + "step": 7109 + }, + { + "epoch": 0.6412048518735627, + "flos": 22897608147360.0, + "grad_norm": 2.136145589831609, + "learning_rate": 1.2049804617992303e-06, + "loss": 0.816, + "num_input_tokens_seen": 231335865, + "step": 7110 + }, + { + "epoch": 0.6412950353970329, + "flos": 54007380678240.0, + "grad_norm": 0.6189331525364683, + "learning_rate": 1.204444425254275e-06, + "loss": 0.5919, + "num_input_tokens_seen": 231422960, + "step": 7111 + }, + { + "epoch": 0.6413852189205033, + "flos": 21112026458400.0, + "grad_norm": 1.9002989759624203, + "learning_rate": 1.203908456590507e-06, + "loss": 0.7298, + "num_input_tokens_seen": 231453345, + "step": 7112 + }, + { + "epoch": 0.6414754024439735, + "flos": 20602237269120.0, + "grad_norm": 1.872285976733059, + "learning_rate": 1.20337255585366e-06, + "loss": 0.7208, + "num_input_tokens_seen": 231479435, + "step": 7113 + }, + { + "epoch": 0.6415655859674437, + "flos": 58975192646400.0, + "grad_norm": 0.7901774168583376, + "learning_rate": 1.2028367230894582e-06, + "loss": 0.5378, + "num_input_tokens_seen": 231565625, + "step": 7114 + }, + { + "epoch": 0.641655769490914, + "flos": 21839962200960.0, + "grad_norm": 2.0448937764828736, + "learning_rate": 1.2023009583436237e-06, + "loss": 0.6666, + "num_input_tokens_seen": 231594015, + "step": 7115 + }, + { + "epoch": 0.6417459530143843, + "flos": 23037129191040.0, + "grad_norm": 2.0268774065867396, + "learning_rate": 1.2017652616618698e-06, + "loss": 0.81, + "num_input_tokens_seen": 231620415, + "step": 7116 + }, + { + "epoch": 0.6418361365378545, + "flos": 22454735993280.0, + "grad_norm": 2.474972476228202, + "learning_rate": 1.2012296330899048e-06, + "loss": 0.7404, + "num_input_tokens_seen": 231648930, + "step": 7117 + }, + { + "epoch": 0.6419263200613248, + "flos": 27159568944000.0, + "grad_norm": 2.3203726096025608, + "learning_rate": 1.200694072673432e-06, + "loss": 0.777, + "num_input_tokens_seen": 231678530, + "step": 7118 + }, + { + "epoch": 0.642016503584795, + "flos": 18924613764000.0, + "grad_norm": 1.8024470967268424, + "learning_rate": 1.200158580458148e-06, + "loss": 0.7762, + "num_input_tokens_seen": 231706385, + "step": 7119 + }, + { + "epoch": 0.6421066871082654, + "flos": 27595453183200.0, + "grad_norm": 2.010348132716216, + "learning_rate": 1.1996231564897448e-06, + "loss": 0.7201, + "num_input_tokens_seen": 231736495, + "step": 7120 + }, + { + "epoch": 0.6421968706317356, + "flos": 21291213369600.0, + "grad_norm": 1.9227583181111771, + "learning_rate": 1.1990878008139062e-06, + "loss": 0.7008, + "num_input_tokens_seen": 231762715, + "step": 7121 + }, + { + "epoch": 0.6422870541552058, + "flos": 48721684223520.0, + "grad_norm": 2.5608287106228222, + "learning_rate": 1.1985525134763132e-06, + "loss": 0.7738, + "num_input_tokens_seen": 231793020, + "step": 7122 + }, + { + "epoch": 0.6423772376786762, + "flos": 29558697823200.0, + "grad_norm": 1.832728141540883, + "learning_rate": 1.1980172945226393e-06, + "loss": 0.7913, + "num_input_tokens_seen": 231823400, + "step": 7123 + }, + { + "epoch": 0.6424674212021464, + "flos": 18561742400640.0, + "grad_norm": 2.139647991784527, + "learning_rate": 1.197482143998551e-06, + "loss": 0.7503, + "num_input_tokens_seen": 231849075, + "step": 7124 + }, + { + "epoch": 0.6425576047256166, + "flos": 20274088195200.0, + "grad_norm": 1.9986047646418297, + "learning_rate": 1.196947061949712e-06, + "loss": 0.7303, + "num_input_tokens_seen": 231876765, + "step": 7125 + }, + { + "epoch": 0.6426477882490869, + "flos": 18708177019680.0, + "grad_norm": 1.7848554562299406, + "learning_rate": 1.1964120484217768e-06, + "loss": 0.8028, + "num_input_tokens_seen": 231903930, + "step": 7126 + }, + { + "epoch": 0.6427379717725572, + "flos": 62195161699200.0, + "grad_norm": 0.6986107127144472, + "learning_rate": 1.1958771034603975e-06, + "loss": 0.5622, + "num_input_tokens_seen": 231988655, + "step": 7127 + }, + { + "epoch": 0.6428281552960274, + "flos": 22749538058880.0, + "grad_norm": 2.1756613679523955, + "learning_rate": 1.1953422271112175e-06, + "loss": 0.6865, + "num_input_tokens_seen": 232018685, + "step": 7128 + }, + { + "epoch": 0.6429183388194977, + "flos": 22205063749920.0, + "grad_norm": 2.1001575410877384, + "learning_rate": 1.1948074194198748e-06, + "loss": 0.7325, + "num_input_tokens_seen": 232045595, + "step": 7129 + }, + { + "epoch": 0.6430085223429679, + "flos": 38777579848320.0, + "grad_norm": 2.4305150625095275, + "learning_rate": 1.1942726804320033e-06, + "loss": 0.6574, + "num_input_tokens_seen": 232078010, + "step": 7130 + }, + { + "epoch": 0.6430987058664382, + "flos": 17395989591360.0, + "grad_norm": 2.234951025065681, + "learning_rate": 1.1937380101932295e-06, + "loss": 0.688, + "num_input_tokens_seen": 232104095, + "step": 7131 + }, + { + "epoch": 0.6431888893899085, + "flos": 21076635113280.0, + "grad_norm": 1.5490589330654083, + "learning_rate": 1.1932034087491745e-06, + "loss": 0.7424, + "num_input_tokens_seen": 232134255, + "step": 7132 + }, + { + "epoch": 0.6432790729133787, + "flos": 32693345076000.0, + "grad_norm": 1.9322004133888389, + "learning_rate": 1.1926688761454531e-06, + "loss": 0.6602, + "num_input_tokens_seen": 232163385, + "step": 7133 + }, + { + "epoch": 0.643369256436849, + "flos": 28945745349120.0, + "grad_norm": 1.6857405779984134, + "learning_rate": 1.1921344124276746e-06, + "loss": 0.7649, + "num_input_tokens_seen": 232195270, + "step": 7134 + }, + { + "epoch": 0.6434594399603193, + "flos": 17752876623360.0, + "grad_norm": 1.8154740104989775, + "learning_rate": 1.1916000176414431e-06, + "loss": 0.7668, + "num_input_tokens_seen": 232221345, + "step": 7135 + }, + { + "epoch": 0.6435496234837895, + "flos": 25625629495680.0, + "grad_norm": 2.0468360955421323, + "learning_rate": 1.1910656918323546e-06, + "loss": 0.7825, + "num_input_tokens_seen": 232248095, + "step": 7136 + }, + { + "epoch": 0.6436398070072598, + "flos": 30220088228160.0, + "grad_norm": 2.1436948274235674, + "learning_rate": 1.1905314350460024e-06, + "loss": 0.7715, + "num_input_tokens_seen": 232279340, + "step": 7137 + }, + { + "epoch": 0.64372999053073, + "flos": 24354520385760.0, + "grad_norm": 1.7077794301285558, + "learning_rate": 1.1899972473279717e-06, + "loss": 0.8008, + "num_input_tokens_seen": 232308490, + "step": 7138 + }, + { + "epoch": 0.6438201740542003, + "flos": 24791854245600.0, + "grad_norm": 1.9699470876955838, + "learning_rate": 1.1894631287238414e-06, + "loss": 0.7359, + "num_input_tokens_seen": 232337615, + "step": 7139 + }, + { + "epoch": 0.6439103575776706, + "flos": 25514846410080.0, + "grad_norm": 1.6555804932755023, + "learning_rate": 1.188929079279187e-06, + "loss": 0.7569, + "num_input_tokens_seen": 232367800, + "step": 7140 + }, + { + "epoch": 0.6440005411011408, + "flos": 21257643342720.0, + "grad_norm": 1.806570284737095, + "learning_rate": 1.1883950990395751e-06, + "loss": 0.6835, + "num_input_tokens_seen": 232395460, + "step": 7141 + }, + { + "epoch": 0.644090724624611, + "flos": 20674729768320.0, + "grad_norm": 2.048386576213775, + "learning_rate": 1.187861188050569e-06, + "loss": 0.6815, + "num_input_tokens_seen": 232424405, + "step": 7142 + }, + { + "epoch": 0.6441809081480814, + "flos": 23079173923200.0, + "grad_norm": 1.7449387521570197, + "learning_rate": 1.187327346357724e-06, + "loss": 0.7326, + "num_input_tokens_seen": 232455705, + "step": 7143 + }, + { + "epoch": 0.6442710916715516, + "flos": 19435964083200.0, + "grad_norm": 2.1102858095812795, + "learning_rate": 1.1867935740065912e-06, + "loss": 0.7695, + "num_input_tokens_seen": 232482935, + "step": 7144 + }, + { + "epoch": 0.6443612751950218, + "flos": 28433056918560.0, + "grad_norm": 2.6091422702457328, + "learning_rate": 1.1862598710427148e-06, + "loss": 0.6347, + "num_input_tokens_seen": 232516005, + "step": 7145 + }, + { + "epoch": 0.6444514587184921, + "flos": 15100247015520.0, + "grad_norm": 3.7016382250723, + "learning_rate": 1.1857262375116328e-06, + "loss": 0.7075, + "num_input_tokens_seen": 232542140, + "step": 7146 + }, + { + "epoch": 0.6445416422419624, + "flos": 28143570128640.0, + "grad_norm": 2.188501837064004, + "learning_rate": 1.1851926734588783e-06, + "loss": 0.7468, + "num_input_tokens_seen": 232572710, + "step": 7147 + }, + { + "epoch": 0.6446318257654327, + "flos": 19873037754720.0, + "grad_norm": 2.0961129817869484, + "learning_rate": 1.184659178929977e-06, + "loss": 0.7671, + "num_input_tokens_seen": 232598580, + "step": 7148 + }, + { + "epoch": 0.6447220092889029, + "flos": 19069412913600.0, + "grad_norm": 1.7808069046099153, + "learning_rate": 1.1841257539704513e-06, + "loss": 0.7414, + "num_input_tokens_seen": 232628715, + "step": 7149 + }, + { + "epoch": 0.6448121928123732, + "flos": 23844471008160.0, + "grad_norm": 2.2108156420012683, + "learning_rate": 1.1835923986258146e-06, + "loss": 0.7642, + "num_input_tokens_seen": 232658020, + "step": 7150 + }, + { + "epoch": 0.6449023763358435, + "flos": 21147863840640.0, + "grad_norm": 1.6167718472954913, + "learning_rate": 1.1830591129415754e-06, + "loss": 0.7062, + "num_input_tokens_seen": 232687020, + "step": 7151 + }, + { + "epoch": 0.6449925598593137, + "flos": 25045801011360.0, + "grad_norm": 2.561870311501041, + "learning_rate": 1.182525896963238e-06, + "loss": 0.6908, + "num_input_tokens_seen": 232716335, + "step": 7152 + }, + { + "epoch": 0.6450827433827839, + "flos": 19180902224640.0, + "grad_norm": 4.294599450721647, + "learning_rate": 1.181992750736298e-06, + "loss": 0.6757, + "num_input_tokens_seen": 232743345, + "step": 7153 + }, + { + "epoch": 0.6451729269062543, + "flos": 21476458951680.0, + "grad_norm": 2.327897335928351, + "learning_rate": 1.1814596743062474e-06, + "loss": 0.7512, + "num_input_tokens_seen": 232769465, + "step": 7154 + }, + { + "epoch": 0.6452631104297245, + "flos": 21543338817120.0, + "grad_norm": 2.569296128629033, + "learning_rate": 1.1809266677185711e-06, + "loss": 0.7689, + "num_input_tokens_seen": 232797435, + "step": 7155 + }, + { + "epoch": 0.6453532939531947, + "flos": 67308813570240.0, + "grad_norm": 0.651454982531398, + "learning_rate": 1.180393731018747e-06, + "loss": 0.6119, + "num_input_tokens_seen": 232896200, + "step": 7156 + }, + { + "epoch": 0.645443477476665, + "flos": 53969536128960.0, + "grad_norm": 0.6502504398149398, + "learning_rate": 1.1798608642522498e-06, + "loss": 0.5509, + "num_input_tokens_seen": 232990835, + "step": 7157 + }, + { + "epoch": 0.6455336610001353, + "flos": 25447520507520.0, + "grad_norm": 3.016863158625378, + "learning_rate": 1.1793280674645454e-06, + "loss": 0.7272, + "num_input_tokens_seen": 233019345, + "step": 7158 + }, + { + "epoch": 0.6456238445236056, + "flos": 17942248048800.0, + "grad_norm": 3.133055717939575, + "learning_rate": 1.1787953407010954e-06, + "loss": 0.8186, + "num_input_tokens_seen": 233046285, + "step": 7159 + }, + { + "epoch": 0.6457140280470758, + "flos": 25702768214880.0, + "grad_norm": 1.9922447046780853, + "learning_rate": 1.1782626840073554e-06, + "loss": 0.7494, + "num_input_tokens_seen": 233075925, + "step": 7160 + }, + { + "epoch": 0.645804211570546, + "flos": 30402025701600.0, + "grad_norm": 2.2959742434183426, + "learning_rate": 1.1777300974287738e-06, + "loss": 0.6504, + "num_input_tokens_seen": 233105390, + "step": 7161 + }, + { + "epoch": 0.6458943950940164, + "flos": 18231586159680.0, + "grad_norm": 5.569416503751604, + "learning_rate": 1.1771975810107947e-06, + "loss": 0.7731, + "num_input_tokens_seen": 233132705, + "step": 7162 + }, + { + "epoch": 0.6459845786174866, + "flos": 16631027034240.0, + "grad_norm": 2.109520462106876, + "learning_rate": 1.1766651347988542e-06, + "loss": 0.8151, + "num_input_tokens_seen": 233159915, + "step": 7163 + }, + { + "epoch": 0.6460747621409568, + "flos": 17942508237120.0, + "grad_norm": 2.939325130882363, + "learning_rate": 1.1761327588383848e-06, + "loss": 0.8574, + "num_input_tokens_seen": 233184995, + "step": 7164 + }, + { + "epoch": 0.6461649456644271, + "flos": 25775706751200.0, + "grad_norm": 9.41068931073716, + "learning_rate": 1.1756004531748105e-06, + "loss": 0.7593, + "num_input_tokens_seen": 233214410, + "step": 7165 + }, + { + "epoch": 0.6462551291878974, + "flos": 26253301194720.0, + "grad_norm": 1.6776839260914074, + "learning_rate": 1.1750682178535521e-06, + "loss": 0.7813, + "num_input_tokens_seen": 233248625, + "step": 7166 + }, + { + "epoch": 0.6463453127113676, + "flos": 23043410880480.0, + "grad_norm": 1.9641789246301942, + "learning_rate": 1.1745360529200218e-06, + "loss": 0.6958, + "num_input_tokens_seen": 233279120, + "step": 7167 + }, + { + "epoch": 0.6464354962348379, + "flos": 24565158647520.0, + "grad_norm": 2.8675250860312547, + "learning_rate": 1.1740039584196261e-06, + "loss": 0.7701, + "num_input_tokens_seen": 233303565, + "step": 7168 + }, + { + "epoch": 0.6465256797583081, + "flos": 20273902346400.0, + "grad_norm": 1.7233831900290435, + "learning_rate": 1.1734719343977683e-06, + "loss": 0.776, + "num_input_tokens_seen": 233330910, + "step": 7169 + }, + { + "epoch": 0.6466158632817784, + "flos": 25517596972320.0, + "grad_norm": 4.021136705618629, + "learning_rate": 1.1729399808998416e-06, + "loss": 0.6892, + "num_input_tokens_seen": 233359505, + "step": 7170 + }, + { + "epoch": 0.6467060468052487, + "flos": 12952165660800.0, + "grad_norm": 2.754653201698391, + "learning_rate": 1.1724080979712368e-06, + "loss": 0.6374, + "num_input_tokens_seen": 233385285, + "step": 7171 + }, + { + "epoch": 0.6467962303287189, + "flos": 28799347899840.0, + "grad_norm": 2.108958676544265, + "learning_rate": 1.1718762856573365e-06, + "loss": 0.8141, + "num_input_tokens_seen": 233413785, + "step": 7172 + }, + { + "epoch": 0.6468864138521893, + "flos": 30178340854080.0, + "grad_norm": 1.6356527639969087, + "learning_rate": 1.1713445440035172e-06, + "loss": 0.7327, + "num_input_tokens_seen": 233445660, + "step": 7173 + }, + { + "epoch": 0.6469765973756595, + "flos": 27669209454240.0, + "grad_norm": 2.0265426081826017, + "learning_rate": 1.1708128730551506e-06, + "loss": 0.7833, + "num_input_tokens_seen": 233473485, + "step": 7174 + }, + { + "epoch": 0.6470667808991297, + "flos": 19363360074720.0, + "grad_norm": 1.7877367596451383, + "learning_rate": 1.1702812728576019e-06, + "loss": 0.7342, + "num_input_tokens_seen": 233502130, + "step": 7175 + }, + { + "epoch": 0.6471569644226, + "flos": 24099012490080.0, + "grad_norm": 1.970519648045751, + "learning_rate": 1.1697497434562303e-06, + "loss": 0.7751, + "num_input_tokens_seen": 233528990, + "step": 7176 + }, + { + "epoch": 0.6472471479460703, + "flos": 20383124302080.0, + "grad_norm": 2.0389633446916338, + "learning_rate": 1.1692182848963885e-06, + "loss": 0.7793, + "num_input_tokens_seen": 233554080, + "step": 7177 + }, + { + "epoch": 0.6473373314695405, + "flos": 22970360834880.0, + "grad_norm": 1.8856377515908536, + "learning_rate": 1.1686868972234227e-06, + "loss": 0.7083, + "num_input_tokens_seen": 233578720, + "step": 7178 + }, + { + "epoch": 0.6474275149930108, + "flos": 23698593935520.0, + "grad_norm": 3.228519021994205, + "learning_rate": 1.1681555804826755e-06, + "loss": 0.7558, + "num_input_tokens_seen": 233609355, + "step": 7179 + }, + { + "epoch": 0.647517698516481, + "flos": 28473949388160.0, + "grad_norm": 2.201552133962189, + "learning_rate": 1.1676243347194806e-06, + "loss": 0.6823, + "num_input_tokens_seen": 233642225, + "step": 7180 + }, + { + "epoch": 0.6476078820399513, + "flos": 24208197276000.0, + "grad_norm": 1.5725916933092825, + "learning_rate": 1.167093159979167e-06, + "loss": 0.7627, + "num_input_tokens_seen": 233672300, + "step": 7181 + }, + { + "epoch": 0.6476980655634216, + "flos": 26467619262720.0, + "grad_norm": 1.893189196752807, + "learning_rate": 1.1665620563070575e-06, + "loss": 0.7727, + "num_input_tokens_seen": 233703340, + "step": 7182 + }, + { + "epoch": 0.6477882490868918, + "flos": 17468444920800.0, + "grad_norm": 1.9606840418206029, + "learning_rate": 1.1660310237484691e-06, + "loss": 0.8574, + "num_input_tokens_seen": 233728540, + "step": 7183 + }, + { + "epoch": 0.647878432610362, + "flos": 25518303197760.0, + "grad_norm": 1.6334182126833394, + "learning_rate": 1.165500062348713e-06, + "loss": 0.6881, + "num_input_tokens_seen": 233758530, + "step": 7184 + }, + { + "epoch": 0.6479686161338324, + "flos": 31342346684640.0, + "grad_norm": 2.6422226265637034, + "learning_rate": 1.164969172153091e-06, + "loss": 0.7864, + "num_input_tokens_seen": 233786935, + "step": 7185 + }, + { + "epoch": 0.6480587996573026, + "flos": 30943303411200.0, + "grad_norm": 1.9079336017035606, + "learning_rate": 1.1644383532069055e-06, + "loss": 0.7627, + "num_input_tokens_seen": 233817010, + "step": 7186 + }, + { + "epoch": 0.6481489831807729, + "flos": 26026010880480.0, + "grad_norm": 1.7412461958143273, + "learning_rate": 1.1639076055554454e-06, + "loss": 0.7855, + "num_input_tokens_seen": 233847580, + "step": 7187 + }, + { + "epoch": 0.6482391667042431, + "flos": 24281916377280.0, + "grad_norm": 2.3343748651833858, + "learning_rate": 1.163376929244e-06, + "loss": 0.7356, + "num_input_tokens_seen": 233875325, + "step": 7188 + }, + { + "epoch": 0.6483293502277134, + "flos": 26357319384000.0, + "grad_norm": 1.6525991256000598, + "learning_rate": 1.1628463243178472e-06, + "loss": 0.7101, + "num_input_tokens_seen": 233905730, + "step": 7189 + }, + { + "epoch": 0.6484195337511837, + "flos": 27698430619200.0, + "grad_norm": 2.0105454662222453, + "learning_rate": 1.1623157908222623e-06, + "loss": 0.8026, + "num_input_tokens_seen": 233934055, + "step": 7190 + }, + { + "epoch": 0.6485097172746539, + "flos": 19691174620800.0, + "grad_norm": 1.6814326956372312, + "learning_rate": 1.1617853288025129e-06, + "loss": 0.7819, + "num_input_tokens_seen": 233962400, + "step": 7191 + }, + { + "epoch": 0.6485999007981241, + "flos": 30111052121280.0, + "grad_norm": 2.033292242583441, + "learning_rate": 1.1612549383038612e-06, + "loss": 0.7212, + "num_input_tokens_seen": 233992435, + "step": 7192 + }, + { + "epoch": 0.6486900843215945, + "flos": 29852719323840.0, + "grad_norm": 1.6873560950061401, + "learning_rate": 1.1607246193715629e-06, + "loss": 0.7231, + "num_input_tokens_seen": 234025940, + "step": 7193 + }, + { + "epoch": 0.6487802678450647, + "flos": 19176032986080.0, + "grad_norm": 3.784857298203095, + "learning_rate": 1.1601943720508684e-06, + "loss": 0.761, + "num_input_tokens_seen": 234053125, + "step": 7194 + }, + { + "epoch": 0.6488704513685349, + "flos": 21688955701440.0, + "grad_norm": 1.7040480465943233, + "learning_rate": 1.1596641963870202e-06, + "loss": 0.7958, + "num_input_tokens_seen": 234080515, + "step": 7195 + }, + { + "epoch": 0.6489606348920053, + "flos": 28213683763200.0, + "grad_norm": 1.610216531823565, + "learning_rate": 1.1591340924252561e-06, + "loss": 0.7603, + "num_input_tokens_seen": 234112220, + "step": 7196 + }, + { + "epoch": 0.6490508184154755, + "flos": 71599549494720.0, + "grad_norm": 0.7487482910900961, + "learning_rate": 1.158604060210808e-06, + "loss": 0.6231, + "num_input_tokens_seen": 234190890, + "step": 7197 + }, + { + "epoch": 0.6491410019389457, + "flos": 19654519503840.0, + "grad_norm": 4.909185418269233, + "learning_rate": 1.1580740997889008e-06, + "loss": 0.8084, + "num_input_tokens_seen": 234218515, + "step": 7198 + }, + { + "epoch": 0.649231185462416, + "flos": 23261743282560.0, + "grad_norm": 2.4483942959204708, + "learning_rate": 1.1575442112047544e-06, + "loss": 0.7492, + "num_input_tokens_seen": 234246765, + "step": 7199 + }, + { + "epoch": 0.6493213689858863, + "flos": 25993816134720.0, + "grad_norm": 2.24428298272909, + "learning_rate": 1.1570143945035797e-06, + "loss": 0.688, + "num_input_tokens_seen": 234275910, + "step": 7200 + }, + { + "epoch": 0.6494115525093566, + "flos": 25733364660960.0, + "grad_norm": 2.4038793602630584, + "learning_rate": 1.1564846497305864e-06, + "loss": 0.7017, + "num_input_tokens_seen": 234302635, + "step": 7201 + }, + { + "epoch": 0.6495017360328268, + "flos": 24390469277280.0, + "grad_norm": 2.2499576140755893, + "learning_rate": 1.1559549769309726e-06, + "loss": 0.7242, + "num_input_tokens_seen": 234330585, + "step": 7202 + }, + { + "epoch": 0.649591919556297, + "flos": 24574897124640.0, + "grad_norm": 1.6917703073758865, + "learning_rate": 1.1554253761499358e-06, + "loss": 0.7404, + "num_input_tokens_seen": 234361475, + "step": 7203 + }, + { + "epoch": 0.6496821030797674, + "flos": 23808261928320.0, + "grad_norm": 2.5186140416559293, + "learning_rate": 1.1548958474326617e-06, + "loss": 0.7824, + "num_input_tokens_seen": 234389005, + "step": 7204 + }, + { + "epoch": 0.6497722866032376, + "flos": 29016713888160.0, + "grad_norm": 1.7379096501947362, + "learning_rate": 1.154366390824334e-06, + "loss": 0.6789, + "num_input_tokens_seen": 234418915, + "step": 7205 + }, + { + "epoch": 0.6498624701267078, + "flos": 66284886329760.0, + "grad_norm": 0.9306731889083111, + "learning_rate": 1.1538370063701287e-06, + "loss": 0.5583, + "num_input_tokens_seen": 234509445, + "step": 7206 + }, + { + "epoch": 0.6499526536501781, + "flos": 23517994573440.0, + "grad_norm": 2.003724088066981, + "learning_rate": 1.1533076941152153e-06, + "loss": 0.703, + "num_input_tokens_seen": 234540130, + "step": 7207 + }, + { + "epoch": 0.6500428371736484, + "flos": 27159494604480.0, + "grad_norm": 2.7506225054763616, + "learning_rate": 1.1527784541047583e-06, + "loss": 0.6758, + "num_input_tokens_seen": 234568895, + "step": 7208 + }, + { + "epoch": 0.6501330206971186, + "flos": 20377920535680.0, + "grad_norm": 1.978207834446899, + "learning_rate": 1.1522492863839152e-06, + "loss": 0.7002, + "num_input_tokens_seen": 234594775, + "step": 7209 + }, + { + "epoch": 0.6502232042205889, + "flos": 20564913096480.0, + "grad_norm": 5.074377561085329, + "learning_rate": 1.1517201909978382e-06, + "loss": 0.7088, + "num_input_tokens_seen": 234618330, + "step": 7210 + }, + { + "epoch": 0.6503133877440591, + "flos": 58190269928160.0, + "grad_norm": 0.5998745061844933, + "learning_rate": 1.151191167991671e-06, + "loss": 0.5779, + "num_input_tokens_seen": 234715760, + "step": 7211 + }, + { + "epoch": 0.6504035712675295, + "flos": 27299982061920.0, + "grad_norm": 1.9752570504093006, + "learning_rate": 1.1506622174105536e-06, + "loss": 0.7532, + "num_input_tokens_seen": 234744505, + "step": 7212 + }, + { + "epoch": 0.6504937547909997, + "flos": 20346580694400.0, + "grad_norm": 1.9424338125492573, + "learning_rate": 1.1501333392996194e-06, + "loss": 0.7746, + "num_input_tokens_seen": 234770895, + "step": 7213 + }, + { + "epoch": 0.6505839383144699, + "flos": 59574615327840.0, + "grad_norm": 0.659842964188155, + "learning_rate": 1.1496045337039943e-06, + "loss": 0.5847, + "num_input_tokens_seen": 234858605, + "step": 7214 + }, + { + "epoch": 0.6506741218379402, + "flos": 21579993934080.0, + "grad_norm": 1.7594686886385442, + "learning_rate": 1.1490758006687995e-06, + "loss": 0.7041, + "num_input_tokens_seen": 234886320, + "step": 7215 + }, + { + "epoch": 0.6507643053614105, + "flos": 25556593784160.0, + "grad_norm": 2.2316491632817366, + "learning_rate": 1.1485471402391502e-06, + "loss": 0.7255, + "num_input_tokens_seen": 234915530, + "step": 7216 + }, + { + "epoch": 0.6508544888848807, + "flos": 19031531194560.0, + "grad_norm": 2.839077642859141, + "learning_rate": 1.1480185524601522e-06, + "loss": 0.7187, + "num_input_tokens_seen": 234943230, + "step": 7217 + }, + { + "epoch": 0.650944672408351, + "flos": 25228147352160.0, + "grad_norm": 3.1147062517659734, + "learning_rate": 1.1474900373769108e-06, + "loss": 0.7498, + "num_input_tokens_seen": 234972830, + "step": 7218 + }, + { + "epoch": 0.6510348559318213, + "flos": 69810659697120.0, + "grad_norm": 0.5682854096305262, + "learning_rate": 1.1469615950345184e-06, + "loss": 0.524, + "num_input_tokens_seen": 235064365, + "step": 7219 + }, + { + "epoch": 0.6511250394552915, + "flos": 67990653076800.0, + "grad_norm": 0.6962061507629755, + "learning_rate": 1.1464332254780678e-06, + "loss": 0.5615, + "num_input_tokens_seen": 235152815, + "step": 7220 + }, + { + "epoch": 0.6512152229787618, + "flos": 20274645741600.0, + "grad_norm": 1.643119006285649, + "learning_rate": 1.1459049287526404e-06, + "loss": 0.7315, + "num_input_tokens_seen": 235179210, + "step": 7221 + }, + { + "epoch": 0.651305406502232, + "flos": 15858147318240.0, + "grad_norm": 1.8912627980769279, + "learning_rate": 1.1453767049033137e-06, + "loss": 0.7044, + "num_input_tokens_seen": 235203825, + "step": 7222 + }, + { + "epoch": 0.6513955900257024, + "flos": 20926855215840.0, + "grad_norm": 1.8591028222658137, + "learning_rate": 1.1448485539751586e-06, + "loss": 0.7872, + "num_input_tokens_seen": 235231140, + "step": 7223 + }, + { + "epoch": 0.6514857735491726, + "flos": 15866101646880.0, + "grad_norm": 1.7514295869569627, + "learning_rate": 1.1443204760132408e-06, + "loss": 0.7181, + "num_input_tokens_seen": 235258065, + "step": 7224 + }, + { + "epoch": 0.6515759570726428, + "flos": 23004562747680.0, + "grad_norm": 2.8236243755602146, + "learning_rate": 1.1437924710626185e-06, + "loss": 0.7242, + "num_input_tokens_seen": 235288920, + "step": 7225 + }, + { + "epoch": 0.651666140596113, + "flos": 63929623500960.0, + "grad_norm": 0.6536728801425498, + "learning_rate": 1.1432645391683429e-06, + "loss": 0.5467, + "num_input_tokens_seen": 235380705, + "step": 7226 + }, + { + "epoch": 0.6517563241195834, + "flos": 23180738908320.0, + "grad_norm": 2.547878044461599, + "learning_rate": 1.1427366803754609e-06, + "loss": 0.7813, + "num_input_tokens_seen": 235409060, + "step": 7227 + }, + { + "epoch": 0.6518465076430536, + "flos": 27482105384160.0, + "grad_norm": 1.6990835552584322, + "learning_rate": 1.142208894729012e-06, + "loss": 0.76, + "num_input_tokens_seen": 235438090, + "step": 7228 + }, + { + "epoch": 0.6519366911665239, + "flos": 28325656281120.0, + "grad_norm": 1.7675978854428946, + "learning_rate": 1.1416811822740301e-06, + "loss": 0.7849, + "num_input_tokens_seen": 235467085, + "step": 7229 + }, + { + "epoch": 0.6520268746899941, + "flos": 23044414464000.0, + "grad_norm": 1.6210155789934524, + "learning_rate": 1.1411535430555428e-06, + "loss": 0.7926, + "num_input_tokens_seen": 235496935, + "step": 7230 + }, + { + "epoch": 0.6521170582134644, + "flos": 15137162320800.0, + "grad_norm": 2.051385118496385, + "learning_rate": 1.1406259771185705e-06, + "loss": 0.8009, + "num_input_tokens_seen": 235522400, + "step": 7231 + }, + { + "epoch": 0.6522072417369347, + "flos": 24753861017280.0, + "grad_norm": 1.6485597827265805, + "learning_rate": 1.1400984845081282e-06, + "loss": 0.7057, + "num_input_tokens_seen": 235554565, + "step": 7232 + }, + { + "epoch": 0.6522974252604049, + "flos": 25119631621920.0, + "grad_norm": 2.1919057892093656, + "learning_rate": 1.139571065269226e-06, + "loss": 0.8216, + "num_input_tokens_seen": 235582475, + "step": 7233 + }, + { + "epoch": 0.6523876087838751, + "flos": 43954691966880.0, + "grad_norm": 2.182120110832716, + "learning_rate": 1.139043719446863e-06, + "loss": 0.7895, + "num_input_tokens_seen": 235614850, + "step": 7234 + }, + { + "epoch": 0.6524777923073455, + "flos": 21221136904800.0, + "grad_norm": 1.7742827833505541, + "learning_rate": 1.1385164470860385e-06, + "loss": 0.8184, + "num_input_tokens_seen": 235645315, + "step": 7235 + }, + { + "epoch": 0.6525679758308157, + "flos": 18853645224960.0, + "grad_norm": 2.396690104886801, + "learning_rate": 1.1379892482317403e-06, + "loss": 0.746, + "num_input_tokens_seen": 235672735, + "step": 7236 + }, + { + "epoch": 0.652658159354286, + "flos": 20565098945280.0, + "grad_norm": 1.9307122231758682, + "learning_rate": 1.1374621229289524e-06, + "loss": 0.7292, + "num_input_tokens_seen": 235699410, + "step": 7237 + }, + { + "epoch": 0.6527483428777562, + "flos": 23294532744480.0, + "grad_norm": 1.8473032963042038, + "learning_rate": 1.1369350712226525e-06, + "loss": 0.73, + "num_input_tokens_seen": 235726440, + "step": 7238 + }, + { + "epoch": 0.6528385264012265, + "flos": 27706347778080.0, + "grad_norm": 1.9863541112829723, + "learning_rate": 1.136408093157811e-06, + "loss": 0.6945, + "num_input_tokens_seen": 235754865, + "step": 7239 + }, + { + "epoch": 0.6529287099246968, + "flos": 23443717925760.0, + "grad_norm": 2.9344596549955964, + "learning_rate": 1.1358811887793935e-06, + "loss": 0.7174, + "num_input_tokens_seen": 235782995, + "step": 7240 + }, + { + "epoch": 0.653018893448167, + "flos": 23335090686240.0, + "grad_norm": 1.9352506686151452, + "learning_rate": 1.135354358132356e-06, + "loss": 0.7857, + "num_input_tokens_seen": 235811555, + "step": 7241 + }, + { + "epoch": 0.6531090769716373, + "flos": 23625989927040.0, + "grad_norm": 3.0077950613333755, + "learning_rate": 1.1348276012616542e-06, + "loss": 0.7843, + "num_input_tokens_seen": 235840425, + "step": 7242 + }, + { + "epoch": 0.6531992604951076, + "flos": 19981702164000.0, + "grad_norm": 2.2323750969098413, + "learning_rate": 1.134300918212231e-06, + "loss": 0.7036, + "num_input_tokens_seen": 235867410, + "step": 7243 + }, + { + "epoch": 0.6532894440185778, + "flos": 35539100255040.0, + "grad_norm": 2.160325552460727, + "learning_rate": 1.133774309029027e-06, + "loss": 0.7561, + "num_input_tokens_seen": 235895515, + "step": 7244 + }, + { + "epoch": 0.653379627542048, + "flos": 23808224758560.0, + "grad_norm": 2.0078342302979597, + "learning_rate": 1.133247773756975e-06, + "loss": 0.7883, + "num_input_tokens_seen": 235923865, + "step": 7245 + }, + { + "epoch": 0.6534698110655184, + "flos": 69272392738080.0, + "grad_norm": 0.6559710590997399, + "learning_rate": 1.1327213124410024e-06, + "loss": 0.6168, + "num_input_tokens_seen": 236009930, + "step": 7246 + }, + { + "epoch": 0.6535599945889886, + "flos": 24172768761120.0, + "grad_norm": 1.8175531519656687, + "learning_rate": 1.1321949251260292e-06, + "loss": 0.7874, + "num_input_tokens_seen": 236039670, + "step": 7247 + }, + { + "epoch": 0.6536501781124588, + "flos": 27888396760800.0, + "grad_norm": 2.2873482909210594, + "learning_rate": 1.1316686118569712e-06, + "loss": 0.6275, + "num_input_tokens_seen": 236068745, + "step": 7248 + }, + { + "epoch": 0.6537403616359291, + "flos": 22605854002080.0, + "grad_norm": 1.7836744974432948, + "learning_rate": 1.1311423726787335e-06, + "loss": 0.657, + "num_input_tokens_seen": 236096510, + "step": 7249 + }, + { + "epoch": 0.6538305451593994, + "flos": 24420433837440.0, + "grad_norm": 1.7958981056067869, + "learning_rate": 1.130616207636221e-06, + "loss": 0.7174, + "num_input_tokens_seen": 236125355, + "step": 7250 + }, + { + "epoch": 0.6539207286828697, + "flos": 20492346257760.0, + "grad_norm": 2.1726567899009916, + "learning_rate": 1.1300901167743263e-06, + "loss": 0.7482, + "num_input_tokens_seen": 236152170, + "step": 7251 + }, + { + "epoch": 0.6540109122063399, + "flos": 23660675046720.0, + "grad_norm": 1.5496706120377965, + "learning_rate": 1.12956410013794e-06, + "loss": 0.7407, + "num_input_tokens_seen": 236183940, + "step": 7252 + }, + { + "epoch": 0.6541010957298101, + "flos": 21403297396800.0, + "grad_norm": 2.490347155840728, + "learning_rate": 1.1290381577719436e-06, + "loss": 0.7316, + "num_input_tokens_seen": 236208660, + "step": 7253 + }, + { + "epoch": 0.6541912792532805, + "flos": 22460051268960.0, + "grad_norm": 2.0188418010235014, + "learning_rate": 1.1285122897212143e-06, + "loss": 0.6515, + "num_input_tokens_seen": 236238365, + "step": 7254 + }, + { + "epoch": 0.6542814627767507, + "flos": 51930082013760.0, + "grad_norm": 0.7479424465982054, + "learning_rate": 1.1279864960306228e-06, + "loss": 0.6171, + "num_input_tokens_seen": 236327280, + "step": 7255 + }, + { + "epoch": 0.6543716463002209, + "flos": 22388190655680.0, + "grad_norm": 1.7995351588531434, + "learning_rate": 1.1274607767450297e-06, + "loss": 0.8227, + "num_input_tokens_seen": 236356645, + "step": 7256 + }, + { + "epoch": 0.6544618298236912, + "flos": 22059298186560.0, + "grad_norm": 1.6245318985681174, + "learning_rate": 1.126935131909296e-06, + "loss": 0.7215, + "num_input_tokens_seen": 236386625, + "step": 7257 + }, + { + "epoch": 0.6545520133471615, + "flos": 20890274438400.0, + "grad_norm": 1.5459949803042623, + "learning_rate": 1.1264095615682693e-06, + "loss": 0.6953, + "num_input_tokens_seen": 236416080, + "step": 7258 + }, + { + "epoch": 0.6546421968706317, + "flos": 20965889197440.0, + "grad_norm": 1.784786194356977, + "learning_rate": 1.1258840657667973e-06, + "loss": 0.6618, + "num_input_tokens_seen": 236444830, + "step": 7259 + }, + { + "epoch": 0.654732380394102, + "flos": 22131865025280.0, + "grad_norm": 1.6372997918659289, + "learning_rate": 1.125358644549716e-06, + "loss": 0.6726, + "num_input_tokens_seen": 236471920, + "step": 7260 + }, + { + "epoch": 0.6548225639175722, + "flos": 59438891333280.0, + "grad_norm": 2.191928631268079, + "learning_rate": 1.1248332979618578e-06, + "loss": 0.6059, + "num_input_tokens_seen": 236509525, + "step": 7261 + }, + { + "epoch": 0.6549127474410426, + "flos": 32551036300320.0, + "grad_norm": 1.5785780152600792, + "learning_rate": 1.1243080260480482e-06, + "loss": 0.6365, + "num_input_tokens_seen": 236544895, + "step": 7262 + }, + { + "epoch": 0.6550029309645128, + "flos": 20528852695680.0, + "grad_norm": 2.8041758801184677, + "learning_rate": 1.1237828288531063e-06, + "loss": 0.8005, + "num_input_tokens_seen": 236569855, + "step": 7263 + }, + { + "epoch": 0.655093114487983, + "flos": 21622559042880.0, + "grad_norm": 1.9296280083385438, + "learning_rate": 1.1232577064218449e-06, + "loss": 0.6597, + "num_input_tokens_seen": 236597700, + "step": 7264 + }, + { + "epoch": 0.6551832980114533, + "flos": 25119371433600.0, + "grad_norm": 2.1616058176731796, + "learning_rate": 1.1227326587990711e-06, + "loss": 0.8216, + "num_input_tokens_seen": 236627790, + "step": 7265 + }, + { + "epoch": 0.6552734815349236, + "flos": 25629197792640.0, + "grad_norm": 2.017678922072207, + "learning_rate": 1.1222076860295832e-06, + "loss": 0.7288, + "num_input_tokens_seen": 236657970, + "step": 7266 + }, + { + "epoch": 0.6553636650583938, + "flos": 22789389775200.0, + "grad_norm": 2.260477993978483, + "learning_rate": 1.1216827881581756e-06, + "loss": 0.7643, + "num_input_tokens_seen": 236686010, + "step": 7267 + }, + { + "epoch": 0.6554538485818641, + "flos": 18592152997920.0, + "grad_norm": 2.0857932508790507, + "learning_rate": 1.1211579652296355e-06, + "loss": 0.7403, + "num_input_tokens_seen": 236711745, + "step": 7268 + }, + { + "epoch": 0.6555440321053344, + "flos": 26394866575200.0, + "grad_norm": 1.9998077061717254, + "learning_rate": 1.1206332172887438e-06, + "loss": 0.8295, + "num_input_tokens_seen": 236740275, + "step": 7269 + }, + { + "epoch": 0.6556342156288046, + "flos": 39292201106400.0, + "grad_norm": 2.1222836579866686, + "learning_rate": 1.1201085443802756e-06, + "loss": 0.7608, + "num_input_tokens_seen": 236770750, + "step": 7270 + }, + { + "epoch": 0.6557243991522749, + "flos": 20602980664320.0, + "grad_norm": 1.744042405527234, + "learning_rate": 1.1195839465489964e-06, + "loss": 0.7126, + "num_input_tokens_seen": 236799385, + "step": 7271 + }, + { + "epoch": 0.6558145826757451, + "flos": 26503902682080.0, + "grad_norm": 2.257031473943449, + "learning_rate": 1.1190594238396708e-06, + "loss": 0.7906, + "num_input_tokens_seen": 236825620, + "step": 7272 + }, + { + "epoch": 0.6559047661992154, + "flos": 29534160048000.0, + "grad_norm": 1.500697053436988, + "learning_rate": 1.1185349762970515e-06, + "loss": 0.7177, + "num_input_tokens_seen": 236860040, + "step": 7273 + }, + { + "epoch": 0.6559949497226857, + "flos": 44756755678080.0, + "grad_norm": 1.8119346789360815, + "learning_rate": 1.1180106039658896e-06, + "loss": 0.7473, + "num_input_tokens_seen": 236895255, + "step": 7274 + }, + { + "epoch": 0.6560851332461559, + "flos": 61896971451840.0, + "grad_norm": 0.5836713328235811, + "learning_rate": 1.117486306890925e-06, + "loss": 0.558, + "num_input_tokens_seen": 236993225, + "step": 7275 + }, + { + "epoch": 0.6561753167696261, + "flos": 27632926034880.0, + "grad_norm": 2.844962060873153, + "learning_rate": 1.116962085116896e-06, + "loss": 0.696, + "num_input_tokens_seen": 237022650, + "step": 7276 + }, + { + "epoch": 0.6562655002930965, + "flos": 68432001270720.0, + "grad_norm": 0.6639760129370397, + "learning_rate": 1.1164379386885302e-06, + "loss": 0.5166, + "num_input_tokens_seen": 237114845, + "step": 7277 + }, + { + "epoch": 0.6563556838165667, + "flos": 17687743736640.0, + "grad_norm": 2.030412706788375, + "learning_rate": 1.1159138676505516e-06, + "loss": 0.7895, + "num_input_tokens_seen": 237141440, + "step": 7278 + }, + { + "epoch": 0.656445867340037, + "flos": 29052328251840.0, + "grad_norm": 1.6622196792955943, + "learning_rate": 1.1153898720476761e-06, + "loss": 0.8034, + "num_input_tokens_seen": 237170710, + "step": 7279 + }, + { + "epoch": 0.6565360508635072, + "flos": 18233295968640.0, + "grad_norm": 3.0463065375151945, + "learning_rate": 1.114865951924615e-06, + "loss": 0.7335, + "num_input_tokens_seen": 237194550, + "step": 7280 + }, + { + "epoch": 0.6566262343869775, + "flos": 19691025941760.0, + "grad_norm": 1.8707266748124534, + "learning_rate": 1.1143421073260721e-06, + "loss": 0.6883, + "num_input_tokens_seen": 237223095, + "step": 7281 + }, + { + "epoch": 0.6567164179104478, + "flos": 17431603955040.0, + "grad_norm": 2.229329100302886, + "learning_rate": 1.1138183382967432e-06, + "loss": 0.6836, + "num_input_tokens_seen": 237248795, + "step": 7282 + }, + { + "epoch": 0.656806601433918, + "flos": 20849902345440.0, + "grad_norm": 2.0010092096378926, + "learning_rate": 1.11329464488132e-06, + "loss": 0.6526, + "num_input_tokens_seen": 237275740, + "step": 7283 + }, + { + "epoch": 0.6568967849573882, + "flos": 33058483794720.0, + "grad_norm": 1.9644446147704782, + "learning_rate": 1.112771027124487e-06, + "loss": 0.7192, + "num_input_tokens_seen": 237307355, + "step": 7284 + }, + { + "epoch": 0.6569869684808586, + "flos": 24135853455840.0, + "grad_norm": 2.936958100676929, + "learning_rate": 1.112247485070922e-06, + "loss": 0.7543, + "num_input_tokens_seen": 237335955, + "step": 7285 + }, + { + "epoch": 0.6570771520043288, + "flos": 19101867847680.0, + "grad_norm": 1.7796827415020888, + "learning_rate": 1.1117240187652968e-06, + "loss": 0.7512, + "num_input_tokens_seen": 237363330, + "step": 7286 + }, + { + "epoch": 0.657167335527799, + "flos": 17614210484160.0, + "grad_norm": 2.2668726149662954, + "learning_rate": 1.1112006282522767e-06, + "loss": 0.8245, + "num_input_tokens_seen": 237390325, + "step": 7287 + }, + { + "epoch": 0.6572575190512693, + "flos": 20491156825440.0, + "grad_norm": 2.0178634870118386, + "learning_rate": 1.1106773135765183e-06, + "loss": 0.7279, + "num_input_tokens_seen": 237418555, + "step": 7288 + }, + { + "epoch": 0.6573477025747396, + "flos": 25558526611680.0, + "grad_norm": 1.9962613806164264, + "learning_rate": 1.110154074782677e-06, + "loss": 0.7206, + "num_input_tokens_seen": 237450550, + "step": 7289 + }, + { + "epoch": 0.6574378860982099, + "flos": 22314806082240.0, + "grad_norm": 2.840311507750439, + "learning_rate": 1.1096309119153948e-06, + "loss": 0.7415, + "num_input_tokens_seen": 237478155, + "step": 7290 + }, + { + "epoch": 0.6575280696216801, + "flos": 68918107589280.0, + "grad_norm": 0.6253976272295841, + "learning_rate": 1.1091078250193145e-06, + "loss": 0.6186, + "num_input_tokens_seen": 237570585, + "step": 7291 + }, + { + "epoch": 0.6576182531451504, + "flos": 18087902102880.0, + "grad_norm": 2.6686908867250176, + "learning_rate": 1.108584814139066e-06, + "loss": 0.7483, + "num_input_tokens_seen": 237596580, + "step": 7292 + }, + { + "epoch": 0.6577084366686207, + "flos": 33972780212160.0, + "grad_norm": 2.169924072484261, + "learning_rate": 1.108061879319276e-06, + "loss": 0.7439, + "num_input_tokens_seen": 237624510, + "step": 7293 + }, + { + "epoch": 0.6577986201920909, + "flos": 21726800250720.0, + "grad_norm": 1.9127755334376664, + "learning_rate": 1.1075390206045648e-06, + "loss": 0.803, + "num_input_tokens_seen": 237652345, + "step": 7294 + }, + { + "epoch": 0.6578888037155611, + "flos": 25811767152000.0, + "grad_norm": 2.1205200768374612, + "learning_rate": 1.1070162380395454e-06, + "loss": 0.7865, + "num_input_tokens_seen": 237678590, + "step": 7295 + }, + { + "epoch": 0.6579789872390315, + "flos": 20965740518400.0, + "grad_norm": 1.901981264393143, + "learning_rate": 1.1064935316688253e-06, + "loss": 0.6881, + "num_input_tokens_seen": 237706130, + "step": 7296 + }, + { + "epoch": 0.6580691707625017, + "flos": 23656809391680.0, + "grad_norm": 1.986258266424571, + "learning_rate": 1.105970901537002e-06, + "loss": 0.8296, + "num_input_tokens_seen": 237731880, + "step": 7297 + }, + { + "epoch": 0.6581593542859719, + "flos": 20199402680160.0, + "grad_norm": 1.865404243123965, + "learning_rate": 1.1054483476886727e-06, + "loss": 0.7893, + "num_input_tokens_seen": 237760970, + "step": 7298 + }, + { + "epoch": 0.6582495378094422, + "flos": 23298138211200.0, + "grad_norm": 1.84394945261971, + "learning_rate": 1.1049258701684222e-06, + "loss": 0.6653, + "num_input_tokens_seen": 237790465, + "step": 7299 + }, + { + "epoch": 0.6583397213329125, + "flos": 18707247775680.0, + "grad_norm": 1.932479753277165, + "learning_rate": 1.1044034690208315e-06, + "loss": 0.695, + "num_input_tokens_seen": 237818340, + "step": 7300 + }, + { + "epoch": 0.6584299048563828, + "flos": 17978791656480.0, + "grad_norm": 4.157415470404021, + "learning_rate": 1.1038811442904755e-06, + "loss": 0.7779, + "num_input_tokens_seen": 237846515, + "step": 7301 + }, + { + "epoch": 0.658520088379853, + "flos": 21185299522560.0, + "grad_norm": 2.096133227977156, + "learning_rate": 1.103358896021921e-06, + "loss": 0.6843, + "num_input_tokens_seen": 237873580, + "step": 7302 + }, + { + "epoch": 0.6586102719033232, + "flos": 23298584248320.0, + "grad_norm": 2.257227415946343, + "learning_rate": 1.1028367242597298e-06, + "loss": 0.7222, + "num_input_tokens_seen": 237903275, + "step": 7303 + }, + { + "epoch": 0.6587004554267936, + "flos": 18561816740160.0, + "grad_norm": 2.254567173359595, + "learning_rate": 1.102314629048457e-06, + "loss": 0.7862, + "num_input_tokens_seen": 237928605, + "step": 7304 + }, + { + "epoch": 0.6587906389502638, + "flos": 21877472222400.0, + "grad_norm": 1.7006772851950178, + "learning_rate": 1.1017926104326484e-06, + "loss": 0.797, + "num_input_tokens_seen": 237958030, + "step": 7305 + }, + { + "epoch": 0.658880822473734, + "flos": 16157484094560.0, + "grad_norm": 2.00022290894936, + "learning_rate": 1.1012706684568483e-06, + "loss": 0.799, + "num_input_tokens_seen": 237984890, + "step": 7306 + }, + { + "epoch": 0.6589710059972043, + "flos": 19727272191360.0, + "grad_norm": 1.719885572353151, + "learning_rate": 1.1007488031655894e-06, + "loss": 0.7605, + "num_input_tokens_seen": 238014630, + "step": 7307 + }, + { + "epoch": 0.6590611895206746, + "flos": 26394755065920.0, + "grad_norm": 2.501548764192897, + "learning_rate": 1.1002270146034013e-06, + "loss": 0.7008, + "num_input_tokens_seen": 238041900, + "step": 7308 + }, + { + "epoch": 0.6591513730441448, + "flos": 28651426490400.0, + "grad_norm": 1.3767386627921914, + "learning_rate": 1.0997053028148052e-06, + "loss": 0.7733, + "num_input_tokens_seen": 238072700, + "step": 7309 + }, + { + "epoch": 0.6592415565676151, + "flos": 26467136055840.0, + "grad_norm": 2.2959976077515902, + "learning_rate": 1.0991836678443173e-06, + "loss": 0.7519, + "num_input_tokens_seen": 238104415, + "step": 7310 + }, + { + "epoch": 0.6593317400910853, + "flos": 24567872040000.0, + "grad_norm": 2.0429694502119977, + "learning_rate": 1.0986621097364465e-06, + "loss": 0.6924, + "num_input_tokens_seen": 238132585, + "step": 7311 + }, + { + "epoch": 0.6594219236145556, + "flos": 21075334171680.0, + "grad_norm": 2.1753647027138308, + "learning_rate": 1.0981406285356932e-06, + "loss": 0.6597, + "num_input_tokens_seen": 238158150, + "step": 7312 + }, + { + "epoch": 0.6595121071380259, + "flos": 23116609605120.0, + "grad_norm": 2.1063502347283873, + "learning_rate": 1.0976192242865554e-06, + "loss": 0.6706, + "num_input_tokens_seen": 238189160, + "step": 7313 + }, + { + "epoch": 0.6596022906614961, + "flos": 18962123785440.0, + "grad_norm": 2.0407969380450335, + "learning_rate": 1.0970978970335202e-06, + "loss": 0.8121, + "num_input_tokens_seen": 238216760, + "step": 7314 + }, + { + "epoch": 0.6596924741849665, + "flos": 27739360258560.0, + "grad_norm": 1.8221222227203229, + "learning_rate": 1.0965766468210714e-06, + "loss": 0.7668, + "num_input_tokens_seen": 238246395, + "step": 7315 + }, + { + "epoch": 0.6597826577084367, + "flos": 22563437572320.0, + "grad_norm": 2.4355020404228385, + "learning_rate": 1.0960554736936843e-06, + "loss": 0.7547, + "num_input_tokens_seen": 238272540, + "step": 7316 + }, + { + "epoch": 0.6598728412319069, + "flos": 22604961927840.0, + "grad_norm": 2.0786940863837673, + "learning_rate": 1.0955343776958283e-06, + "loss": 0.6547, + "num_input_tokens_seen": 238298570, + "step": 7317 + }, + { + "epoch": 0.6599630247553772, + "flos": 22059632714400.0, + "grad_norm": 2.0677068471703044, + "learning_rate": 1.0950133588719665e-06, + "loss": 0.7603, + "num_input_tokens_seen": 238327600, + "step": 7318 + }, + { + "epoch": 0.6600532082788475, + "flos": 21544379570400.0, + "grad_norm": 2.4269880895522773, + "learning_rate": 1.0944924172665551e-06, + "loss": 0.7709, + "num_input_tokens_seen": 238354130, + "step": 7319 + }, + { + "epoch": 0.6601433918023177, + "flos": 19030899308640.0, + "grad_norm": 1.9352215321216977, + "learning_rate": 1.0939715529240437e-06, + "loss": 0.7854, + "num_input_tokens_seen": 238380060, + "step": 7320 + }, + { + "epoch": 0.660233575325788, + "flos": 27815309545440.0, + "grad_norm": 5.146743796829708, + "learning_rate": 1.0934507658888755e-06, + "loss": 0.777, + "num_input_tokens_seen": 238409565, + "step": 7321 + }, + { + "epoch": 0.6603237588492582, + "flos": 25338001193760.0, + "grad_norm": 1.6990431420797034, + "learning_rate": 1.092930056205486e-06, + "loss": 0.7178, + "num_input_tokens_seen": 238441710, + "step": 7322 + }, + { + "epoch": 0.6604139423727285, + "flos": 15283596939840.0, + "grad_norm": 2.624617594158045, + "learning_rate": 1.092409423918306e-06, + "loss": 0.7135, + "num_input_tokens_seen": 238468450, + "step": 7323 + }, + { + "epoch": 0.6605041258961988, + "flos": 23551081393440.0, + "grad_norm": 1.62660959195759, + "learning_rate": 1.0918888690717581e-06, + "loss": 0.6689, + "num_input_tokens_seen": 238500195, + "step": 7324 + }, + { + "epoch": 0.660594309419669, + "flos": 25658047260000.0, + "grad_norm": 1.7973522050179973, + "learning_rate": 1.091368391710259e-06, + "loss": 0.8119, + "num_input_tokens_seen": 238527505, + "step": 7325 + }, + { + "epoch": 0.6606844929431392, + "flos": 14153049626880.0, + "grad_norm": 1.7443170234014378, + "learning_rate": 1.0908479918782198e-06, + "loss": 0.6919, + "num_input_tokens_seen": 238552985, + "step": 7326 + }, + { + "epoch": 0.6607746764666096, + "flos": 22896418715040.0, + "grad_norm": 1.7395910236241956, + "learning_rate": 1.0903276696200413e-06, + "loss": 0.7673, + "num_input_tokens_seen": 238582255, + "step": 7327 + }, + { + "epoch": 0.6608648599900798, + "flos": 26031177477120.0, + "grad_norm": 2.180734581155177, + "learning_rate": 1.0898074249801234e-06, + "loss": 0.7515, + "num_input_tokens_seen": 238611775, + "step": 7328 + }, + { + "epoch": 0.6609550435135501, + "flos": 25519975836960.0, + "grad_norm": 1.5504317574440247, + "learning_rate": 1.0892872580028533e-06, + "loss": 0.7635, + "num_input_tokens_seen": 238643545, + "step": 7329 + }, + { + "epoch": 0.6610452270370203, + "flos": 25076100099360.0, + "grad_norm": 1.7678169215637445, + "learning_rate": 1.0887671687326178e-06, + "loss": 0.7793, + "num_input_tokens_seen": 238673360, + "step": 7330 + }, + { + "epoch": 0.6611354105604906, + "flos": 16303621355520.0, + "grad_norm": 2.091990286525546, + "learning_rate": 1.0882471572137908e-06, + "loss": 0.7146, + "num_input_tokens_seen": 238699885, + "step": 7331 + }, + { + "epoch": 0.6612255940839609, + "flos": 20674729768320.0, + "grad_norm": 1.886286399697392, + "learning_rate": 1.087727223490744e-06, + "loss": 0.7906, + "num_input_tokens_seen": 238729125, + "step": 7332 + }, + { + "epoch": 0.6613157776074311, + "flos": 20235202892640.0, + "grad_norm": 1.8751364976153952, + "learning_rate": 1.0872073676078405e-06, + "loss": 0.8064, + "num_input_tokens_seen": 238757115, + "step": 7333 + }, + { + "epoch": 0.6614059611309013, + "flos": 26134117743360.0, + "grad_norm": 5.00585994653479, + "learning_rate": 1.0866875896094375e-06, + "loss": 0.7004, + "num_input_tokens_seen": 238785055, + "step": 7334 + }, + { + "epoch": 0.6614961446543717, + "flos": 22896976261440.0, + "grad_norm": 1.8110982143915446, + "learning_rate": 1.0861678895398854e-06, + "loss": 0.7653, + "num_input_tokens_seen": 238813795, + "step": 7335 + }, + { + "epoch": 0.6615863281778419, + "flos": 16809173192160.0, + "grad_norm": 2.2604932874593824, + "learning_rate": 1.0856482674435286e-06, + "loss": 0.7111, + "num_input_tokens_seen": 238840830, + "step": 7336 + }, + { + "epoch": 0.6616765117013121, + "flos": 71324484571680.0, + "grad_norm": 0.6667063199608425, + "learning_rate": 1.0851287233647024e-06, + "loss": 0.5741, + "num_input_tokens_seen": 238926505, + "step": 7337 + }, + { + "epoch": 0.6617666952247825, + "flos": 21367311335520.0, + "grad_norm": 3.106660664543062, + "learning_rate": 1.084609257347738e-06, + "loss": 0.802, + "num_input_tokens_seen": 238953080, + "step": 7338 + }, + { + "epoch": 0.6618568787482527, + "flos": 25301197397760.0, + "grad_norm": 2.981814099571242, + "learning_rate": 1.0840898694369594e-06, + "loss": 0.7226, + "num_input_tokens_seen": 238984090, + "step": 7339 + }, + { + "epoch": 0.661947062271723, + "flos": 27123954580320.0, + "grad_norm": 1.9427402142592554, + "learning_rate": 1.083570559676683e-06, + "loss": 0.7369, + "num_input_tokens_seen": 239014865, + "step": 7340 + }, + { + "epoch": 0.6620372457951932, + "flos": 24974014737600.0, + "grad_norm": 1.7920837212658003, + "learning_rate": 1.08305132811122e-06, + "loss": 0.7236, + "num_input_tokens_seen": 239043425, + "step": 7341 + }, + { + "epoch": 0.6621274293186635, + "flos": 35466421907040.0, + "grad_norm": 1.7399334888242788, + "learning_rate": 1.0825321747848735e-06, + "loss": 0.749, + "num_input_tokens_seen": 239073805, + "step": 7342 + }, + { + "epoch": 0.6622176128421338, + "flos": 62742163551840.0, + "grad_norm": 0.7091818290222158, + "learning_rate": 1.0820130997419417e-06, + "loss": 0.6439, + "num_input_tokens_seen": 239155635, + "step": 7343 + }, + { + "epoch": 0.662307796365604, + "flos": 21396829858560.0, + "grad_norm": 2.174147151490494, + "learning_rate": 1.0814941030267123e-06, + "loss": 0.7372, + "num_input_tokens_seen": 239182350, + "step": 7344 + }, + { + "epoch": 0.6623979798890742, + "flos": 66033206919360.0, + "grad_norm": 0.5822494661010155, + "learning_rate": 1.080975184683472e-06, + "loss": 0.5622, + "num_input_tokens_seen": 239282460, + "step": 7345 + }, + { + "epoch": 0.6624881634125446, + "flos": 24496531803360.0, + "grad_norm": 4.86148879022119, + "learning_rate": 1.0804563447564948e-06, + "loss": 0.8058, + "num_input_tokens_seen": 239312585, + "step": 7346 + }, + { + "epoch": 0.6625783469360148, + "flos": 46031470254720.0, + "grad_norm": 2.4106714820089326, + "learning_rate": 1.0799375832900545e-06, + "loss": 0.6935, + "num_input_tokens_seen": 239341295, + "step": 7347 + }, + { + "epoch": 0.662668530459485, + "flos": 21797025394560.0, + "grad_norm": 2.0760674867557194, + "learning_rate": 1.0794189003284118e-06, + "loss": 0.7833, + "num_input_tokens_seen": 239367940, + "step": 7348 + }, + { + "epoch": 0.6627587139829553, + "flos": 31673915376480.0, + "grad_norm": 1.8914638581419165, + "learning_rate": 1.0789002959158242e-06, + "loss": 0.7597, + "num_input_tokens_seen": 239396340, + "step": 7349 + }, + { + "epoch": 0.6628488975064256, + "flos": 13970926304640.0, + "grad_norm": 6.625945689523691, + "learning_rate": 1.0783817700965428e-06, + "loss": 0.7001, + "num_input_tokens_seen": 239422925, + "step": 7350 + }, + { + "epoch": 0.6629390810298958, + "flos": 26791456644480.0, + "grad_norm": 1.7145515488389664, + "learning_rate": 1.0778633229148102e-06, + "loss": 0.7959, + "num_input_tokens_seen": 239453815, + "step": 7351 + }, + { + "epoch": 0.6630292645533661, + "flos": 28143309940320.0, + "grad_norm": 2.0255350559866905, + "learning_rate": 1.0773449544148645e-06, + "loss": 0.7508, + "num_input_tokens_seen": 239481900, + "step": 7352 + }, + { + "epoch": 0.6631194480768363, + "flos": 25338447230880.0, + "grad_norm": 1.821910042532074, + "learning_rate": 1.076826664640934e-06, + "loss": 0.7873, + "num_input_tokens_seen": 239512325, + "step": 7353 + }, + { + "epoch": 0.6632096316003067, + "flos": 50603881585920.0, + "grad_norm": 0.7443938137466656, + "learning_rate": 1.0763084536372424e-06, + "loss": 0.6033, + "num_input_tokens_seen": 239590555, + "step": 7354 + }, + { + "epoch": 0.6632998151237769, + "flos": 22236998307360.0, + "grad_norm": 1.960416200200619, + "learning_rate": 1.0757903214480068e-06, + "loss": 0.7555, + "num_input_tokens_seen": 239616765, + "step": 7355 + }, + { + "epoch": 0.6633899986472471, + "flos": 27013877720160.0, + "grad_norm": 1.796014198111166, + "learning_rate": 1.0752722681174376e-06, + "loss": 0.693, + "num_input_tokens_seen": 239645300, + "step": 7356 + }, + { + "epoch": 0.6634801821707174, + "flos": 27159866302080.0, + "grad_norm": 21.325399751482898, + "learning_rate": 1.074754293689737e-06, + "loss": 0.6698, + "num_input_tokens_seen": 239675420, + "step": 7357 + }, + { + "epoch": 0.6635703656941877, + "flos": 26248877993280.0, + "grad_norm": 2.2710456982159144, + "learning_rate": 1.0742363982091023e-06, + "loss": 0.7517, + "num_input_tokens_seen": 239704220, + "step": 7358 + }, + { + "epoch": 0.6636605492176579, + "flos": 20018877657600.0, + "grad_norm": 1.936266044712759, + "learning_rate": 1.0737185817197215e-06, + "loss": 0.7032, + "num_input_tokens_seen": 239730770, + "step": 7359 + }, + { + "epoch": 0.6637507327411282, + "flos": 21294075441120.0, + "grad_norm": 2.3779156506259214, + "learning_rate": 1.0732008442657803e-06, + "loss": 0.7093, + "num_input_tokens_seen": 239758355, + "step": 7360 + }, + { + "epoch": 0.6638409162645985, + "flos": 26322411245760.0, + "grad_norm": 1.9068103029371666, + "learning_rate": 1.0726831858914516e-06, + "loss": 0.7285, + "num_input_tokens_seen": 239789070, + "step": 7361 + }, + { + "epoch": 0.6639310997880687, + "flos": 21840594086880.0, + "grad_norm": 2.330310077745269, + "learning_rate": 1.0721656066409084e-06, + "loss": 0.7654, + "num_input_tokens_seen": 239817600, + "step": 7362 + }, + { + "epoch": 0.664021283311539, + "flos": 21439878174240.0, + "grad_norm": 2.648695221460972, + "learning_rate": 1.0716481065583108e-06, + "loss": 0.7576, + "num_input_tokens_seen": 239844950, + "step": 7363 + }, + { + "epoch": 0.6641114668350092, + "flos": 65536575825120.0, + "grad_norm": 0.6355959113112559, + "learning_rate": 1.071130685687816e-06, + "loss": 0.5735, + "num_input_tokens_seen": 239937100, + "step": 7364 + }, + { + "epoch": 0.6642016503584796, + "flos": 24936170188320.0, + "grad_norm": 1.9333909670142762, + "learning_rate": 1.0706133440735723e-06, + "loss": 0.6826, + "num_input_tokens_seen": 239968435, + "step": 7365 + }, + { + "epoch": 0.6642918338819498, + "flos": 27013952059680.0, + "grad_norm": 2.327591188408404, + "learning_rate": 1.070096081759723e-06, + "loss": 0.6938, + "num_input_tokens_seen": 239997130, + "step": 7366 + }, + { + "epoch": 0.66438201740542, + "flos": 19617864386880.0, + "grad_norm": 2.0842148886614162, + "learning_rate": 1.069578898790404e-06, + "loss": 0.8476, + "num_input_tokens_seen": 240024235, + "step": 7367 + }, + { + "epoch": 0.6644722009288903, + "flos": 25698976899360.0, + "grad_norm": 2.6106619944655653, + "learning_rate": 1.0690617952097424e-06, + "loss": 0.7289, + "num_input_tokens_seen": 240052620, + "step": 7368 + }, + { + "epoch": 0.6645623844523606, + "flos": 22383990472800.0, + "grad_norm": 1.9105373384670177, + "learning_rate": 1.068544771061863e-06, + "loss": 0.7934, + "num_input_tokens_seen": 240080125, + "step": 7369 + }, + { + "epoch": 0.6646525679758308, + "flos": 27518202954720.0, + "grad_norm": 2.4766156404491952, + "learning_rate": 1.0680278263908787e-06, + "loss": 0.7848, + "num_input_tokens_seen": 240105985, + "step": 7370 + }, + { + "epoch": 0.6647427514993011, + "flos": 67523837863680.0, + "grad_norm": 0.7308286300545538, + "learning_rate": 1.0675109612408991e-06, + "loss": 0.5836, + "num_input_tokens_seen": 240197690, + "step": 7371 + }, + { + "epoch": 0.6648329350227713, + "flos": 24828174834720.0, + "grad_norm": 2.0391685487494016, + "learning_rate": 1.0669941756560264e-06, + "loss": 0.715, + "num_input_tokens_seen": 240225940, + "step": 7372 + }, + { + "epoch": 0.6649231185462416, + "flos": 21581666573280.0, + "grad_norm": 2.8688271638885197, + "learning_rate": 1.0664774696803548e-06, + "loss": 0.6928, + "num_input_tokens_seen": 240252770, + "step": 7373 + }, + { + "epoch": 0.6650133020697119, + "flos": 23225199674880.0, + "grad_norm": 1.8632529074634556, + "learning_rate": 1.065960843357973e-06, + "loss": 0.7175, + "num_input_tokens_seen": 240281810, + "step": 7374 + }, + { + "epoch": 0.6651034855931821, + "flos": 15938333957760.0, + "grad_norm": 2.0397770239948256, + "learning_rate": 1.065444296732963e-06, + "loss": 0.7605, + "num_input_tokens_seen": 240308345, + "step": 7375 + }, + { + "epoch": 0.6651936691166523, + "flos": 65633568929760.0, + "grad_norm": 0.671305725760997, + "learning_rate": 1.064927829849397e-06, + "loss": 0.5858, + "num_input_tokens_seen": 240398335, + "step": 7376 + }, + { + "epoch": 0.6652838526401227, + "flos": 19945493084160.0, + "grad_norm": 2.276328853642803, + "learning_rate": 1.0644114427513465e-06, + "loss": 0.7404, + "num_input_tokens_seen": 240424060, + "step": 7377 + }, + { + "epoch": 0.6653740361635929, + "flos": 26248246107360.0, + "grad_norm": 2.141366426472666, + "learning_rate": 1.0638951354828693e-06, + "loss": 0.6756, + "num_input_tokens_seen": 240456640, + "step": 7378 + }, + { + "epoch": 0.6654642196870632, + "flos": 25336848931200.0, + "grad_norm": 2.8019590574781668, + "learning_rate": 1.063378908088021e-06, + "loss": 0.7999, + "num_input_tokens_seen": 240485160, + "step": 7379 + }, + { + "epoch": 0.6655544032105334, + "flos": 21112323816480.0, + "grad_norm": 2.3080520123351396, + "learning_rate": 1.0628627606108486e-06, + "loss": 0.7372, + "num_input_tokens_seen": 240512690, + "step": 7380 + }, + { + "epoch": 0.6656445867340037, + "flos": 23695843373280.0, + "grad_norm": 2.536458061787449, + "learning_rate": 1.062346693095393e-06, + "loss": 0.8095, + "num_input_tokens_seen": 240539680, + "step": 7381 + }, + { + "epoch": 0.665734770257474, + "flos": 18742304592960.0, + "grad_norm": 1.8268884242396883, + "learning_rate": 1.0618307055856882e-06, + "loss": 0.6851, + "num_input_tokens_seen": 240568120, + "step": 7382 + }, + { + "epoch": 0.6658249537809442, + "flos": 20164420202400.0, + "grad_norm": 1.722703721134115, + "learning_rate": 1.061314798125759e-06, + "loss": 0.7975, + "num_input_tokens_seen": 240593575, + "step": 7383 + }, + { + "epoch": 0.6659151373044144, + "flos": 26321593511040.0, + "grad_norm": 1.6514771987121744, + "learning_rate": 1.0607989707596293e-06, + "loss": 0.7029, + "num_input_tokens_seen": 240626000, + "step": 7384 + }, + { + "epoch": 0.6660053208278848, + "flos": 25699683124800.0, + "grad_norm": 1.7037423790966268, + "learning_rate": 1.0602832235313078e-06, + "loss": 0.8691, + "num_input_tokens_seen": 240655165, + "step": 7385 + }, + { + "epoch": 0.666095504351355, + "flos": 21872194116480.0, + "grad_norm": 2.6145930397694475, + "learning_rate": 1.0597675564848053e-06, + "loss": 0.736, + "num_input_tokens_seen": 240681565, + "step": 7386 + }, + { + "epoch": 0.6661856878748252, + "flos": 21658470764640.0, + "grad_norm": 2.584353881159224, + "learning_rate": 1.059251969664118e-06, + "loss": 0.813, + "num_input_tokens_seen": 240707900, + "step": 7387 + }, + { + "epoch": 0.6662758713982956, + "flos": 25957458375840.0, + "grad_norm": 2.122257707153806, + "learning_rate": 1.0587364631132402e-06, + "loss": 0.7553, + "num_input_tokens_seen": 240736465, + "step": 7388 + }, + { + "epoch": 0.6663660549217658, + "flos": 26935958436000.0, + "grad_norm": 1.71324740340324, + "learning_rate": 1.0582210368761573e-06, + "loss": 0.7848, + "num_input_tokens_seen": 240765155, + "step": 7389 + }, + { + "epoch": 0.666456238445236, + "flos": 24390766635360.0, + "grad_norm": 3.028764472577338, + "learning_rate": 1.0577056909968485e-06, + "loss": 0.7495, + "num_input_tokens_seen": 240792405, + "step": 7390 + }, + { + "epoch": 0.6665464219687063, + "flos": 28793029040640.0, + "grad_norm": 2.398046505150189, + "learning_rate": 1.0571904255192857e-06, + "loss": 0.7636, + "num_input_tokens_seen": 240822410, + "step": 7391 + }, + { + "epoch": 0.6666366054921766, + "flos": 14662913155680.0, + "grad_norm": 5.651171685992255, + "learning_rate": 1.0566752404874354e-06, + "loss": 0.7759, + "num_input_tokens_seen": 240848155, + "step": 7392 + }, + { + "epoch": 0.6667267890156469, + "flos": 19727569549440.0, + "grad_norm": 13.278796303672033, + "learning_rate": 1.0561601359452543e-06, + "loss": 0.7426, + "num_input_tokens_seen": 240873510, + "step": 7393 + }, + { + "epoch": 0.6668169725391171, + "flos": 35750184553920.0, + "grad_norm": 2.646215716587251, + "learning_rate": 1.0556451119366947e-06, + "loss": 0.6801, + "num_input_tokens_seen": 240903630, + "step": 7394 + }, + { + "epoch": 0.6669071560625873, + "flos": 42785668218720.0, + "grad_norm": 1.8321257934901733, + "learning_rate": 1.0551301685057011e-06, + "loss": 0.6847, + "num_input_tokens_seen": 240936350, + "step": 7395 + }, + { + "epoch": 0.6669973395860577, + "flos": 26279697457920.0, + "grad_norm": 1.8269140683465543, + "learning_rate": 1.0546153056962117e-06, + "loss": 0.8295, + "num_input_tokens_seen": 240964325, + "step": 7396 + }, + { + "epoch": 0.6670875231095279, + "flos": 20783951724000.0, + "grad_norm": 2.0297827483184236, + "learning_rate": 1.0541005235521578e-06, + "loss": 0.7417, + "num_input_tokens_seen": 240993685, + "step": 7397 + }, + { + "epoch": 0.6671777066329981, + "flos": 18962867180640.0, + "grad_norm": 2.023610582644707, + "learning_rate": 1.0535858221174614e-06, + "loss": 0.7577, + "num_input_tokens_seen": 241018955, + "step": 7398 + }, + { + "epoch": 0.6672678901564684, + "flos": 19872703226880.0, + "grad_norm": 1.5761382939615085, + "learning_rate": 1.0530712014360426e-06, + "loss": 0.7827, + "num_input_tokens_seen": 241047145, + "step": 7399 + }, + { + "epoch": 0.6673580736799387, + "flos": 22346257432800.0, + "grad_norm": 1.7832896952177182, + "learning_rate": 1.0525566615518088e-06, + "loss": 0.7967, + "num_input_tokens_seen": 241074460, + "step": 7400 + }, + { + "epoch": 0.667448257203409, + "flos": 23691866208960.0, + "grad_norm": 1.719225384252961, + "learning_rate": 1.0520422025086662e-06, + "loss": 0.7595, + "num_input_tokens_seen": 241102940, + "step": 7401 + }, + { + "epoch": 0.6675384407268792, + "flos": 27124140429120.0, + "grad_norm": 1.9787751730825587, + "learning_rate": 1.0515278243505092e-06, + "loss": 0.7982, + "num_input_tokens_seen": 241132810, + "step": 7402 + }, + { + "epoch": 0.6676286242503494, + "flos": 19254286798080.0, + "grad_norm": 1.4672866999003944, + "learning_rate": 1.0510135271212278e-06, + "loss": 0.7717, + "num_input_tokens_seen": 241161110, + "step": 7403 + }, + { + "epoch": 0.6677188077738198, + "flos": 25737155976480.0, + "grad_norm": 1.6030205765115688, + "learning_rate": 1.0504993108647052e-06, + "loss": 0.8109, + "num_input_tokens_seen": 241192110, + "step": 7404 + }, + { + "epoch": 0.66780899129729, + "flos": 47488122304800.0, + "grad_norm": 1.6162508403542821, + "learning_rate": 1.0499851756248168e-06, + "loss": 0.7168, + "num_input_tokens_seen": 241226180, + "step": 7405 + }, + { + "epoch": 0.6678991748207602, + "flos": 26392747898880.0, + "grad_norm": 2.705564952244804, + "learning_rate": 1.0494711214454316e-06, + "loss": 0.7287, + "num_input_tokens_seen": 241257785, + "step": 7406 + }, + { + "epoch": 0.6679893583442305, + "flos": 28653024790080.0, + "grad_norm": 2.1512269932404933, + "learning_rate": 1.0489571483704111e-06, + "loss": 0.6641, + "num_input_tokens_seen": 241287760, + "step": 7407 + }, + { + "epoch": 0.6680795418677008, + "flos": 29527692509760.0, + "grad_norm": 2.2036712185803964, + "learning_rate": 1.048443256443612e-06, + "loss": 0.6616, + "num_input_tokens_seen": 241316180, + "step": 7408 + }, + { + "epoch": 0.668169725391171, + "flos": 19764410515200.0, + "grad_norm": 1.9636955167228602, + "learning_rate": 1.0479294457088801e-06, + "loss": 0.7631, + "num_input_tokens_seen": 241343990, + "step": 7409 + }, + { + "epoch": 0.6682599089146413, + "flos": 25593509089440.0, + "grad_norm": 1.7384026302863707, + "learning_rate": 1.0474157162100574e-06, + "loss": 0.8679, + "num_input_tokens_seen": 241373725, + "step": 7410 + }, + { + "epoch": 0.6683500924381116, + "flos": 14335804835040.0, + "grad_norm": 1.8987054831916022, + "learning_rate": 1.0469020679909786e-06, + "loss": 0.6834, + "num_input_tokens_seen": 241397790, + "step": 7411 + }, + { + "epoch": 0.6684402759615818, + "flos": 30365890961280.0, + "grad_norm": 2.1388287212391757, + "learning_rate": 1.0463885010954705e-06, + "loss": 0.6716, + "num_input_tokens_seen": 241426580, + "step": 7412 + }, + { + "epoch": 0.6685304594850521, + "flos": 20528220809760.0, + "grad_norm": 1.9355811184376246, + "learning_rate": 1.0458750155673536e-06, + "loss": 0.8161, + "num_input_tokens_seen": 241452850, + "step": 7413 + }, + { + "epoch": 0.6686206430085223, + "flos": 31786110912960.0, + "grad_norm": 1.9985019915019429, + "learning_rate": 1.0453616114504421e-06, + "loss": 0.7008, + "num_input_tokens_seen": 241482335, + "step": 7414 + }, + { + "epoch": 0.6687108265319927, + "flos": 20419630740000.0, + "grad_norm": 2.3384160699796586, + "learning_rate": 1.0448482887885406e-06, + "loss": 0.7794, + "num_input_tokens_seen": 241510355, + "step": 7415 + }, + { + "epoch": 0.6688010100554629, + "flos": 21690554001120.0, + "grad_norm": 2.0709090614927783, + "learning_rate": 1.044335047625451e-06, + "loss": 0.7509, + "num_input_tokens_seen": 241539530, + "step": 7416 + }, + { + "epoch": 0.6688911935789331, + "flos": 34738151636640.0, + "grad_norm": 4.2125130220206275, + "learning_rate": 1.0438218880049637e-06, + "loss": 0.7567, + "num_input_tokens_seen": 241570950, + "step": 7417 + }, + { + "epoch": 0.6689813771024034, + "flos": 22642583458560.0, + "grad_norm": 2.1089885216333215, + "learning_rate": 1.0433088099708653e-06, + "loss": 0.8032, + "num_input_tokens_seen": 241596905, + "step": 7418 + }, + { + "epoch": 0.6690715606258737, + "flos": 67236060882720.0, + "grad_norm": 0.5928267780167096, + "learning_rate": 1.0427958135669346e-06, + "loss": 0.5117, + "num_input_tokens_seen": 241692270, + "step": 7419 + }, + { + "epoch": 0.6691617441493439, + "flos": 24827951816160.0, + "grad_norm": 1.961579662209633, + "learning_rate": 1.0422828988369428e-06, + "loss": 0.6805, + "num_input_tokens_seen": 241722055, + "step": 7420 + }, + { + "epoch": 0.6692519276728142, + "flos": 20893917074880.0, + "grad_norm": 1.6444679664885562, + "learning_rate": 1.041770065824655e-06, + "loss": 0.8159, + "num_input_tokens_seen": 241749695, + "step": 7421 + }, + { + "epoch": 0.6693421111962844, + "flos": 35248201014240.0, + "grad_norm": 1.8802860768766, + "learning_rate": 1.0412573145738287e-06, + "loss": 0.6007, + "num_input_tokens_seen": 241782485, + "step": 7422 + }, + { + "epoch": 0.6694322947197547, + "flos": 22823851876320.0, + "grad_norm": 2.056732668361157, + "learning_rate": 1.040744645128216e-06, + "loss": 0.7217, + "num_input_tokens_seen": 241809065, + "step": 7423 + }, + { + "epoch": 0.669522478243225, + "flos": 24464262718080.0, + "grad_norm": 1.9558263789236017, + "learning_rate": 1.040232057531558e-06, + "loss": 0.8042, + "num_input_tokens_seen": 241838130, + "step": 7424 + }, + { + "epoch": 0.6696126617666952, + "flos": 15682974741120.0, + "grad_norm": 3.4509897664354194, + "learning_rate": 1.0397195518275932e-06, + "loss": 0.7217, + "num_input_tokens_seen": 241863150, + "step": 7425 + }, + { + "epoch": 0.6697028452901654, + "flos": 49526795855040.0, + "grad_norm": 2.151423927559759, + "learning_rate": 1.0392071280600512e-06, + "loss": 0.7166, + "num_input_tokens_seen": 241897530, + "step": 7426 + }, + { + "epoch": 0.6697930288136358, + "flos": 20602162929600.0, + "grad_norm": 2.0983063952974415, + "learning_rate": 1.0386947862726549e-06, + "loss": 0.7462, + "num_input_tokens_seen": 241926005, + "step": 7427 + }, + { + "epoch": 0.669883212337106, + "flos": 41368570526880.0, + "grad_norm": 2.1510202870141613, + "learning_rate": 1.0381825265091197e-06, + "loss": 0.6684, + "num_input_tokens_seen": 241956370, + "step": 7428 + }, + { + "epoch": 0.6699733958605762, + "flos": 22569273224640.0, + "grad_norm": 2.073026670160158, + "learning_rate": 1.037670348813155e-06, + "loss": 0.6669, + "num_input_tokens_seen": 241985650, + "step": 7429 + }, + { + "epoch": 0.6700635793840465, + "flos": 27593706204480.0, + "grad_norm": 1.9230759132822761, + "learning_rate": 1.0371582532284624e-06, + "loss": 0.7768, + "num_input_tokens_seen": 242015130, + "step": 7430 + }, + { + "epoch": 0.6701537629075168, + "flos": 55095480125280.0, + "grad_norm": 1.8936008331672503, + "learning_rate": 1.0366462397987375e-06, + "loss": 0.6791, + "num_input_tokens_seen": 242048155, + "step": 7431 + }, + { + "epoch": 0.6702439464309871, + "flos": 22274471159040.0, + "grad_norm": 2.5679611999753535, + "learning_rate": 1.0361343085676665e-06, + "loss": 0.7537, + "num_input_tokens_seen": 242074750, + "step": 7432 + }, + { + "epoch": 0.6703341299544573, + "flos": 21112472495520.0, + "grad_norm": 1.8472164993707385, + "learning_rate": 1.0356224595789309e-06, + "loss": 0.8627, + "num_input_tokens_seen": 242103190, + "step": 7433 + }, + { + "epoch": 0.6704243134779276, + "flos": 22455925425600.0, + "grad_norm": 1.8688242613667683, + "learning_rate": 1.0351106928762046e-06, + "loss": 0.8038, + "num_input_tokens_seen": 242130835, + "step": 7434 + }, + { + "epoch": 0.6705144970013979, + "flos": 12805470853440.0, + "grad_norm": 2.538392057668365, + "learning_rate": 1.034599008503154e-06, + "loss": 0.818, + "num_input_tokens_seen": 242156770, + "step": 7435 + }, + { + "epoch": 0.6706046805248681, + "flos": 21367943221440.0, + "grad_norm": 2.471052134594895, + "learning_rate": 1.0340874065034406e-06, + "loss": 0.7364, + "num_input_tokens_seen": 242187135, + "step": 7436 + }, + { + "epoch": 0.6706948640483383, + "flos": 22641802893600.0, + "grad_norm": 1.9333066785005335, + "learning_rate": 1.0335758869207137e-06, + "loss": 0.6727, + "num_input_tokens_seen": 242214230, + "step": 7437 + }, + { + "epoch": 0.6707850475718087, + "flos": 20233455913920.0, + "grad_norm": 2.932943340587936, + "learning_rate": 1.0330644497986227e-06, + "loss": 0.7638, + "num_input_tokens_seen": 242242650, + "step": 7438 + }, + { + "epoch": 0.6708752310952789, + "flos": 28106543314080.0, + "grad_norm": 1.767357845174307, + "learning_rate": 1.0325530951808029e-06, + "loss": 0.7854, + "num_input_tokens_seen": 242269885, + "step": 7439 + }, + { + "epoch": 0.6709654146187491, + "flos": 33237075989760.0, + "grad_norm": 1.959117272567137, + "learning_rate": 1.0320418231108887e-06, + "loss": 0.7551, + "num_input_tokens_seen": 242300560, + "step": 7440 + }, + { + "epoch": 0.6710555981422194, + "flos": 22892478720480.0, + "grad_norm": 1.8052095260953696, + "learning_rate": 1.0315306336325028e-06, + "loss": 0.7877, + "num_input_tokens_seen": 242329955, + "step": 7441 + }, + { + "epoch": 0.6711457816656897, + "flos": 18700371370080.0, + "grad_norm": 1.9670069410922097, + "learning_rate": 1.0310195267892635e-06, + "loss": 0.7013, + "num_input_tokens_seen": 242356210, + "step": 7442 + }, + { + "epoch": 0.67123596518916, + "flos": 27705976080480.0, + "grad_norm": 1.9256516852134768, + "learning_rate": 1.030508502624781e-06, + "loss": 0.6918, + "num_input_tokens_seen": 242387210, + "step": 7443 + }, + { + "epoch": 0.6713261487126302, + "flos": 21768213096960.0, + "grad_norm": 1.8412336214350973, + "learning_rate": 1.0299975611826587e-06, + "loss": 0.6958, + "num_input_tokens_seen": 242418745, + "step": 7444 + }, + { + "epoch": 0.6714163322361004, + "flos": 20671310150400.0, + "grad_norm": 1.6329109855774313, + "learning_rate": 1.0294867025064928e-06, + "loss": 0.7545, + "num_input_tokens_seen": 242448095, + "step": 7445 + }, + { + "epoch": 0.6715065157595708, + "flos": 22094838210720.0, + "grad_norm": 1.9486064675492474, + "learning_rate": 1.028975926639874e-06, + "loss": 0.755, + "num_input_tokens_seen": 242479270, + "step": 7446 + }, + { + "epoch": 0.671596699283041, + "flos": 13096890470880.0, + "grad_norm": 2.5887363189232606, + "learning_rate": 1.0284652336263823e-06, + "loss": 0.6654, + "num_input_tokens_seen": 242503245, + "step": 7447 + }, + { + "epoch": 0.6716868828065112, + "flos": 15684424361760.0, + "grad_norm": 1.882936465665831, + "learning_rate": 1.0279546235095938e-06, + "loss": 0.674, + "num_input_tokens_seen": 242531045, + "step": 7448 + }, + { + "epoch": 0.6717770663299815, + "flos": 66412024109760.0, + "grad_norm": 0.7347430666310519, + "learning_rate": 1.0274440963330768e-06, + "loss": 0.608, + "num_input_tokens_seen": 242618995, + "step": 7449 + }, + { + "epoch": 0.6718672498534518, + "flos": 22017104775360.0, + "grad_norm": 1.9226712073992487, + "learning_rate": 1.0269336521403919e-06, + "loss": 0.7471, + "num_input_tokens_seen": 242647095, + "step": 7450 + }, + { + "epoch": 0.671957433376922, + "flos": 26066829010560.0, + "grad_norm": 1.9717726739852026, + "learning_rate": 1.0264232909750936e-06, + "loss": 0.8747, + "num_input_tokens_seen": 242675290, + "step": 7451 + }, + { + "epoch": 0.6720476169003923, + "flos": 26393268275520.0, + "grad_norm": 1.7301020868863732, + "learning_rate": 1.025913012880728e-06, + "loss": 0.8098, + "num_input_tokens_seen": 242706655, + "step": 7452 + }, + { + "epoch": 0.6721378004238625, + "flos": 20855180451360.0, + "grad_norm": 1.5913120397323843, + "learning_rate": 1.0254028179008362e-06, + "loss": 0.6561, + "num_input_tokens_seen": 242734640, + "step": 7453 + }, + { + "epoch": 0.6722279839473329, + "flos": 28432796730240.0, + "grad_norm": 2.2318134696979923, + "learning_rate": 1.0248927060789483e-06, + "loss": 0.6402, + "num_input_tokens_seen": 242762445, + "step": 7454 + }, + { + "epoch": 0.6723181674708031, + "flos": 18306866390880.0, + "grad_norm": 1.664066798244889, + "learning_rate": 1.0243826774585928e-06, + "loss": 0.7053, + "num_input_tokens_seen": 242791010, + "step": 7455 + }, + { + "epoch": 0.6724083509942733, + "flos": 25156881455040.0, + "grad_norm": 2.136122932561213, + "learning_rate": 1.0238727320832854e-06, + "loss": 0.7333, + "num_input_tokens_seen": 242822440, + "step": 7456 + }, + { + "epoch": 0.6724985345177437, + "flos": 23662273346400.0, + "grad_norm": 4.537867062016491, + "learning_rate": 1.0233628699965403e-06, + "loss": 0.7495, + "num_input_tokens_seen": 242850795, + "step": 7457 + }, + { + "epoch": 0.6725887180412139, + "flos": 23039954092800.0, + "grad_norm": 1.785090205266762, + "learning_rate": 1.0228530912418594e-06, + "loss": 0.7701, + "num_input_tokens_seen": 242878360, + "step": 7458 + }, + { + "epoch": 0.6726789015646841, + "flos": 28577187012480.0, + "grad_norm": 3.1415268850892826, + "learning_rate": 1.0223433958627404e-06, + "loss": 0.7181, + "num_input_tokens_seen": 242907180, + "step": 7459 + }, + { + "epoch": 0.6727690850881544, + "flos": 30620729801280.0, + "grad_norm": 3.077304788901149, + "learning_rate": 1.021833783902674e-06, + "loss": 0.7696, + "num_input_tokens_seen": 242934270, + "step": 7460 + }, + { + "epoch": 0.6728592686116247, + "flos": 15173557249440.0, + "grad_norm": 1.9612190682747934, + "learning_rate": 1.0213242554051427e-06, + "loss": 0.7524, + "num_input_tokens_seen": 242960615, + "step": 7461 + }, + { + "epoch": 0.6729494521350949, + "flos": 19726826154240.0, + "grad_norm": 1.7858317826988666, + "learning_rate": 1.0208148104136229e-06, + "loss": 0.6879, + "num_input_tokens_seen": 242989130, + "step": 7462 + }, + { + "epoch": 0.6730396356585652, + "flos": 14918160863040.0, + "grad_norm": 2.0213916702021884, + "learning_rate": 1.020305448971582e-06, + "loss": 0.7202, + "num_input_tokens_seen": 243015710, + "step": 7463 + }, + { + "epoch": 0.6731298191820354, + "flos": 25880654184480.0, + "grad_norm": 1.8567153116265145, + "learning_rate": 1.0197961711224824e-06, + "loss": 0.7122, + "num_input_tokens_seen": 243045525, + "step": 7464 + }, + { + "epoch": 0.6732200027055057, + "flos": 21109313065920.0, + "grad_norm": 2.0656263758298414, + "learning_rate": 1.0192869769097777e-06, + "loss": 0.7577, + "num_input_tokens_seen": 243073830, + "step": 7465 + }, + { + "epoch": 0.673310186228976, + "flos": 20274534232320.0, + "grad_norm": 2.9833707867478867, + "learning_rate": 1.018777866376916e-06, + "loss": 0.7514, + "num_input_tokens_seen": 243101060, + "step": 7466 + }, + { + "epoch": 0.6734003697524462, + "flos": 19253246044800.0, + "grad_norm": 5.055230655407339, + "learning_rate": 1.0182688395673374e-06, + "loss": 0.6159, + "num_input_tokens_seen": 243128770, + "step": 7467 + }, + { + "epoch": 0.6734905532759164, + "flos": 21512928219840.0, + "grad_norm": 1.8179875535570384, + "learning_rate": 1.017759896524475e-06, + "loss": 0.7148, + "num_input_tokens_seen": 243159155, + "step": 7468 + }, + { + "epoch": 0.6735807367993868, + "flos": 26576283672000.0, + "grad_norm": 2.2222414082122786, + "learning_rate": 1.0172510372917528e-06, + "loss": 0.7812, + "num_input_tokens_seen": 243184310, + "step": 7469 + }, + { + "epoch": 0.673670920322857, + "flos": 17317772949120.0, + "grad_norm": 2.5058250276228944, + "learning_rate": 1.0167422619125925e-06, + "loss": 0.6559, + "num_input_tokens_seen": 243209845, + "step": 7470 + }, + { + "epoch": 0.6737611038463273, + "flos": 24318683003520.0, + "grad_norm": 2.5475018437821566, + "learning_rate": 1.0162335704304026e-06, + "loss": 0.7507, + "num_input_tokens_seen": 243239290, + "step": 7471 + }, + { + "epoch": 0.6738512873697975, + "flos": 67190893890720.0, + "grad_norm": 0.6905072288728675, + "learning_rate": 1.0157249628885903e-06, + "loss": 0.5916, + "num_input_tokens_seen": 243325750, + "step": 7472 + }, + { + "epoch": 0.6739414708932678, + "flos": 27925683763680.0, + "grad_norm": 1.3371637082017065, + "learning_rate": 1.0152164393305506e-06, + "loss": 0.6553, + "num_input_tokens_seen": 243357070, + "step": 7473 + }, + { + "epoch": 0.6740316544167381, + "flos": 22566113795040.0, + "grad_norm": 2.2371563439744784, + "learning_rate": 1.0147079997996746e-06, + "loss": 0.7418, + "num_input_tokens_seen": 243384475, + "step": 7474 + }, + { + "epoch": 0.6741218379402083, + "flos": 22423730679840.0, + "grad_norm": 2.195650122260171, + "learning_rate": 1.0141996443393446e-06, + "loss": 0.7776, + "num_input_tokens_seen": 243412110, + "step": 7475 + }, + { + "epoch": 0.6742120214636785, + "flos": 15574905048000.0, + "grad_norm": 1.9738870841045284, + "learning_rate": 1.0136913729929369e-06, + "loss": 0.7315, + "num_input_tokens_seen": 243439365, + "step": 7476 + }, + { + "epoch": 0.6743022049871489, + "flos": 27232544650080.0, + "grad_norm": 1.9041088774389252, + "learning_rate": 1.0131831858038203e-06, + "loss": 0.7041, + "num_input_tokens_seen": 243467720, + "step": 7477 + }, + { + "epoch": 0.6743923885106191, + "flos": 31383722361120.0, + "grad_norm": 1.7918297301615655, + "learning_rate": 1.0126750828153538e-06, + "loss": 0.8078, + "num_input_tokens_seen": 243498795, + "step": 7478 + }, + { + "epoch": 0.6744825720340893, + "flos": 14663582211360.0, + "grad_norm": 2.1666378850736967, + "learning_rate": 1.012167064070895e-06, + "loss": 0.7872, + "num_input_tokens_seen": 243523740, + "step": 7479 + }, + { + "epoch": 0.6745727555575597, + "flos": 21580254122400.0, + "grad_norm": 1.99812330172517, + "learning_rate": 1.0116591296137885e-06, + "loss": 0.7735, + "num_input_tokens_seen": 243550260, + "step": 7480 + }, + { + "epoch": 0.6746629390810299, + "flos": 25993778964960.0, + "grad_norm": 1.8340915571424123, + "learning_rate": 1.0111512794873746e-06, + "loss": 0.7723, + "num_input_tokens_seen": 243583135, + "step": 7481 + }, + { + "epoch": 0.6747531226045002, + "flos": 27737576110080.0, + "grad_norm": 1.5302223338228258, + "learning_rate": 1.010643513734986e-06, + "loss": 0.7462, + "num_input_tokens_seen": 243613885, + "step": 7482 + }, + { + "epoch": 0.6748433061279704, + "flos": 21913383944160.0, + "grad_norm": 1.8009841374361155, + "learning_rate": 1.010135832399948e-06, + "loss": 0.7851, + "num_input_tokens_seen": 243642135, + "step": 7483 + }, + { + "epoch": 0.6749334896514407, + "flos": 20020252938720.0, + "grad_norm": 1.5345156354607903, + "learning_rate": 1.0096282355255792e-06, + "loss": 0.7517, + "num_input_tokens_seen": 243672000, + "step": 7484 + }, + { + "epoch": 0.675023673174911, + "flos": 18197309907360.0, + "grad_norm": 2.498721845047431, + "learning_rate": 1.0091207231551905e-06, + "loss": 0.8037, + "num_input_tokens_seen": 243696860, + "step": 7485 + }, + { + "epoch": 0.6751138566983812, + "flos": 23152038120000.0, + "grad_norm": 1.7839409428858535, + "learning_rate": 1.0086132953320842e-06, + "loss": 0.7114, + "num_input_tokens_seen": 243724985, + "step": 7486 + }, + { + "epoch": 0.6752040402218514, + "flos": 29634981637920.0, + "grad_norm": 2.094018719637759, + "learning_rate": 1.0081059520995591e-06, + "loss": 0.7778, + "num_input_tokens_seen": 243753430, + "step": 7487 + }, + { + "epoch": 0.6752942237453218, + "flos": 25666261776960.0, + "grad_norm": 2.0017206419889, + "learning_rate": 1.0075986935009028e-06, + "loss": 0.8177, + "num_input_tokens_seen": 243781460, + "step": 7488 + }, + { + "epoch": 0.675384407268792, + "flos": 21768361776000.0, + "grad_norm": 2.0508525145727914, + "learning_rate": 1.0070915195793982e-06, + "loss": 0.7585, + "num_input_tokens_seen": 243808790, + "step": 7489 + }, + { + "epoch": 0.6754745907922622, + "flos": 22168185614400.0, + "grad_norm": 2.7991742960205164, + "learning_rate": 1.0065844303783197e-06, + "loss": 0.7229, + "num_input_tokens_seen": 243835660, + "step": 7490 + }, + { + "epoch": 0.6755647743157325, + "flos": 23258695362240.0, + "grad_norm": 1.9279420193940646, + "learning_rate": 1.0060774259409356e-06, + "loss": 0.6645, + "num_input_tokens_seen": 243864545, + "step": 7491 + }, + { + "epoch": 0.6756549578392028, + "flos": 22166921842560.0, + "grad_norm": 2.2679230709383815, + "learning_rate": 1.0055705063105065e-06, + "loss": 0.782, + "num_input_tokens_seen": 243894810, + "step": 7492 + }, + { + "epoch": 0.675745141362673, + "flos": 19217520171840.0, + "grad_norm": 2.186083525769458, + "learning_rate": 1.0050636715302837e-06, + "loss": 0.7398, + "num_input_tokens_seen": 243923255, + "step": 7493 + }, + { + "epoch": 0.6758353248861433, + "flos": 23078988074400.0, + "grad_norm": 2.3599660777224085, + "learning_rate": 1.0045569216435157e-06, + "loss": 0.7603, + "num_input_tokens_seen": 243950790, + "step": 7494 + }, + { + "epoch": 0.6759255084096135, + "flos": 21982568334720.0, + "grad_norm": 1.7233919056371505, + "learning_rate": 1.0040502566934384e-06, + "loss": 0.8269, + "num_input_tokens_seen": 243978815, + "step": 7495 + }, + { + "epoch": 0.6760156919330839, + "flos": 22751210698080.0, + "grad_norm": 2.28139401334009, + "learning_rate": 1.0035436767232866e-06, + "loss": 0.6843, + "num_input_tokens_seen": 244005105, + "step": 7496 + }, + { + "epoch": 0.6761058754565541, + "flos": 23079173923200.0, + "grad_norm": 1.639362321601734, + "learning_rate": 1.0030371817762816e-06, + "loss": 0.7603, + "num_input_tokens_seen": 244032300, + "step": 7497 + }, + { + "epoch": 0.6761960589800243, + "flos": 20930497852320.0, + "grad_norm": 1.727523122210411, + "learning_rate": 1.0025307718956417e-06, + "loss": 0.7546, + "num_input_tokens_seen": 244061180, + "step": 7498 + }, + { + "epoch": 0.6762862425034946, + "flos": 25374730650240.0, + "grad_norm": 1.6718966045944552, + "learning_rate": 1.0020244471245765e-06, + "loss": 0.7695, + "num_input_tokens_seen": 244091370, + "step": 7499 + }, + { + "epoch": 0.6763764260269649, + "flos": 20820904199040.0, + "grad_norm": 2.374359646853227, + "learning_rate": 1.001518207506288e-06, + "loss": 0.7028, + "num_input_tokens_seen": 244116630, + "step": 7500 + }, + { + "epoch": 0.6764666095504351, + "flos": 23442193965600.0, + "grad_norm": 1.6960863026262611, + "learning_rate": 1.0010120530839717e-06, + "loss": 0.8119, + "num_input_tokens_seen": 244145615, + "step": 7501 + }, + { + "epoch": 0.6765567930739054, + "flos": 25775037695520.0, + "grad_norm": 1.9569283646818705, + "learning_rate": 1.0005059839008161e-06, + "loss": 0.7712, + "num_input_tokens_seen": 244177510, + "step": 7502 + }, + { + "epoch": 0.6766469765973756, + "flos": 22715187467040.0, + "grad_norm": 2.09672210181395, + "learning_rate": 1.0000000000000004e-06, + "loss": 0.7001, + "num_input_tokens_seen": 244206740, + "step": 7503 + }, + { + "epoch": 0.676737160120846, + "flos": 19873669640640.0, + "grad_norm": 1.6778906859300053, + "learning_rate": 9.994941014246985e-07, + "loss": 0.793, + "num_input_tokens_seen": 244234315, + "step": 7504 + }, + { + "epoch": 0.6768273436443162, + "flos": 25299636267840.0, + "grad_norm": 1.8231656876274949, + "learning_rate": 9.989882882180766e-07, + "loss": 0.6391, + "num_input_tokens_seen": 244265645, + "step": 7505 + }, + { + "epoch": 0.6769175271677864, + "flos": 18707954001120.0, + "grad_norm": 1.9145761851556753, + "learning_rate": 9.984825604232938e-07, + "loss": 0.7657, + "num_input_tokens_seen": 244291905, + "step": 7506 + }, + { + "epoch": 0.6770077106912568, + "flos": 22926903651840.0, + "grad_norm": 2.503632576636673, + "learning_rate": 9.97976918083502e-07, + "loss": 0.688, + "num_input_tokens_seen": 244320390, + "step": 7507 + }, + { + "epoch": 0.677097894214727, + "flos": 33171645744960.0, + "grad_norm": 1.7792720855761488, + "learning_rate": 9.974713612418427e-07, + "loss": 0.82, + "num_input_tokens_seen": 244353460, + "step": 7508 + }, + { + "epoch": 0.6771880777381972, + "flos": 27378719080800.0, + "grad_norm": 2.0430469388569024, + "learning_rate": 9.969658899414563e-07, + "loss": 0.7608, + "num_input_tokens_seen": 244383820, + "step": 7509 + }, + { + "epoch": 0.6772782612616675, + "flos": 22496037330240.0, + "grad_norm": 1.930188442181945, + "learning_rate": 9.964605042254696e-07, + "loss": 0.6654, + "num_input_tokens_seen": 244411690, + "step": 7510 + }, + { + "epoch": 0.6773684447851378, + "flos": 19650170641920.0, + "grad_norm": 1.895443537397169, + "learning_rate": 9.959552041370076e-07, + "loss": 0.6879, + "num_input_tokens_seen": 244439805, + "step": 7511 + }, + { + "epoch": 0.677458628308608, + "flos": 25447148809920.0, + "grad_norm": 1.9174745004443021, + "learning_rate": 9.954499897191824e-07, + "loss": 0.6661, + "num_input_tokens_seen": 244469160, + "step": 7512 + }, + { + "epoch": 0.6775488118320783, + "flos": 39801432749280.0, + "grad_norm": 4.094081020803964, + "learning_rate": 9.949448610151043e-07, + "loss": 0.7, + "num_input_tokens_seen": 244497165, + "step": 7513 + }, + { + "epoch": 0.6776389953555485, + "flos": 34409445016320.0, + "grad_norm": 8.69186577085802, + "learning_rate": 9.944398180678719e-07, + "loss": 0.7256, + "num_input_tokens_seen": 244530540, + "step": 7514 + }, + { + "epoch": 0.6777291788790188, + "flos": 28726223514720.0, + "grad_norm": 5.176530385054509, + "learning_rate": 9.939348609205789e-07, + "loss": 0.6981, + "num_input_tokens_seen": 244559815, + "step": 7515 + }, + { + "epoch": 0.6778193624024891, + "flos": 25994001983520.0, + "grad_norm": 2.1057560146172976, + "learning_rate": 9.93429989616311e-07, + "loss": 0.7556, + "num_input_tokens_seen": 244590575, + "step": 7516 + }, + { + "epoch": 0.6779095459259593, + "flos": 20820420992160.0, + "grad_norm": 1.656110865668901, + "learning_rate": 9.929252041981464e-07, + "loss": 0.663, + "num_input_tokens_seen": 244619555, + "step": 7517 + }, + { + "epoch": 0.6779997294494295, + "flos": 28283054002560.0, + "grad_norm": 2.3967778431126736, + "learning_rate": 9.924205047091572e-07, + "loss": 0.7195, + "num_input_tokens_seen": 244648860, + "step": 7518 + }, + { + "epoch": 0.6780899129728999, + "flos": 22235400007680.0, + "grad_norm": 4.70944873572366, + "learning_rate": 9.919158911924056e-07, + "loss": 0.8688, + "num_input_tokens_seen": 244675905, + "step": 7519 + }, + { + "epoch": 0.6781800964963701, + "flos": 21731669489280.0, + "grad_norm": 1.7179535841261586, + "learning_rate": 9.914113636909483e-07, + "loss": 0.7821, + "num_input_tokens_seen": 244705985, + "step": 7520 + }, + { + "epoch": 0.6782702800198404, + "flos": 58867792008960.0, + "grad_norm": 0.7343695620071682, + "learning_rate": 9.90906922247835e-07, + "loss": 0.626, + "num_input_tokens_seen": 244799865, + "step": 7521 + }, + { + "epoch": 0.6783604635433106, + "flos": 28506404322240.0, + "grad_norm": 2.345746525117605, + "learning_rate": 9.904025669061072e-07, + "loss": 0.7747, + "num_input_tokens_seen": 244832440, + "step": 7522 + }, + { + "epoch": 0.6784506470667809, + "flos": 31092711611040.0, + "grad_norm": 1.7698617251852289, + "learning_rate": 9.89898297708799e-07, + "loss": 0.7444, + "num_input_tokens_seen": 244864425, + "step": 7523 + }, + { + "epoch": 0.6785408305902512, + "flos": 25957272527040.0, + "grad_norm": 8.370622544192502, + "learning_rate": 9.893941146989388e-07, + "loss": 0.6977, + "num_input_tokens_seen": 244894490, + "step": 7524 + }, + { + "epoch": 0.6786310141137214, + "flos": 38964014862720.0, + "grad_norm": 2.24234603693523, + "learning_rate": 9.888900179195437e-07, + "loss": 0.7117, + "num_input_tokens_seen": 244925005, + "step": 7525 + }, + { + "epoch": 0.6787211976371916, + "flos": 19903968728640.0, + "grad_norm": 2.07193608949107, + "learning_rate": 9.883860074136285e-07, + "loss": 0.8157, + "num_input_tokens_seen": 244952590, + "step": 7526 + }, + { + "epoch": 0.678811381160662, + "flos": 27880107904320.0, + "grad_norm": 2.2790407953833305, + "learning_rate": 9.87882083224196e-07, + "loss": 0.7389, + "num_input_tokens_seen": 244977700, + "step": 7527 + }, + { + "epoch": 0.6789015646841322, + "flos": 68503267167840.0, + "grad_norm": 0.9637916790451236, + "learning_rate": 9.873782453942462e-07, + "loss": 0.6037, + "num_input_tokens_seen": 245072410, + "step": 7528 + }, + { + "epoch": 0.6789917482076024, + "flos": 20820346652640.0, + "grad_norm": 11.744966768142243, + "learning_rate": 9.868744939667676e-07, + "loss": 0.6733, + "num_input_tokens_seen": 245099260, + "step": 7529 + }, + { + "epoch": 0.6790819317310728, + "flos": 25447000130880.0, + "grad_norm": 2.1244259227138595, + "learning_rate": 9.863708289847432e-07, + "loss": 0.6312, + "num_input_tokens_seen": 245127085, + "step": 7530 + }, + { + "epoch": 0.679172115254543, + "flos": 25047585159840.0, + "grad_norm": 2.41730259607127, + "learning_rate": 9.85867250491149e-07, + "loss": 0.655, + "num_input_tokens_seen": 245156495, + "step": 7531 + }, + { + "epoch": 0.6792622987780133, + "flos": 19253840760960.0, + "grad_norm": 2.0694587530655566, + "learning_rate": 9.853637585289528e-07, + "loss": 0.8563, + "num_input_tokens_seen": 245183555, + "step": 7532 + }, + { + "epoch": 0.6793524823014835, + "flos": 28434989746080.0, + "grad_norm": 2.187024637070437, + "learning_rate": 9.848603531411159e-07, + "loss": 0.7342, + "num_input_tokens_seen": 245215595, + "step": 7533 + }, + { + "epoch": 0.6794426658249538, + "flos": 18416348534880.0, + "grad_norm": 2.212042758052418, + "learning_rate": 9.843570343705899e-07, + "loss": 0.757, + "num_input_tokens_seen": 245241840, + "step": 7534 + }, + { + "epoch": 0.6795328493484241, + "flos": 27194105384640.0, + "grad_norm": 1.4803577874705023, + "learning_rate": 9.83853802260323e-07, + "loss": 0.8191, + "num_input_tokens_seen": 245272550, + "step": 7535 + }, + { + "epoch": 0.6796230328718943, + "flos": 25521090929760.0, + "grad_norm": 1.9442458081587124, + "learning_rate": 9.833506568532524e-07, + "loss": 0.6887, + "num_input_tokens_seen": 245301470, + "step": 7536 + }, + { + "epoch": 0.6797132163953645, + "flos": 19363657432800.0, + "grad_norm": 2.069373151616475, + "learning_rate": 9.828475981923093e-07, + "loss": 0.5911, + "num_input_tokens_seen": 245330295, + "step": 7537 + }, + { + "epoch": 0.6798033999188349, + "flos": 20856704411520.0, + "grad_norm": 2.6433064767339878, + "learning_rate": 9.823446263204175e-07, + "loss": 0.7239, + "num_input_tokens_seen": 245356530, + "step": 7538 + }, + { + "epoch": 0.6798935834423051, + "flos": 23589818016960.0, + "grad_norm": 1.824578782684489, + "learning_rate": 9.818417412804937e-07, + "loss": 0.7596, + "num_input_tokens_seen": 245385260, + "step": 7539 + }, + { + "epoch": 0.6799837669657753, + "flos": 22242164904000.0, + "grad_norm": 1.7800884140216207, + "learning_rate": 9.813389431154463e-07, + "loss": 0.8291, + "num_input_tokens_seen": 245412950, + "step": 7540 + }, + { + "epoch": 0.6800739504892456, + "flos": 23808187588800.0, + "grad_norm": 2.218935246860697, + "learning_rate": 9.808362318681783e-07, + "loss": 0.7748, + "num_input_tokens_seen": 245439675, + "step": 7541 + }, + { + "epoch": 0.6801641340127159, + "flos": 16739171066880.0, + "grad_norm": 2.227040771515408, + "learning_rate": 9.803336075815807e-07, + "loss": 0.7689, + "num_input_tokens_seen": 245467570, + "step": 7542 + }, + { + "epoch": 0.6802543175361861, + "flos": 28070928950400.0, + "grad_norm": 1.906022542922252, + "learning_rate": 9.79831070298544e-07, + "loss": 0.6762, + "num_input_tokens_seen": 245499125, + "step": 7543 + }, + { + "epoch": 0.6803445010596564, + "flos": 30729542889600.0, + "grad_norm": 2.3478191401736677, + "learning_rate": 9.793286200619443e-07, + "loss": 0.7161, + "num_input_tokens_seen": 245527045, + "step": 7544 + }, + { + "epoch": 0.6804346845831266, + "flos": 22857682091520.0, + "grad_norm": 1.8256013193132412, + "learning_rate": 9.78826256914655e-07, + "loss": 0.8089, + "num_input_tokens_seen": 245556975, + "step": 7545 + }, + { + "epoch": 0.680524868106597, + "flos": 20529335902560.0, + "grad_norm": 1.7480182325691123, + "learning_rate": 9.7832398089954e-07, + "loss": 0.7936, + "num_input_tokens_seen": 245584420, + "step": 7546 + }, + { + "epoch": 0.6806150516300672, + "flos": 19721994085440.0, + "grad_norm": 2.0279323774421147, + "learning_rate": 9.778217920594565e-07, + "loss": 0.8521, + "num_input_tokens_seen": 245609085, + "step": 7547 + }, + { + "epoch": 0.6807052351535374, + "flos": 19909358343840.0, + "grad_norm": 2.027098418516489, + "learning_rate": 9.773196904372547e-07, + "loss": 0.7893, + "num_input_tokens_seen": 245637200, + "step": 7548 + }, + { + "epoch": 0.6807954186770077, + "flos": 21877323543360.0, + "grad_norm": 1.9969607745211713, + "learning_rate": 9.768176760757742e-07, + "loss": 0.8312, + "num_input_tokens_seen": 245664105, + "step": 7549 + }, + { + "epoch": 0.680885602200478, + "flos": 19618644951840.0, + "grad_norm": 1.8928804421854548, + "learning_rate": 9.76315749017853e-07, + "loss": 0.7285, + "num_input_tokens_seen": 245692905, + "step": 7550 + }, + { + "epoch": 0.6809757857239482, + "flos": 21069424179840.0, + "grad_norm": 1.90221998372174, + "learning_rate": 9.758139093063161e-07, + "loss": 0.751, + "num_input_tokens_seen": 245719735, + "step": 7551 + }, + { + "epoch": 0.6810659692474185, + "flos": 23333232198240.0, + "grad_norm": 2.3498773461656897, + "learning_rate": 9.753121569839834e-07, + "loss": 0.8133, + "num_input_tokens_seen": 245750445, + "step": 7552 + }, + { + "epoch": 0.6811561527708888, + "flos": 58054874727840.0, + "grad_norm": 0.7360829316949355, + "learning_rate": 9.748104920936678e-07, + "loss": 0.548, + "num_input_tokens_seen": 245825130, + "step": 7553 + }, + { + "epoch": 0.681246336294359, + "flos": 29234563083360.0, + "grad_norm": 1.9287909150743858, + "learning_rate": 9.743089146781738e-07, + "loss": 0.7257, + "num_input_tokens_seen": 245854885, + "step": 7554 + }, + { + "epoch": 0.6813365198178293, + "flos": 22861547746560.0, + "grad_norm": 1.6532245569952229, + "learning_rate": 9.738074247802988e-07, + "loss": 0.6733, + "num_input_tokens_seen": 245884235, + "step": 7555 + }, + { + "epoch": 0.6814267033412995, + "flos": 26065416559680.0, + "grad_norm": 2.0568682915002365, + "learning_rate": 9.733060224428325e-07, + "loss": 0.766, + "num_input_tokens_seen": 245911700, + "step": 7556 + }, + { + "epoch": 0.6815168868647699, + "flos": 18489547259520.0, + "grad_norm": 2.1398982464138063, + "learning_rate": 9.728047077085577e-07, + "loss": 0.753, + "num_input_tokens_seen": 245938105, + "step": 7557 + }, + { + "epoch": 0.6816070703882401, + "flos": 24063398126400.0, + "grad_norm": 2.99457695675071, + "learning_rate": 9.723034806202497e-07, + "loss": 0.6545, + "num_input_tokens_seen": 245964915, + "step": 7558 + }, + { + "epoch": 0.6816972539117103, + "flos": 17432347350240.0, + "grad_norm": 3.5154361928467086, + "learning_rate": 9.718023412206748e-07, + "loss": 0.7688, + "num_input_tokens_seen": 245991075, + "step": 7559 + }, + { + "epoch": 0.6817874374351806, + "flos": 25153275988320.0, + "grad_norm": 1.663335793722062, + "learning_rate": 9.713012895525935e-07, + "loss": 0.7689, + "num_input_tokens_seen": 246022440, + "step": 7560 + }, + { + "epoch": 0.6818776209586509, + "flos": 20092559589120.0, + "grad_norm": 2.0871452806808772, + "learning_rate": 9.708003256587584e-07, + "loss": 0.7365, + "num_input_tokens_seen": 246050270, + "step": 7561 + }, + { + "epoch": 0.6819678044821211, + "flos": 29673941280000.0, + "grad_norm": 1.9904836511656812, + "learning_rate": 9.702994495819147e-07, + "loss": 0.7964, + "num_input_tokens_seen": 246081800, + "step": 7562 + }, + { + "epoch": 0.6820579880055914, + "flos": 13743970518240.0, + "grad_norm": 2.542627830330136, + "learning_rate": 9.697986613647999e-07, + "loss": 0.6515, + "num_input_tokens_seen": 246103995, + "step": 7563 + }, + { + "epoch": 0.6821481715290616, + "flos": 63703411109760.0, + "grad_norm": 0.6973872786536859, + "learning_rate": 9.692979610501425e-07, + "loss": 0.6125, + "num_input_tokens_seen": 246192515, + "step": 7564 + }, + { + "epoch": 0.6822383550525319, + "flos": 22672659528000.0, + "grad_norm": 1.850101006817697, + "learning_rate": 9.68797348680668e-07, + "loss": 0.8205, + "num_input_tokens_seen": 246218545, + "step": 7565 + }, + { + "epoch": 0.6823285385760022, + "flos": 22239897548640.0, + "grad_norm": 1.6358051527060393, + "learning_rate": 9.682968242990878e-07, + "loss": 0.717, + "num_input_tokens_seen": 246248645, + "step": 7566 + }, + { + "epoch": 0.6824187220994724, + "flos": 28504471494720.0, + "grad_norm": 1.844321275560069, + "learning_rate": 9.677963879481132e-07, + "loss": 0.7975, + "num_input_tokens_seen": 246277815, + "step": 7567 + }, + { + "epoch": 0.6825089056229426, + "flos": 16667087435040.0, + "grad_norm": 2.23049044890797, + "learning_rate": 9.672960396704416e-07, + "loss": 0.7378, + "num_input_tokens_seen": 246303630, + "step": 7568 + }, + { + "epoch": 0.682599089146413, + "flos": 21253852027200.0, + "grad_norm": 1.9850834460459095, + "learning_rate": 9.667957795087657e-07, + "loss": 0.7522, + "num_input_tokens_seen": 246331810, + "step": 7569 + }, + { + "epoch": 0.6826892726698832, + "flos": 19764447684960.0, + "grad_norm": 2.495569690499378, + "learning_rate": 9.662956075057712e-07, + "loss": 0.8094, + "num_input_tokens_seen": 246359895, + "step": 7570 + }, + { + "epoch": 0.6827794561933535, + "flos": 20559337632480.0, + "grad_norm": 4.068183796244266, + "learning_rate": 9.657955237041354e-07, + "loss": 0.8156, + "num_input_tokens_seen": 246384480, + "step": 7571 + }, + { + "epoch": 0.6828696397168237, + "flos": 21214297668960.0, + "grad_norm": 1.748260587276442, + "learning_rate": 9.652955281465278e-07, + "loss": 0.726, + "num_input_tokens_seen": 246412820, + "step": 7572 + }, + { + "epoch": 0.682959823240294, + "flos": 17504951358720.0, + "grad_norm": 2.1142458591607824, + "learning_rate": 9.64795620875612e-07, + "loss": 0.7541, + "num_input_tokens_seen": 246440500, + "step": 7573 + }, + { + "epoch": 0.6830500067637643, + "flos": 28356921782880.0, + "grad_norm": 2.164732460952725, + "learning_rate": 9.64295801934041e-07, + "loss": 0.6944, + "num_input_tokens_seen": 246471675, + "step": 7574 + }, + { + "epoch": 0.6831401902872345, + "flos": 65082850101120.0, + "grad_norm": 0.6690801999611874, + "learning_rate": 9.63796071364463e-07, + "loss": 0.5866, + "num_input_tokens_seen": 246554080, + "step": 7575 + }, + { + "epoch": 0.6832303738107048, + "flos": 19108484064960.0, + "grad_norm": 1.9672020776040369, + "learning_rate": 9.632964292095179e-07, + "loss": 0.8345, + "num_input_tokens_seen": 246581630, + "step": 7576 + }, + { + "epoch": 0.6833205573341751, + "flos": 21440175532320.0, + "grad_norm": 2.286667127062396, + "learning_rate": 9.627968755118374e-07, + "loss": 0.7361, + "num_input_tokens_seen": 246610060, + "step": 7577 + }, + { + "epoch": 0.6834107408576453, + "flos": 56167839563040.0, + "grad_norm": 0.6677770506857864, + "learning_rate": 9.622974103140468e-07, + "loss": 0.5662, + "num_input_tokens_seen": 246691495, + "step": 7578 + }, + { + "epoch": 0.6835009243811155, + "flos": 24241915981920.0, + "grad_norm": 1.6028128604512917, + "learning_rate": 9.617980336587632e-07, + "loss": 0.8113, + "num_input_tokens_seen": 246721220, + "step": 7579 + }, + { + "epoch": 0.6835911079045859, + "flos": 26612715770400.0, + "grad_norm": 2.27708891658209, + "learning_rate": 9.612987455885964e-07, + "loss": 0.629, + "num_input_tokens_seen": 246752900, + "step": 7580 + }, + { + "epoch": 0.6836812914280561, + "flos": 30292989594720.0, + "grad_norm": 1.7760437623251206, + "learning_rate": 9.607995461461467e-07, + "loss": 0.7541, + "num_input_tokens_seen": 246782490, + "step": 7581 + }, + { + "epoch": 0.6837714749515263, + "flos": 40347914225280.0, + "grad_norm": 1.9117603392572722, + "learning_rate": 9.603004353740111e-07, + "loss": 0.6501, + "num_input_tokens_seen": 246815795, + "step": 7582 + }, + { + "epoch": 0.6838616584749966, + "flos": 62886330815520.0, + "grad_norm": 0.7629538279895786, + "learning_rate": 9.598014133147738e-07, + "loss": 0.6217, + "num_input_tokens_seen": 246895975, + "step": 7583 + }, + { + "epoch": 0.6839518419984669, + "flos": 26797589654880.0, + "grad_norm": 1.5415375457853493, + "learning_rate": 9.59302480011017e-07, + "loss": 0.7089, + "num_input_tokens_seen": 246929010, + "step": 7584 + }, + { + "epoch": 0.6840420255219372, + "flos": 63809362126560.0, + "grad_norm": 0.7029255054176702, + "learning_rate": 9.588036355053102e-07, + "loss": 0.5981, + "num_input_tokens_seen": 247011820, + "step": 7585 + }, + { + "epoch": 0.6841322090454074, + "flos": 20492755125120.0, + "grad_norm": 19.82341015900219, + "learning_rate": 9.583048798402182e-07, + "loss": 0.6152, + "num_input_tokens_seen": 247039690, + "step": 7586 + }, + { + "epoch": 0.6842223925688776, + "flos": 19327522692480.0, + "grad_norm": 1.7292730536154697, + "learning_rate": 9.57806213058298e-07, + "loss": 0.7413, + "num_input_tokens_seen": 247068345, + "step": 7587 + }, + { + "epoch": 0.684312576092348, + "flos": 60470283961920.0, + "grad_norm": 0.6886737946793517, + "learning_rate": 9.57307635202098e-07, + "loss": 0.6156, + "num_input_tokens_seen": 247162105, + "step": 7588 + }, + { + "epoch": 0.6844027596158182, + "flos": 19035248170560.0, + "grad_norm": 2.0630982427240006, + "learning_rate": 9.568091463141607e-07, + "loss": 0.8518, + "num_input_tokens_seen": 247188995, + "step": 7589 + }, + { + "epoch": 0.6844929431392884, + "flos": 22278299644320.0, + "grad_norm": 1.6909021062324479, + "learning_rate": 9.563107464370187e-07, + "loss": 0.7116, + "num_input_tokens_seen": 247216900, + "step": 7590 + }, + { + "epoch": 0.6845831266627587, + "flos": 19758091656000.0, + "grad_norm": 2.045322674054981, + "learning_rate": 9.558124356131982e-07, + "loss": 0.7662, + "num_input_tokens_seen": 247244260, + "step": 7591 + }, + { + "epoch": 0.684673310186229, + "flos": 23990310911040.0, + "grad_norm": 8.125381695511535, + "learning_rate": 9.553142138852187e-07, + "loss": 0.6803, + "num_input_tokens_seen": 247272185, + "step": 7592 + }, + { + "epoch": 0.6847634937096992, + "flos": 24682149083040.0, + "grad_norm": 7.522404077843171, + "learning_rate": 9.548160812955905e-07, + "loss": 0.6563, + "num_input_tokens_seen": 247302135, + "step": 7593 + }, + { + "epoch": 0.6848536772331695, + "flos": 27269422785600.0, + "grad_norm": 1.8855260392249524, + "learning_rate": 9.543180378868175e-07, + "loss": 0.5696, + "num_input_tokens_seen": 247332525, + "step": 7594 + }, + { + "epoch": 0.6849438607566397, + "flos": 23909975592480.0, + "grad_norm": 12.490612930395095, + "learning_rate": 9.538200837013962e-07, + "loss": 0.784, + "num_input_tokens_seen": 247359800, + "step": 7595 + }, + { + "epoch": 0.68503404428011, + "flos": 64395695318880.0, + "grad_norm": 0.6592974748279868, + "learning_rate": 9.533222187818122e-07, + "loss": 0.6132, + "num_input_tokens_seen": 247449695, + "step": 7596 + }, + { + "epoch": 0.6851242278035803, + "flos": 16849693964160.0, + "grad_norm": 2.436404935902124, + "learning_rate": 9.528244431705492e-07, + "loss": 0.7568, + "num_input_tokens_seen": 247475495, + "step": 7597 + }, + { + "epoch": 0.6852144113270505, + "flos": 23768967758400.0, + "grad_norm": 1.8280400313495409, + "learning_rate": 9.523267569100774e-07, + "loss": 0.7558, + "num_input_tokens_seen": 247505000, + "step": 7598 + }, + { + "epoch": 0.6853045948505209, + "flos": 24786464630400.0, + "grad_norm": 1.840916541541507, + "learning_rate": 9.518291600428652e-07, + "loss": 0.7017, + "num_input_tokens_seen": 247533200, + "step": 7599 + }, + { + "epoch": 0.6853947783739911, + "flos": 17833286281440.0, + "grad_norm": 2.08794143943222, + "learning_rate": 9.513316526113677e-07, + "loss": 0.7729, + "num_input_tokens_seen": 247560100, + "step": 7600 + }, + { + "epoch": 0.6854849618974613, + "flos": 21732078356640.0, + "grad_norm": 2.0967231252050262, + "learning_rate": 9.50834234658036e-07, + "loss": 0.7719, + "num_input_tokens_seen": 247586475, + "step": 7601 + }, + { + "epoch": 0.6855751454209316, + "flos": 26171070218400.0, + "grad_norm": 1.6827751700990263, + "learning_rate": 9.503369062253123e-07, + "loss": 0.762, + "num_input_tokens_seen": 247616170, + "step": 7602 + }, + { + "epoch": 0.6856653289444019, + "flos": 66368046550080.0, + "grad_norm": 0.6370317356836649, + "learning_rate": 9.498396673556317e-07, + "loss": 0.5893, + "num_input_tokens_seen": 247705415, + "step": 7603 + }, + { + "epoch": 0.6857555124678721, + "flos": 24969182668800.0, + "grad_norm": 1.9377659645884628, + "learning_rate": 9.493425180914219e-07, + "loss": 0.7828, + "num_input_tokens_seen": 247735405, + "step": 7604 + }, + { + "epoch": 0.6858456959913424, + "flos": 25702136328960.0, + "grad_norm": 1.826886662417605, + "learning_rate": 9.488454584751e-07, + "loss": 0.6786, + "num_input_tokens_seen": 247765840, + "step": 7605 + }, + { + "epoch": 0.6859358795148126, + "flos": 48109029107520.0, + "grad_norm": 1.7292427904156042, + "learning_rate": 9.483484885490813e-07, + "loss": 0.7516, + "num_input_tokens_seen": 247798695, + "step": 7606 + }, + { + "epoch": 0.686026063038283, + "flos": 25193016195360.0, + "grad_norm": 1.9131607111152935, + "learning_rate": 9.478516083557675e-07, + "loss": 0.7609, + "num_input_tokens_seen": 247824835, + "step": 7607 + }, + { + "epoch": 0.6861162465617532, + "flos": 27084548901120.0, + "grad_norm": 2.398254586368002, + "learning_rate": 9.473548179375561e-07, + "loss": 0.7652, + "num_input_tokens_seen": 247852690, + "step": 7608 + }, + { + "epoch": 0.6862064300852234, + "flos": 60742858511040.0, + "grad_norm": 0.6901738665428669, + "learning_rate": 9.468581173368358e-07, + "loss": 0.6164, + "num_input_tokens_seen": 247936530, + "step": 7609 + }, + { + "epoch": 0.6862966136086937, + "flos": 25407222754080.0, + "grad_norm": 1.5802895082744042, + "learning_rate": 9.463615065959878e-07, + "loss": 0.7553, + "num_input_tokens_seen": 247967290, + "step": 7610 + }, + { + "epoch": 0.686386797132164, + "flos": 22636078750560.0, + "grad_norm": 4.141684572934613, + "learning_rate": 9.458649857573857e-07, + "loss": 0.8046, + "num_input_tokens_seen": 247995235, + "step": 7611 + }, + { + "epoch": 0.6864769806556342, + "flos": 24719064388320.0, + "grad_norm": 1.7913058013636374, + "learning_rate": 9.453685548633963e-07, + "loss": 0.7589, + "num_input_tokens_seen": 248023160, + "step": 7612 + }, + { + "epoch": 0.6865671641791045, + "flos": 39365213982240.0, + "grad_norm": 1.9337714836046715, + "learning_rate": 9.448722139563756e-07, + "loss": 0.6967, + "num_input_tokens_seen": 248054895, + "step": 7613 + }, + { + "epoch": 0.6866573477025747, + "flos": 29965249388160.0, + "grad_norm": 1.7260226963668273, + "learning_rate": 9.443759630786769e-07, + "loss": 0.6429, + "num_input_tokens_seen": 248087815, + "step": 7614 + }, + { + "epoch": 0.686747531226045, + "flos": 23442305474880.0, + "grad_norm": 3.339950113666338, + "learning_rate": 9.438798022726408e-07, + "loss": 0.8216, + "num_input_tokens_seen": 248115170, + "step": 7615 + }, + { + "epoch": 0.6868377147495153, + "flos": 20346506354880.0, + "grad_norm": 1.7835637508213769, + "learning_rate": 9.433837315806037e-07, + "loss": 0.7008, + "num_input_tokens_seen": 248141340, + "step": 7616 + }, + { + "epoch": 0.6869278982729855, + "flos": 32114520175200.0, + "grad_norm": 3.3553078804932923, + "learning_rate": 9.428877510448925e-07, + "loss": 0.7052, + "num_input_tokens_seen": 248168170, + "step": 7617 + }, + { + "epoch": 0.6870180817964557, + "flos": 21473522540640.0, + "grad_norm": 1.776612183640245, + "learning_rate": 9.423918607078272e-07, + "loss": 0.6767, + "num_input_tokens_seen": 248197185, + "step": 7618 + }, + { + "epoch": 0.6871082653199261, + "flos": 21256491080160.0, + "grad_norm": 2.0564756553687347, + "learning_rate": 9.418960606117208e-07, + "loss": 0.76, + "num_input_tokens_seen": 248224545, + "step": 7619 + }, + { + "epoch": 0.6871984488433963, + "flos": 23297580664800.0, + "grad_norm": 2.702306722810256, + "learning_rate": 9.414003507988752e-07, + "loss": 0.6905, + "num_input_tokens_seen": 248252685, + "step": 7620 + }, + { + "epoch": 0.6872886323668665, + "flos": 14481124361280.0, + "grad_norm": 2.678939581502642, + "learning_rate": 9.409047313115904e-07, + "loss": 0.6832, + "num_input_tokens_seen": 248279390, + "step": 7621 + }, + { + "epoch": 0.6873788158903368, + "flos": 25482540155040.0, + "grad_norm": 2.671941031370653, + "learning_rate": 9.404092021921521e-07, + "loss": 0.6863, + "num_input_tokens_seen": 248309095, + "step": 7622 + }, + { + "epoch": 0.6874689994138071, + "flos": 23807704381920.0, + "grad_norm": 2.0208063287667595, + "learning_rate": 9.399137634828447e-07, + "loss": 0.73, + "num_input_tokens_seen": 248337845, + "step": 7623 + }, + { + "epoch": 0.6875591829372774, + "flos": 20129066027040.0, + "grad_norm": 1.7589342395229175, + "learning_rate": 9.394184152259396e-07, + "loss": 0.7699, + "num_input_tokens_seen": 248363840, + "step": 7624 + }, + { + "epoch": 0.6876493664607476, + "flos": 23661641460480.0, + "grad_norm": 2.832277568760072, + "learning_rate": 9.389231574637033e-07, + "loss": 0.8516, + "num_input_tokens_seen": 248394805, + "step": 7625 + }, + { + "epoch": 0.6877395499842179, + "flos": 19799578841760.0, + "grad_norm": 2.069668536530605, + "learning_rate": 9.384279902383938e-07, + "loss": 0.8366, + "num_input_tokens_seen": 248419525, + "step": 7626 + }, + { + "epoch": 0.6878297335076882, + "flos": 29814688925760.0, + "grad_norm": 1.8951716333734097, + "learning_rate": 9.379329135922615e-07, + "loss": 0.7651, + "num_input_tokens_seen": 248448620, + "step": 7627 + }, + { + "epoch": 0.6879199170311584, + "flos": 36340792268640.0, + "grad_norm": 2.8998511637937145, + "learning_rate": 9.374379275675495e-07, + "loss": 0.7047, + "num_input_tokens_seen": 248480400, + "step": 7628 + }, + { + "epoch": 0.6880101005546286, + "flos": 24422292325440.0, + "grad_norm": 2.044671306543068, + "learning_rate": 9.369430322064931e-07, + "loss": 0.7468, + "num_input_tokens_seen": 248508395, + "step": 7629 + }, + { + "epoch": 0.688100284078099, + "flos": 19545483396960.0, + "grad_norm": 3.150076692646651, + "learning_rate": 9.364482275513179e-07, + "loss": 0.8262, + "num_input_tokens_seen": 248535155, + "step": 7630 + }, + { + "epoch": 0.6881904676015692, + "flos": 20201484186720.0, + "grad_norm": 1.688735156668169, + "learning_rate": 9.359535136442444e-07, + "loss": 0.8147, + "num_input_tokens_seen": 248562285, + "step": 7631 + }, + { + "epoch": 0.6882806511250394, + "flos": 23733390564480.0, + "grad_norm": 2.2126734089110287, + "learning_rate": 9.354588905274843e-07, + "loss": 0.6836, + "num_input_tokens_seen": 248591190, + "step": 7632 + }, + { + "epoch": 0.6883708346485097, + "flos": 23443271888640.0, + "grad_norm": 2.0480637920174987, + "learning_rate": 9.349643582432414e-07, + "loss": 0.7804, + "num_input_tokens_seen": 248620035, + "step": 7633 + }, + { + "epoch": 0.68846101817198, + "flos": 23626250115360.0, + "grad_norm": 1.600183851884765, + "learning_rate": 9.344699168337127e-07, + "loss": 0.7379, + "num_input_tokens_seen": 248647180, + "step": 7634 + }, + { + "epoch": 0.6885512016954503, + "flos": 19181682789600.0, + "grad_norm": 1.8894534303650323, + "learning_rate": 9.339755663410845e-07, + "loss": 0.7094, + "num_input_tokens_seen": 248675000, + "step": 7635 + }, + { + "epoch": 0.6886413852189205, + "flos": 19217966208960.0, + "grad_norm": 1.8471574493116734, + "learning_rate": 9.334813068075405e-07, + "loss": 0.7802, + "num_input_tokens_seen": 248703170, + "step": 7636 + }, + { + "epoch": 0.6887315687423907, + "flos": 24391361351520.0, + "grad_norm": 1.57867709030005, + "learning_rate": 9.329871382752506e-07, + "loss": 0.7515, + "num_input_tokens_seen": 248731885, + "step": 7637 + }, + { + "epoch": 0.6888217522658611, + "flos": 25191009028320.0, + "grad_norm": 1.890878770679359, + "learning_rate": 9.32493060786383e-07, + "loss": 0.7379, + "num_input_tokens_seen": 248761040, + "step": 7638 + }, + { + "epoch": 0.6889119357893313, + "flos": 34333384220160.0, + "grad_norm": 1.5558357024349299, + "learning_rate": 9.31999074383093e-07, + "loss": 0.8066, + "num_input_tokens_seen": 248793120, + "step": 7639 + }, + { + "epoch": 0.6890021193128015, + "flos": 24718767030240.0, + "grad_norm": 3.141934705341114, + "learning_rate": 9.315051791075308e-07, + "loss": 0.7397, + "num_input_tokens_seen": 248821120, + "step": 7640 + }, + { + "epoch": 0.6890923028362718, + "flos": 24497089349760.0, + "grad_norm": 1.5986099194918928, + "learning_rate": 9.310113750018382e-07, + "loss": 0.7405, + "num_input_tokens_seen": 248850545, + "step": 7641 + }, + { + "epoch": 0.6891824863597421, + "flos": 16521210362400.0, + "grad_norm": 2.225315534527733, + "learning_rate": 9.305176621081496e-07, + "loss": 0.8146, + "num_input_tokens_seen": 248876710, + "step": 7642 + }, + { + "epoch": 0.6892726698832123, + "flos": 26030619930720.0, + "grad_norm": 2.426205903506988, + "learning_rate": 9.300240404685911e-07, + "loss": 0.7819, + "num_input_tokens_seen": 248902805, + "step": 7643 + }, + { + "epoch": 0.6893628534066826, + "flos": 24100053243360.0, + "grad_norm": 1.8548671028055985, + "learning_rate": 9.295305101252812e-07, + "loss": 0.7492, + "num_input_tokens_seen": 248930955, + "step": 7644 + }, + { + "epoch": 0.6894530369301528, + "flos": 22495962990720.0, + "grad_norm": 2.0060529903285467, + "learning_rate": 9.290370711203314e-07, + "loss": 0.7289, + "num_input_tokens_seen": 248956975, + "step": 7645 + }, + { + "epoch": 0.6895432204536232, + "flos": 20559858009120.0, + "grad_norm": 1.7556667540471567, + "learning_rate": 9.285437234958433e-07, + "loss": 0.7434, + "num_input_tokens_seen": 248984040, + "step": 7646 + }, + { + "epoch": 0.6896334039770934, + "flos": 27811146532320.0, + "grad_norm": 2.1497761942570413, + "learning_rate": 9.280504672939124e-07, + "loss": 0.7378, + "num_input_tokens_seen": 249011100, + "step": 7647 + }, + { + "epoch": 0.6897235875005636, + "flos": 25335696668640.0, + "grad_norm": 1.8636978931165993, + "learning_rate": 9.275573025566266e-07, + "loss": 0.788, + "num_input_tokens_seen": 249040900, + "step": 7648 + }, + { + "epoch": 0.689813771024034, + "flos": 27995500040160.0, + "grad_norm": 1.6580110939722283, + "learning_rate": 9.27064229326065e-07, + "loss": 0.7823, + "num_input_tokens_seen": 249071880, + "step": 7649 + }, + { + "epoch": 0.6899039545475042, + "flos": 24209052180480.0, + "grad_norm": 2.4519326780887845, + "learning_rate": 9.265712476442995e-07, + "loss": 0.7172, + "num_input_tokens_seen": 249099730, + "step": 7650 + }, + { + "epoch": 0.6899941380709744, + "flos": 24683264175840.0, + "grad_norm": 1.60969281067784, + "learning_rate": 9.260783575533949e-07, + "loss": 0.6765, + "num_input_tokens_seen": 249129495, + "step": 7651 + }, + { + "epoch": 0.6900843215944447, + "flos": 21804533686080.0, + "grad_norm": 2.2073202028420695, + "learning_rate": 9.255855590954045e-07, + "loss": 0.6972, + "num_input_tokens_seen": 249158030, + "step": 7652 + }, + { + "epoch": 0.690174505117915, + "flos": 22934226094560.0, + "grad_norm": 2.247511122933807, + "learning_rate": 9.250928523123802e-07, + "loss": 0.8168, + "num_input_tokens_seen": 249185785, + "step": 7653 + }, + { + "epoch": 0.6902646886413852, + "flos": 26577250085760.0, + "grad_norm": 10.949561917673458, + "learning_rate": 9.24600237246359e-07, + "loss": 0.7578, + "num_input_tokens_seen": 249217495, + "step": 7654 + }, + { + "epoch": 0.6903548721648555, + "flos": 20784620779680.0, + "grad_norm": 2.2549095631330616, + "learning_rate": 9.241077139393769e-07, + "loss": 0.755, + "num_input_tokens_seen": 249245150, + "step": 7655 + }, + { + "epoch": 0.6904450556883257, + "flos": 22387038393120.0, + "grad_norm": 2.1644906952768785, + "learning_rate": 9.236152824334564e-07, + "loss": 0.7798, + "num_input_tokens_seen": 249274770, + "step": 7656 + }, + { + "epoch": 0.690535239211796, + "flos": 26504014191360.0, + "grad_norm": 1.8104871556907054, + "learning_rate": 9.231229427706151e-07, + "loss": 0.6475, + "num_input_tokens_seen": 249303830, + "step": 7657 + }, + { + "epoch": 0.6906254227352663, + "flos": 23152372647840.0, + "grad_norm": 2.713082819612752, + "learning_rate": 9.226306949928622e-07, + "loss": 0.7548, + "num_input_tokens_seen": 249333285, + "step": 7658 + }, + { + "epoch": 0.6907156062587365, + "flos": 16593145315200.0, + "grad_norm": 2.6308908212845488, + "learning_rate": 9.221385391421988e-07, + "loss": 0.8225, + "num_input_tokens_seen": 249358285, + "step": 7659 + }, + { + "epoch": 0.6908057897822067, + "flos": 23589520658880.0, + "grad_norm": 1.8569376381472193, + "learning_rate": 9.216464752606192e-07, + "loss": 0.7639, + "num_input_tokens_seen": 249388335, + "step": 7660 + }, + { + "epoch": 0.6908959733056771, + "flos": 17755441336800.0, + "grad_norm": 1.6105636355249418, + "learning_rate": 9.211545033901078e-07, + "loss": 0.8283, + "num_input_tokens_seen": 249414815, + "step": 7661 + }, + { + "epoch": 0.6909861568291473, + "flos": 21578916011040.0, + "grad_norm": 2.3850123522296665, + "learning_rate": 9.206626235726426e-07, + "loss": 0.7761, + "num_input_tokens_seen": 249441955, + "step": 7662 + }, + { + "epoch": 0.6910763403526176, + "flos": 27050235479040.0, + "grad_norm": 4.086140556468512, + "learning_rate": 9.20170835850194e-07, + "loss": 0.686, + "num_input_tokens_seen": 249469385, + "step": 7663 + }, + { + "epoch": 0.6911665238760878, + "flos": 68590293043200.0, + "grad_norm": 0.6372196123189265, + "learning_rate": 9.196791402647237e-07, + "loss": 0.5726, + "num_input_tokens_seen": 249565480, + "step": 7664 + }, + { + "epoch": 0.6912567073995581, + "flos": 25191083367840.0, + "grad_norm": 1.7816478887174474, + "learning_rate": 9.191875368581861e-07, + "loss": 0.672, + "num_input_tokens_seen": 249596255, + "step": 7665 + }, + { + "epoch": 0.6913468909230284, + "flos": 31457032595040.0, + "grad_norm": 2.2600820150094316, + "learning_rate": 9.186960256725271e-07, + "loss": 0.7575, + "num_input_tokens_seen": 249626345, + "step": 7666 + }, + { + "epoch": 0.6914370744464986, + "flos": 22746899005920.0, + "grad_norm": 1.7254450016724343, + "learning_rate": 9.182046067496856e-07, + "loss": 0.8121, + "num_input_tokens_seen": 249655115, + "step": 7667 + }, + { + "epoch": 0.6915272579699688, + "flos": 26867926308000.0, + "grad_norm": 2.1946113525633755, + "learning_rate": 9.177132801315927e-07, + "loss": 0.834, + "num_input_tokens_seen": 249683675, + "step": 7668 + }, + { + "epoch": 0.6916174414934392, + "flos": 19139229190080.0, + "grad_norm": 1.6751595435382785, + "learning_rate": 9.172220458601692e-07, + "loss": 0.7239, + "num_input_tokens_seen": 249708490, + "step": 7669 + }, + { + "epoch": 0.6917076250169094, + "flos": 38964200711520.0, + "grad_norm": 2.401394887720218, + "learning_rate": 9.167309039773324e-07, + "loss": 0.7161, + "num_input_tokens_seen": 249737660, + "step": 7670 + }, + { + "epoch": 0.6917978085403796, + "flos": 33132128556480.0, + "grad_norm": 1.430776629397255, + "learning_rate": 9.162398545249872e-07, + "loss": 0.6476, + "num_input_tokens_seen": 249770015, + "step": 7671 + }, + { + "epoch": 0.69188799206385, + "flos": 26539888743360.0, + "grad_norm": 1.4501879142803713, + "learning_rate": 9.157488975450334e-07, + "loss": 0.8027, + "num_input_tokens_seen": 249803500, + "step": 7672 + }, + { + "epoch": 0.6919781755873202, + "flos": 41946689202240.0, + "grad_norm": 1.6865088059630589, + "learning_rate": 9.15258033079362e-07, + "loss": 0.582, + "num_input_tokens_seen": 249836790, + "step": 7673 + }, + { + "epoch": 0.6920683591107905, + "flos": 22679461594080.0, + "grad_norm": 1.7986041166676041, + "learning_rate": 9.147672611698567e-07, + "loss": 0.7378, + "num_input_tokens_seen": 249867045, + "step": 7674 + }, + { + "epoch": 0.6921585426342607, + "flos": 17576477444160.0, + "grad_norm": 3.135777162293274, + "learning_rate": 9.142765818583933e-07, + "loss": 0.7376, + "num_input_tokens_seen": 249893505, + "step": 7675 + }, + { + "epoch": 0.692248726157731, + "flos": 26103818655360.0, + "grad_norm": 1.6195478674941906, + "learning_rate": 9.13785995186837e-07, + "loss": 0.7266, + "num_input_tokens_seen": 249922725, + "step": 7676 + }, + { + "epoch": 0.6923389096812013, + "flos": 25010372496480.0, + "grad_norm": 1.8005087557354396, + "learning_rate": 9.132955011970502e-07, + "loss": 0.7139, + "num_input_tokens_seen": 249951710, + "step": 7677 + }, + { + "epoch": 0.6924290932046715, + "flos": 39614960565120.0, + "grad_norm": 1.3906817478216809, + "learning_rate": 9.128050999308827e-07, + "loss": 0.738, + "num_input_tokens_seen": 249983275, + "step": 7678 + }, + { + "epoch": 0.6925192767281417, + "flos": 24755310637920.0, + "grad_norm": 2.2909131169455796, + "learning_rate": 9.123147914301789e-07, + "loss": 0.8386, + "num_input_tokens_seen": 250009940, + "step": 7679 + }, + { + "epoch": 0.6926094602516121, + "flos": 16265776806240.0, + "grad_norm": 1.7671933255828975, + "learning_rate": 9.118245757367745e-07, + "loss": 0.7918, + "num_input_tokens_seen": 250036300, + "step": 7680 + }, + { + "epoch": 0.6926996437750823, + "flos": 65434384687680.0, + "grad_norm": 0.6222168227393834, + "learning_rate": 9.113344528924973e-07, + "loss": 0.5592, + "num_input_tokens_seen": 250118870, + "step": 7681 + }, + { + "epoch": 0.6927898272985525, + "flos": 38598318597600.0, + "grad_norm": 2.208657750705939, + "learning_rate": 9.108444229391676e-07, + "loss": 0.73, + "num_input_tokens_seen": 250150710, + "step": 7682 + }, + { + "epoch": 0.6928800108220228, + "flos": 18671261714400.0, + "grad_norm": 2.1614978342391775, + "learning_rate": 9.103544859185979e-07, + "loss": 0.7234, + "num_input_tokens_seen": 250177285, + "step": 7683 + }, + { + "epoch": 0.6929701943454931, + "flos": 23662793723040.0, + "grad_norm": 1.8472391508596688, + "learning_rate": 9.098646418725902e-07, + "loss": 0.8359, + "num_input_tokens_seen": 250204840, + "step": 7684 + }, + { + "epoch": 0.6930603778689634, + "flos": 64853329601280.0, + "grad_norm": 0.6363070040125346, + "learning_rate": 9.093748908429437e-07, + "loss": 0.6025, + "num_input_tokens_seen": 250293995, + "step": 7685 + }, + { + "epoch": 0.6931505613924336, + "flos": 20747259437280.0, + "grad_norm": 2.205968874562175, + "learning_rate": 9.088852328714444e-07, + "loss": 0.737, + "num_input_tokens_seen": 250321535, + "step": 7686 + }, + { + "epoch": 0.6932407449159038, + "flos": 35244930075360.0, + "grad_norm": 1.6383781670030222, + "learning_rate": 9.083956679998735e-07, + "loss": 0.6561, + "num_input_tokens_seen": 250351290, + "step": 7687 + }, + { + "epoch": 0.6933309284393742, + "flos": 19618830800640.0, + "grad_norm": 2.3733894509028093, + "learning_rate": 9.079061962700032e-07, + "loss": 0.7181, + "num_input_tokens_seen": 250381030, + "step": 7688 + }, + { + "epoch": 0.6934211119628444, + "flos": 25375436875680.0, + "grad_norm": 1.5545133560259656, + "learning_rate": 9.074168177235979e-07, + "loss": 0.7703, + "num_input_tokens_seen": 250411020, + "step": 7689 + }, + { + "epoch": 0.6935112954863146, + "flos": 23043002013120.0, + "grad_norm": 2.8032040724132514, + "learning_rate": 9.069275324024151e-07, + "loss": 0.7628, + "num_input_tokens_seen": 250439225, + "step": 7690 + }, + { + "epoch": 0.6936014790097849, + "flos": 23990125062240.0, + "grad_norm": 2.468591226228416, + "learning_rate": 9.064383403482005e-07, + "loss": 0.6814, + "num_input_tokens_seen": 250467550, + "step": 7691 + }, + { + "epoch": 0.6936916625332552, + "flos": 24973791719040.0, + "grad_norm": 2.068252748040977, + "learning_rate": 9.059492416026983e-07, + "loss": 0.7124, + "num_input_tokens_seen": 250497255, + "step": 7692 + }, + { + "epoch": 0.6937818460567254, + "flos": 34880460412320.0, + "grad_norm": 2.378778463605292, + "learning_rate": 9.054602362076378e-07, + "loss": 0.7452, + "num_input_tokens_seen": 250529560, + "step": 7693 + }, + { + "epoch": 0.6938720295801957, + "flos": 23223824393760.0, + "grad_norm": 2.5100924861632126, + "learning_rate": 9.049713242047468e-07, + "loss": 0.7387, + "num_input_tokens_seen": 250556575, + "step": 7694 + }, + { + "epoch": 0.693962213103666, + "flos": 18671113035360.0, + "grad_norm": 1.9044878585119949, + "learning_rate": 9.044825056357395e-07, + "loss": 0.7223, + "num_input_tokens_seen": 250584460, + "step": 7695 + }, + { + "epoch": 0.6940523966271362, + "flos": 21548988620640.0, + "grad_norm": 1.7831269798441638, + "learning_rate": 9.039937805423255e-07, + "loss": 0.6998, + "num_input_tokens_seen": 250610590, + "step": 7696 + }, + { + "epoch": 0.6941425801506065, + "flos": 69654741055680.0, + "grad_norm": 0.5926014180845248, + "learning_rate": 9.035051489662051e-07, + "loss": 0.5167, + "num_input_tokens_seen": 250703210, + "step": 7697 + }, + { + "epoch": 0.6942327636740767, + "flos": 61437229960320.0, + "grad_norm": 1.9666083055123562, + "learning_rate": 9.030166109490718e-07, + "loss": 0.6948, + "num_input_tokens_seen": 250736275, + "step": 7698 + }, + { + "epoch": 0.6943229471975471, + "flos": 14700051479520.0, + "grad_norm": 3.508232108594156, + "learning_rate": 9.025281665326099e-07, + "loss": 0.7511, + "num_input_tokens_seen": 250760280, + "step": 7699 + }, + { + "epoch": 0.6944131307210173, + "flos": 22096325001120.0, + "grad_norm": 2.7257828349193316, + "learning_rate": 9.020398157584967e-07, + "loss": 0.7555, + "num_input_tokens_seen": 250789635, + "step": 7700 + }, + { + "epoch": 0.6945033142444875, + "flos": 20237321568960.0, + "grad_norm": 3.1002935701232013, + "learning_rate": 9.015515586684002e-07, + "loss": 0.6503, + "num_input_tokens_seen": 250816675, + "step": 7701 + }, + { + "epoch": 0.6945934977679578, + "flos": 69819580439520.0, + "grad_norm": 0.6525135618751652, + "learning_rate": 9.010633953039812e-07, + "loss": 0.592, + "num_input_tokens_seen": 250908505, + "step": 7702 + }, + { + "epoch": 0.6946836812914281, + "flos": 26935252210560.0, + "grad_norm": 1.7362826741537942, + "learning_rate": 9.005753257068929e-07, + "loss": 0.7612, + "num_input_tokens_seen": 250935885, + "step": 7703 + }, + { + "epoch": 0.6947738648148983, + "flos": 19210383577920.0, + "grad_norm": 2.009461434553549, + "learning_rate": 9.000873499187797e-07, + "loss": 0.7645, + "num_input_tokens_seen": 250963310, + "step": 7704 + }, + { + "epoch": 0.6948640483383686, + "flos": 20930163324480.0, + "grad_norm": 1.8300314198074032, + "learning_rate": 8.995994679812797e-07, + "loss": 0.7803, + "num_input_tokens_seen": 250991485, + "step": 7705 + }, + { + "epoch": 0.6949542318618388, + "flos": 26503382305440.0, + "grad_norm": 1.6997592110365338, + "learning_rate": 8.991116799360192e-07, + "loss": 0.7689, + "num_input_tokens_seen": 251021390, + "step": 7706 + }, + { + "epoch": 0.6950444153853091, + "flos": 22711247472480.0, + "grad_norm": 2.5856476128584998, + "learning_rate": 8.986239858246217e-07, + "loss": 0.7399, + "num_input_tokens_seen": 251050655, + "step": 7707 + }, + { + "epoch": 0.6951345989087794, + "flos": 27159680453280.0, + "grad_norm": 2.27013490301232, + "learning_rate": 8.981363856886972e-07, + "loss": 0.8276, + "num_input_tokens_seen": 251077405, + "step": 7708 + }, + { + "epoch": 0.6952247824322496, + "flos": 29124746411520.0, + "grad_norm": 5.5031658077947325, + "learning_rate": 8.976488795698533e-07, + "loss": 0.7266, + "num_input_tokens_seen": 251108995, + "step": 7709 + }, + { + "epoch": 0.6953149659557198, + "flos": 22931326853280.0, + "grad_norm": 2.2144570526072065, + "learning_rate": 8.971614675096841e-07, + "loss": 0.7622, + "num_input_tokens_seen": 251137105, + "step": 7710 + }, + { + "epoch": 0.6954051494791902, + "flos": 23261780452320.0, + "grad_norm": 2.4297890595924145, + "learning_rate": 8.966741495497807e-07, + "loss": 0.8119, + "num_input_tokens_seen": 251163570, + "step": 7711 + }, + { + "epoch": 0.6954953330026604, + "flos": 26176943040480.0, + "grad_norm": 1.56553675237235, + "learning_rate": 8.961869257317218e-07, + "loss": 0.7122, + "num_input_tokens_seen": 251194120, + "step": 7712 + }, + { + "epoch": 0.6955855165261307, + "flos": 18669812093760.0, + "grad_norm": 1.8588810112246565, + "learning_rate": 8.956997960970809e-07, + "loss": 0.7298, + "num_input_tokens_seen": 251220400, + "step": 7713 + }, + { + "epoch": 0.6956757000496009, + "flos": 30438755158080.0, + "grad_norm": 1.878505798936046, + "learning_rate": 8.952127606874224e-07, + "loss": 0.6535, + "num_input_tokens_seen": 251250745, + "step": 7714 + }, + { + "epoch": 0.6957658835730712, + "flos": 20092113552000.0, + "grad_norm": 1.726396923842708, + "learning_rate": 8.947258195443028e-07, + "loss": 0.7499, + "num_input_tokens_seen": 251279170, + "step": 7715 + }, + { + "epoch": 0.6958560670965415, + "flos": 18263632226400.0, + "grad_norm": 1.8799625845493362, + "learning_rate": 8.942389727092716e-07, + "loss": 0.7237, + "num_input_tokens_seen": 251306745, + "step": 7716 + }, + { + "epoch": 0.6959462506200117, + "flos": 58740096682560.0, + "grad_norm": 0.6782881703053399, + "learning_rate": 8.937522202238677e-07, + "loss": 0.5637, + "num_input_tokens_seen": 251395855, + "step": 7717 + }, + { + "epoch": 0.696036434143482, + "flos": 23111703196800.0, + "grad_norm": 2.641306681929721, + "learning_rate": 8.932655621296239e-07, + "loss": 0.7653, + "num_input_tokens_seen": 251423355, + "step": 7718 + }, + { + "epoch": 0.6961266176669523, + "flos": 46901751942720.0, + "grad_norm": 1.5067213785033267, + "learning_rate": 8.927789984680649e-07, + "loss": 0.7417, + "num_input_tokens_seen": 251458915, + "step": 7719 + }, + { + "epoch": 0.6962168011904225, + "flos": 44827687047360.0, + "grad_norm": 1.603608996406073, + "learning_rate": 8.922925292807068e-07, + "loss": 0.6148, + "num_input_tokens_seen": 251495750, + "step": 7720 + }, + { + "epoch": 0.6963069847138927, + "flos": 22386741035040.0, + "grad_norm": 1.9995672830725906, + "learning_rate": 8.91806154609058e-07, + "loss": 0.8314, + "num_input_tokens_seen": 251524455, + "step": 7721 + }, + { + "epoch": 0.6963971682373631, + "flos": 21877360713120.0, + "grad_norm": 2.0840742955207925, + "learning_rate": 8.913198744946195e-07, + "loss": 0.7171, + "num_input_tokens_seen": 251551355, + "step": 7722 + }, + { + "epoch": 0.6964873517608333, + "flos": 23879156127840.0, + "grad_norm": 1.7838822708329203, + "learning_rate": 8.908336889788807e-07, + "loss": 0.7707, + "num_input_tokens_seen": 251580955, + "step": 7723 + }, + { + "epoch": 0.6965775352843036, + "flos": 28066840276800.0, + "grad_norm": 2.3250377077715996, + "learning_rate": 8.903475981033293e-07, + "loss": 0.628, + "num_input_tokens_seen": 251612215, + "step": 7724 + }, + { + "epoch": 0.6966677188077738, + "flos": 58556560909440.0, + "grad_norm": 0.7390559738548492, + "learning_rate": 8.898616019094376e-07, + "loss": 0.632, + "num_input_tokens_seen": 251691535, + "step": 7725 + }, + { + "epoch": 0.6967579023312441, + "flos": 20529224393280.0, + "grad_norm": 2.1601868225472733, + "learning_rate": 8.89375700438677e-07, + "loss": 0.783, + "num_input_tokens_seen": 251719425, + "step": 7726 + }, + { + "epoch": 0.6968480858547144, + "flos": 20783765875200.0, + "grad_norm": 1.9878335085140835, + "learning_rate": 8.888898937325047e-07, + "loss": 0.7114, + "num_input_tokens_seen": 251746865, + "step": 7727 + }, + { + "epoch": 0.6969382693781846, + "flos": 24281098642560.0, + "grad_norm": 3.2657409355354914, + "learning_rate": 8.884041818323733e-07, + "loss": 0.7125, + "num_input_tokens_seen": 251773360, + "step": 7728 + }, + { + "epoch": 0.6970284529016548, + "flos": 27305148658560.0, + "grad_norm": 1.7520153838357537, + "learning_rate": 8.879185647797262e-07, + "loss": 0.7825, + "num_input_tokens_seen": 251804325, + "step": 7729 + }, + { + "epoch": 0.6971186364251252, + "flos": 23261520264000.0, + "grad_norm": 1.8740504085474476, + "learning_rate": 8.874330426159993e-07, + "loss": 0.7267, + "num_input_tokens_seen": 251833885, + "step": 7730 + }, + { + "epoch": 0.6972088199485954, + "flos": 20343384095040.0, + "grad_norm": 1.759015769263233, + "learning_rate": 8.869476153826205e-07, + "loss": 0.7702, + "num_input_tokens_seen": 251861880, + "step": 7731 + }, + { + "epoch": 0.6972990034720656, + "flos": 20812169305440.0, + "grad_norm": 1.6901965319331471, + "learning_rate": 8.864622831210071e-07, + "loss": 0.7487, + "num_input_tokens_seen": 251888385, + "step": 7732 + }, + { + "epoch": 0.6973891869955359, + "flos": 24208494634080.0, + "grad_norm": 2.0595493595944023, + "learning_rate": 8.85977045872573e-07, + "loss": 0.7534, + "num_input_tokens_seen": 251914190, + "step": 7733 + }, + { + "epoch": 0.6974793705190062, + "flos": 26831977416480.0, + "grad_norm": 1.7344735607288546, + "learning_rate": 8.854919036787194e-07, + "loss": 0.7228, + "num_input_tokens_seen": 251944300, + "step": 7734 + }, + { + "epoch": 0.6975695540424764, + "flos": 21509285583360.0, + "grad_norm": 1.9421301860132416, + "learning_rate": 8.850068565808417e-07, + "loss": 0.7105, + "num_input_tokens_seen": 251974235, + "step": 7735 + }, + { + "epoch": 0.6976597375659467, + "flos": 26322002378400.0, + "grad_norm": 2.21124815359087, + "learning_rate": 8.845219046203271e-07, + "loss": 0.7968, + "num_input_tokens_seen": 252001175, + "step": 7736 + }, + { + "epoch": 0.6977499210894169, + "flos": 17541680815200.0, + "grad_norm": 2.8120221897731117, + "learning_rate": 8.840370478385544e-07, + "loss": 0.7746, + "num_input_tokens_seen": 252027475, + "step": 7737 + }, + { + "epoch": 0.6978401046128873, + "flos": 26828074591680.0, + "grad_norm": 4.396540556268649, + "learning_rate": 8.83552286276894e-07, + "loss": 0.7689, + "num_input_tokens_seen": 252059295, + "step": 7738 + }, + { + "epoch": 0.6979302881363575, + "flos": 18233444647680.0, + "grad_norm": 1.9014370302481172, + "learning_rate": 8.830676199767095e-07, + "loss": 0.7653, + "num_input_tokens_seen": 252085755, + "step": 7739 + }, + { + "epoch": 0.6980204716598277, + "flos": 24354817743840.0, + "grad_norm": 1.7379359213031378, + "learning_rate": 8.825830489793527e-07, + "loss": 0.7296, + "num_input_tokens_seen": 252115865, + "step": 7740 + }, + { + "epoch": 0.698110655183298, + "flos": 28724773894080.0, + "grad_norm": 1.7268862873803348, + "learning_rate": 8.820985733261732e-07, + "loss": 0.6973, + "num_input_tokens_seen": 252146915, + "step": 7741 + }, + { + "epoch": 0.6982008387067683, + "flos": 23660675046720.0, + "grad_norm": 3.145519030366523, + "learning_rate": 8.816141930585066e-07, + "loss": 0.657, + "num_input_tokens_seen": 252174240, + "step": 7742 + }, + { + "epoch": 0.6982910222302385, + "flos": 67206021983040.0, + "grad_norm": 0.6834460507639838, + "learning_rate": 8.811299082176837e-07, + "loss": 0.5886, + "num_input_tokens_seen": 252273790, + "step": 7743 + }, + { + "epoch": 0.6983812057537088, + "flos": 24425005717920.0, + "grad_norm": 2.3937897522018474, + "learning_rate": 8.806457188450265e-07, + "loss": 0.709, + "num_input_tokens_seen": 252304860, + "step": 7744 + }, + { + "epoch": 0.6984713892771791, + "flos": 22752214281600.0, + "grad_norm": 1.6687970215534027, + "learning_rate": 8.801616249818487e-07, + "loss": 0.7835, + "num_input_tokens_seen": 252333525, + "step": 7745 + }, + { + "epoch": 0.6985615728006493, + "flos": 20638520688480.0, + "grad_norm": 1.9095153981736903, + "learning_rate": 8.796776266694564e-07, + "loss": 0.7291, + "num_input_tokens_seen": 252361850, + "step": 7746 + }, + { + "epoch": 0.6986517563241196, + "flos": 33462842343840.0, + "grad_norm": 1.4232152972685301, + "learning_rate": 8.79193723949145e-07, + "loss": 0.692, + "num_input_tokens_seen": 252393000, + "step": 7747 + }, + { + "epoch": 0.6987419398475898, + "flos": 35605199555520.0, + "grad_norm": 1.9297017085366959, + "learning_rate": 8.787099168622063e-07, + "loss": 0.6438, + "num_input_tokens_seen": 252422925, + "step": 7748 + }, + { + "epoch": 0.6988321233710602, + "flos": 31746668064000.0, + "grad_norm": 2.1641444662387843, + "learning_rate": 8.782262054499199e-07, + "loss": 0.6877, + "num_input_tokens_seen": 252452090, + "step": 7749 + }, + { + "epoch": 0.6989223068945304, + "flos": 23261148566400.0, + "grad_norm": 1.4952327950180893, + "learning_rate": 8.777425897535588e-07, + "loss": 0.7755, + "num_input_tokens_seen": 252481365, + "step": 7750 + }, + { + "epoch": 0.6990124904180006, + "flos": 22751247867840.0, + "grad_norm": 2.0743543792462855, + "learning_rate": 8.77259069814388e-07, + "loss": 0.751, + "num_input_tokens_seen": 252509825, + "step": 7751 + }, + { + "epoch": 0.6991026739414709, + "flos": 28107732746400.0, + "grad_norm": 1.915913212612445, + "learning_rate": 8.767756456736641e-07, + "loss": 0.7462, + "num_input_tokens_seen": 252542500, + "step": 7752 + }, + { + "epoch": 0.6991928574649412, + "flos": 22823740367040.0, + "grad_norm": 1.7342553999620018, + "learning_rate": 8.762923173726358e-07, + "loss": 0.6711, + "num_input_tokens_seen": 252572830, + "step": 7753 + }, + { + "epoch": 0.6992830409884114, + "flos": 18304338847200.0, + "grad_norm": 1.6477374719286155, + "learning_rate": 8.758090849525428e-07, + "loss": 0.8405, + "num_input_tokens_seen": 252598430, + "step": 7754 + }, + { + "epoch": 0.6993732245118817, + "flos": 19654965540960.0, + "grad_norm": 2.0063868110570047, + "learning_rate": 8.753259484546174e-07, + "loss": 0.6842, + "num_input_tokens_seen": 252623470, + "step": 7755 + }, + { + "epoch": 0.6994634080353519, + "flos": 12113372493120.0, + "grad_norm": 2.0940466144040633, + "learning_rate": 8.748429079200841e-07, + "loss": 0.7028, + "num_input_tokens_seen": 252648745, + "step": 7756 + }, + { + "epoch": 0.6995535915588222, + "flos": 19581841155840.0, + "grad_norm": 2.5644206780780046, + "learning_rate": 8.743599633901575e-07, + "loss": 0.8025, + "num_input_tokens_seen": 252674335, + "step": 7757 + }, + { + "epoch": 0.6996437750822925, + "flos": 23838746865120.0, + "grad_norm": 1.8087916090897607, + "learning_rate": 8.738771149060453e-07, + "loss": 0.7637, + "num_input_tokens_seen": 252700420, + "step": 7758 + }, + { + "epoch": 0.6997339586057627, + "flos": 25225173771360.0, + "grad_norm": 1.696733011448735, + "learning_rate": 8.73394362508947e-07, + "loss": 0.7105, + "num_input_tokens_seen": 252728910, + "step": 7759 + }, + { + "epoch": 0.6998241421292329, + "flos": 53719683770400.0, + "grad_norm": 1.9927098804819496, + "learning_rate": 8.72911706240054e-07, + "loss": 0.709, + "num_input_tokens_seen": 252765070, + "step": 7760 + }, + { + "epoch": 0.6999143256527033, + "flos": 42673955889120.0, + "grad_norm": 2.001363527888964, + "learning_rate": 8.724291461405493e-07, + "loss": 0.6199, + "num_input_tokens_seen": 252798245, + "step": 7761 + }, + { + "epoch": 0.7000045091761735, + "flos": 25295064387360.0, + "grad_norm": 1.918816093772718, + "learning_rate": 8.71946682251606e-07, + "loss": 0.7117, + "num_input_tokens_seen": 252825975, + "step": 7762 + }, + { + "epoch": 0.7000946926996438, + "flos": 24750961776000.0, + "grad_norm": 2.62579324634993, + "learning_rate": 8.714643146143932e-07, + "loss": 0.7424, + "num_input_tokens_seen": 252854725, + "step": 7763 + }, + { + "epoch": 0.700184876223114, + "flos": 23698742614560.0, + "grad_norm": 2.153773558095368, + "learning_rate": 8.709820432700663e-07, + "loss": 0.7015, + "num_input_tokens_seen": 252882150, + "step": 7764 + }, + { + "epoch": 0.7002750597465843, + "flos": 24681851724960.0, + "grad_norm": 1.683382662752222, + "learning_rate": 8.704998682597784e-07, + "loss": 0.6607, + "num_input_tokens_seen": 252911960, + "step": 7765 + }, + { + "epoch": 0.7003652432700546, + "flos": 22970249325600.0, + "grad_norm": 1.987174274936274, + "learning_rate": 8.700177896246688e-07, + "loss": 0.763, + "num_input_tokens_seen": 252941580, + "step": 7766 + }, + { + "epoch": 0.7004554267935248, + "flos": 19728089926080.0, + "grad_norm": 2.068792051135707, + "learning_rate": 8.695358074058721e-07, + "loss": 0.6727, + "num_input_tokens_seen": 252967370, + "step": 7767 + }, + { + "epoch": 0.7005456103169951, + "flos": 16157744282880.0, + "grad_norm": 2.0557853753545103, + "learning_rate": 8.690539216445136e-07, + "loss": 0.7831, + "num_input_tokens_seen": 252994240, + "step": 7768 + }, + { + "epoch": 0.7006357938404654, + "flos": 23844879875520.0, + "grad_norm": 19.96496565275345, + "learning_rate": 8.685721323817106e-07, + "loss": 0.6502, + "num_input_tokens_seen": 253024410, + "step": 7769 + }, + { + "epoch": 0.7007259773639356, + "flos": 24675755884320.0, + "grad_norm": 2.3136744223378893, + "learning_rate": 8.680904396585718e-07, + "loss": 0.7829, + "num_input_tokens_seen": 253053000, + "step": 7770 + }, + { + "epoch": 0.7008161608874058, + "flos": 36887199405120.0, + "grad_norm": 2.136746422067157, + "learning_rate": 8.676088435161988e-07, + "loss": 0.6871, + "num_input_tokens_seen": 253082765, + "step": 7771 + }, + { + "epoch": 0.7009063444108762, + "flos": 22200194511360.0, + "grad_norm": 2.0845982970999994, + "learning_rate": 8.671273439956824e-07, + "loss": 0.7636, + "num_input_tokens_seen": 253107740, + "step": 7772 + }, + { + "epoch": 0.7009965279343464, + "flos": 24574005050400.0, + "grad_norm": 1.7233961942535307, + "learning_rate": 8.666459411381075e-07, + "loss": 0.7453, + "num_input_tokens_seen": 253136685, + "step": 7773 + }, + { + "epoch": 0.7010867114578166, + "flos": 28614176657280.0, + "grad_norm": 2.2372701601438783, + "learning_rate": 8.661646349845501e-07, + "loss": 0.7727, + "num_input_tokens_seen": 253167335, + "step": 7774 + }, + { + "epoch": 0.7011768949812869, + "flos": 13279162472160.0, + "grad_norm": 1.9686314986153957, + "learning_rate": 8.656834255760783e-07, + "loss": 0.8002, + "num_input_tokens_seen": 253191010, + "step": 7775 + }, + { + "epoch": 0.7012670785047572, + "flos": 22094875380480.0, + "grad_norm": 1.9287194051435956, + "learning_rate": 8.652023129537509e-07, + "loss": 0.7999, + "num_input_tokens_seen": 253220125, + "step": 7776 + }, + { + "epoch": 0.7013572620282275, + "flos": 25775483732640.0, + "grad_norm": 2.1059844848626446, + "learning_rate": 8.647212971586195e-07, + "loss": 0.7735, + "num_input_tokens_seen": 253250435, + "step": 7777 + }, + { + "epoch": 0.7014474455516977, + "flos": 17432904896640.0, + "grad_norm": 2.089565707810961, + "learning_rate": 8.642403782317275e-07, + "loss": 0.6952, + "num_input_tokens_seen": 253279410, + "step": 7778 + }, + { + "epoch": 0.7015376290751679, + "flos": 21505717286400.0, + "grad_norm": 2.3251390189991046, + "learning_rate": 8.637595562141075e-07, + "loss": 0.7502, + "num_input_tokens_seen": 253306185, + "step": 7779 + }, + { + "epoch": 0.7016278125986383, + "flos": 25994559529920.0, + "grad_norm": 2.283394524062969, + "learning_rate": 8.632788311467889e-07, + "loss": 0.7129, + "num_input_tokens_seen": 253339910, + "step": 7780 + }, + { + "epoch": 0.7017179961221085, + "flos": 23697999219360.0, + "grad_norm": 2.102489508059492, + "learning_rate": 8.627982030707867e-07, + "loss": 0.7409, + "num_input_tokens_seen": 253372265, + "step": 7781 + }, + { + "epoch": 0.7018081796455787, + "flos": 20602385948160.0, + "grad_norm": 2.069810684134653, + "learning_rate": 8.623176720271139e-07, + "loss": 0.8094, + "num_input_tokens_seen": 253397355, + "step": 7782 + }, + { + "epoch": 0.701898363169049, + "flos": 19501134139680.0, + "grad_norm": 2.172503483355751, + "learning_rate": 8.618372380567696e-07, + "loss": 0.8145, + "num_input_tokens_seen": 253419945, + "step": 7783 + }, + { + "epoch": 0.7019885466925193, + "flos": 25115319929760.0, + "grad_norm": 2.2185740871271395, + "learning_rate": 8.613569012007478e-07, + "loss": 0.699, + "num_input_tokens_seen": 253447425, + "step": 7784 + }, + { + "epoch": 0.7020787302159895, + "flos": 16594669275360.0, + "grad_norm": 1.9933281120287196, + "learning_rate": 8.608766615000338e-07, + "loss": 0.7836, + "num_input_tokens_seen": 253474785, + "step": 7785 + }, + { + "epoch": 0.7021689137394598, + "flos": 17759790198720.0, + "grad_norm": 2.0217464258740927, + "learning_rate": 8.603965189956039e-07, + "loss": 0.7661, + "num_input_tokens_seen": 253500605, + "step": 7786 + }, + { + "epoch": 0.70225909726293, + "flos": 25848236420160.0, + "grad_norm": 2.178809423199861, + "learning_rate": 8.599164737284276e-07, + "loss": 0.7502, + "num_input_tokens_seen": 253526365, + "step": 7787 + }, + { + "epoch": 0.7023492807864004, + "flos": 22605891171840.0, + "grad_norm": 1.7970218764275327, + "learning_rate": 8.594365257394634e-07, + "loss": 0.7042, + "num_input_tokens_seen": 253554565, + "step": 7788 + }, + { + "epoch": 0.7024394643098706, + "flos": 26139581698080.0, + "grad_norm": 2.092650167347708, + "learning_rate": 8.589566750696637e-07, + "loss": 0.6901, + "num_input_tokens_seen": 253584255, + "step": 7789 + }, + { + "epoch": 0.7025296478333408, + "flos": 32478729649920.0, + "grad_norm": 2.0290236963850234, + "learning_rate": 8.584769217599721e-07, + "loss": 0.6782, + "num_input_tokens_seen": 253614845, + "step": 7790 + }, + { + "epoch": 0.7026198313568112, + "flos": 17759864538240.0, + "grad_norm": 3.7419367250264353, + "learning_rate": 8.579972658513239e-07, + "loss": 0.694, + "num_input_tokens_seen": 253640785, + "step": 7791 + }, + { + "epoch": 0.7027100148802814, + "flos": 21293294876160.0, + "grad_norm": 2.3445808463871236, + "learning_rate": 8.57517707384646e-07, + "loss": 0.7518, + "num_input_tokens_seen": 253667425, + "step": 7792 + }, + { + "epoch": 0.7028001984037516, + "flos": 24609656583840.0, + "grad_norm": 2.252554608867491, + "learning_rate": 8.570382464008574e-07, + "loss": 0.7844, + "num_input_tokens_seen": 253697225, + "step": 7793 + }, + { + "epoch": 0.7028903819272219, + "flos": 24063100768320.0, + "grad_norm": 2.8600313027542854, + "learning_rate": 8.565588829408665e-07, + "loss": 0.7087, + "num_input_tokens_seen": 253726040, + "step": 7794 + }, + { + "epoch": 0.7029805654506922, + "flos": 23954436359040.0, + "grad_norm": 2.0747437468251917, + "learning_rate": 8.560796170455782e-07, + "loss": 0.722, + "num_input_tokens_seen": 253756430, + "step": 7795 + }, + { + "epoch": 0.7030707489741624, + "flos": 22788832228800.0, + "grad_norm": 1.9301072254932483, + "learning_rate": 8.556004487558828e-07, + "loss": 0.7846, + "num_input_tokens_seen": 253786885, + "step": 7796 + }, + { + "epoch": 0.7031609324976327, + "flos": 42278443742880.0, + "grad_norm": 1.86758029269787, + "learning_rate": 8.55121378112669e-07, + "loss": 0.66, + "num_input_tokens_seen": 253819295, + "step": 7797 + }, + { + "epoch": 0.7032511160211029, + "flos": 20930051815200.0, + "grad_norm": 2.2652394986213933, + "learning_rate": 8.546424051568111e-07, + "loss": 0.723, + "num_input_tokens_seen": 253848640, + "step": 7798 + }, + { + "epoch": 0.7033412995445733, + "flos": 15787364628000.0, + "grad_norm": 2.210655167416681, + "learning_rate": 8.541635299291785e-07, + "loss": 0.7959, + "num_input_tokens_seen": 253873005, + "step": 7799 + }, + { + "epoch": 0.7034314830680435, + "flos": 70959643211040.0, + "grad_norm": 0.597721395494976, + "learning_rate": 8.536847524706317e-07, + "loss": 0.5991, + "num_input_tokens_seen": 253972645, + "step": 7800 + }, + { + "epoch": 0.7035216665915137, + "flos": 32660852972160.0, + "grad_norm": 1.6940974086872482, + "learning_rate": 8.532060728220225e-07, + "loss": 0.6743, + "num_input_tokens_seen": 254002825, + "step": 7801 + }, + { + "epoch": 0.703611850114984, + "flos": 23188767576480.0, + "grad_norm": 2.3957843830802434, + "learning_rate": 8.527274910241955e-07, + "loss": 0.7555, + "num_input_tokens_seen": 254032065, + "step": 7802 + }, + { + "epoch": 0.7037020336384543, + "flos": 22456334292960.0, + "grad_norm": 2.00585818638901, + "learning_rate": 8.522490071179833e-07, + "loss": 0.7396, + "num_input_tokens_seen": 254060525, + "step": 7803 + }, + { + "epoch": 0.7037922171619245, + "flos": 41294814255840.0, + "grad_norm": 1.6553306907272192, + "learning_rate": 8.517706211442159e-07, + "loss": 0.6239, + "num_input_tokens_seen": 254094825, + "step": 7804 + }, + { + "epoch": 0.7038824006853948, + "flos": 23986073558400.0, + "grad_norm": 1.8679696204719127, + "learning_rate": 8.512923331437097e-07, + "loss": 0.8239, + "num_input_tokens_seen": 254122375, + "step": 7805 + }, + { + "epoch": 0.703972584208865, + "flos": 23298324060000.0, + "grad_norm": 1.5820293661977596, + "learning_rate": 8.508141431572755e-07, + "loss": 0.7112, + "num_input_tokens_seen": 254151790, + "step": 7806 + }, + { + "epoch": 0.7040627677323353, + "flos": 22314322875360.0, + "grad_norm": 2.485199701432413, + "learning_rate": 8.503360512257152e-07, + "loss": 0.731, + "num_input_tokens_seen": 254180075, + "step": 7807 + }, + { + "epoch": 0.7041529512558056, + "flos": 30328492449120.0, + "grad_norm": 2.0355624856974326, + "learning_rate": 8.498580573898219e-07, + "loss": 0.7115, + "num_input_tokens_seen": 254209595, + "step": 7808 + }, + { + "epoch": 0.7042431347792758, + "flos": 16484109208320.0, + "grad_norm": 2.7138347699294045, + "learning_rate": 8.493801616903813e-07, + "loss": 0.751, + "num_input_tokens_seen": 254233945, + "step": 7809 + }, + { + "epoch": 0.704333318302746, + "flos": 18125486463840.0, + "grad_norm": 1.7070062491465121, + "learning_rate": 8.489023641681705e-07, + "loss": 0.7627, + "num_input_tokens_seen": 254261755, + "step": 7810 + }, + { + "epoch": 0.7044235018262164, + "flos": 27742928555520.0, + "grad_norm": 1.8687808209687615, + "learning_rate": 8.484246648639555e-07, + "loss": 0.7927, + "num_input_tokens_seen": 254293600, + "step": 7811 + }, + { + "epoch": 0.7045136853496866, + "flos": 16230199612320.0, + "grad_norm": 2.579427031982873, + "learning_rate": 8.479470638184994e-07, + "loss": 0.6191, + "num_input_tokens_seen": 254321455, + "step": 7812 + }, + { + "epoch": 0.7046038688731568, + "flos": 24172136875200.0, + "grad_norm": 2.8303935189163765, + "learning_rate": 8.474695610725513e-07, + "loss": 0.6625, + "num_input_tokens_seen": 254348220, + "step": 7813 + }, + { + "epoch": 0.7046940523966272, + "flos": 37175422423200.0, + "grad_norm": 1.5874564878252648, + "learning_rate": 8.469921566668552e-07, + "loss": 0.8174, + "num_input_tokens_seen": 254377800, + "step": 7814 + }, + { + "epoch": 0.7047842359200974, + "flos": 24713600433600.0, + "grad_norm": 4.576643876706417, + "learning_rate": 8.46514850642146e-07, + "loss": 0.7926, + "num_input_tokens_seen": 254408840, + "step": 7815 + }, + { + "epoch": 0.7048744194435677, + "flos": 22970174986080.0, + "grad_norm": 2.137014197752669, + "learning_rate": 8.460376430391499e-07, + "loss": 0.7282, + "num_input_tokens_seen": 254437795, + "step": 7816 + }, + { + "epoch": 0.7049646029670379, + "flos": 24464225548320.0, + "grad_norm": 1.7408581096492382, + "learning_rate": 8.455605338985858e-07, + "loss": 0.7723, + "num_input_tokens_seen": 254470785, + "step": 7817 + }, + { + "epoch": 0.7050547864905082, + "flos": 37434758804160.0, + "grad_norm": 1.7611548627251632, + "learning_rate": 8.45083523261161e-07, + "loss": 0.5993, + "num_input_tokens_seen": 254503725, + "step": 7818 + }, + { + "epoch": 0.7051449700139785, + "flos": 25702805384640.0, + "grad_norm": 1.873609790404074, + "learning_rate": 8.446066111675796e-07, + "loss": 0.7495, + "num_input_tokens_seen": 254531695, + "step": 7819 + }, + { + "epoch": 0.7052351535374487, + "flos": 27922598673600.0, + "grad_norm": 2.01413559407372, + "learning_rate": 8.441297976585314e-07, + "loss": 0.6362, + "num_input_tokens_seen": 254561405, + "step": 7820 + }, + { + "epoch": 0.7053253370609189, + "flos": 21221434262880.0, + "grad_norm": 1.814041403057167, + "learning_rate": 8.436530827747037e-07, + "loss": 0.8295, + "num_input_tokens_seen": 254589365, + "step": 7821 + }, + { + "epoch": 0.7054155205843893, + "flos": 21652486433280.0, + "grad_norm": 2.310577830915835, + "learning_rate": 8.431764665567704e-07, + "loss": 0.8798, + "num_input_tokens_seen": 254615105, + "step": 7822 + }, + { + "epoch": 0.7055057041078595, + "flos": 21039645468480.0, + "grad_norm": 2.5327212920289757, + "learning_rate": 8.426999490453996e-07, + "loss": 0.6866, + "num_input_tokens_seen": 254643245, + "step": 7823 + }, + { + "epoch": 0.7055958876313297, + "flos": 18816506901120.0, + "grad_norm": 1.9646887898247491, + "learning_rate": 8.422235302812504e-07, + "loss": 0.7532, + "num_input_tokens_seen": 254669310, + "step": 7824 + }, + { + "epoch": 0.7056860711548, + "flos": 25406851056480.0, + "grad_norm": 1.7664462179920308, + "learning_rate": 8.417472103049734e-07, + "loss": 0.8165, + "num_input_tokens_seen": 254697550, + "step": 7825 + }, + { + "epoch": 0.7057762546782703, + "flos": 21726540062400.0, + "grad_norm": 1.9404064651274606, + "learning_rate": 8.412709891572112e-07, + "loss": 0.7771, + "num_input_tokens_seen": 254726155, + "step": 7826 + }, + { + "epoch": 0.7058664382017406, + "flos": 19290087010560.0, + "grad_norm": 2.1792626188997484, + "learning_rate": 8.407948668785978e-07, + "loss": 0.78, + "num_input_tokens_seen": 254753635, + "step": 7827 + }, + { + "epoch": 0.7059566217252108, + "flos": 23735026033920.0, + "grad_norm": 5.464863295175913, + "learning_rate": 8.403188435097576e-07, + "loss": 0.7375, + "num_input_tokens_seen": 254782755, + "step": 7828 + }, + { + "epoch": 0.706046805248681, + "flos": 25005689106720.0, + "grad_norm": 1.705505107888765, + "learning_rate": 8.398429190913081e-07, + "loss": 0.8152, + "num_input_tokens_seen": 254812290, + "step": 7829 + }, + { + "epoch": 0.7061369887721514, + "flos": 26540632138560.0, + "grad_norm": 1.8326734381146232, + "learning_rate": 8.393670936638578e-07, + "loss": 0.7451, + "num_input_tokens_seen": 254840540, + "step": 7830 + }, + { + "epoch": 0.7062271722956216, + "flos": 22496520537120.0, + "grad_norm": 2.125693376400946, + "learning_rate": 8.388913672680067e-07, + "loss": 0.7412, + "num_input_tokens_seen": 254870650, + "step": 7831 + }, + { + "epoch": 0.7063173558190918, + "flos": 24676945316640.0, + "grad_norm": 2.2666051388785875, + "learning_rate": 8.384157399443472e-07, + "loss": 0.6901, + "num_input_tokens_seen": 254898680, + "step": 7832 + }, + { + "epoch": 0.7064075393425621, + "flos": 19909730041440.0, + "grad_norm": 1.6734699947789005, + "learning_rate": 8.379402117334601e-07, + "loss": 0.7672, + "num_input_tokens_seen": 254925180, + "step": 7833 + }, + { + "epoch": 0.7064977228660324, + "flos": 22782736388160.0, + "grad_norm": 1.7760285258622717, + "learning_rate": 8.374647826759232e-07, + "loss": 0.7432, + "num_input_tokens_seen": 254954055, + "step": 7834 + }, + { + "epoch": 0.7065879063895026, + "flos": 24492703318080.0, + "grad_norm": 3.2259120331682554, + "learning_rate": 8.369894528122998e-07, + "loss": 0.7437, + "num_input_tokens_seen": 254982435, + "step": 7835 + }, + { + "epoch": 0.7066780899129729, + "flos": 24245632957920.0, + "grad_norm": 2.1217653675999344, + "learning_rate": 8.365142221831505e-07, + "loss": 0.6075, + "num_input_tokens_seen": 255012800, + "step": 7836 + }, + { + "epoch": 0.7067682734364432, + "flos": 23407025639040.0, + "grad_norm": 2.26431394329929, + "learning_rate": 8.360390908290222e-07, + "loss": 0.7055, + "num_input_tokens_seen": 255040720, + "step": 7837 + }, + { + "epoch": 0.7068584569599135, + "flos": 21986880026880.0, + "grad_norm": 2.28415348841992, + "learning_rate": 8.355640587904569e-07, + "loss": 0.7246, + "num_input_tokens_seen": 255067640, + "step": 7838 + }, + { + "epoch": 0.7069486404833837, + "flos": 20237061380640.0, + "grad_norm": 2.5973873142622623, + "learning_rate": 8.350891261079866e-07, + "loss": 0.754, + "num_input_tokens_seen": 255093035, + "step": 7839 + }, + { + "epoch": 0.7070388240068539, + "flos": 21148347047520.0, + "grad_norm": 2.0207734115473337, + "learning_rate": 8.346142928221356e-07, + "loss": 0.7866, + "num_input_tokens_seen": 255121085, + "step": 7840 + }, + { + "epoch": 0.7071290075303243, + "flos": 18743717043840.0, + "grad_norm": 2.182180203039519, + "learning_rate": 8.341395589734189e-07, + "loss": 0.5863, + "num_input_tokens_seen": 255148800, + "step": 7841 + }, + { + "epoch": 0.7072191910537945, + "flos": 15647211698400.0, + "grad_norm": 2.5336377220325614, + "learning_rate": 8.336649246023433e-07, + "loss": 0.6957, + "num_input_tokens_seen": 255173900, + "step": 7842 + }, + { + "epoch": 0.7073093745772647, + "flos": 21185299522560.0, + "grad_norm": 1.7477085915692359, + "learning_rate": 8.331903897494084e-07, + "loss": 0.6452, + "num_input_tokens_seen": 255202455, + "step": 7843 + }, + { + "epoch": 0.707399558100735, + "flos": 20747668304640.0, + "grad_norm": 2.0986545112521204, + "learning_rate": 8.327159544551024e-07, + "loss": 0.757, + "num_input_tokens_seen": 255229265, + "step": 7844 + }, + { + "epoch": 0.7074897416242053, + "flos": 26572232168160.0, + "grad_norm": 2.061385576128163, + "learning_rate": 8.322416187599073e-07, + "loss": 0.7048, + "num_input_tokens_seen": 255257220, + "step": 7845 + }, + { + "epoch": 0.7075799251476755, + "flos": 25264728129600.0, + "grad_norm": 2.4478083052222432, + "learning_rate": 8.317673827042963e-07, + "loss": 0.7664, + "num_input_tokens_seen": 255283535, + "step": 7846 + }, + { + "epoch": 0.7076701086711458, + "flos": 39905339429280.0, + "grad_norm": 1.9501477552335442, + "learning_rate": 8.312932463287339e-07, + "loss": 0.7292, + "num_input_tokens_seen": 255311855, + "step": 7847 + }, + { + "epoch": 0.707760292194616, + "flos": 27816015770880.0, + "grad_norm": 2.7182133462862983, + "learning_rate": 8.308192096736759e-07, + "loss": 0.7134, + "num_input_tokens_seen": 255341365, + "step": 7848 + }, + { + "epoch": 0.7078504757180863, + "flos": 20237358738720.0, + "grad_norm": 1.7229400885334165, + "learning_rate": 8.303452727795703e-07, + "loss": 0.7225, + "num_input_tokens_seen": 255367470, + "step": 7849 + }, + { + "epoch": 0.7079406592415566, + "flos": 18889408267680.0, + "grad_norm": 2.3275451962255023, + "learning_rate": 8.298714356868542e-07, + "loss": 0.8212, + "num_input_tokens_seen": 255394225, + "step": 7850 + }, + { + "epoch": 0.7080308427650268, + "flos": 29673792600960.0, + "grad_norm": 1.8761224565861723, + "learning_rate": 8.293976984359605e-07, + "loss": 0.7592, + "num_input_tokens_seen": 255424395, + "step": 7851 + }, + { + "epoch": 0.708121026288497, + "flos": 21731595149760.0, + "grad_norm": 3.643636481216393, + "learning_rate": 8.289240610673092e-07, + "loss": 0.7479, + "num_input_tokens_seen": 255453165, + "step": 7852 + }, + { + "epoch": 0.7082112098119674, + "flos": 20560415555520.0, + "grad_norm": 2.083756467446472, + "learning_rate": 8.284505236213144e-07, + "loss": 0.7753, + "num_input_tokens_seen": 255481570, + "step": 7853 + }, + { + "epoch": 0.7083013933354376, + "flos": 27775941036000.0, + "grad_norm": 1.7956211525291281, + "learning_rate": 8.279770861383806e-07, + "loss": 0.7371, + "num_input_tokens_seen": 255511520, + "step": 7854 + }, + { + "epoch": 0.7083915768589079, + "flos": 21258721265760.0, + "grad_norm": 1.8230744720456127, + "learning_rate": 8.275037486589042e-07, + "loss": 0.7394, + "num_input_tokens_seen": 255540080, + "step": 7855 + }, + { + "epoch": 0.7084817603823781, + "flos": 63376042600800.0, + "grad_norm": 0.67175555830893, + "learning_rate": 8.270305112232739e-07, + "loss": 0.5582, + "num_input_tokens_seen": 255629015, + "step": 7856 + }, + { + "epoch": 0.7085719439058484, + "flos": 25663585554240.0, + "grad_norm": 2.1820156557131627, + "learning_rate": 8.265573738718665e-07, + "loss": 0.7417, + "num_input_tokens_seen": 255659005, + "step": 7857 + }, + { + "epoch": 0.7086621274293187, + "flos": 23949975987840.0, + "grad_norm": 2.038683625900221, + "learning_rate": 8.260843366450559e-07, + "loss": 0.7031, + "num_input_tokens_seen": 255686680, + "step": 7858 + }, + { + "epoch": 0.7087523109527889, + "flos": 22606039850880.0, + "grad_norm": 1.8669893806086038, + "learning_rate": 8.256113995832017e-07, + "loss": 0.74, + "num_input_tokens_seen": 255715195, + "step": 7859 + }, + { + "epoch": 0.7088424944762591, + "flos": 25813922998080.0, + "grad_norm": 1.837802410542361, + "learning_rate": 8.251385627266583e-07, + "loss": 0.7875, + "num_input_tokens_seen": 255748685, + "step": 7860 + }, + { + "epoch": 0.7089326779997295, + "flos": 23589074621760.0, + "grad_norm": 2.33278675716681, + "learning_rate": 8.24665826115771e-07, + "loss": 0.7564, + "num_input_tokens_seen": 255776895, + "step": 7861 + }, + { + "epoch": 0.7090228615231997, + "flos": 23403754700160.0, + "grad_norm": 1.753149060306918, + "learning_rate": 8.241931897908763e-07, + "loss": 0.7602, + "num_input_tokens_seen": 255806135, + "step": 7862 + }, + { + "epoch": 0.7091130450466699, + "flos": 24932787740160.0, + "grad_norm": 1.7207870203148492, + "learning_rate": 8.237206537923016e-07, + "loss": 0.7655, + "num_input_tokens_seen": 255834785, + "step": 7863 + }, + { + "epoch": 0.7092032285701403, + "flos": 19290198519840.0, + "grad_norm": 4.650567600514803, + "learning_rate": 8.232482181603671e-07, + "loss": 0.734, + "num_input_tokens_seen": 255862190, + "step": 7864 + }, + { + "epoch": 0.7092934120936105, + "flos": 21219612944640.0, + "grad_norm": 1.9648080143797693, + "learning_rate": 8.227758829353828e-07, + "loss": 0.7219, + "num_input_tokens_seen": 255892505, + "step": 7865 + }, + { + "epoch": 0.7093835956170808, + "flos": 22423879358880.0, + "grad_norm": 1.8893407447932502, + "learning_rate": 8.223036481576522e-07, + "loss": 0.7129, + "num_input_tokens_seen": 255920725, + "step": 7866 + }, + { + "epoch": 0.709473779140551, + "flos": 37645248386880.0, + "grad_norm": 2.302097119136171, + "learning_rate": 8.218315138674672e-07, + "loss": 0.7466, + "num_input_tokens_seen": 255950725, + "step": 7867 + }, + { + "epoch": 0.7095639626640213, + "flos": 60895909347360.0, + "grad_norm": 0.6184238171328773, + "learning_rate": 8.21359480105114e-07, + "loss": 0.5758, + "num_input_tokens_seen": 256037025, + "step": 7868 + }, + { + "epoch": 0.7096541461874916, + "flos": 27045923786880.0, + "grad_norm": 1.9473934456857465, + "learning_rate": 8.208875469108689e-07, + "loss": 0.7763, + "num_input_tokens_seen": 256065790, + "step": 7869 + }, + { + "epoch": 0.7097443297109618, + "flos": 27013654701600.0, + "grad_norm": 3.060671980880725, + "learning_rate": 8.204157143249997e-07, + "loss": 0.7295, + "num_input_tokens_seen": 256093345, + "step": 7870 + }, + { + "epoch": 0.709834513234432, + "flos": 31092600101760.0, + "grad_norm": 1.7380671337142324, + "learning_rate": 8.199439823877668e-07, + "loss": 0.7179, + "num_input_tokens_seen": 256125310, + "step": 7871 + }, + { + "epoch": 0.7099246967579024, + "flos": 27779434993440.0, + "grad_norm": 1.8056729751292682, + "learning_rate": 8.194723511394186e-07, + "loss": 0.7269, + "num_input_tokens_seen": 256155645, + "step": 7872 + }, + { + "epoch": 0.7100148802813726, + "flos": 14044125029280.0, + "grad_norm": 2.2363533918127767, + "learning_rate": 8.190008206202002e-07, + "loss": 0.6901, + "num_input_tokens_seen": 256181695, + "step": 7873 + }, + { + "epoch": 0.7101050638048428, + "flos": 12294863929440.0, + "grad_norm": 2.3449118214114817, + "learning_rate": 8.185293908703423e-07, + "loss": 0.7912, + "num_input_tokens_seen": 256206720, + "step": 7874 + }, + { + "epoch": 0.7101952473283131, + "flos": 17360152209120.0, + "grad_norm": 1.7202731232809694, + "learning_rate": 8.180580619300727e-07, + "loss": 0.7059, + "num_input_tokens_seen": 256233615, + "step": 7875 + }, + { + "epoch": 0.7102854308517834, + "flos": 43882125128160.0, + "grad_norm": 1.56184396322919, + "learning_rate": 8.175868338396057e-07, + "loss": 0.6948, + "num_input_tokens_seen": 256269635, + "step": 7876 + }, + { + "epoch": 0.7103756143752537, + "flos": 24282139395840.0, + "grad_norm": 2.168896395884396, + "learning_rate": 8.171157066391499e-07, + "loss": 0.7137, + "num_input_tokens_seen": 256299585, + "step": 7877 + }, + { + "epoch": 0.7104657978987239, + "flos": 23553200069760.0, + "grad_norm": 2.42688297432516, + "learning_rate": 8.166446803689045e-07, + "loss": 0.7642, + "num_input_tokens_seen": 256327680, + "step": 7878 + }, + { + "epoch": 0.7105559814221941, + "flos": 65276607558240.0, + "grad_norm": 0.6696469156959705, + "learning_rate": 8.161737550690595e-07, + "loss": 0.5826, + "num_input_tokens_seen": 256420500, + "step": 7879 + }, + { + "epoch": 0.7106461649456645, + "flos": 23516359104000.0, + "grad_norm": 4.7528699938143255, + "learning_rate": 8.157029307797976e-07, + "loss": 0.7657, + "num_input_tokens_seen": 256450760, + "step": 7880 + }, + { + "epoch": 0.7107363484691347, + "flos": 23512381939680.0, + "grad_norm": 1.5420489309837402, + "learning_rate": 8.152322075412925e-07, + "loss": 0.7971, + "num_input_tokens_seen": 256479430, + "step": 7881 + }, + { + "epoch": 0.7108265319926049, + "flos": 26026382578080.0, + "grad_norm": 1.5706936957155262, + "learning_rate": 8.147615853937073e-07, + "loss": 0.6949, + "num_input_tokens_seen": 256511220, + "step": 7882 + }, + { + "epoch": 0.7109167155160752, + "flos": 23918004260640.0, + "grad_norm": 1.947739033751034, + "learning_rate": 8.142910643771992e-07, + "loss": 0.6656, + "num_input_tokens_seen": 256541705, + "step": 7883 + }, + { + "epoch": 0.7110068990395455, + "flos": 21913309604640.0, + "grad_norm": 2.1017133807507293, + "learning_rate": 8.138206445319152e-07, + "loss": 0.7024, + "num_input_tokens_seen": 256567760, + "step": 7884 + }, + { + "epoch": 0.7110970825630157, + "flos": 27888248081760.0, + "grad_norm": 1.7620884408047666, + "learning_rate": 8.133503258979944e-07, + "loss": 0.6422, + "num_input_tokens_seen": 256595680, + "step": 7885 + }, + { + "epoch": 0.711187266086486, + "flos": 20636773709760.0, + "grad_norm": 5.384783689781994, + "learning_rate": 8.12880108515567e-07, + "loss": 0.6672, + "num_input_tokens_seen": 256624645, + "step": 7886 + }, + { + "epoch": 0.7112774496099563, + "flos": 24901224880320.0, + "grad_norm": 1.6595419790734451, + "learning_rate": 8.124099924247543e-07, + "loss": 0.7118, + "num_input_tokens_seen": 256653475, + "step": 7887 + }, + { + "epoch": 0.7113676331334265, + "flos": 25549791718080.0, + "grad_norm": 1.6778599222479549, + "learning_rate": 8.119399776656701e-07, + "loss": 0.6466, + "num_input_tokens_seen": 256681915, + "step": 7888 + }, + { + "epoch": 0.7114578166568968, + "flos": 26866179329280.0, + "grad_norm": 1.7435754285270766, + "learning_rate": 8.114700642784167e-07, + "loss": 0.7382, + "num_input_tokens_seen": 256711110, + "step": 7889 + }, + { + "epoch": 0.711548000180367, + "flos": 25769164873440.0, + "grad_norm": 1.668810451141285, + "learning_rate": 8.110002523030921e-07, + "loss": 0.7997, + "num_input_tokens_seen": 256739620, + "step": 7890 + }, + { + "epoch": 0.7116381837038374, + "flos": 22786825061760.0, + "grad_norm": 1.7967719007022223, + "learning_rate": 8.105305417797808e-07, + "loss": 0.7615, + "num_input_tokens_seen": 256768985, + "step": 7891 + }, + { + "epoch": 0.7117283672273076, + "flos": 26686137513600.0, + "grad_norm": 2.104815493219785, + "learning_rate": 8.100609327485635e-07, + "loss": 0.7799, + "num_input_tokens_seen": 256797520, + "step": 7892 + }, + { + "epoch": 0.7118185507507778, + "flos": 22854522661920.0, + "grad_norm": 4.008081378803495, + "learning_rate": 8.095914252495082e-07, + "loss": 0.8075, + "num_input_tokens_seen": 256823450, + "step": 7893 + }, + { + "epoch": 0.7119087342742481, + "flos": 20849976684960.0, + "grad_norm": 2.2685096003846565, + "learning_rate": 8.091220193226762e-07, + "loss": 0.7095, + "num_input_tokens_seen": 256849945, + "step": 7894 + }, + { + "epoch": 0.7119989177977184, + "flos": 28108885008960.0, + "grad_norm": 1.622035476889255, + "learning_rate": 8.0865271500812e-07, + "loss": 0.7271, + "num_input_tokens_seen": 256880180, + "step": 7895 + }, + { + "epoch": 0.7120891013211886, + "flos": 21221211244320.0, + "grad_norm": 3.362909063819745, + "learning_rate": 8.081835123458831e-07, + "loss": 0.7786, + "num_input_tokens_seen": 256908520, + "step": 7896 + }, + { + "epoch": 0.7121792848446589, + "flos": 22824520932000.0, + "grad_norm": 2.2069757235133434, + "learning_rate": 8.077144113760013e-07, + "loss": 0.7244, + "num_input_tokens_seen": 256936880, + "step": 7897 + }, + { + "epoch": 0.7122694683681291, + "flos": 35975393361600.0, + "grad_norm": 1.562323674743392, + "learning_rate": 8.072454121384995e-07, + "loss": 0.7201, + "num_input_tokens_seen": 256968690, + "step": 7898 + }, + { + "epoch": 0.7123596518915994, + "flos": 28799050541760.0, + "grad_norm": 2.071565414474999, + "learning_rate": 8.067765146733958e-07, + "loss": 0.7863, + "num_input_tokens_seen": 256998480, + "step": 7899 + }, + { + "epoch": 0.7124498354150697, + "flos": 24171876686880.0, + "grad_norm": 4.241850950263685, + "learning_rate": 8.063077190206993e-07, + "loss": 0.816, + "num_input_tokens_seen": 257026825, + "step": 7900 + }, + { + "epoch": 0.7125400189385399, + "flos": 60137637347040.0, + "grad_norm": 0.6050419195091585, + "learning_rate": 8.058390252204101e-07, + "loss": 0.4932, + "num_input_tokens_seen": 257122420, + "step": 7901 + }, + { + "epoch": 0.7126302024620101, + "flos": 27303810547200.0, + "grad_norm": 2.0516342859427006, + "learning_rate": 8.0537043331252e-07, + "loss": 0.7346, + "num_input_tokens_seen": 257155320, + "step": 7902 + }, + { + "epoch": 0.7127203859854805, + "flos": 30255888440640.0, + "grad_norm": 2.0857756476884646, + "learning_rate": 8.049019433370121e-07, + "loss": 0.6787, + "num_input_tokens_seen": 257183760, + "step": 7903 + }, + { + "epoch": 0.7128105695089507, + "flos": 35862380090400.0, + "grad_norm": 1.8556133655608649, + "learning_rate": 8.044335553338588e-07, + "loss": 0.743, + "num_input_tokens_seen": 257213490, + "step": 7904 + }, + { + "epoch": 0.712900753032421, + "flos": 21840854275200.0, + "grad_norm": 1.976528061879614, + "learning_rate": 8.039652693430281e-07, + "loss": 0.7534, + "num_input_tokens_seen": 257243250, + "step": 7905 + }, + { + "epoch": 0.7129909365558912, + "flos": 25623956856480.0, + "grad_norm": 3.1739986432558753, + "learning_rate": 8.034970854044742e-07, + "loss": 0.7176, + "num_input_tokens_seen": 257272035, + "step": 7906 + }, + { + "epoch": 0.7130811200793615, + "flos": 24313181879040.0, + "grad_norm": 6.026670308294856, + "learning_rate": 8.03029003558148e-07, + "loss": 0.8381, + "num_input_tokens_seen": 257300515, + "step": 7907 + }, + { + "epoch": 0.7131713036028318, + "flos": 13095998396640.0, + "grad_norm": 2.200860449111025, + "learning_rate": 8.025610238439864e-07, + "loss": 0.805, + "num_input_tokens_seen": 257322890, + "step": 7908 + }, + { + "epoch": 0.713261487126302, + "flos": 20164494541920.0, + "grad_norm": 2.4486694693455426, + "learning_rate": 8.020931463019207e-07, + "loss": 0.8178, + "num_input_tokens_seen": 257349955, + "step": 7909 + }, + { + "epoch": 0.7133516706497723, + "flos": 16448717863200.0, + "grad_norm": 1.9993039350409958, + "learning_rate": 8.016253709718732e-07, + "loss": 0.6875, + "num_input_tokens_seen": 257377435, + "step": 7910 + }, + { + "epoch": 0.7134418541732426, + "flos": 29382633171840.0, + "grad_norm": 5.264738557271008, + "learning_rate": 8.011576978937567e-07, + "loss": 0.65, + "num_input_tokens_seen": 257404635, + "step": 7911 + }, + { + "epoch": 0.7135320376967128, + "flos": 17577369518400.0, + "grad_norm": 1.8562444410658292, + "learning_rate": 8.006901271074764e-07, + "loss": 0.7388, + "num_input_tokens_seen": 257431815, + "step": 7912 + }, + { + "epoch": 0.713622221220183, + "flos": 21652300584480.0, + "grad_norm": 3.1231238264440835, + "learning_rate": 8.002226586529261e-07, + "loss": 0.7806, + "num_input_tokens_seen": 257460510, + "step": 7913 + }, + { + "epoch": 0.7137124047436534, + "flos": 28979575564320.0, + "grad_norm": 1.9581067447925498, + "learning_rate": 7.997552925699956e-07, + "loss": 0.82, + "num_input_tokens_seen": 257490140, + "step": 7914 + }, + { + "epoch": 0.7138025882671236, + "flos": 25483878266400.0, + "grad_norm": 2.057642215527661, + "learning_rate": 7.992880288985606e-07, + "loss": 0.8358, + "num_input_tokens_seen": 257518685, + "step": 7915 + }, + { + "epoch": 0.7138927717905938, + "flos": 23328325789920.0, + "grad_norm": 1.834363676747819, + "learning_rate": 7.988208676784918e-07, + "loss": 0.7512, + "num_input_tokens_seen": 257547520, + "step": 7916 + }, + { + "epoch": 0.7139829553140641, + "flos": 36450237242880.0, + "grad_norm": 4.076245806708544, + "learning_rate": 7.983538089496497e-07, + "loss": 0.7373, + "num_input_tokens_seen": 257576825, + "step": 7917 + }, + { + "epoch": 0.7140731388375344, + "flos": 29091696761280.0, + "grad_norm": 1.9234003976052558, + "learning_rate": 7.978868527518864e-07, + "loss": 0.7001, + "num_input_tokens_seen": 257607735, + "step": 7918 + }, + { + "epoch": 0.7141633223610047, + "flos": 27378867759840.0, + "grad_norm": 1.7057609807510425, + "learning_rate": 7.974199991250455e-07, + "loss": 0.717, + "num_input_tokens_seen": 257638960, + "step": 7919 + }, + { + "epoch": 0.7142535058844749, + "flos": 23371151087040.0, + "grad_norm": 2.4614369687834676, + "learning_rate": 7.969532481089616e-07, + "loss": 0.8399, + "num_input_tokens_seen": 257666965, + "step": 7920 + }, + { + "epoch": 0.7143436894079451, + "flos": 23224902316800.0, + "grad_norm": 1.8240317140549323, + "learning_rate": 7.964865997434589e-07, + "loss": 0.7564, + "num_input_tokens_seen": 257694935, + "step": 7921 + }, + { + "epoch": 0.7144338729314155, + "flos": 27123620052480.0, + "grad_norm": 1.8294021188428633, + "learning_rate": 7.96020054068357e-07, + "loss": 0.8132, + "num_input_tokens_seen": 257723625, + "step": 7922 + }, + { + "epoch": 0.7145240564548857, + "flos": 34337621572800.0, + "grad_norm": 1.5606634409199105, + "learning_rate": 7.95553611123462e-07, + "loss": 0.6979, + "num_input_tokens_seen": 257759715, + "step": 7923 + }, + { + "epoch": 0.7146142399783559, + "flos": 22460237117760.0, + "grad_norm": 3.9596798798225326, + "learning_rate": 7.950872709485741e-07, + "loss": 0.9198, + "num_input_tokens_seen": 257786680, + "step": 7924 + }, + { + "epoch": 0.7147044235018262, + "flos": 20820458161920.0, + "grad_norm": 2.9317746471697155, + "learning_rate": 7.946210335834842e-07, + "loss": 0.7973, + "num_input_tokens_seen": 257813975, + "step": 7925 + }, + { + "epoch": 0.7147946070252965, + "flos": 21913606962720.0, + "grad_norm": 1.5598699178012971, + "learning_rate": 7.94154899067974e-07, + "loss": 0.667, + "num_input_tokens_seen": 257844845, + "step": 7926 + }, + { + "epoch": 0.7148847905487667, + "flos": 24500694816480.0, + "grad_norm": 1.9340936188036495, + "learning_rate": 7.936888674418177e-07, + "loss": 0.6701, + "num_input_tokens_seen": 257875000, + "step": 7927 + }, + { + "epoch": 0.714974974072237, + "flos": 16995385188000.0, + "grad_norm": 2.3285763706218354, + "learning_rate": 7.932229387447771e-07, + "loss": 0.7455, + "num_input_tokens_seen": 257900655, + "step": 7928 + }, + { + "epoch": 0.7150651575957072, + "flos": 20747333776800.0, + "grad_norm": 2.814642704568792, + "learning_rate": 7.927571130166109e-07, + "loss": 0.7326, + "num_input_tokens_seen": 257925010, + "step": 7929 + }, + { + "epoch": 0.7151553411191776, + "flos": 67449003669600.0, + "grad_norm": 0.6266213605825653, + "learning_rate": 7.922913902970632e-07, + "loss": 0.5669, + "num_input_tokens_seen": 258024305, + "step": 7930 + }, + { + "epoch": 0.7152455246426478, + "flos": 21293815252800.0, + "grad_norm": 2.066186442519814, + "learning_rate": 7.918257706258744e-07, + "loss": 0.7323, + "num_input_tokens_seen": 258052565, + "step": 7931 + }, + { + "epoch": 0.715335708166118, + "flos": 22896567394080.0, + "grad_norm": 3.6655621743570452, + "learning_rate": 7.913602540427724e-07, + "loss": 0.7448, + "num_input_tokens_seen": 258080800, + "step": 7932 + }, + { + "epoch": 0.7154258916895884, + "flos": 23079471281280.0, + "grad_norm": 1.781530993868377, + "learning_rate": 7.908948405874775e-07, + "loss": 0.7386, + "num_input_tokens_seen": 258108090, + "step": 7933 + }, + { + "epoch": 0.7155160752130586, + "flos": 34956075171360.0, + "grad_norm": 1.7452406879565139, + "learning_rate": 7.904295302997019e-07, + "loss": 0.6671, + "num_input_tokens_seen": 258136290, + "step": 7934 + }, + { + "epoch": 0.7156062587365288, + "flos": 21872305625760.0, + "grad_norm": 1.9698810658642598, + "learning_rate": 7.899643232191484e-07, + "loss": 0.7546, + "num_input_tokens_seen": 258163135, + "step": 7935 + }, + { + "epoch": 0.7156964422599991, + "flos": 22715187467040.0, + "grad_norm": 1.6460767718844034, + "learning_rate": 7.894992193855108e-07, + "loss": 0.7262, + "num_input_tokens_seen": 258191560, + "step": 7936 + }, + { + "epoch": 0.7157866257834694, + "flos": 22493249598240.0, + "grad_norm": 2.151159497251933, + "learning_rate": 7.890342188384751e-07, + "loss": 0.6912, + "num_input_tokens_seen": 258220770, + "step": 7937 + }, + { + "epoch": 0.7158768093069396, + "flos": 21838475410560.0, + "grad_norm": 2.4521364783721222, + "learning_rate": 7.885693216177165e-07, + "loss": 0.6829, + "num_input_tokens_seen": 258250055, + "step": 7938 + }, + { + "epoch": 0.7159669928304099, + "flos": 23150476990080.0, + "grad_norm": 1.7488042969962376, + "learning_rate": 7.88104527762903e-07, + "loss": 0.7373, + "num_input_tokens_seen": 258279690, + "step": 7939 + }, + { + "epoch": 0.7160571763538801, + "flos": 20528778356160.0, + "grad_norm": 2.563783370776815, + "learning_rate": 7.876398373136936e-07, + "loss": 0.6975, + "num_input_tokens_seen": 258308615, + "step": 7940 + }, + { + "epoch": 0.7161473598773505, + "flos": 28362385737600.0, + "grad_norm": 1.7588806297955513, + "learning_rate": 7.87175250309738e-07, + "loss": 0.7389, + "num_input_tokens_seen": 258339580, + "step": 7941 + }, + { + "epoch": 0.7162375434008207, + "flos": 22343952907680.0, + "grad_norm": 1.805153724591379, + "learning_rate": 7.867107667906785e-07, + "loss": 0.8142, + "num_input_tokens_seen": 258366475, + "step": 7942 + }, + { + "epoch": 0.7163277269242909, + "flos": 24789326701920.0, + "grad_norm": 2.653270261851962, + "learning_rate": 7.862463867961446e-07, + "loss": 0.7406, + "num_input_tokens_seen": 258397625, + "step": 7943 + }, + { + "epoch": 0.7164179104477612, + "flos": 25186176959520.0, + "grad_norm": 1.518195420724965, + "learning_rate": 7.857821103657632e-07, + "loss": 0.8043, + "num_input_tokens_seen": 258425950, + "step": 7944 + }, + { + "epoch": 0.7165080939712315, + "flos": 22093165571520.0, + "grad_norm": 1.7085251894269788, + "learning_rate": 7.853179375391459e-07, + "loss": 0.7725, + "num_input_tokens_seen": 258454490, + "step": 7945 + }, + { + "epoch": 0.7165982774947017, + "flos": 26613793693440.0, + "grad_norm": 2.2269766439130385, + "learning_rate": 7.848538683559012e-07, + "loss": 0.7296, + "num_input_tokens_seen": 258483825, + "step": 7946 + }, + { + "epoch": 0.716688461018172, + "flos": 23585989531680.0, + "grad_norm": 2.152758328202963, + "learning_rate": 7.843899028556238e-07, + "loss": 0.7819, + "num_input_tokens_seen": 258511260, + "step": 7947 + }, + { + "epoch": 0.7167786445416422, + "flos": 23735657919840.0, + "grad_norm": 3.5413581863396817, + "learning_rate": 7.839260410779029e-07, + "loss": 0.7341, + "num_input_tokens_seen": 258543525, + "step": 7948 + }, + { + "epoch": 0.7168688280651125, + "flos": 24719324576640.0, + "grad_norm": 1.7227467623436497, + "learning_rate": 7.834622830623175e-07, + "loss": 0.7621, + "num_input_tokens_seen": 258572730, + "step": 7949 + }, + { + "epoch": 0.7169590115885828, + "flos": 22496334688320.0, + "grad_norm": 2.1143228780764702, + "learning_rate": 7.82998628848438e-07, + "loss": 0.7126, + "num_input_tokens_seen": 258600705, + "step": 7950 + }, + { + "epoch": 0.717049195112053, + "flos": 16996202922720.0, + "grad_norm": 1.5692286760365146, + "learning_rate": 7.825350784758261e-07, + "loss": 0.7314, + "num_input_tokens_seen": 258629260, + "step": 7951 + }, + { + "epoch": 0.7171393786355232, + "flos": 24609061867680.0, + "grad_norm": 1.6679007094216542, + "learning_rate": 7.820716319840342e-07, + "loss": 0.7055, + "num_input_tokens_seen": 258658640, + "step": 7952 + }, + { + "epoch": 0.7172295621589936, + "flos": 25301940792960.0, + "grad_norm": 2.8173918672677285, + "learning_rate": 7.816082894126069e-07, + "loss": 0.8265, + "num_input_tokens_seen": 258688140, + "step": 7953 + }, + { + "epoch": 0.7173197456824638, + "flos": 24718990048800.0, + "grad_norm": 2.158277470615684, + "learning_rate": 7.811450508010778e-07, + "loss": 0.7133, + "num_input_tokens_seen": 258715385, + "step": 7954 + }, + { + "epoch": 0.717409929205934, + "flos": 25010223817440.0, + "grad_norm": 4.785335615260814, + "learning_rate": 7.806819161889737e-07, + "loss": 0.7128, + "num_input_tokens_seen": 258744355, + "step": 7955 + }, + { + "epoch": 0.7175001127294043, + "flos": 20012484458880.0, + "grad_norm": 2.0164639405219456, + "learning_rate": 7.802188856158119e-07, + "loss": 0.774, + "num_input_tokens_seen": 258771725, + "step": 7956 + }, + { + "epoch": 0.7175902962528746, + "flos": 21549360318240.0, + "grad_norm": 2.0594363722968736, + "learning_rate": 7.797559591211002e-07, + "loss": 0.7275, + "num_input_tokens_seen": 258799210, + "step": 7957 + }, + { + "epoch": 0.7176804797763449, + "flos": 27779509332960.0, + "grad_norm": 1.8862479749988483, + "learning_rate": 7.79293136744339e-07, + "loss": 0.7061, + "num_input_tokens_seen": 258828780, + "step": 7958 + }, + { + "epoch": 0.7177706632998151, + "flos": 18889371097920.0, + "grad_norm": 1.9830157813801887, + "learning_rate": 7.788304185250185e-07, + "loss": 0.7375, + "num_input_tokens_seen": 258855400, + "step": 7959 + }, + { + "epoch": 0.7178608468232854, + "flos": 21184147260000.0, + "grad_norm": 1.9191273221968599, + "learning_rate": 7.78367804502619e-07, + "loss": 0.7677, + "num_input_tokens_seen": 258882585, + "step": 7960 + }, + { + "epoch": 0.7179510303467557, + "flos": 12586134867840.0, + "grad_norm": 2.449998843912903, + "learning_rate": 7.779052947166156e-07, + "loss": 0.749, + "num_input_tokens_seen": 258907385, + "step": 7961 + }, + { + "epoch": 0.7180412138702259, + "flos": 26212706083200.0, + "grad_norm": 3.990254451124238, + "learning_rate": 7.774428892064697e-07, + "loss": 0.6959, + "num_input_tokens_seen": 258936090, + "step": 7962 + }, + { + "epoch": 0.7181313973936961, + "flos": 24208717652640.0, + "grad_norm": 1.7550522218852034, + "learning_rate": 7.769805880116391e-07, + "loss": 0.7333, + "num_input_tokens_seen": 258963900, + "step": 7963 + }, + { + "epoch": 0.7182215809171665, + "flos": 13716050294880.0, + "grad_norm": 2.393434556540126, + "learning_rate": 7.765183911715678e-07, + "loss": 0.7187, + "num_input_tokens_seen": 258989075, + "step": 7964 + }, + { + "epoch": 0.7183117644406367, + "flos": 70548042292320.0, + "grad_norm": 4.287431250374118, + "learning_rate": 7.760562987256933e-07, + "loss": 0.6411, + "num_input_tokens_seen": 259023990, + "step": 7965 + }, + { + "epoch": 0.718401947964107, + "flos": 17250707234880.0, + "grad_norm": 1.8812417090350486, + "learning_rate": 7.755943107134444e-07, + "loss": 0.7462, + "num_input_tokens_seen": 259051785, + "step": 7966 + }, + { + "epoch": 0.7184921314875772, + "flos": 59095980131040.0, + "grad_norm": 0.633055863856241, + "learning_rate": 7.751324271742401e-07, + "loss": 0.5674, + "num_input_tokens_seen": 259135375, + "step": 7967 + }, + { + "epoch": 0.7185823150110475, + "flos": 25665741400320.0, + "grad_norm": 1.8083047656543632, + "learning_rate": 7.746706481474916e-07, + "loss": 0.7526, + "num_input_tokens_seen": 259166260, + "step": 7968 + }, + { + "epoch": 0.7186724985345178, + "flos": 68629141176000.0, + "grad_norm": 0.5984691999918862, + "learning_rate": 7.742089736725992e-07, + "loss": 0.5361, + "num_input_tokens_seen": 259262425, + "step": 7969 + }, + { + "epoch": 0.718762682057988, + "flos": 24573001466880.0, + "grad_norm": 6.74366180107697, + "learning_rate": 7.737474037889559e-07, + "loss": 0.6595, + "num_input_tokens_seen": 259290800, + "step": 7970 + }, + { + "epoch": 0.7188528655814582, + "flos": 31820015467680.0, + "grad_norm": 1.7614607865286207, + "learning_rate": 7.732859385359458e-07, + "loss": 0.7124, + "num_input_tokens_seen": 259320710, + "step": 7971 + }, + { + "epoch": 0.7189430491049286, + "flos": 26503828342560.0, + "grad_norm": 2.836399513631275, + "learning_rate": 7.728245779529434e-07, + "loss": 0.6609, + "num_input_tokens_seen": 259348615, + "step": 7972 + }, + { + "epoch": 0.7190332326283988, + "flos": 19436298611040.0, + "grad_norm": 1.8218234753397686, + "learning_rate": 7.723633220793146e-07, + "loss": 0.7683, + "num_input_tokens_seen": 259374630, + "step": 7973 + }, + { + "epoch": 0.719123416151869, + "flos": 27087150784320.0, + "grad_norm": 2.389154218630855, + "learning_rate": 7.719021709544162e-07, + "loss": 0.6324, + "num_input_tokens_seen": 259402895, + "step": 7974 + }, + { + "epoch": 0.7192135996753393, + "flos": 23297729343840.0, + "grad_norm": 1.7795527888649254, + "learning_rate": 7.714411246175964e-07, + "loss": 0.8291, + "num_input_tokens_seen": 259431290, + "step": 7975 + }, + { + "epoch": 0.7193037831988096, + "flos": 24753861017280.0, + "grad_norm": 1.7782426175066548, + "learning_rate": 7.709801831081946e-07, + "loss": 0.6586, + "num_input_tokens_seen": 259460200, + "step": 7976 + }, + { + "epoch": 0.7193939667222798, + "flos": 28179853548000.0, + "grad_norm": 2.3914442466252055, + "learning_rate": 7.705193464655391e-07, + "loss": 0.6971, + "num_input_tokens_seen": 259489925, + "step": 7977 + }, + { + "epoch": 0.7194841502457501, + "flos": 23367396941280.0, + "grad_norm": 1.4510629441051572, + "learning_rate": 7.700586147289534e-07, + "loss": 0.7176, + "num_input_tokens_seen": 259519520, + "step": 7978 + }, + { + "epoch": 0.7195743337692203, + "flos": 20055978811680.0, + "grad_norm": 1.734751961097428, + "learning_rate": 7.695979879377481e-07, + "loss": 0.698, + "num_input_tokens_seen": 259546930, + "step": 7979 + }, + { + "epoch": 0.7196645172926907, + "flos": 26359623909120.0, + "grad_norm": 1.8280684633560416, + "learning_rate": 7.691374661312266e-07, + "loss": 0.7593, + "num_input_tokens_seen": 259578210, + "step": 7980 + }, + { + "epoch": 0.7197547008161609, + "flos": 37649485739520.0, + "grad_norm": 1.776510860930638, + "learning_rate": 7.686770493486834e-07, + "loss": 0.7138, + "num_input_tokens_seen": 259610165, + "step": 7981 + }, + { + "epoch": 0.7198448843396311, + "flos": 22641394026240.0, + "grad_norm": 2.408569803035351, + "learning_rate": 7.68216737629404e-07, + "loss": 0.6806, + "num_input_tokens_seen": 259638720, + "step": 7982 + }, + { + "epoch": 0.7199350678631015, + "flos": 26613570674880.0, + "grad_norm": 1.6009702750549422, + "learning_rate": 7.67756531012665e-07, + "loss": 0.8078, + "num_input_tokens_seen": 259669095, + "step": 7983 + }, + { + "epoch": 0.7200252513865717, + "flos": 25008402499200.0, + "grad_norm": 3.0179996924378054, + "learning_rate": 7.67296429537732e-07, + "loss": 0.7974, + "num_input_tokens_seen": 259699150, + "step": 7984 + }, + { + "epoch": 0.7201154349100419, + "flos": 19582287192960.0, + "grad_norm": 2.1752300009299708, + "learning_rate": 7.668364332438661e-07, + "loss": 0.8035, + "num_input_tokens_seen": 259727265, + "step": 7985 + }, + { + "epoch": 0.7202056184335122, + "flos": 22386815374560.0, + "grad_norm": 2.025170839426522, + "learning_rate": 7.663765421703145e-07, + "loss": 0.7514, + "num_input_tokens_seen": 259758020, + "step": 7986 + }, + { + "epoch": 0.7202958019569825, + "flos": 23808745135200.0, + "grad_norm": 1.9583940767106691, + "learning_rate": 7.659167563563187e-07, + "loss": 0.6767, + "num_input_tokens_seen": 259788690, + "step": 7987 + }, + { + "epoch": 0.7203859854804527, + "flos": 19728052756320.0, + "grad_norm": 2.574582341960472, + "learning_rate": 7.654570758411096e-07, + "loss": 0.7621, + "num_input_tokens_seen": 259817825, + "step": 7988 + }, + { + "epoch": 0.720476169003923, + "flos": 19982817256800.0, + "grad_norm": 1.9778608196960228, + "learning_rate": 7.649975006639103e-07, + "loss": 0.8, + "num_input_tokens_seen": 259845465, + "step": 7989 + }, + { + "epoch": 0.7205663525273932, + "flos": 66426934917120.0, + "grad_norm": 0.6229004882276565, + "learning_rate": 7.645380308639337e-07, + "loss": 0.5501, + "num_input_tokens_seen": 259928775, + "step": 7990 + }, + { + "epoch": 0.7206565360508635, + "flos": 23334458800320.0, + "grad_norm": 1.7581434150325352, + "learning_rate": 7.640786664803853e-07, + "loss": 0.7857, + "num_input_tokens_seen": 259956540, + "step": 7991 + }, + { + "epoch": 0.7207467195743338, + "flos": 53210080429920.0, + "grad_norm": 2.0543255505298603, + "learning_rate": 7.636194075524587e-07, + "loss": 0.6622, + "num_input_tokens_seen": 259992640, + "step": 7992 + }, + { + "epoch": 0.720836903097804, + "flos": 29965435236960.0, + "grad_norm": 1.985420912597642, + "learning_rate": 7.631602541193429e-07, + "loss": 0.6942, + "num_input_tokens_seen": 260024095, + "step": 7993 + }, + { + "epoch": 0.7209270866212742, + "flos": 25183909604160.0, + "grad_norm": 2.870725688567946, + "learning_rate": 7.627012062202132e-07, + "loss": 0.7945, + "num_input_tokens_seen": 260049885, + "step": 7994 + }, + { + "epoch": 0.7210172701447446, + "flos": 14517296271360.0, + "grad_norm": 2.34814427741278, + "learning_rate": 7.622422638942391e-07, + "loss": 0.7388, + "num_input_tokens_seen": 260077300, + "step": 7995 + }, + { + "epoch": 0.7211074536682148, + "flos": 24134961381600.0, + "grad_norm": 2.065627068361294, + "learning_rate": 7.617834271805801e-07, + "loss": 0.8043, + "num_input_tokens_seen": 260108095, + "step": 7996 + }, + { + "epoch": 0.7211976371916851, + "flos": 69946197842880.0, + "grad_norm": 0.6901836901256198, + "learning_rate": 7.613246961183863e-07, + "loss": 0.5645, + "num_input_tokens_seen": 260196985, + "step": 7997 + }, + { + "epoch": 0.7212878207151553, + "flos": 22497301102080.0, + "grad_norm": 1.8031859653684932, + "learning_rate": 7.608660707468002e-07, + "loss": 0.6978, + "num_input_tokens_seen": 260227785, + "step": 7998 + }, + { + "epoch": 0.7213780042386256, + "flos": 30982783429920.0, + "grad_norm": 2.567763342793632, + "learning_rate": 7.604075511049522e-07, + "loss": 0.6334, + "num_input_tokens_seen": 260260280, + "step": 7999 + }, + { + "epoch": 0.7214681877620959, + "flos": 32332666728480.0, + "grad_norm": 1.844750278496143, + "learning_rate": 7.599491372319682e-07, + "loss": 0.6851, + "num_input_tokens_seen": 260291985, + "step": 8000 + }, + { + "epoch": 0.7215583712855661, + "flos": 26501337968640.0, + "grad_norm": 1.3779698723570992, + "learning_rate": 7.594908291669601e-07, + "loss": 0.7376, + "num_input_tokens_seen": 260324810, + "step": 8001 + }, + { + "epoch": 0.7216485548090363, + "flos": 19506746773440.0, + "grad_norm": 3.0935530166337273, + "learning_rate": 7.590326269490359e-07, + "loss": 0.7746, + "num_input_tokens_seen": 260353540, + "step": 8002 + }, + { + "epoch": 0.7217387383325067, + "flos": 22679275745280.0, + "grad_norm": 1.572311419655285, + "learning_rate": 7.585745306172899e-07, + "loss": 0.6987, + "num_input_tokens_seen": 260383690, + "step": 8003 + }, + { + "epoch": 0.7218289218559769, + "flos": 21003287709600.0, + "grad_norm": 3.8584952027302286, + "learning_rate": 7.5811654021081e-07, + "loss": 0.7183, + "num_input_tokens_seen": 260411765, + "step": 8004 + }, + { + "epoch": 0.7219191053794471, + "flos": 24821707296480.0, + "grad_norm": 1.6213512537858896, + "learning_rate": 7.576586557686748e-07, + "loss": 0.7692, + "num_input_tokens_seen": 260440095, + "step": 8005 + }, + { + "epoch": 0.7220092889029175, + "flos": 30731401377600.0, + "grad_norm": 2.078900529523145, + "learning_rate": 7.572008773299531e-07, + "loss": 0.7117, + "num_input_tokens_seen": 260471085, + "step": 8006 + }, + { + "epoch": 0.7220994724263877, + "flos": 29709295455360.0, + "grad_norm": 1.7407524459451336, + "learning_rate": 7.567432049337055e-07, + "loss": 0.7757, + "num_input_tokens_seen": 260502345, + "step": 8007 + }, + { + "epoch": 0.722189655949858, + "flos": 21440249871840.0, + "grad_norm": 2.121977283108105, + "learning_rate": 7.562856386189834e-07, + "loss": 0.6758, + "num_input_tokens_seen": 260530410, + "step": 8008 + }, + { + "epoch": 0.7222798394733282, + "flos": 21841263142560.0, + "grad_norm": 3.112882543135214, + "learning_rate": 7.558281784248275e-07, + "loss": 0.7512, + "num_input_tokens_seen": 260558375, + "step": 8009 + }, + { + "epoch": 0.7223700229967985, + "flos": 19654222145760.0, + "grad_norm": 1.907248751291018, + "learning_rate": 7.553708243902721e-07, + "loss": 0.7267, + "num_input_tokens_seen": 260584320, + "step": 8010 + }, + { + "epoch": 0.7224602065202688, + "flos": 31350524031840.0, + "grad_norm": 1.7266430728115356, + "learning_rate": 7.549135765543404e-07, + "loss": 0.6278, + "num_input_tokens_seen": 260617350, + "step": 8011 + }, + { + "epoch": 0.722550390043739, + "flos": 20894028584160.0, + "grad_norm": 3.1353366034710577, + "learning_rate": 7.544564349560481e-07, + "loss": 0.7624, + "num_input_tokens_seen": 260642580, + "step": 8012 + }, + { + "epoch": 0.7226405735672092, + "flos": 28036057981920.0, + "grad_norm": 13.157665475554603, + "learning_rate": 7.539993996344009e-07, + "loss": 0.723, + "num_input_tokens_seen": 260674490, + "step": 8013 + }, + { + "epoch": 0.7227307570906796, + "flos": 21659028311040.0, + "grad_norm": 1.815392586676269, + "learning_rate": 7.535424706283941e-07, + "loss": 0.7169, + "num_input_tokens_seen": 260702605, + "step": 8014 + }, + { + "epoch": 0.7228209406141498, + "flos": 26136831135840.0, + "grad_norm": 2.2580223118006604, + "learning_rate": 7.530856479770181e-07, + "loss": 0.6987, + "num_input_tokens_seen": 260732400, + "step": 8015 + }, + { + "epoch": 0.72291112413762, + "flos": 27340577173440.0, + "grad_norm": 1.6431127669211678, + "learning_rate": 7.526289317192484e-07, + "loss": 0.8172, + "num_input_tokens_seen": 260759960, + "step": 8016 + }, + { + "epoch": 0.7230013076610903, + "flos": 18780037632960.0, + "grad_norm": 2.1931397896260103, + "learning_rate": 7.521723218940579e-07, + "loss": 0.735, + "num_input_tokens_seen": 260785050, + "step": 8017 + }, + { + "epoch": 0.7230914911845606, + "flos": 22460125608480.0, + "grad_norm": 2.8212516256656452, + "learning_rate": 7.517158185404038e-07, + "loss": 0.799, + "num_input_tokens_seen": 260812365, + "step": 8018 + }, + { + "epoch": 0.7231816747080309, + "flos": 26758927370880.0, + "grad_norm": 2.1717152708308274, + "learning_rate": 7.512594216972403e-07, + "loss": 0.7626, + "num_input_tokens_seen": 260843900, + "step": 8019 + }, + { + "epoch": 0.7232718582315011, + "flos": 22021193448960.0, + "grad_norm": 1.638474790158588, + "learning_rate": 7.508031314035078e-07, + "loss": 0.7736, + "num_input_tokens_seen": 260873330, + "step": 8020 + }, + { + "epoch": 0.7233620417549713, + "flos": 22860804351360.0, + "grad_norm": 1.8323100514345307, + "learning_rate": 7.503469476981401e-07, + "loss": 0.7589, + "num_input_tokens_seen": 260901490, + "step": 8021 + }, + { + "epoch": 0.7234522252784417, + "flos": 34190963935200.0, + "grad_norm": 2.9218919722936207, + "learning_rate": 7.498908706200613e-07, + "loss": 0.6112, + "num_input_tokens_seen": 260932195, + "step": 8022 + }, + { + "epoch": 0.7235424088019119, + "flos": 22935080999040.0, + "grad_norm": 1.7031116310084877, + "learning_rate": 7.494349002081866e-07, + "loss": 0.7865, + "num_input_tokens_seen": 260962045, + "step": 8023 + }, + { + "epoch": 0.7236325923253821, + "flos": 20126129616000.0, + "grad_norm": 2.391038646939281, + "learning_rate": 7.489790365014224e-07, + "loss": 0.766, + "num_input_tokens_seen": 260990360, + "step": 8024 + }, + { + "epoch": 0.7237227758488524, + "flos": 27743597611200.0, + "grad_norm": 2.659491502157664, + "learning_rate": 7.485232795386642e-07, + "loss": 0.757, + "num_input_tokens_seen": 261021300, + "step": 8025 + }, + { + "epoch": 0.7238129593723227, + "flos": 22858202468160.0, + "grad_norm": 1.9256146031946841, + "learning_rate": 7.480676293588002e-07, + "loss": 0.762, + "num_input_tokens_seen": 261049125, + "step": 8026 + }, + { + "epoch": 0.7239031428957929, + "flos": 67559526566880.0, + "grad_norm": 0.7625747956659631, + "learning_rate": 7.476120860007093e-07, + "loss": 0.5958, + "num_input_tokens_seen": 261133175, + "step": 8027 + }, + { + "epoch": 0.7239933264192632, + "flos": 22932070248480.0, + "grad_norm": 2.1994467256324404, + "learning_rate": 7.471566495032608e-07, + "loss": 0.7803, + "num_input_tokens_seen": 261156875, + "step": 8028 + }, + { + "epoch": 0.7240835099427335, + "flos": 37979344622400.0, + "grad_norm": 2.0235967978439136, + "learning_rate": 7.467013199053152e-07, + "loss": 0.6737, + "num_input_tokens_seen": 261187110, + "step": 8029 + }, + { + "epoch": 0.7241736934662037, + "flos": 19071866117760.0, + "grad_norm": 1.9934662548239195, + "learning_rate": 7.46246097245724e-07, + "loss": 0.7359, + "num_input_tokens_seen": 261213825, + "step": 8030 + }, + { + "epoch": 0.724263876989674, + "flos": 21870707326080.0, + "grad_norm": 2.061836945274215, + "learning_rate": 7.457909815633276e-07, + "loss": 0.7419, + "num_input_tokens_seen": 261240155, + "step": 8031 + }, + { + "epoch": 0.7243540605131442, + "flos": 31386918960480.0, + "grad_norm": 2.29383254149558, + "learning_rate": 7.453359728969618e-07, + "loss": 0.7314, + "num_input_tokens_seen": 261269005, + "step": 8032 + }, + { + "epoch": 0.7244442440366146, + "flos": 35794385132160.0, + "grad_norm": 1.6843917619370128, + "learning_rate": 7.448810712854475e-07, + "loss": 0.6654, + "num_input_tokens_seen": 261301810, + "step": 8033 + }, + { + "epoch": 0.7245344275600848, + "flos": 21804533686080.0, + "grad_norm": 1.7518928372866251, + "learning_rate": 7.444262767676022e-07, + "loss": 0.7786, + "num_input_tokens_seen": 261330235, + "step": 8034 + }, + { + "epoch": 0.724624611083555, + "flos": 22781026579200.0, + "grad_norm": 1.8967663352044686, + "learning_rate": 7.439715893822296e-07, + "loss": 0.812, + "num_input_tokens_seen": 261357205, + "step": 8035 + }, + { + "epoch": 0.7247147946070253, + "flos": 18562969002720.0, + "grad_norm": 7.292207127477871, + "learning_rate": 7.435170091681264e-07, + "loss": 0.7617, + "num_input_tokens_seen": 261386780, + "step": 8036 + }, + { + "epoch": 0.7248049781304956, + "flos": 24969257008320.0, + "grad_norm": 2.3890800719927885, + "learning_rate": 7.430625361640803e-07, + "loss": 0.6967, + "num_input_tokens_seen": 261415345, + "step": 8037 + }, + { + "epoch": 0.7248951616539658, + "flos": 20638334839680.0, + "grad_norm": 2.3097093600783296, + "learning_rate": 7.426081704088694e-07, + "loss": 0.7823, + "num_input_tokens_seen": 261442500, + "step": 8038 + }, + { + "epoch": 0.7249853451774361, + "flos": 18962495483040.0, + "grad_norm": 1.9188226170420228, + "learning_rate": 7.42153911941263e-07, + "loss": 0.8096, + "num_input_tokens_seen": 261469350, + "step": 8039 + }, + { + "epoch": 0.7250755287009063, + "flos": 18121583639040.0, + "grad_norm": 1.7709874777930061, + "learning_rate": 7.416997608000192e-07, + "loss": 0.78, + "num_input_tokens_seen": 261496165, + "step": 8040 + }, + { + "epoch": 0.7251657122243766, + "flos": 19108000858080.0, + "grad_norm": 2.06325297626968, + "learning_rate": 7.412457170238918e-07, + "loss": 0.7215, + "num_input_tokens_seen": 261522760, + "step": 8041 + }, + { + "epoch": 0.7252558957478469, + "flos": 21549917864640.0, + "grad_norm": 1.831421122188309, + "learning_rate": 7.407917806516193e-07, + "loss": 0.6859, + "num_input_tokens_seen": 261551405, + "step": 8042 + }, + { + "epoch": 0.7253460792713171, + "flos": 22379753120160.0, + "grad_norm": 2.3179540496761173, + "learning_rate": 7.403379517219354e-07, + "loss": 0.7318, + "num_input_tokens_seen": 261579260, + "step": 8043 + }, + { + "epoch": 0.7254362627947873, + "flos": 24317976778080.0, + "grad_norm": 2.215365878126795, + "learning_rate": 7.398842302735636e-07, + "loss": 0.7668, + "num_input_tokens_seen": 261604595, + "step": 8044 + }, + { + "epoch": 0.7255264463182577, + "flos": 67731242356320.0, + "grad_norm": 0.6675913229631562, + "learning_rate": 7.394306163452171e-07, + "loss": 0.5453, + "num_input_tokens_seen": 261695115, + "step": 8045 + }, + { + "epoch": 0.7256166298417279, + "flos": 24317902438560.0, + "grad_norm": 2.195710169928989, + "learning_rate": 7.38977109975601e-07, + "loss": 0.7169, + "num_input_tokens_seen": 261725155, + "step": 8046 + }, + { + "epoch": 0.7257068133651982, + "flos": 28470715619040.0, + "grad_norm": 2.444936705162024, + "learning_rate": 7.385237112034119e-07, + "loss": 0.6507, + "num_input_tokens_seen": 261753800, + "step": 8047 + }, + { + "epoch": 0.7257969968886684, + "flos": 17723804137440.0, + "grad_norm": 2.2163236003602163, + "learning_rate": 7.380704200673342e-07, + "loss": 0.8031, + "num_input_tokens_seen": 261781375, + "step": 8048 + }, + { + "epoch": 0.7258871804121387, + "flos": 24535194087360.0, + "grad_norm": 1.5131364823054998, + "learning_rate": 7.376172366060478e-07, + "loss": 0.7636, + "num_input_tokens_seen": 261810865, + "step": 8049 + }, + { + "epoch": 0.725977363935609, + "flos": 53154128473920.0, + "grad_norm": 0.6638582252016444, + "learning_rate": 7.371641608582187e-07, + "loss": 0.5114, + "num_input_tokens_seen": 261896210, + "step": 8050 + }, + { + "epoch": 0.7260675474590792, + "flos": 20272527065280.0, + "grad_norm": 2.682925211827706, + "learning_rate": 7.367111928625067e-07, + "loss": 0.6066, + "num_input_tokens_seen": 261923995, + "step": 8051 + }, + { + "epoch": 0.7261577309825495, + "flos": 27888508270080.0, + "grad_norm": 1.9278460531148482, + "learning_rate": 7.362583326575613e-07, + "loss": 0.6515, + "num_input_tokens_seen": 261953365, + "step": 8052 + }, + { + "epoch": 0.7262479145060198, + "flos": 23808336267840.0, + "grad_norm": 2.2242131202661346, + "learning_rate": 7.358055802820234e-07, + "loss": 0.7285, + "num_input_tokens_seen": 261981395, + "step": 8053 + }, + { + "epoch": 0.72633809802949, + "flos": 27043470582720.0, + "grad_norm": 2.8907257887306304, + "learning_rate": 7.353529357745245e-07, + "loss": 0.7082, + "num_input_tokens_seen": 262009570, + "step": 8054 + }, + { + "epoch": 0.7264282815529602, + "flos": 20419779419040.0, + "grad_norm": 2.1027948673057213, + "learning_rate": 7.349003991736851e-07, + "loss": 0.7781, + "num_input_tokens_seen": 262035300, + "step": 8055 + }, + { + "epoch": 0.7265184650764306, + "flos": 29090395819680.0, + "grad_norm": 1.8589822049738465, + "learning_rate": 7.344479705181206e-07, + "loss": 0.5889, + "num_input_tokens_seen": 262063025, + "step": 8056 + }, + { + "epoch": 0.7266086485999008, + "flos": 22934114585280.0, + "grad_norm": 2.87380840652004, + "learning_rate": 7.339956498464322e-07, + "loss": 0.6895, + "num_input_tokens_seen": 262092790, + "step": 8057 + }, + { + "epoch": 0.726698832123371, + "flos": 20565582152160.0, + "grad_norm": 1.9471793045608583, + "learning_rate": 7.335434371972169e-07, + "loss": 0.7859, + "num_input_tokens_seen": 262120850, + "step": 8058 + }, + { + "epoch": 0.7267890156468413, + "flos": 23626027096800.0, + "grad_norm": 1.7277607621884554, + "learning_rate": 7.33091332609058e-07, + "loss": 0.7279, + "num_input_tokens_seen": 262149015, + "step": 8059 + }, + { + "epoch": 0.7268791991703116, + "flos": 18407948169120.0, + "grad_norm": 2.26963209758728, + "learning_rate": 7.326393361205323e-07, + "loss": 0.7556, + "num_input_tokens_seen": 262175205, + "step": 8060 + }, + { + "epoch": 0.7269693826937819, + "flos": 14955224847360.0, + "grad_norm": 1.9652831418608605, + "learning_rate": 7.321874477702068e-07, + "loss": 0.8191, + "num_input_tokens_seen": 262201715, + "step": 8061 + }, + { + "epoch": 0.7270595662172521, + "flos": 69163170782400.0, + "grad_norm": 0.7537350633194565, + "learning_rate": 7.317356675966386e-07, + "loss": 0.5911, + "num_input_tokens_seen": 262291740, + "step": 8062 + }, + { + "epoch": 0.7271497497407223, + "flos": 20237247229440.0, + "grad_norm": 1.8143129161130014, + "learning_rate": 7.312839956383765e-07, + "loss": 0.7718, + "num_input_tokens_seen": 262322020, + "step": 8063 + }, + { + "epoch": 0.7272399332641927, + "flos": 22021639486080.0, + "grad_norm": 1.8763098909602547, + "learning_rate": 7.308324319339603e-07, + "loss": 0.7182, + "num_input_tokens_seen": 262351970, + "step": 8064 + }, + { + "epoch": 0.7273301167876629, + "flos": 31890129102240.0, + "grad_norm": 2.8594271962555777, + "learning_rate": 7.303809765219182e-07, + "loss": 0.741, + "num_input_tokens_seen": 262382480, + "step": 8065 + }, + { + "epoch": 0.7274203003111331, + "flos": 24536978235840.0, + "grad_norm": 1.7386090944569015, + "learning_rate": 7.299296294407719e-07, + "loss": 0.7857, + "num_input_tokens_seen": 262410250, + "step": 8066 + }, + { + "epoch": 0.7275104838346034, + "flos": 25081080847200.0, + "grad_norm": 1.994046947883108, + "learning_rate": 7.294783907290327e-07, + "loss": 0.7802, + "num_input_tokens_seen": 262439105, + "step": 8067 + }, + { + "epoch": 0.7276006673580737, + "flos": 37287692299200.0, + "grad_norm": 2.3220741443279955, + "learning_rate": 7.290272604252028e-07, + "loss": 0.6061, + "num_input_tokens_seen": 262472730, + "step": 8068 + }, + { + "epoch": 0.727690850881544, + "flos": 37941946110240.0, + "grad_norm": 1.9729050281131688, + "learning_rate": 7.285762385677758e-07, + "loss": 0.6448, + "num_input_tokens_seen": 262501640, + "step": 8069 + }, + { + "epoch": 0.7277810344050142, + "flos": 20490413430240.0, + "grad_norm": 1.8885371762737986, + "learning_rate": 7.281253251952335e-07, + "loss": 0.7784, + "num_input_tokens_seen": 262530375, + "step": 8070 + }, + { + "epoch": 0.7278712179284844, + "flos": 22492803561120.0, + "grad_norm": 1.6540811121008956, + "learning_rate": 7.276745203460526e-07, + "loss": 0.7194, + "num_input_tokens_seen": 262559950, + "step": 8071 + }, + { + "epoch": 0.7279614014519548, + "flos": 69868650256320.0, + "grad_norm": 0.7226999218091003, + "learning_rate": 7.272238240586959e-07, + "loss": 0.6315, + "num_input_tokens_seen": 262653920, + "step": 8072 + }, + { + "epoch": 0.728051584975425, + "flos": 31204349601120.0, + "grad_norm": 2.3669836378285436, + "learning_rate": 7.267732363716219e-07, + "loss": 0.6744, + "num_input_tokens_seen": 262687305, + "step": 8073 + }, + { + "epoch": 0.7281417684988952, + "flos": 15501408965280.0, + "grad_norm": 1.6204953479505864, + "learning_rate": 7.263227573232753e-07, + "loss": 0.6938, + "num_input_tokens_seen": 262715670, + "step": 8074 + }, + { + "epoch": 0.7282319520223655, + "flos": 26577175746240.0, + "grad_norm": 3.0549330356582214, + "learning_rate": 7.258723869520937e-07, + "loss": 0.828, + "num_input_tokens_seen": 262742790, + "step": 8075 + }, + { + "epoch": 0.7283221355458358, + "flos": 19796642430720.0, + "grad_norm": 1.6539567870292093, + "learning_rate": 7.254221252965059e-07, + "loss": 0.7837, + "num_input_tokens_seen": 262770620, + "step": 8076 + }, + { + "epoch": 0.728412319069306, + "flos": 25374210273600.0, + "grad_norm": 2.224265322316575, + "learning_rate": 7.249719723949301e-07, + "loss": 0.6559, + "num_input_tokens_seen": 262799485, + "step": 8077 + }, + { + "epoch": 0.7285025025927763, + "flos": 23295759346560.0, + "grad_norm": 1.7303293830774071, + "learning_rate": 7.245219282857761e-07, + "loss": 0.7614, + "num_input_tokens_seen": 262829110, + "step": 8078 + }, + { + "epoch": 0.7285926861162466, + "flos": 27044585675520.0, + "grad_norm": 2.011326308998804, + "learning_rate": 7.240719930074442e-07, + "loss": 0.7796, + "num_input_tokens_seen": 262857910, + "step": 8079 + }, + { + "epoch": 0.7286828696397168, + "flos": 23655731468640.0, + "grad_norm": 10.465400729391963, + "learning_rate": 7.236221665983257e-07, + "loss": 0.7674, + "num_input_tokens_seen": 262885275, + "step": 8080 + }, + { + "epoch": 0.7287730531631871, + "flos": 61491274791360.0, + "grad_norm": 0.7242963171183117, + "learning_rate": 7.231724490968012e-07, + "loss": 0.5566, + "num_input_tokens_seen": 262973625, + "step": 8081 + }, + { + "epoch": 0.7288632366866573, + "flos": 23079954488160.0, + "grad_norm": 1.8857501527133047, + "learning_rate": 7.227228405412438e-07, + "loss": 0.7088, + "num_input_tokens_seen": 263001905, + "step": 8082 + }, + { + "epoch": 0.7289534202101277, + "flos": 19830286797120.0, + "grad_norm": 1.9251631234152113, + "learning_rate": 7.222733409700165e-07, + "loss": 0.6974, + "num_input_tokens_seen": 263029795, + "step": 8083 + }, + { + "epoch": 0.7290436037335979, + "flos": 26650411640640.0, + "grad_norm": 1.5750304748977684, + "learning_rate": 7.21823950421473e-07, + "loss": 0.7498, + "num_input_tokens_seen": 263062200, + "step": 8084 + }, + { + "epoch": 0.7291337872570681, + "flos": 21148161198720.0, + "grad_norm": 2.001013022407582, + "learning_rate": 7.213746689339577e-07, + "loss": 0.7076, + "num_input_tokens_seen": 263089845, + "step": 8085 + }, + { + "epoch": 0.7292239707805384, + "flos": 29747363023200.0, + "grad_norm": 2.506195783147597, + "learning_rate": 7.20925496545807e-07, + "loss": 0.7422, + "num_input_tokens_seen": 263120215, + "step": 8086 + }, + { + "epoch": 0.7293141543040087, + "flos": 58710026346720.0, + "grad_norm": 1.73136516972941, + "learning_rate": 7.20476433295344e-07, + "loss": 0.6469, + "num_input_tokens_seen": 263154360, + "step": 8087 + }, + { + "epoch": 0.7294043378274789, + "flos": 25374842159520.0, + "grad_norm": 4.390786164607717, + "learning_rate": 7.200274792208882e-07, + "loss": 0.7786, + "num_input_tokens_seen": 263185160, + "step": 8088 + }, + { + "epoch": 0.7294945213509492, + "flos": 27050718685920.0, + "grad_norm": 2.0128154246820276, + "learning_rate": 7.195786343607444e-07, + "loss": 0.6292, + "num_input_tokens_seen": 263214310, + "step": 8089 + }, + { + "epoch": 0.7295847048744194, + "flos": 21658284915840.0, + "grad_norm": 1.9146461686818714, + "learning_rate": 7.191298987532131e-07, + "loss": 0.7325, + "num_input_tokens_seen": 263243065, + "step": 8090 + }, + { + "epoch": 0.7296748883978897, + "flos": 27779360653920.0, + "grad_norm": 1.8724196162369158, + "learning_rate": 7.186812724365805e-07, + "loss": 0.7449, + "num_input_tokens_seen": 263275955, + "step": 8091 + }, + { + "epoch": 0.72976507192136, + "flos": 24755384977440.0, + "grad_norm": 4.692664594654598, + "learning_rate": 7.182327554491272e-07, + "loss": 0.7466, + "num_input_tokens_seen": 263303475, + "step": 8092 + }, + { + "epoch": 0.7298552554448302, + "flos": 62473566167040.0, + "grad_norm": 0.6877508789425735, + "learning_rate": 7.177843478291225e-07, + "loss": 0.5863, + "num_input_tokens_seen": 263398145, + "step": 8093 + }, + { + "epoch": 0.7299454389683004, + "flos": 39255620328960.0, + "grad_norm": 2.1062483609246914, + "learning_rate": 7.173360496148276e-07, + "loss": 0.7647, + "num_input_tokens_seen": 263428570, + "step": 8094 + }, + { + "epoch": 0.7300356224917708, + "flos": 27524112946560.0, + "grad_norm": 1.9959386477893393, + "learning_rate": 7.168878608444939e-07, + "loss": 0.7531, + "num_input_tokens_seen": 263458525, + "step": 8095 + }, + { + "epoch": 0.730125806015241, + "flos": 21877583731680.0, + "grad_norm": 2.6171503320588414, + "learning_rate": 7.164397815563623e-07, + "loss": 0.7435, + "num_input_tokens_seen": 263485125, + "step": 8096 + }, + { + "epoch": 0.7302159895387113, + "flos": 25775074865280.0, + "grad_norm": 2.3123433859089118, + "learning_rate": 7.159918117886661e-07, + "loss": 0.7458, + "num_input_tokens_seen": 263512430, + "step": 8097 + }, + { + "epoch": 0.7303061730621815, + "flos": 16121126335680.0, + "grad_norm": 4.246062755032918, + "learning_rate": 7.155439515796284e-07, + "loss": 0.7566, + "num_input_tokens_seen": 263538945, + "step": 8098 + }, + { + "epoch": 0.7303963565856518, + "flos": 15647397547200.0, + "grad_norm": 2.4837892429609414, + "learning_rate": 7.150962009674633e-07, + "loss": 0.7407, + "num_input_tokens_seen": 263567520, + "step": 8099 + }, + { + "epoch": 0.7304865401091221, + "flos": 26680673558880.0, + "grad_norm": 76.53733961582901, + "learning_rate": 7.146485599903751e-07, + "loss": 0.7889, + "num_input_tokens_seen": 263594290, + "step": 8100 + }, + { + "epoch": 0.7305767236325923, + "flos": 23078579207040.0, + "grad_norm": 1.7550833305724882, + "learning_rate": 7.142010286865592e-07, + "loss": 0.8122, + "num_input_tokens_seen": 263623695, + "step": 8101 + }, + { + "epoch": 0.7306669071560626, + "flos": 66227633432160.0, + "grad_norm": 0.5496424542350075, + "learning_rate": 7.137536070942012e-07, + "loss": 0.5469, + "num_input_tokens_seen": 263712975, + "step": 8102 + }, + { + "epoch": 0.7307570906795329, + "flos": 25374767820000.0, + "grad_norm": 2.045631205511194, + "learning_rate": 7.133062952514786e-07, + "loss": 0.7348, + "num_input_tokens_seen": 263745725, + "step": 8103 + }, + { + "epoch": 0.7308472742030031, + "flos": 65313337014720.0, + "grad_norm": 0.6448699292543493, + "learning_rate": 7.128590931965562e-07, + "loss": 0.6023, + "num_input_tokens_seen": 263843400, + "step": 8104 + }, + { + "epoch": 0.7309374577264733, + "flos": 22022977597440.0, + "grad_norm": 2.1354789414668183, + "learning_rate": 7.124120009675945e-07, + "loss": 0.8031, + "num_input_tokens_seen": 263873160, + "step": 8105 + }, + { + "epoch": 0.7310276412499437, + "flos": 24683003987520.0, + "grad_norm": 1.719290284288776, + "learning_rate": 7.119650186027399e-07, + "loss": 0.7664, + "num_input_tokens_seen": 263900630, + "step": 8106 + }, + { + "epoch": 0.7311178247734139, + "flos": 24528875228160.0, + "grad_norm": 2.1047086741537084, + "learning_rate": 7.11518146140132e-07, + "loss": 0.8605, + "num_input_tokens_seen": 263927335, + "step": 8107 + }, + { + "epoch": 0.7312080082968841, + "flos": 17067691838400.0, + "grad_norm": 2.903234414460468, + "learning_rate": 7.110713836179007e-07, + "loss": 0.7005, + "num_input_tokens_seen": 263954310, + "step": 8108 + }, + { + "epoch": 0.7312981918203544, + "flos": 20311077840000.0, + "grad_norm": 1.8269556984711002, + "learning_rate": 7.106247310741659e-07, + "loss": 0.6669, + "num_input_tokens_seen": 263982050, + "step": 8109 + }, + { + "epoch": 0.7313883753438247, + "flos": 23698519596000.0, + "grad_norm": 1.905157119899699, + "learning_rate": 7.101781885470393e-07, + "loss": 0.8029, + "num_input_tokens_seen": 264010380, + "step": 8110 + }, + { + "epoch": 0.731478558867295, + "flos": 23730045286080.0, + "grad_norm": 2.4987185054776977, + "learning_rate": 7.097317560746203e-07, + "loss": 0.7792, + "num_input_tokens_seen": 264038710, + "step": 8111 + }, + { + "epoch": 0.7315687423907652, + "flos": 26025713522400.0, + "grad_norm": 3.243930871878419, + "learning_rate": 7.092854336950036e-07, + "loss": 0.7993, + "num_input_tokens_seen": 264067640, + "step": 8112 + }, + { + "epoch": 0.7316589259142354, + "flos": 24860443920000.0, + "grad_norm": 1.6718490455965835, + "learning_rate": 7.0883922144627e-07, + "loss": 0.7965, + "num_input_tokens_seen": 264094985, + "step": 8113 + }, + { + "epoch": 0.7317491094377058, + "flos": 21767581211040.0, + "grad_norm": 1.7966668713321228, + "learning_rate": 7.083931193664934e-07, + "loss": 0.7798, + "num_input_tokens_seen": 264120610, + "step": 8114 + }, + { + "epoch": 0.731839292961176, + "flos": 23258844041280.0, + "grad_norm": 2.1133940661314834, + "learning_rate": 7.079471274937378e-07, + "loss": 0.789, + "num_input_tokens_seen": 264149870, + "step": 8115 + }, + { + "epoch": 0.7319294764846462, + "flos": 28503096213600.0, + "grad_norm": 3.15229065577383, + "learning_rate": 7.075012458660574e-07, + "loss": 0.7269, + "num_input_tokens_seen": 264178770, + "step": 8116 + }, + { + "epoch": 0.7320196600081165, + "flos": 60761592070080.0, + "grad_norm": 0.6566699137784904, + "learning_rate": 7.070554745214976e-07, + "loss": 0.5747, + "num_input_tokens_seen": 264276625, + "step": 8117 + }, + { + "epoch": 0.7321098435315868, + "flos": 27415708725600.0, + "grad_norm": 2.256023926121067, + "learning_rate": 7.066098134980947e-07, + "loss": 0.6995, + "num_input_tokens_seen": 264307265, + "step": 8118 + }, + { + "epoch": 0.732200027055057, + "flos": 22821324332640.0, + "grad_norm": 2.7104406423678715, + "learning_rate": 7.061642628338727e-07, + "loss": 0.7436, + "num_input_tokens_seen": 264336700, + "step": 8119 + }, + { + "epoch": 0.7322902105785273, + "flos": 23585023117920.0, + "grad_norm": 2.2366456984869783, + "learning_rate": 7.057188225668513e-07, + "loss": 0.7398, + "num_input_tokens_seen": 264367315, + "step": 8120 + }, + { + "epoch": 0.7323803941019975, + "flos": 26248840823520.0, + "grad_norm": 2.0332863797045633, + "learning_rate": 7.052734927350358e-07, + "loss": 0.89, + "num_input_tokens_seen": 264397675, + "step": 8121 + }, + { + "epoch": 0.7324705776254679, + "flos": 25042418563200.0, + "grad_norm": 3.218869994065664, + "learning_rate": 7.048282733764252e-07, + "loss": 0.7647, + "num_input_tokens_seen": 264426900, + "step": 8122 + }, + { + "epoch": 0.7325607611489381, + "flos": 22131753516000.0, + "grad_norm": 2.264631698171479, + "learning_rate": 7.043831645290077e-07, + "loss": 0.8224, + "num_input_tokens_seen": 264454415, + "step": 8123 + }, + { + "epoch": 0.7326509446724083, + "flos": 17396509968000.0, + "grad_norm": 2.7435829546223025, + "learning_rate": 7.039381662307624e-07, + "loss": 0.712, + "num_input_tokens_seen": 264480710, + "step": 8124 + }, + { + "epoch": 0.7327411281958787, + "flos": 22856678508000.0, + "grad_norm": 2.549344042014664, + "learning_rate": 7.034932785196601e-07, + "loss": 0.7235, + "num_input_tokens_seen": 264511155, + "step": 8125 + }, + { + "epoch": 0.7328313117193489, + "flos": 21768659134080.0, + "grad_norm": 5.624735847204745, + "learning_rate": 7.030485014336585e-07, + "loss": 0.7262, + "num_input_tokens_seen": 264538895, + "step": 8126 + }, + { + "epoch": 0.7329214952428191, + "flos": 16666864416480.0, + "grad_norm": 2.1951899918553286, + "learning_rate": 7.026038350107118e-07, + "loss": 0.7002, + "num_input_tokens_seen": 264565140, + "step": 8127 + }, + { + "epoch": 0.7330116787662894, + "flos": 29416277538240.0, + "grad_norm": 1.6615261753625175, + "learning_rate": 7.021592792887579e-07, + "loss": 0.8711, + "num_input_tokens_seen": 264594930, + "step": 8128 + }, + { + "epoch": 0.7331018622897597, + "flos": 23698110728640.0, + "grad_norm": 1.7449011773051193, + "learning_rate": 7.01714834305732e-07, + "loss": 0.7944, + "num_input_tokens_seen": 264626195, + "step": 8129 + }, + { + "epoch": 0.7331920458132299, + "flos": 28720648050720.0, + "grad_norm": 2.228010382787321, + "learning_rate": 7.012705000995544e-07, + "loss": 0.7007, + "num_input_tokens_seen": 264656390, + "step": 8130 + }, + { + "epoch": 0.7332822293367002, + "flos": 22600204198560.0, + "grad_norm": 1.7386529760762703, + "learning_rate": 7.008262767081392e-07, + "loss": 0.7361, + "num_input_tokens_seen": 264685630, + "step": 8131 + }, + { + "epoch": 0.7333724128601704, + "flos": 22677974803680.0, + "grad_norm": 4.028868689000426, + "learning_rate": 7.003821641693892e-07, + "loss": 0.7967, + "num_input_tokens_seen": 264710905, + "step": 8132 + }, + { + "epoch": 0.7334625963836408, + "flos": 21440918927520.0, + "grad_norm": 1.753925824640211, + "learning_rate": 6.999381625211993e-07, + "loss": 0.7312, + "num_input_tokens_seen": 264740025, + "step": 8133 + }, + { + "epoch": 0.733552779907111, + "flos": 28101934263840.0, + "grad_norm": 1.893361341925202, + "learning_rate": 6.994942718014536e-07, + "loss": 0.6644, + "num_input_tokens_seen": 264772805, + "step": 8134 + }, + { + "epoch": 0.7336429634305812, + "flos": 15901269973440.0, + "grad_norm": 2.5281086019354335, + "learning_rate": 6.990504920480282e-07, + "loss": 0.6545, + "num_input_tokens_seen": 264799795, + "step": 8135 + }, + { + "epoch": 0.7337331469540515, + "flos": 15902124877920.0, + "grad_norm": 4.048043218660898, + "learning_rate": 6.986068232987879e-07, + "loss": 0.7392, + "num_input_tokens_seen": 264827185, + "step": 8136 + }, + { + "epoch": 0.7338233304775218, + "flos": 24899700920160.0, + "grad_norm": 1.5176867589767702, + "learning_rate": 6.981632655915888e-07, + "loss": 0.6961, + "num_input_tokens_seen": 264856880, + "step": 8137 + }, + { + "epoch": 0.733913514000992, + "flos": 25847939062080.0, + "grad_norm": 1.837737933473564, + "learning_rate": 6.977198189642783e-07, + "loss": 0.8043, + "num_input_tokens_seen": 264886685, + "step": 8138 + }, + { + "epoch": 0.7340036975244623, + "flos": 21440138362560.0, + "grad_norm": 2.061599766173403, + "learning_rate": 6.972764834546935e-07, + "loss": 0.7697, + "num_input_tokens_seen": 264914930, + "step": 8139 + }, + { + "epoch": 0.7340938810479325, + "flos": 23079954488160.0, + "grad_norm": 1.846627334022067, + "learning_rate": 6.96833259100663e-07, + "loss": 0.759, + "num_input_tokens_seen": 264944560, + "step": 8140 + }, + { + "epoch": 0.7341840645714028, + "flos": 19690542734880.0, + "grad_norm": 4.531589048131733, + "learning_rate": 6.96390145940003e-07, + "loss": 0.7171, + "num_input_tokens_seen": 264971335, + "step": 8141 + }, + { + "epoch": 0.7342742480948731, + "flos": 23333603895840.0, + "grad_norm": 2.025617801944193, + "learning_rate": 6.959471440105253e-07, + "loss": 0.6955, + "num_input_tokens_seen": 265002045, + "step": 8142 + }, + { + "epoch": 0.7343644316183433, + "flos": 19217557341600.0, + "grad_norm": 2.9260167433285083, + "learning_rate": 6.955042533500261e-07, + "loss": 0.7748, + "num_input_tokens_seen": 265029120, + "step": 8143 + }, + { + "epoch": 0.7344546151418135, + "flos": 22305141944640.0, + "grad_norm": 1.9028954303253376, + "learning_rate": 6.950614739962986e-07, + "loss": 0.7003, + "num_input_tokens_seen": 265054765, + "step": 8144 + }, + { + "epoch": 0.7345447986652839, + "flos": 24751444982880.0, + "grad_norm": 1.8866063433979388, + "learning_rate": 6.946188059871198e-07, + "loss": 0.6315, + "num_input_tokens_seen": 265084565, + "step": 8145 + }, + { + "epoch": 0.7346349821887541, + "flos": 20816964204480.0, + "grad_norm": 1.9184524951627782, + "learning_rate": 6.941762493602638e-07, + "loss": 0.8622, + "num_input_tokens_seen": 265113160, + "step": 8146 + }, + { + "epoch": 0.7347251657122243, + "flos": 25119891810240.0, + "grad_norm": 2.483904140164603, + "learning_rate": 6.937338041534899e-07, + "loss": 0.6824, + "num_input_tokens_seen": 265143110, + "step": 8147 + }, + { + "epoch": 0.7348153492356947, + "flos": 26573644619040.0, + "grad_norm": 1.6896089162273387, + "learning_rate": 6.932914704045505e-07, + "loss": 0.7401, + "num_input_tokens_seen": 265174220, + "step": 8148 + }, + { + "epoch": 0.7349055327591649, + "flos": 25662507631200.0, + "grad_norm": 2.044352717288366, + "learning_rate": 6.928492481511878e-07, + "loss": 0.7698, + "num_input_tokens_seen": 265202005, + "step": 8149 + }, + { + "epoch": 0.7349957162826352, + "flos": 21257606172960.0, + "grad_norm": 3.0375241502343377, + "learning_rate": 6.924071374311349e-07, + "loss": 0.6932, + "num_input_tokens_seen": 265229845, + "step": 8150 + }, + { + "epoch": 0.7350858998061054, + "flos": 52519240107840.0, + "grad_norm": 0.5987537504568484, + "learning_rate": 6.919651382821157e-07, + "loss": 0.5579, + "num_input_tokens_seen": 265317645, + "step": 8151 + }, + { + "epoch": 0.7351760833295757, + "flos": 24496197275520.0, + "grad_norm": 2.2317047168696713, + "learning_rate": 6.915232507418425e-07, + "loss": 0.7126, + "num_input_tokens_seen": 265347885, + "step": 8152 + }, + { + "epoch": 0.735266266853046, + "flos": 22310234201760.0, + "grad_norm": 2.1726763228141097, + "learning_rate": 6.910814748480204e-07, + "loss": 0.6257, + "num_input_tokens_seen": 265377200, + "step": 8153 + }, + { + "epoch": 0.7353564503765162, + "flos": 30651326247360.0, + "grad_norm": 2.2945123055403034, + "learning_rate": 6.906398106383445e-07, + "loss": 0.7582, + "num_input_tokens_seen": 265403655, + "step": 8154 + }, + { + "epoch": 0.7354466338999864, + "flos": 20821015708320.0, + "grad_norm": 1.6305061368946674, + "learning_rate": 6.901982581504994e-07, + "loss": 0.7362, + "num_input_tokens_seen": 265432205, + "step": 8155 + }, + { + "epoch": 0.7355368174234568, + "flos": 45735664605600.0, + "grad_norm": 3.1418651959926667, + "learning_rate": 6.897568174221611e-07, + "loss": 0.6123, + "num_input_tokens_seen": 265465225, + "step": 8156 + }, + { + "epoch": 0.735627000946927, + "flos": 20632387678080.0, + "grad_norm": 2.0797812379824396, + "learning_rate": 6.893154884909966e-07, + "loss": 0.8108, + "num_input_tokens_seen": 265493405, + "step": 8157 + }, + { + "epoch": 0.7357171844703972, + "flos": 68631557210400.0, + "grad_norm": 0.6048125125482527, + "learning_rate": 6.888742713946602e-07, + "loss": 0.5831, + "num_input_tokens_seen": 265589210, + "step": 8158 + }, + { + "epoch": 0.7358073679938675, + "flos": 21585903925920.0, + "grad_norm": 2.4654159244520977, + "learning_rate": 6.884331661708018e-07, + "loss": 0.7428, + "num_input_tokens_seen": 265617060, + "step": 8159 + }, + { + "epoch": 0.7358975515173378, + "flos": 20926892385600.0, + "grad_norm": 1.7046643221474405, + "learning_rate": 6.879921728570561e-07, + "loss": 0.7388, + "num_input_tokens_seen": 265646970, + "step": 8160 + }, + { + "epoch": 0.735987735040808, + "flos": 22749538058880.0, + "grad_norm": 1.6568642682764363, + "learning_rate": 6.875512914910539e-07, + "loss": 0.8182, + "num_input_tokens_seen": 265676470, + "step": 8161 + }, + { + "epoch": 0.7360779185642783, + "flos": 24098343434400.0, + "grad_norm": 1.9481176044909954, + "learning_rate": 6.871105221104119e-07, + "loss": 0.6893, + "num_input_tokens_seen": 265707050, + "step": 8162 + }, + { + "epoch": 0.7361681020877485, + "flos": 23550226488960.0, + "grad_norm": 2.29363723186641, + "learning_rate": 6.866698647527391e-07, + "loss": 0.7234, + "num_input_tokens_seen": 265737415, + "step": 8163 + }, + { + "epoch": 0.7362582856112189, + "flos": 20853210454080.0, + "grad_norm": 2.4014721289404717, + "learning_rate": 6.862293194556353e-07, + "loss": 0.7631, + "num_input_tokens_seen": 265765385, + "step": 8164 + }, + { + "epoch": 0.7363484691346891, + "flos": 19873297943040.0, + "grad_norm": 2.384827809065896, + "learning_rate": 6.857888862566896e-07, + "loss": 0.6863, + "num_input_tokens_seen": 265795080, + "step": 8165 + }, + { + "epoch": 0.7364386526581593, + "flos": 26244937998720.0, + "grad_norm": 1.8418444468563042, + "learning_rate": 6.853485651934836e-07, + "loss": 0.7056, + "num_input_tokens_seen": 265824675, + "step": 8166 + }, + { + "epoch": 0.7365288361816296, + "flos": 35949443135520.0, + "grad_norm": 2.692621275172662, + "learning_rate": 6.849083563035855e-07, + "loss": 0.5948, + "num_input_tokens_seen": 265862780, + "step": 8167 + }, + { + "epoch": 0.7366190197050999, + "flos": 65718253110240.0, + "grad_norm": 0.7100073629991456, + "learning_rate": 6.844682596245592e-07, + "loss": 0.6606, + "num_input_tokens_seen": 265948935, + "step": 8168 + }, + { + "epoch": 0.7367092032285701, + "flos": 20778004562400.0, + "grad_norm": 1.9239202641918765, + "learning_rate": 6.840282751939539e-07, + "loss": 0.7996, + "num_input_tokens_seen": 265975010, + "step": 8169 + }, + { + "epoch": 0.7367993867520404, + "flos": 19910399097120.0, + "grad_norm": 2.6279535480324716, + "learning_rate": 6.835884030493126e-07, + "loss": 0.7874, + "num_input_tokens_seen": 266001430, + "step": 8170 + }, + { + "epoch": 0.7368895702755107, + "flos": 26430889806240.0, + "grad_norm": 2.051512030324264, + "learning_rate": 6.831486432281672e-07, + "loss": 0.7654, + "num_input_tokens_seen": 266030860, + "step": 8171 + }, + { + "epoch": 0.736979753798981, + "flos": 17724361683840.0, + "grad_norm": 2.2564881344395986, + "learning_rate": 6.827089957680407e-07, + "loss": 0.7526, + "num_input_tokens_seen": 266058275, + "step": 8172 + }, + { + "epoch": 0.7370699373224512, + "flos": 29268764996160.0, + "grad_norm": 1.8086390314551175, + "learning_rate": 6.822694607064461e-07, + "loss": 0.8505, + "num_input_tokens_seen": 266087535, + "step": 8173 + }, + { + "epoch": 0.7371601208459214, + "flos": 25703362931040.0, + "grad_norm": 1.98513076687805, + "learning_rate": 6.818300380808877e-07, + "loss": 0.7646, + "num_input_tokens_seen": 266116870, + "step": 8174 + }, + { + "epoch": 0.7372503043693918, + "flos": 24026705839680.0, + "grad_norm": 1.7991232697184298, + "learning_rate": 6.813907279288574e-07, + "loss": 0.7329, + "num_input_tokens_seen": 266146240, + "step": 8175 + }, + { + "epoch": 0.737340487892862, + "flos": 17354539575360.0, + "grad_norm": 1.9322606278743049, + "learning_rate": 6.809515302878422e-07, + "loss": 0.7691, + "num_input_tokens_seen": 266172495, + "step": 8176 + }, + { + "epoch": 0.7374306714163322, + "flos": 26831680058400.0, + "grad_norm": 3.141222996502688, + "learning_rate": 6.80512445195315e-07, + "loss": 0.7351, + "num_input_tokens_seen": 266200710, + "step": 8177 + }, + { + "epoch": 0.7375208549398025, + "flos": 25262795302080.0, + "grad_norm": 3.0455109501743443, + "learning_rate": 6.800734726887416e-07, + "loss": 0.7601, + "num_input_tokens_seen": 266232025, + "step": 8178 + }, + { + "epoch": 0.7376110384632728, + "flos": 16994976320640.0, + "grad_norm": 1.911108984625666, + "learning_rate": 6.796346128055775e-07, + "loss": 0.7461, + "num_input_tokens_seen": 266257455, + "step": 8179 + }, + { + "epoch": 0.737701221986743, + "flos": 22781658465120.0, + "grad_norm": 2.077799032933297, + "learning_rate": 6.791958655832684e-07, + "loss": 0.7118, + "num_input_tokens_seen": 266285430, + "step": 8180 + }, + { + "epoch": 0.7377914055102133, + "flos": 33607418474880.0, + "grad_norm": 2.1173579861737526, + "learning_rate": 6.787572310592518e-07, + "loss": 0.7063, + "num_input_tokens_seen": 266314860, + "step": 8181 + }, + { + "epoch": 0.7378815890336835, + "flos": 26394755065920.0, + "grad_norm": 2.104877226051265, + "learning_rate": 6.783187092709521e-07, + "loss": 0.8255, + "num_input_tokens_seen": 266343260, + "step": 8182 + }, + { + "epoch": 0.7379717725571538, + "flos": 18343261319520.0, + "grad_norm": 5.761922996909491, + "learning_rate": 6.778803002557891e-07, + "loss": 0.7528, + "num_input_tokens_seen": 266369040, + "step": 8183 + }, + { + "epoch": 0.7380619560806241, + "flos": 24209126520000.0, + "grad_norm": 2.0601615364663637, + "learning_rate": 6.774420040511686e-07, + "loss": 0.7785, + "num_input_tokens_seen": 266399295, + "step": 8184 + }, + { + "epoch": 0.7381521396040943, + "flos": 24464857434240.0, + "grad_norm": 2.679876559958953, + "learning_rate": 6.770038206944886e-07, + "loss": 0.7449, + "num_input_tokens_seen": 266428800, + "step": 8185 + }, + { + "epoch": 0.7382423231275645, + "flos": 26977780149600.0, + "grad_norm": 1.819230572633334, + "learning_rate": 6.765657502231375e-07, + "loss": 0.7571, + "num_input_tokens_seen": 266459870, + "step": 8186 + }, + { + "epoch": 0.7383325066510349, + "flos": 22096101982560.0, + "grad_norm": 6.8000410751897356, + "learning_rate": 6.761277926744939e-07, + "loss": 0.8093, + "num_input_tokens_seen": 266490655, + "step": 8187 + }, + { + "epoch": 0.7384226901745051, + "flos": 26243079510720.0, + "grad_norm": 2.0599408311586904, + "learning_rate": 6.756899480859268e-07, + "loss": 0.732, + "num_input_tokens_seen": 266519920, + "step": 8188 + }, + { + "epoch": 0.7385128736979754, + "flos": 22824186404160.0, + "grad_norm": 2.4482884442616095, + "learning_rate": 6.752522164947956e-07, + "loss": 0.7858, + "num_input_tokens_seen": 266546330, + "step": 8189 + }, + { + "epoch": 0.7386030572214456, + "flos": 24863826368160.0, + "grad_norm": 1.4939145783614256, + "learning_rate": 6.748145979384498e-07, + "loss": 0.7982, + "num_input_tokens_seen": 266576990, + "step": 8190 + }, + { + "epoch": 0.7386932407449159, + "flos": 24281842037760.0, + "grad_norm": 1.882743232687536, + "learning_rate": 6.743770924542303e-07, + "loss": 0.6847, + "num_input_tokens_seen": 266605525, + "step": 8191 + }, + { + "epoch": 0.7387834242683862, + "flos": 58716711169920.0, + "grad_norm": 0.6361054819552996, + "learning_rate": 6.739397000794658e-07, + "loss": 0.5688, + "num_input_tokens_seen": 266690805, + "step": 8192 + }, + { + "epoch": 0.7388736077918564, + "flos": 21145076108640.0, + "grad_norm": 1.9667074719199067, + "learning_rate": 6.735024208514782e-07, + "loss": 0.7187, + "num_input_tokens_seen": 266717585, + "step": 8193 + }, + { + "epoch": 0.7389637913153266, + "flos": 35062843922880.0, + "grad_norm": 1.515044897692047, + "learning_rate": 6.73065254807578e-07, + "loss": 0.6604, + "num_input_tokens_seen": 266749885, + "step": 8194 + }, + { + "epoch": 0.739053974838797, + "flos": 22677417257280.0, + "grad_norm": 1.749139165324369, + "learning_rate": 6.726282019850669e-07, + "loss": 0.7959, + "num_input_tokens_seen": 266777840, + "step": 8195 + }, + { + "epoch": 0.7391441583622672, + "flos": 24100053243360.0, + "grad_norm": 1.8777619682238378, + "learning_rate": 6.721912624212376e-07, + "loss": 0.8125, + "num_input_tokens_seen": 266810010, + "step": 8196 + }, + { + "epoch": 0.7392343418857374, + "flos": 36262687135680.0, + "grad_norm": 1.8647885413881133, + "learning_rate": 6.717544361533696e-07, + "loss": 0.6438, + "num_input_tokens_seen": 266839685, + "step": 8197 + }, + { + "epoch": 0.7393245254092078, + "flos": 23990756948160.0, + "grad_norm": 2.922234987328533, + "learning_rate": 6.713177232187386e-07, + "loss": 0.7244, + "num_input_tokens_seen": 266868555, + "step": 8198 + }, + { + "epoch": 0.739414708932678, + "flos": 23150402650560.0, + "grad_norm": 1.5441289861246712, + "learning_rate": 6.708811236546041e-07, + "loss": 0.7355, + "num_input_tokens_seen": 266897930, + "step": 8199 + }, + { + "epoch": 0.7395048924561483, + "flos": 24201023512320.0, + "grad_norm": 1.988836664293209, + "learning_rate": 6.704446374982224e-07, + "loss": 0.7092, + "num_input_tokens_seen": 266924195, + "step": 8200 + }, + { + "epoch": 0.7395950759796185, + "flos": 16813001677440.0, + "grad_norm": 2.115380335850155, + "learning_rate": 6.700082647868346e-07, + "loss": 0.776, + "num_input_tokens_seen": 266950685, + "step": 8201 + }, + { + "epoch": 0.7396852595030888, + "flos": 26322485585280.0, + "grad_norm": 2.0734334202534397, + "learning_rate": 6.695720055576751e-07, + "loss": 0.7971, + "num_input_tokens_seen": 266980220, + "step": 8202 + }, + { + "epoch": 0.7397754430265591, + "flos": 24172025365920.0, + "grad_norm": 1.9737727565311156, + "learning_rate": 6.691358598479679e-07, + "loss": 0.7529, + "num_input_tokens_seen": 267008105, + "step": 8203 + }, + { + "epoch": 0.7398656265500293, + "flos": 22420162382880.0, + "grad_norm": 1.5509133078513637, + "learning_rate": 6.686998276949276e-07, + "loss": 0.7334, + "num_input_tokens_seen": 267037130, + "step": 8204 + }, + { + "epoch": 0.7399558100734995, + "flos": 65869296779520.0, + "grad_norm": 0.6359791819341195, + "learning_rate": 6.682639091357587e-07, + "loss": 0.5853, + "num_input_tokens_seen": 267132235, + "step": 8205 + }, + { + "epoch": 0.7400459935969699, + "flos": 20857336297440.0, + "grad_norm": 1.9099516651768358, + "learning_rate": 6.678281042076568e-07, + "loss": 0.7355, + "num_input_tokens_seen": 267161530, + "step": 8206 + }, + { + "epoch": 0.7401361771204401, + "flos": 24172136875200.0, + "grad_norm": 4.95959281419033, + "learning_rate": 6.673924129478059e-07, + "loss": 0.7473, + "num_input_tokens_seen": 267191795, + "step": 8207 + }, + { + "epoch": 0.7402263606439103, + "flos": 27268976748480.0, + "grad_norm": 1.6676247306739447, + "learning_rate": 6.669568353933824e-07, + "loss": 0.6647, + "num_input_tokens_seen": 267223040, + "step": 8208 + }, + { + "epoch": 0.7403165441673806, + "flos": 19509274317120.0, + "grad_norm": 1.5584225668152503, + "learning_rate": 6.665213715815519e-07, + "loss": 0.6783, + "num_input_tokens_seen": 267252315, + "step": 8209 + }, + { + "epoch": 0.7404067276908509, + "flos": 36843928070880.0, + "grad_norm": 2.434132328106848, + "learning_rate": 6.660860215494706e-07, + "loss": 0.7565, + "num_input_tokens_seen": 267280370, + "step": 8210 + }, + { + "epoch": 0.7404969112143212, + "flos": 27774751603680.0, + "grad_norm": 1.9945564098093644, + "learning_rate": 6.656507853342852e-07, + "loss": 0.6969, + "num_input_tokens_seen": 267309405, + "step": 8211 + }, + { + "epoch": 0.7405870947377914, + "flos": 26868335175360.0, + "grad_norm": 1.631680694080773, + "learning_rate": 6.652156629731323e-07, + "loss": 0.7549, + "num_input_tokens_seen": 267341670, + "step": 8212 + }, + { + "epoch": 0.7406772782612616, + "flos": 21689699096640.0, + "grad_norm": 2.02527573203539, + "learning_rate": 6.647806545031396e-07, + "loss": 0.8086, + "num_input_tokens_seen": 267369760, + "step": 8213 + }, + { + "epoch": 0.740767461784732, + "flos": 22861547746560.0, + "grad_norm": 1.7810176898649654, + "learning_rate": 6.643457599614224e-07, + "loss": 0.6669, + "num_input_tokens_seen": 267395940, + "step": 8214 + }, + { + "epoch": 0.7408576453082022, + "flos": 24755050449600.0, + "grad_norm": 3.5818979951955416, + "learning_rate": 6.63910979385091e-07, + "loss": 0.8017, + "num_input_tokens_seen": 267424410, + "step": 8215 + }, + { + "epoch": 0.7409478288316724, + "flos": 19254435477120.0, + "grad_norm": 1.817932340997514, + "learning_rate": 6.634763128112409e-07, + "loss": 0.8266, + "num_input_tokens_seen": 267451295, + "step": 8216 + }, + { + "epoch": 0.7410380123551427, + "flos": 23261520264000.0, + "grad_norm": 2.1483500514873235, + "learning_rate": 6.630417602769622e-07, + "loss": 0.7295, + "num_input_tokens_seen": 267478820, + "step": 8217 + }, + { + "epoch": 0.741128195878613, + "flos": 35827843649760.0, + "grad_norm": 1.9173170174995218, + "learning_rate": 6.62607321819332e-07, + "loss": 0.7483, + "num_input_tokens_seen": 267509195, + "step": 8218 + }, + { + "epoch": 0.7412183794020832, + "flos": 20346617864160.0, + "grad_norm": 1.4849358553149776, + "learning_rate": 6.621729974754196e-07, + "loss": 0.741, + "num_input_tokens_seen": 267539660, + "step": 8219 + }, + { + "epoch": 0.7413085629255535, + "flos": 22455144860640.0, + "grad_norm": 2.2417991381599047, + "learning_rate": 6.617387872822835e-07, + "loss": 0.6928, + "num_input_tokens_seen": 267568440, + "step": 8220 + }, + { + "epoch": 0.7413987464490238, + "flos": 23516433443520.0, + "grad_norm": 2.404467628545078, + "learning_rate": 6.613046912769731e-07, + "loss": 0.7138, + "num_input_tokens_seen": 267596120, + "step": 8221 + }, + { + "epoch": 0.741488929972494, + "flos": 25011227400960.0, + "grad_norm": 2.4793732394112906, + "learning_rate": 6.608707094965289e-07, + "loss": 0.8119, + "num_input_tokens_seen": 267623675, + "step": 8222 + }, + { + "epoch": 0.7415791134959643, + "flos": 15537766724160.0, + "grad_norm": 2.825495725364874, + "learning_rate": 6.604368419779787e-07, + "loss": 0.6799, + "num_input_tokens_seen": 267649110, + "step": 8223 + }, + { + "epoch": 0.7416692970194345, + "flos": 25700538029280.0, + "grad_norm": 2.0375559418844413, + "learning_rate": 6.600030887583434e-07, + "loss": 0.7547, + "num_input_tokens_seen": 267677260, + "step": 8224 + }, + { + "epoch": 0.7417594805429049, + "flos": 27811183702080.0, + "grad_norm": 1.9799993135470346, + "learning_rate": 6.595694498746336e-07, + "loss": 0.8491, + "num_input_tokens_seen": 267706180, + "step": 8225 + }, + { + "epoch": 0.7418496640663751, + "flos": 31420117289760.0, + "grad_norm": 2.2468819736165577, + "learning_rate": 6.591359253638491e-07, + "loss": 0.7118, + "num_input_tokens_seen": 267739600, + "step": 8226 + }, + { + "epoch": 0.7419398475898453, + "flos": 21731000433600.0, + "grad_norm": 4.782809333588363, + "learning_rate": 6.587025152629808e-07, + "loss": 0.6343, + "num_input_tokens_seen": 267767750, + "step": 8227 + }, + { + "epoch": 0.7420300311133156, + "flos": 15829112002080.0, + "grad_norm": 3.064134900970454, + "learning_rate": 6.582692196090107e-07, + "loss": 0.7826, + "num_input_tokens_seen": 267793865, + "step": 8228 + }, + { + "epoch": 0.7421202146367859, + "flos": 27773822359680.0, + "grad_norm": 1.7556795870482547, + "learning_rate": 6.578360384389074e-07, + "loss": 0.7572, + "num_input_tokens_seen": 267821460, + "step": 8229 + }, + { + "epoch": 0.7422103981602561, + "flos": 27633557920800.0, + "grad_norm": 2.66985365048623, + "learning_rate": 6.574029717896355e-07, + "loss": 0.724, + "num_input_tokens_seen": 267851935, + "step": 8230 + }, + { + "epoch": 0.7423005816837264, + "flos": 21731223452160.0, + "grad_norm": 2.6819961434435444, + "learning_rate": 6.569700196981436e-07, + "loss": 0.7341, + "num_input_tokens_seen": 267880205, + "step": 8231 + }, + { + "epoch": 0.7423907652071966, + "flos": 24609656583840.0, + "grad_norm": 3.6102609579177427, + "learning_rate": 6.565371822013763e-07, + "loss": 0.6461, + "num_input_tokens_seen": 267909200, + "step": 8232 + }, + { + "epoch": 0.742480948730667, + "flos": 19035954396000.0, + "grad_norm": 2.136189899168339, + "learning_rate": 6.561044593362636e-07, + "loss": 0.7986, + "num_input_tokens_seen": 267936450, + "step": 8233 + }, + { + "epoch": 0.7425711322541372, + "flos": 23334458800320.0, + "grad_norm": 2.0485644175350033, + "learning_rate": 6.556718511397288e-07, + "loss": 0.7272, + "num_input_tokens_seen": 267964045, + "step": 8234 + }, + { + "epoch": 0.7426613157776074, + "flos": 20197209664320.0, + "grad_norm": 3.067678627340962, + "learning_rate": 6.552393576486843e-07, + "loss": 0.8617, + "num_input_tokens_seen": 267990965, + "step": 8235 + }, + { + "epoch": 0.7427514993010776, + "flos": 12951236416800.0, + "grad_norm": 2.0900362015285645, + "learning_rate": 6.548069789000325e-07, + "loss": 0.7572, + "num_input_tokens_seen": 268017645, + "step": 8236 + }, + { + "epoch": 0.742841682824548, + "flos": 30114323060160.0, + "grad_norm": 2.017778886996864, + "learning_rate": 6.543747149306673e-07, + "loss": 0.7329, + "num_input_tokens_seen": 268051965, + "step": 8237 + }, + { + "epoch": 0.7429318663480182, + "flos": 30182206509120.0, + "grad_norm": 1.7019258529792967, + "learning_rate": 6.5394256577747e-07, + "loss": 0.7919, + "num_input_tokens_seen": 268081620, + "step": 8238 + }, + { + "epoch": 0.7430220498714885, + "flos": 18445346681280.0, + "grad_norm": 1.701337243505148, + "learning_rate": 6.535105314773161e-07, + "loss": 0.7737, + "num_input_tokens_seen": 268106400, + "step": 8239 + }, + { + "epoch": 0.7431122333949587, + "flos": 31312159105920.0, + "grad_norm": 2.0811329347604817, + "learning_rate": 6.530786120670677e-07, + "loss": 0.6757, + "num_input_tokens_seen": 268136360, + "step": 8240 + }, + { + "epoch": 0.743202416918429, + "flos": 20995147532160.0, + "grad_norm": 2.142087673362674, + "learning_rate": 6.526468075835787e-07, + "loss": 0.7981, + "num_input_tokens_seen": 268160870, + "step": 8241 + }, + { + "epoch": 0.7432926004418993, + "flos": 24057079267200.0, + "grad_norm": 2.335096329002865, + "learning_rate": 6.522151180636937e-07, + "loss": 0.7911, + "num_input_tokens_seen": 268188690, + "step": 8242 + }, + { + "epoch": 0.7433827839653695, + "flos": 29928148234080.0, + "grad_norm": 1.8295727727219337, + "learning_rate": 6.517835435442461e-07, + "loss": 0.7077, + "num_input_tokens_seen": 268219350, + "step": 8243 + }, + { + "epoch": 0.7434729674888398, + "flos": 22787271098880.0, + "grad_norm": 3.6335566589128994, + "learning_rate": 6.513520840620606e-07, + "loss": 0.7694, + "num_input_tokens_seen": 268244115, + "step": 8244 + }, + { + "epoch": 0.7435631510123101, + "flos": 22205769975360.0, + "grad_norm": 4.001469580581776, + "learning_rate": 6.509207396539525e-07, + "loss": 0.7065, + "num_input_tokens_seen": 268273875, + "step": 8245 + }, + { + "epoch": 0.7436533345357803, + "flos": 62457694679520.0, + "grad_norm": 0.620401320356259, + "learning_rate": 6.50489510356724e-07, + "loss": 0.5516, + "num_input_tokens_seen": 268364685, + "step": 8246 + }, + { + "epoch": 0.7437435180592505, + "flos": 17869235172960.0, + "grad_norm": 2.3808267292291174, + "learning_rate": 6.500583962071732e-07, + "loss": 0.6714, + "num_input_tokens_seen": 268392595, + "step": 8247 + }, + { + "epoch": 0.7438337015827209, + "flos": 23996109393600.0, + "grad_norm": 1.5749391180714618, + "learning_rate": 6.496273972420827e-07, + "loss": 0.7722, + "num_input_tokens_seen": 268424990, + "step": 8248 + }, + { + "epoch": 0.7439238851061911, + "flos": 20855031772320.0, + "grad_norm": 1.890998488176973, + "learning_rate": 6.491965134982287e-07, + "loss": 0.7921, + "num_input_tokens_seen": 268447610, + "step": 8249 + }, + { + "epoch": 0.7440140686296614, + "flos": 22675707448320.0, + "grad_norm": 2.0964946691844224, + "learning_rate": 6.487657450123765e-07, + "loss": 0.7148, + "num_input_tokens_seen": 268476360, + "step": 8250 + }, + { + "epoch": 0.7441042521531316, + "flos": 22386666695520.0, + "grad_norm": 1.5583913571304095, + "learning_rate": 6.483350918212814e-07, + "loss": 0.7588, + "num_input_tokens_seen": 268504775, + "step": 8251 + }, + { + "epoch": 0.7441944356766019, + "flos": 67407033276960.0, + "grad_norm": 0.7031043791396244, + "learning_rate": 6.479045539616898e-07, + "loss": 0.5836, + "num_input_tokens_seen": 268595940, + "step": 8252 + }, + { + "epoch": 0.7442846192000722, + "flos": 23763169275840.0, + "grad_norm": 2.101780876527282, + "learning_rate": 6.474741314703358e-07, + "loss": 0.7472, + "num_input_tokens_seen": 268618595, + "step": 8253 + }, + { + "epoch": 0.7443748027235424, + "flos": 18779777444640.0, + "grad_norm": 3.91435341125119, + "learning_rate": 6.47043824383948e-07, + "loss": 0.8044, + "num_input_tokens_seen": 268644260, + "step": 8254 + }, + { + "epoch": 0.7444649862470126, + "flos": 25623399310080.0, + "grad_norm": 2.150501937826313, + "learning_rate": 6.466136327392399e-07, + "loss": 0.7849, + "num_input_tokens_seen": 268667360, + "step": 8255 + }, + { + "epoch": 0.744555169770483, + "flos": 18742527611520.0, + "grad_norm": 3.6086466000591764, + "learning_rate": 6.461835565729206e-07, + "loss": 0.7798, + "num_input_tokens_seen": 268695555, + "step": 8256 + }, + { + "epoch": 0.7446453532939532, + "flos": 18702267027840.0, + "grad_norm": 1.8411726960880308, + "learning_rate": 6.457535959216844e-07, + "loss": 0.7751, + "num_input_tokens_seen": 268722365, + "step": 8257 + }, + { + "epoch": 0.7447355368174234, + "flos": 23006309726400.0, + "grad_norm": 2.205065258383908, + "learning_rate": 6.453237508222186e-07, + "loss": 0.7056, + "num_input_tokens_seen": 268750240, + "step": 8258 + }, + { + "epoch": 0.7448257203408937, + "flos": 23441339061120.0, + "grad_norm": 1.5947953330242222, + "learning_rate": 6.448940213112e-07, + "loss": 0.7668, + "num_input_tokens_seen": 268781360, + "step": 8259 + }, + { + "epoch": 0.744915903864364, + "flos": 22202052999360.0, + "grad_norm": 1.6720715701331637, + "learning_rate": 6.444644074252954e-07, + "loss": 0.8281, + "num_input_tokens_seen": 268809520, + "step": 8260 + }, + { + "epoch": 0.7450060873878342, + "flos": 22096845377760.0, + "grad_norm": 1.702375544565804, + "learning_rate": 6.440349092011628e-07, + "loss": 0.8155, + "num_input_tokens_seen": 268838145, + "step": 8261 + }, + { + "epoch": 0.7450962709113045, + "flos": 19437413703840.0, + "grad_norm": 1.7672636816736613, + "learning_rate": 6.436055266754475e-07, + "loss": 0.7364, + "num_input_tokens_seen": 268865355, + "step": 8262 + }, + { + "epoch": 0.7451864544347747, + "flos": 26642122784160.0, + "grad_norm": 2.218456947423046, + "learning_rate": 6.431762598847879e-07, + "loss": 0.7246, + "num_input_tokens_seen": 268896445, + "step": 8263 + }, + { + "epoch": 0.7452766379582451, + "flos": 27411508542720.0, + "grad_norm": 2.0412382597623617, + "learning_rate": 6.427471088658111e-07, + "loss": 0.7218, + "num_input_tokens_seen": 268925100, + "step": 8264 + }, + { + "epoch": 0.7453668214817153, + "flos": 36231087106080.0, + "grad_norm": 2.167236534394528, + "learning_rate": 6.42318073655135e-07, + "loss": 0.7444, + "num_input_tokens_seen": 268954735, + "step": 8265 + }, + { + "epoch": 0.7454570050051855, + "flos": 26684836572000.0, + "grad_norm": 1.9042953099953506, + "learning_rate": 6.41889154289367e-07, + "loss": 0.7089, + "num_input_tokens_seen": 268986260, + "step": 8266 + }, + { + "epoch": 0.7455471885286559, + "flos": 18525050113920.0, + "grad_norm": 2.432458113557692, + "learning_rate": 6.414603508051055e-07, + "loss": 0.8032, + "num_input_tokens_seen": 269012530, + "step": 8267 + }, + { + "epoch": 0.7456373720521261, + "flos": 59433681833280.0, + "grad_norm": 0.637159464145909, + "learning_rate": 6.410316632389365e-07, + "loss": 0.6036, + "num_input_tokens_seen": 269105605, + "step": 8268 + }, + { + "epoch": 0.7457275555755963, + "flos": 22715038788000.0, + "grad_norm": 2.4914071227509056, + "learning_rate": 6.406030916274406e-07, + "loss": 0.7548, + "num_input_tokens_seen": 269133405, + "step": 8269 + }, + { + "epoch": 0.7458177390990666, + "flos": 21944723785440.0, + "grad_norm": 2.30059865541098, + "learning_rate": 6.401746360071831e-07, + "loss": 0.7046, + "num_input_tokens_seen": 269161630, + "step": 8270 + }, + { + "epoch": 0.7459079226225369, + "flos": 28832732077920.0, + "grad_norm": 2.380103600755562, + "learning_rate": 6.397462964147251e-07, + "loss": 0.7095, + "num_input_tokens_seen": 269192270, + "step": 8271 + }, + { + "epoch": 0.7459981061460071, + "flos": 22095730284960.0, + "grad_norm": 1.8602001378867206, + "learning_rate": 6.393180728866128e-07, + "loss": 0.7355, + "num_input_tokens_seen": 269218960, + "step": 8272 + }, + { + "epoch": 0.7460882896694774, + "flos": 26060324302560.0, + "grad_norm": 1.9595526558434069, + "learning_rate": 6.388899654593853e-07, + "loss": 0.7809, + "num_input_tokens_seen": 269245555, + "step": 8273 + }, + { + "epoch": 0.7461784731929476, + "flos": 23769153607200.0, + "grad_norm": 2.242818604017575, + "learning_rate": 6.384619741695709e-07, + "loss": 0.7726, + "num_input_tokens_seen": 269275430, + "step": 8274 + }, + { + "epoch": 0.746268656716418, + "flos": 23880568578720.0, + "grad_norm": 1.295799728808262, + "learning_rate": 6.380340990536883e-07, + "loss": 0.7458, + "num_input_tokens_seen": 269307655, + "step": 8275 + }, + { + "epoch": 0.7463588402398882, + "flos": 27268605050880.0, + "grad_norm": 2.790273195242601, + "learning_rate": 6.37606340148247e-07, + "loss": 0.7513, + "num_input_tokens_seen": 269337070, + "step": 8276 + }, + { + "epoch": 0.7464490237633584, + "flos": 23297469155520.0, + "grad_norm": 2.1916615818482024, + "learning_rate": 6.371786974897433e-07, + "loss": 0.6566, + "num_input_tokens_seen": 269361570, + "step": 8277 + }, + { + "epoch": 0.7465392072868287, + "flos": 22199302437120.0, + "grad_norm": 1.6084936747238352, + "learning_rate": 6.367511711146691e-07, + "loss": 0.7761, + "num_input_tokens_seen": 269389225, + "step": 8278 + }, + { + "epoch": 0.746629390810299, + "flos": 25010595515040.0, + "grad_norm": 2.007114392938363, + "learning_rate": 6.363237610595014e-07, + "loss": 0.6904, + "num_input_tokens_seen": 269417175, + "step": 8279 + }, + { + "epoch": 0.7467195743337692, + "flos": 20413200371520.0, + "grad_norm": 2.132093706680399, + "learning_rate": 6.358964673607094e-07, + "loss": 0.7278, + "num_input_tokens_seen": 269444120, + "step": 8280 + }, + { + "epoch": 0.7468097578572395, + "flos": 24829587285600.0, + "grad_norm": 2.4909012204573036, + "learning_rate": 6.354692900547525e-07, + "loss": 0.7795, + "num_input_tokens_seen": 269474215, + "step": 8281 + }, + { + "epoch": 0.7468999413807097, + "flos": 20820346652640.0, + "grad_norm": 1.8091891731936838, + "learning_rate": 6.350422291780797e-07, + "loss": 0.7207, + "num_input_tokens_seen": 269502340, + "step": 8282 + }, + { + "epoch": 0.74699012490418, + "flos": 19070342157600.0, + "grad_norm": 2.3899750667532387, + "learning_rate": 6.346152847671302e-07, + "loss": 0.8453, + "num_input_tokens_seen": 269526210, + "step": 8283 + }, + { + "epoch": 0.7470803084276503, + "flos": 21003659407200.0, + "grad_norm": 2.0796573123704185, + "learning_rate": 6.34188456858334e-07, + "loss": 0.7377, + "num_input_tokens_seen": 269554965, + "step": 8284 + }, + { + "epoch": 0.7471704919511205, + "flos": 25187961108000.0, + "grad_norm": 2.2204177391243887, + "learning_rate": 6.337617454881081e-07, + "loss": 0.6989, + "num_input_tokens_seen": 269584190, + "step": 8285 + }, + { + "epoch": 0.7472606754745907, + "flos": 40858372470240.0, + "grad_norm": 1.7556712354232782, + "learning_rate": 6.333351506928651e-07, + "loss": 0.6567, + "num_input_tokens_seen": 269615975, + "step": 8286 + }, + { + "epoch": 0.7473508589980611, + "flos": 35794682490240.0, + "grad_norm": 1.8498628436615232, + "learning_rate": 6.329086725090018e-07, + "loss": 0.717, + "num_input_tokens_seen": 269646645, + "step": 8287 + }, + { + "epoch": 0.7474410425215313, + "flos": 23225980239840.0, + "grad_norm": 2.441121565433172, + "learning_rate": 6.324823109729087e-07, + "loss": 0.7103, + "num_input_tokens_seen": 269675710, + "step": 8288 + }, + { + "epoch": 0.7475312260450016, + "flos": 34151260897920.0, + "grad_norm": 4.029029412020926, + "learning_rate": 6.320560661209653e-07, + "loss": 0.6809, + "num_input_tokens_seen": 269707795, + "step": 8289 + }, + { + "epoch": 0.7476214095684719, + "flos": 22492692051840.0, + "grad_norm": 1.836028790740621, + "learning_rate": 6.316299379895411e-07, + "loss": 0.6302, + "num_input_tokens_seen": 269736650, + "step": 8290 + }, + { + "epoch": 0.7477115930919421, + "flos": 18525310302240.0, + "grad_norm": 1.8941451160946448, + "learning_rate": 6.312039266149965e-07, + "loss": 0.7821, + "num_input_tokens_seen": 269764015, + "step": 8291 + }, + { + "epoch": 0.7478017766154124, + "flos": 25663920082080.0, + "grad_norm": 1.7654347275017255, + "learning_rate": 6.307780320336789e-07, + "loss": 0.7774, + "num_input_tokens_seen": 269794480, + "step": 8292 + }, + { + "epoch": 0.7478919601388826, + "flos": 19072237815360.0, + "grad_norm": 2.069761267638635, + "learning_rate": 6.303522542819306e-07, + "loss": 0.8043, + "num_input_tokens_seen": 269822110, + "step": 8293 + }, + { + "epoch": 0.7479821436623529, + "flos": 19909135325280.0, + "grad_norm": 3.1290773875930884, + "learning_rate": 6.299265933960796e-07, + "loss": 0.7175, + "num_input_tokens_seen": 269849745, + "step": 8294 + }, + { + "epoch": 0.7480723271858232, + "flos": 31311044013120.0, + "grad_norm": 2.3447224951807706, + "learning_rate": 6.295010494124462e-07, + "loss": 0.7002, + "num_input_tokens_seen": 269874730, + "step": 8295 + }, + { + "epoch": 0.7481625107092934, + "flos": 23477027764320.0, + "grad_norm": 1.8515286341922024, + "learning_rate": 6.290756223673399e-07, + "loss": 0.723, + "num_input_tokens_seen": 269904290, + "step": 8296 + }, + { + "epoch": 0.7482526942327636, + "flos": 28470641279520.0, + "grad_norm": 1.6964836470189335, + "learning_rate": 6.28650312297061e-07, + "loss": 0.7768, + "num_input_tokens_seen": 269935165, + "step": 8297 + }, + { + "epoch": 0.748342877756234, + "flos": 11747787737280.0, + "grad_norm": 4.101748507686996, + "learning_rate": 6.282251192378987e-07, + "loss": 0.6629, + "num_input_tokens_seen": 269959230, + "step": 8298 + }, + { + "epoch": 0.7484330612797042, + "flos": 25739088804000.0, + "grad_norm": 1.7571147975165458, + "learning_rate": 6.278000432261334e-07, + "loss": 0.6733, + "num_input_tokens_seen": 269991475, + "step": 8299 + }, + { + "epoch": 0.7485232448031744, + "flos": 16771923359040.0, + "grad_norm": 2.0728935994409534, + "learning_rate": 6.273750842980345e-07, + "loss": 0.7671, + "num_input_tokens_seen": 270018105, + "step": 8300 + }, + { + "epoch": 0.7486134283266447, + "flos": 21221248414080.0, + "grad_norm": 6.341860635458657, + "learning_rate": 6.269502424898625e-07, + "loss": 0.7693, + "num_input_tokens_seen": 270045685, + "step": 8301 + }, + { + "epoch": 0.748703611850115, + "flos": 18488580845760.0, + "grad_norm": 2.074942309944932, + "learning_rate": 6.265255178378663e-07, + "loss": 0.7876, + "num_input_tokens_seen": 270071405, + "step": 8302 + }, + { + "epoch": 0.7487937953735853, + "flos": 28471347504960.0, + "grad_norm": 1.8116033243419614, + "learning_rate": 6.261009103782861e-07, + "loss": 0.6409, + "num_input_tokens_seen": 270102690, + "step": 8303 + }, + { + "epoch": 0.7488839788970555, + "flos": 18452037238080.0, + "grad_norm": 6.331219705281359, + "learning_rate": 6.256764201473519e-07, + "loss": 0.7913, + "num_input_tokens_seen": 270129455, + "step": 8304 + }, + { + "epoch": 0.7489741624205257, + "flos": 58970657935680.0, + "grad_norm": 0.62644255576919, + "learning_rate": 6.252520471812835e-07, + "loss": 0.5714, + "num_input_tokens_seen": 270217705, + "step": 8305 + }, + { + "epoch": 0.7490643459439961, + "flos": 23188470218400.0, + "grad_norm": 2.110800706340418, + "learning_rate": 6.248277915162912e-07, + "loss": 0.7602, + "num_input_tokens_seen": 270246885, + "step": 8306 + }, + { + "epoch": 0.7491545294674663, + "flos": 23808261928320.0, + "grad_norm": 1.9737310219952158, + "learning_rate": 6.244036531885731e-07, + "loss": 0.5992, + "num_input_tokens_seen": 270276460, + "step": 8307 + }, + { + "epoch": 0.7492447129909365, + "flos": 45411938733120.0, + "grad_norm": 1.928884176851827, + "learning_rate": 6.239796322343216e-07, + "loss": 0.6785, + "num_input_tokens_seen": 270310125, + "step": 8308 + }, + { + "epoch": 0.7493348965144068, + "flos": 20930237664000.0, + "grad_norm": 2.3026292823837413, + "learning_rate": 6.235557286897137e-07, + "loss": 0.6833, + "num_input_tokens_seen": 270339545, + "step": 8309 + }, + { + "epoch": 0.7494250800378771, + "flos": 23880828767040.0, + "grad_norm": 1.8552995785338249, + "learning_rate": 6.231319425909223e-07, + "loss": 0.743, + "num_input_tokens_seen": 270369280, + "step": 8310 + }, + { + "epoch": 0.7495152635613473, + "flos": 22934151755040.0, + "grad_norm": 1.7643485392343996, + "learning_rate": 6.227082739741045e-07, + "loss": 0.7777, + "num_input_tokens_seen": 270398540, + "step": 8311 + }, + { + "epoch": 0.7496054470848176, + "flos": 48031370011680.0, + "grad_norm": 1.8126052299733006, + "learning_rate": 6.222847228754113e-07, + "loss": 0.7719, + "num_input_tokens_seen": 270431265, + "step": 8312 + }, + { + "epoch": 0.7496956306082878, + "flos": 20961651844800.0, + "grad_norm": 42.22585640598364, + "learning_rate": 6.218612893309823e-07, + "loss": 0.8403, + "num_input_tokens_seen": 270455585, + "step": 8313 + }, + { + "epoch": 0.7497858141317582, + "flos": 25152792781440.0, + "grad_norm": 1.8924368381416568, + "learning_rate": 6.214379733769468e-07, + "loss": 0.6864, + "num_input_tokens_seen": 270487530, + "step": 8314 + }, + { + "epoch": 0.7498759976552284, + "flos": 20347249750080.0, + "grad_norm": 1.744385841961923, + "learning_rate": 6.21014775049425e-07, + "loss": 0.7756, + "num_input_tokens_seen": 270515220, + "step": 8315 + }, + { + "epoch": 0.7499661811786986, + "flos": 65103330638880.0, + "grad_norm": 0.5761569263652704, + "learning_rate": 6.205916943845267e-07, + "loss": 0.5638, + "num_input_tokens_seen": 270620125, + "step": 8316 + }, + { + "epoch": 0.750056364702169, + "flos": 15136567604640.0, + "grad_norm": 2.0667734667141495, + "learning_rate": 6.201687314183504e-07, + "loss": 0.7907, + "num_input_tokens_seen": 270644675, + "step": 8317 + }, + { + "epoch": 0.7501465482256392, + "flos": 70516733887200.0, + "grad_norm": 0.6516566284029316, + "learning_rate": 6.197458861869862e-07, + "loss": 0.6107, + "num_input_tokens_seen": 270734215, + "step": 8318 + }, + { + "epoch": 0.7502367317491094, + "flos": 28542278874240.0, + "grad_norm": 1.9703109245082235, + "learning_rate": 6.193231587265138e-07, + "loss": 0.7908, + "num_input_tokens_seen": 270765880, + "step": 8319 + }, + { + "epoch": 0.7503269152725797, + "flos": 20310297275040.0, + "grad_norm": 2.2809209716967347, + "learning_rate": 6.189005490730024e-07, + "loss": 0.7266, + "num_input_tokens_seen": 270790735, + "step": 8320 + }, + { + "epoch": 0.75041709879605, + "flos": 33787757648640.0, + "grad_norm": 2.4454929663027216, + "learning_rate": 6.184780572625115e-07, + "loss": 0.7677, + "num_input_tokens_seen": 270821745, + "step": 8321 + }, + { + "epoch": 0.7505072823195202, + "flos": 39619383766560.0, + "grad_norm": 1.9910619211305924, + "learning_rate": 6.180556833310902e-07, + "loss": 0.6469, + "num_input_tokens_seen": 270852385, + "step": 8322 + }, + { + "epoch": 0.7505974658429905, + "flos": 20669117134560.0, + "grad_norm": 2.4438312175552266, + "learning_rate": 6.176334273147788e-07, + "loss": 0.7829, + "num_input_tokens_seen": 270878040, + "step": 8323 + }, + { + "epoch": 0.7506876493664607, + "flos": 24026482821120.0, + "grad_norm": 2.6072001273626606, + "learning_rate": 6.172112892496042e-07, + "loss": 0.6627, + "num_input_tokens_seen": 270903445, + "step": 8324 + }, + { + "epoch": 0.750777832889931, + "flos": 22420050873600.0, + "grad_norm": 1.9207514758538202, + "learning_rate": 6.167892691715883e-07, + "loss": 0.7886, + "num_input_tokens_seen": 270931330, + "step": 8325 + }, + { + "epoch": 0.7508680164134013, + "flos": 25302349660320.0, + "grad_norm": 1.9594631047494901, + "learning_rate": 6.163673671167378e-07, + "loss": 0.7257, + "num_input_tokens_seen": 270960740, + "step": 8326 + }, + { + "epoch": 0.7509581999368715, + "flos": 23181854001120.0, + "grad_norm": 1.7153482238196365, + "learning_rate": 6.15945583121054e-07, + "loss": 0.7848, + "num_input_tokens_seen": 270989345, + "step": 8327 + }, + { + "epoch": 0.7510483834603418, + "flos": 22896790412640.0, + "grad_norm": 2.461613928458722, + "learning_rate": 6.15523917220524e-07, + "loss": 0.695, + "num_input_tokens_seen": 271014690, + "step": 8328 + }, + { + "epoch": 0.7511385669838121, + "flos": 17574135749280.0, + "grad_norm": 1.843594884004817, + "learning_rate": 6.151023694511273e-07, + "loss": 0.8109, + "num_input_tokens_seen": 271041205, + "step": 8329 + }, + { + "epoch": 0.7512287505072823, + "flos": 21877435052640.0, + "grad_norm": 1.728264707467687, + "learning_rate": 6.146809398488328e-07, + "loss": 0.6912, + "num_input_tokens_seen": 271069515, + "step": 8330 + }, + { + "epoch": 0.7513189340307526, + "flos": 23404721113920.0, + "grad_norm": 1.8767094766579702, + "learning_rate": 6.142596284495989e-07, + "loss": 0.7803, + "num_input_tokens_seen": 271099915, + "step": 8331 + }, + { + "epoch": 0.7514091175542228, + "flos": 37506545077920.0, + "grad_norm": 1.8011368340463465, + "learning_rate": 6.138384352893751e-07, + "loss": 0.7016, + "num_input_tokens_seen": 271133365, + "step": 8332 + }, + { + "epoch": 0.7514993010776931, + "flos": 17104346955360.0, + "grad_norm": 2.6103043414581757, + "learning_rate": 6.134173604040987e-07, + "loss": 0.7362, + "num_input_tokens_seen": 271160320, + "step": 8333 + }, + { + "epoch": 0.7515894846011634, + "flos": 19472396181600.0, + "grad_norm": 2.03515214974545, + "learning_rate": 6.129964038296984e-07, + "loss": 0.7489, + "num_input_tokens_seen": 271188020, + "step": 8334 + }, + { + "epoch": 0.7516796681246336, + "flos": 23807890230720.0, + "grad_norm": 2.083996044294297, + "learning_rate": 6.12575565602093e-07, + "loss": 0.8148, + "num_input_tokens_seen": 271217725, + "step": 8335 + }, + { + "epoch": 0.7517698516481038, + "flos": 25048254215520.0, + "grad_norm": 1.6482337717292777, + "learning_rate": 6.121548457571905e-07, + "loss": 0.8172, + "num_input_tokens_seen": 271251790, + "step": 8336 + }, + { + "epoch": 0.7518600351715742, + "flos": 35244855735840.0, + "grad_norm": 2.8056154548956833, + "learning_rate": 6.11734244330889e-07, + "loss": 0.6907, + "num_input_tokens_seen": 271285620, + "step": 8337 + }, + { + "epoch": 0.7519502186950444, + "flos": 27487197641280.0, + "grad_norm": 3.2202753693255284, + "learning_rate": 6.113137613590773e-07, + "loss": 0.7565, + "num_input_tokens_seen": 271314735, + "step": 8338 + }, + { + "epoch": 0.7520404022185146, + "flos": 21185076504000.0, + "grad_norm": 1.71613515137593, + "learning_rate": 6.108933968776313e-07, + "loss": 0.8013, + "num_input_tokens_seen": 271341805, + "step": 8339 + }, + { + "epoch": 0.752130585741985, + "flos": 23732758678560.0, + "grad_norm": 2.1343707865147032, + "learning_rate": 6.104731509224212e-07, + "loss": 0.7444, + "num_input_tokens_seen": 271370305, + "step": 8340 + }, + { + "epoch": 0.7522207692654552, + "flos": 36086325126240.0, + "grad_norm": 1.6827368447414803, + "learning_rate": 6.100530235293027e-07, + "loss": 0.6769, + "num_input_tokens_seen": 271401380, + "step": 8341 + }, + { + "epoch": 0.7523109527889255, + "flos": 28835817168000.0, + "grad_norm": 4.044733489773572, + "learning_rate": 6.096330147341253e-07, + "loss": 0.7679, + "num_input_tokens_seen": 271430660, + "step": 8342 + }, + { + "epoch": 0.7524011363123957, + "flos": 19173877140000.0, + "grad_norm": 2.881165072498935, + "learning_rate": 6.09213124572725e-07, + "loss": 0.8022, + "num_input_tokens_seen": 271452985, + "step": 8343 + }, + { + "epoch": 0.752491319835866, + "flos": 21367794542400.0, + "grad_norm": 1.6873758764888511, + "learning_rate": 6.087933530809297e-07, + "loss": 0.8205, + "num_input_tokens_seen": 271482140, + "step": 8344 + }, + { + "epoch": 0.7525815033593363, + "flos": 13206335445120.0, + "grad_norm": 1.7744005587981972, + "learning_rate": 6.083737002945566e-07, + "loss": 0.7307, + "num_input_tokens_seen": 271508845, + "step": 8345 + }, + { + "epoch": 0.7526716868828065, + "flos": 45521904084000.0, + "grad_norm": 2.118298003559642, + "learning_rate": 6.079541662494126e-07, + "loss": 0.6961, + "num_input_tokens_seen": 271540500, + "step": 8346 + }, + { + "epoch": 0.7527618704062767, + "flos": 23622421630080.0, + "grad_norm": 1.8131605457396969, + "learning_rate": 6.075347509812954e-07, + "loss": 0.7491, + "num_input_tokens_seen": 271569065, + "step": 8347 + }, + { + "epoch": 0.7528520539297471, + "flos": 22601244951840.0, + "grad_norm": 1.6832555444451307, + "learning_rate": 6.0711545452599e-07, + "loss": 0.6986, + "num_input_tokens_seen": 271595315, + "step": 8348 + }, + { + "epoch": 0.7529422374532173, + "flos": 37360036119360.0, + "grad_norm": 4.662745099618214, + "learning_rate": 6.066962769192756e-07, + "loss": 0.6698, + "num_input_tokens_seen": 271629420, + "step": 8349 + }, + { + "epoch": 0.7530324209766875, + "flos": 18231995027040.0, + "grad_norm": 3.7446799369913273, + "learning_rate": 6.062772181969167e-07, + "loss": 0.6801, + "num_input_tokens_seen": 271654025, + "step": 8350 + }, + { + "epoch": 0.7531226045001578, + "flos": 22638977991840.0, + "grad_norm": 1.7524895670712226, + "learning_rate": 6.058582783946706e-07, + "loss": 0.8134, + "num_input_tokens_seen": 271681815, + "step": 8351 + }, + { + "epoch": 0.7532127880236281, + "flos": 22569347564160.0, + "grad_norm": 2.2259854823870375, + "learning_rate": 6.054394575482833e-07, + "loss": 0.7351, + "num_input_tokens_seen": 271711350, + "step": 8352 + }, + { + "epoch": 0.7533029715470984, + "flos": 21075222662400.0, + "grad_norm": 1.8882399719044725, + "learning_rate": 6.05020755693491e-07, + "loss": 0.66, + "num_input_tokens_seen": 271737915, + "step": 8353 + }, + { + "epoch": 0.7533931550705686, + "flos": 32879482732320.0, + "grad_norm": 1.8578851324694565, + "learning_rate": 6.046021728660198e-07, + "loss": 0.6571, + "num_input_tokens_seen": 271767845, + "step": 8354 + }, + { + "epoch": 0.7534833385940388, + "flos": 21002804502720.0, + "grad_norm": 2.476810671529033, + "learning_rate": 6.041837091015858e-07, + "loss": 0.7819, + "num_input_tokens_seen": 271796435, + "step": 8355 + }, + { + "epoch": 0.7535735221175092, + "flos": 34154680515840.0, + "grad_norm": 1.67519199352774, + "learning_rate": 6.037653644358931e-07, + "loss": 0.7953, + "num_input_tokens_seen": 271826815, + "step": 8356 + }, + { + "epoch": 0.7536637056409794, + "flos": 23079396941760.0, + "grad_norm": 2.6424612849782005, + "learning_rate": 6.033471389046393e-07, + "loss": 0.7553, + "num_input_tokens_seen": 271857775, + "step": 8357 + }, + { + "epoch": 0.7537538891644496, + "flos": 25083608390880.0, + "grad_norm": 1.7928551199483125, + "learning_rate": 6.029290325435084e-07, + "loss": 0.7006, + "num_input_tokens_seen": 271889190, + "step": 8358 + }, + { + "epoch": 0.7538440726879199, + "flos": 22314954761280.0, + "grad_norm": 2.9248971252608826, + "learning_rate": 6.025110453881756e-07, + "loss": 0.8065, + "num_input_tokens_seen": 271915105, + "step": 8359 + }, + { + "epoch": 0.7539342562113902, + "flos": 28762321085280.0, + "grad_norm": 2.2080013337476823, + "learning_rate": 6.020931774743061e-07, + "loss": 0.6807, + "num_input_tokens_seen": 271943725, + "step": 8360 + }, + { + "epoch": 0.7540244397348604, + "flos": 24646311700800.0, + "grad_norm": 2.558490966596045, + "learning_rate": 6.016754288375546e-07, + "loss": 0.6684, + "num_input_tokens_seen": 271970305, + "step": 8361 + }, + { + "epoch": 0.7541146232583307, + "flos": 20966446743840.0, + "grad_norm": 2.0151539333413306, + "learning_rate": 6.012577995135665e-07, + "loss": 0.7375, + "num_input_tokens_seen": 271998520, + "step": 8362 + }, + { + "epoch": 0.754204806781801, + "flos": 18416348534880.0, + "grad_norm": 2.2895100126828556, + "learning_rate": 6.008402895379743e-07, + "loss": 0.6904, + "num_input_tokens_seen": 272024835, + "step": 8363 + }, + { + "epoch": 0.7542949903052713, + "flos": 23880754427520.0, + "grad_norm": 4.14037647581113, + "learning_rate": 6.004228989464047e-07, + "loss": 0.8318, + "num_input_tokens_seen": 272049870, + "step": 8364 + }, + { + "epoch": 0.7543851738287415, + "flos": 18416199855840.0, + "grad_norm": 2.2938659056605415, + "learning_rate": 6.000056277744692e-07, + "loss": 0.7677, + "num_input_tokens_seen": 272079220, + "step": 8365 + }, + { + "epoch": 0.7544753573522117, + "flos": 21401327399520.0, + "grad_norm": 1.6896180177163198, + "learning_rate": 5.995884760577745e-07, + "loss": 0.7144, + "num_input_tokens_seen": 272109685, + "step": 8366 + }, + { + "epoch": 0.7545655408756821, + "flos": 64537632396960.0, + "grad_norm": 0.622662967085654, + "learning_rate": 5.99171443831912e-07, + "loss": 0.5643, + "num_input_tokens_seen": 272203145, + "step": 8367 + }, + { + "epoch": 0.7546557243991523, + "flos": 23990794117920.0, + "grad_norm": 1.5711318578134141, + "learning_rate": 5.98754531132466e-07, + "loss": 0.743, + "num_input_tokens_seen": 272234515, + "step": 8368 + }, + { + "epoch": 0.7547459079226225, + "flos": 25083534051360.0, + "grad_norm": 2.078721957404667, + "learning_rate": 5.983377379950099e-07, + "loss": 0.7457, + "num_input_tokens_seen": 272262530, + "step": 8369 + }, + { + "epoch": 0.7548360914460928, + "flos": 21804496516320.0, + "grad_norm": 2.8875557766105766, + "learning_rate": 5.979210644551067e-07, + "loss": 0.6824, + "num_input_tokens_seen": 272290570, + "step": 8370 + }, + { + "epoch": 0.7549262749695631, + "flos": 20820272313120.0, + "grad_norm": 2.6368403653955514, + "learning_rate": 5.975045105483091e-07, + "loss": 0.7431, + "num_input_tokens_seen": 272314635, + "step": 8371 + }, + { + "epoch": 0.7550164584930333, + "flos": 69322912175520.0, + "grad_norm": 0.60221767664684, + "learning_rate": 5.970880763101607e-07, + "loss": 0.587, + "num_input_tokens_seen": 272413170, + "step": 8372 + }, + { + "epoch": 0.7551066420165036, + "flos": 17936523905760.0, + "grad_norm": 2.3984668245293532, + "learning_rate": 5.966717617761925e-07, + "loss": 0.7686, + "num_input_tokens_seen": 272438790, + "step": 8373 + }, + { + "epoch": 0.7551968255399738, + "flos": 22059669884160.0, + "grad_norm": 4.159074784241453, + "learning_rate": 5.962555669819276e-07, + "loss": 0.7259, + "num_input_tokens_seen": 272469395, + "step": 8374 + }, + { + "epoch": 0.7552870090634441, + "flos": 27705158345760.0, + "grad_norm": 1.4553255216488175, + "learning_rate": 5.958394919628777e-07, + "loss": 0.7654, + "num_input_tokens_seen": 272501060, + "step": 8375 + }, + { + "epoch": 0.7553771925869144, + "flos": 25301569095360.0, + "grad_norm": 1.4719015357992724, + "learning_rate": 5.954235367545451e-07, + "loss": 0.6841, + "num_input_tokens_seen": 272533425, + "step": 8376 + }, + { + "epoch": 0.7554673761103846, + "flos": 22565147381280.0, + "grad_norm": 2.0380332950945776, + "learning_rate": 5.950077013924213e-07, + "loss": 0.7397, + "num_input_tokens_seen": 272562615, + "step": 8377 + }, + { + "epoch": 0.7555575596338548, + "flos": 19654891201440.0, + "grad_norm": 2.158720950321607, + "learning_rate": 5.945919859119865e-07, + "loss": 0.8285, + "num_input_tokens_seen": 272589100, + "step": 8378 + }, + { + "epoch": 0.7556477431573252, + "flos": 21586089774720.0, + "grad_norm": 1.928869467871666, + "learning_rate": 5.94176390348714e-07, + "loss": 0.6335, + "num_input_tokens_seen": 272618220, + "step": 8379 + }, + { + "epoch": 0.7557379266807954, + "flos": 24427979298720.0, + "grad_norm": 1.4138051425827898, + "learning_rate": 5.937609147380622e-07, + "loss": 0.8506, + "num_input_tokens_seen": 272649635, + "step": 8380 + }, + { + "epoch": 0.7558281102042657, + "flos": 21622744891680.0, + "grad_norm": 1.72982121530526, + "learning_rate": 5.933455591154844e-07, + "loss": 0.6887, + "num_input_tokens_seen": 272676840, + "step": 8381 + }, + { + "epoch": 0.7559182937277359, + "flos": 21913644132480.0, + "grad_norm": 2.5767140544598517, + "learning_rate": 5.929303235164191e-07, + "loss": 0.7165, + "num_input_tokens_seen": 272703840, + "step": 8382 + }, + { + "epoch": 0.7560084772512062, + "flos": 23370965238240.0, + "grad_norm": 2.2211029041811443, + "learning_rate": 5.92515207976297e-07, + "loss": 0.7165, + "num_input_tokens_seen": 272731765, + "step": 8383 + }, + { + "epoch": 0.7560986607746765, + "flos": 28253684158560.0, + "grad_norm": 1.724509490113374, + "learning_rate": 5.921002125305383e-07, + "loss": 0.735, + "num_input_tokens_seen": 272761255, + "step": 8384 + }, + { + "epoch": 0.7561888442981467, + "flos": 25409155581600.0, + "grad_norm": 2.552998769153936, + "learning_rate": 5.916853372145525e-07, + "loss": 0.6656, + "num_input_tokens_seen": 272788210, + "step": 8385 + }, + { + "epoch": 0.756279027821617, + "flos": 25264765299360.0, + "grad_norm": 1.895389574815643, + "learning_rate": 5.912705820637389e-07, + "loss": 0.6737, + "num_input_tokens_seen": 272816560, + "step": 8386 + }, + { + "epoch": 0.7563692113450873, + "flos": 30619540368960.0, + "grad_norm": 1.6556499041840365, + "learning_rate": 5.908559471134871e-07, + "loss": 0.8113, + "num_input_tokens_seen": 272848745, + "step": 8387 + }, + { + "epoch": 0.7564593948685575, + "flos": 23840419504320.0, + "grad_norm": 1.8330087880216837, + "learning_rate": 5.904414323991764e-07, + "loss": 0.6827, + "num_input_tokens_seen": 272875625, + "step": 8388 + }, + { + "epoch": 0.7565495783920277, + "flos": 23293975198080.0, + "grad_norm": 1.8420772399778553, + "learning_rate": 5.900270379561743e-07, + "loss": 0.7513, + "num_input_tokens_seen": 272903345, + "step": 8389 + }, + { + "epoch": 0.7566397619154981, + "flos": 20711050357440.0, + "grad_norm": 2.005830253827983, + "learning_rate": 5.896127638198399e-07, + "loss": 0.7635, + "num_input_tokens_seen": 272931010, + "step": 8390 + }, + { + "epoch": 0.7567299454389683, + "flos": 17686926001920.0, + "grad_norm": 2.030633451350863, + "learning_rate": 5.89198610025521e-07, + "loss": 0.7475, + "num_input_tokens_seen": 272955005, + "step": 8391 + }, + { + "epoch": 0.7568201289624386, + "flos": 25666224607200.0, + "grad_norm": 2.212262407035062, + "learning_rate": 5.887845766085559e-07, + "loss": 0.8116, + "num_input_tokens_seen": 272986640, + "step": 8392 + }, + { + "epoch": 0.7569103124859088, + "flos": 20378738270400.0, + "grad_norm": 2.203275959443277, + "learning_rate": 5.883706636042722e-07, + "loss": 0.7067, + "num_input_tokens_seen": 273014425, + "step": 8393 + }, + { + "epoch": 0.7570004960093791, + "flos": 23516433443520.0, + "grad_norm": 3.1477292307793463, + "learning_rate": 5.879568710479879e-07, + "loss": 0.6861, + "num_input_tokens_seen": 273041105, + "step": 8394 + }, + { + "epoch": 0.7570906795328494, + "flos": 27335633595360.0, + "grad_norm": 1.7698079076077529, + "learning_rate": 5.875431989750078e-07, + "loss": 0.8344, + "num_input_tokens_seen": 273072730, + "step": 8395 + }, + { + "epoch": 0.7571808630563196, + "flos": 20856927430080.0, + "grad_norm": 1.5820673432177128, + "learning_rate": 5.871296474206313e-07, + "loss": 0.6808, + "num_input_tokens_seen": 273101790, + "step": 8396 + }, + { + "epoch": 0.7572710465797898, + "flos": 21877323543360.0, + "grad_norm": 1.6223459577981687, + "learning_rate": 5.867162164201427e-07, + "loss": 0.7504, + "num_input_tokens_seen": 273130445, + "step": 8397 + }, + { + "epoch": 0.7573612301032602, + "flos": 21512482182720.0, + "grad_norm": 1.898334139298203, + "learning_rate": 5.863029060088205e-07, + "loss": 0.7748, + "num_input_tokens_seen": 273159285, + "step": 8398 + }, + { + "epoch": 0.7574514136267304, + "flos": 20420559984000.0, + "grad_norm": 1.7710416633879087, + "learning_rate": 5.858897162219289e-07, + "loss": 0.6433, + "num_input_tokens_seen": 273185100, + "step": 8399 + }, + { + "epoch": 0.7575415971502006, + "flos": 24026445651360.0, + "grad_norm": 2.0066521404961457, + "learning_rate": 5.854766470947238e-07, + "loss": 0.7927, + "num_input_tokens_seen": 273212410, + "step": 8400 + }, + { + "epoch": 0.7576317806736709, + "flos": 29231217804960.0, + "grad_norm": 1.6700825574878246, + "learning_rate": 5.850636986624511e-07, + "loss": 0.7362, + "num_input_tokens_seen": 273242095, + "step": 8401 + }, + { + "epoch": 0.7577219641971412, + "flos": 23406653941440.0, + "grad_norm": 2.3936896533181966, + "learning_rate": 5.846508709603453e-07, + "loss": 0.7553, + "num_input_tokens_seen": 273268580, + "step": 8402 + }, + { + "epoch": 0.7578121477206115, + "flos": 24242101830720.0, + "grad_norm": 1.7411408295266149, + "learning_rate": 5.842381640236318e-07, + "loss": 0.7215, + "num_input_tokens_seen": 273300680, + "step": 8403 + }, + { + "epoch": 0.7579023312440817, + "flos": 24792263112960.0, + "grad_norm": 1.6616272636162266, + "learning_rate": 5.838255778875242e-07, + "loss": 0.7442, + "num_input_tokens_seen": 273330705, + "step": 8404 + }, + { + "epoch": 0.7579925147675519, + "flos": 28362014040000.0, + "grad_norm": 2.395323978404035, + "learning_rate": 5.83413112587227e-07, + "loss": 0.7633, + "num_input_tokens_seen": 273358650, + "step": 8405 + }, + { + "epoch": 0.7580826982910223, + "flos": 36923668673280.0, + "grad_norm": 2.100810316969269, + "learning_rate": 5.830007681579338e-07, + "loss": 0.7319, + "num_input_tokens_seen": 273389885, + "step": 8406 + }, + { + "epoch": 0.7581728818144925, + "flos": 18116714400480.0, + "grad_norm": 1.9007941794749452, + "learning_rate": 5.825885446348284e-07, + "loss": 0.7603, + "num_input_tokens_seen": 273412375, + "step": 8407 + }, + { + "epoch": 0.7582630653379627, + "flos": 35682784311840.0, + "grad_norm": 1.7784065171222305, + "learning_rate": 5.821764420530842e-07, + "loss": 0.6809, + "num_input_tokens_seen": 273445370, + "step": 8408 + }, + { + "epoch": 0.7583532488614331, + "flos": 22022940427680.0, + "grad_norm": 1.6822527791381683, + "learning_rate": 5.817644604478633e-07, + "loss": 0.7625, + "num_input_tokens_seen": 273476405, + "step": 8409 + }, + { + "epoch": 0.7584434323849033, + "flos": 21549062960160.0, + "grad_norm": 1.8165695937349733, + "learning_rate": 5.81352599854319e-07, + "loss": 0.6909, + "num_input_tokens_seen": 273505360, + "step": 8410 + }, + { + "epoch": 0.7585336159083735, + "flos": 21289838088480.0, + "grad_norm": 1.6147218788345374, + "learning_rate": 5.809408603075938e-07, + "loss": 0.7764, + "num_input_tokens_seen": 273531660, + "step": 8411 + }, + { + "epoch": 0.7586237994318438, + "flos": 23917669732800.0, + "grad_norm": 2.2679627032434246, + "learning_rate": 5.805292418428176e-07, + "loss": 0.7257, + "num_input_tokens_seen": 273559955, + "step": 8412 + }, + { + "epoch": 0.7587139829553141, + "flos": 22933185341280.0, + "grad_norm": 2.4988653469131092, + "learning_rate": 5.801177444951148e-07, + "loss": 0.7353, + "num_input_tokens_seen": 273585930, + "step": 8413 + }, + { + "epoch": 0.7588041664787843, + "flos": 23954324849760.0, + "grad_norm": 1.7188413422328068, + "learning_rate": 5.797063682995944e-07, + "loss": 0.7754, + "num_input_tokens_seen": 273616660, + "step": 8414 + }, + { + "epoch": 0.7588943500022546, + "flos": 23698110728640.0, + "grad_norm": 1.99392606382358, + "learning_rate": 5.792951132913584e-07, + "loss": 0.8103, + "num_input_tokens_seen": 273643380, + "step": 8415 + }, + { + "epoch": 0.7589845335257248, + "flos": 22059409695840.0, + "grad_norm": 2.3270909618046605, + "learning_rate": 5.788839795054968e-07, + "loss": 0.7981, + "num_input_tokens_seen": 273672095, + "step": 8416 + }, + { + "epoch": 0.7590747170491952, + "flos": 21694531165440.0, + "grad_norm": 1.7280849459043472, + "learning_rate": 5.784729669770898e-07, + "loss": 0.6615, + "num_input_tokens_seen": 273702550, + "step": 8417 + }, + { + "epoch": 0.7591649005726654, + "flos": 21692895696000.0, + "grad_norm": 1.5019778880842238, + "learning_rate": 5.780620757412084e-07, + "loss": 0.742, + "num_input_tokens_seen": 273732730, + "step": 8418 + }, + { + "epoch": 0.7592550840961356, + "flos": 21986508329280.0, + "grad_norm": 1.9647833206257828, + "learning_rate": 5.776513058329098e-07, + "loss": 0.7636, + "num_input_tokens_seen": 273759255, + "step": 8419 + }, + { + "epoch": 0.7593452676196059, + "flos": 18994838907840.0, + "grad_norm": 1.7259770432113595, + "learning_rate": 5.772406572872459e-07, + "loss": 0.7687, + "num_input_tokens_seen": 273786900, + "step": 8420 + }, + { + "epoch": 0.7594354511430762, + "flos": 13534298670240.0, + "grad_norm": 2.334236367640053, + "learning_rate": 5.768301301392535e-07, + "loss": 0.7736, + "num_input_tokens_seen": 273812305, + "step": 8421 + }, + { + "epoch": 0.7595256346665464, + "flos": 26176088136000.0, + "grad_norm": 3.8712291212692116, + "learning_rate": 5.764197244239615e-07, + "loss": 0.7544, + "num_input_tokens_seen": 273841520, + "step": 8422 + }, + { + "epoch": 0.7596158181900167, + "flos": 26686880908800.0, + "grad_norm": 1.854471295938637, + "learning_rate": 5.760094401763884e-07, + "loss": 0.7836, + "num_input_tokens_seen": 273871815, + "step": 8423 + }, + { + "epoch": 0.7597060017134869, + "flos": 32474826825120.0, + "grad_norm": 4.50646550566625, + "learning_rate": 5.755992774315414e-07, + "loss": 0.7325, + "num_input_tokens_seen": 273905185, + "step": 8424 + }, + { + "epoch": 0.7597961852369572, + "flos": 21913123755840.0, + "grad_norm": 2.987391999020307, + "learning_rate": 5.751892362244183e-07, + "loss": 0.7345, + "num_input_tokens_seen": 273935525, + "step": 8425 + }, + { + "epoch": 0.7598863687604275, + "flos": 18379842096960.0, + "grad_norm": 2.0834931757325186, + "learning_rate": 5.747793165900065e-07, + "loss": 0.8011, + "num_input_tokens_seen": 273961700, + "step": 8426 + }, + { + "epoch": 0.7599765522838977, + "flos": 19509237147360.0, + "grad_norm": 1.8606092700030648, + "learning_rate": 5.743695185632806e-07, + "loss": 0.8017, + "num_input_tokens_seen": 273989665, + "step": 8427 + }, + { + "epoch": 0.7600667358073679, + "flos": 26321965208640.0, + "grad_norm": 2.015837902091774, + "learning_rate": 5.739598421792098e-07, + "loss": 0.7846, + "num_input_tokens_seen": 274014205, + "step": 8428 + }, + { + "epoch": 0.7601569193308383, + "flos": 22096473680160.0, + "grad_norm": 2.187652364601645, + "learning_rate": 5.735502874727474e-07, + "loss": 0.7278, + "num_input_tokens_seen": 274044195, + "step": 8429 + }, + { + "epoch": 0.7602471028543085, + "flos": 29194153820640.0, + "grad_norm": 1.7361466577917462, + "learning_rate": 5.731408544788398e-07, + "loss": 0.8749, + "num_input_tokens_seen": 274073325, + "step": 8430 + }, + { + "epoch": 0.7603372863777788, + "flos": 26683572800160.0, + "grad_norm": 2.1640057607986574, + "learning_rate": 5.727315432324225e-07, + "loss": 0.6616, + "num_input_tokens_seen": 274106830, + "step": 8431 + }, + { + "epoch": 0.760427469901249, + "flos": 20309776898400.0, + "grad_norm": 2.1021971471073773, + "learning_rate": 5.723223537684196e-07, + "loss": 0.7312, + "num_input_tokens_seen": 274133935, + "step": 8432 + }, + { + "epoch": 0.7605176534247193, + "flos": 18481667270400.0, + "grad_norm": 2.1465754537376194, + "learning_rate": 5.719132861217462e-07, + "loss": 0.8245, + "num_input_tokens_seen": 274160890, + "step": 8433 + }, + { + "epoch": 0.7606078369481896, + "flos": 44355073351680.0, + "grad_norm": 4.016762182806886, + "learning_rate": 5.715043403273044e-07, + "loss": 0.7125, + "num_input_tokens_seen": 274196410, + "step": 8434 + }, + { + "epoch": 0.7606980204716598, + "flos": 24093920232960.0, + "grad_norm": 3.3961901019012397, + "learning_rate": 5.710955164199902e-07, + "loss": 0.699, + "num_input_tokens_seen": 274219720, + "step": 8435 + }, + { + "epoch": 0.7607882039951301, + "flos": 20015569548960.0, + "grad_norm": 1.8973927704036542, + "learning_rate": 5.706868144346841e-07, + "loss": 0.7658, + "num_input_tokens_seen": 274246830, + "step": 8436 + }, + { + "epoch": 0.7608783875186004, + "flos": 23881609332000.0, + "grad_norm": 1.8016156230945313, + "learning_rate": 5.702782344062613e-07, + "loss": 0.7388, + "num_input_tokens_seen": 274276500, + "step": 8437 + }, + { + "epoch": 0.7609685710420706, + "flos": 28216397155680.0, + "grad_norm": 1.9705382808402159, + "learning_rate": 5.698697763695826e-07, + "loss": 0.7231, + "num_input_tokens_seen": 274306380, + "step": 8438 + }, + { + "epoch": 0.7610587545655408, + "flos": 21840817105440.0, + "grad_norm": 1.9275699997596427, + "learning_rate": 5.694614403595002e-07, + "loss": 0.7672, + "num_input_tokens_seen": 274334090, + "step": 8439 + }, + { + "epoch": 0.7611489380890112, + "flos": 34440487499520.0, + "grad_norm": 2.0118026567756515, + "learning_rate": 5.690532264108554e-07, + "loss": 0.6558, + "num_input_tokens_seen": 274364860, + "step": 8440 + }, + { + "epoch": 0.7612391216124814, + "flos": 22346480451360.0, + "grad_norm": 1.9443077955216341, + "learning_rate": 5.686451345584795e-07, + "loss": 0.7784, + "num_input_tokens_seen": 274393910, + "step": 8441 + }, + { + "epoch": 0.7613293051359517, + "flos": 25592802864000.0, + "grad_norm": 2.3429479625424645, + "learning_rate": 5.682371648371933e-07, + "loss": 0.842, + "num_input_tokens_seen": 274420970, + "step": 8442 + }, + { + "epoch": 0.7614194886594219, + "flos": 22017067605600.0, + "grad_norm": 1.9462967186761417, + "learning_rate": 5.678293172818074e-07, + "loss": 0.7059, + "num_input_tokens_seen": 274448655, + "step": 8443 + }, + { + "epoch": 0.7615096721828922, + "flos": 26206238544960.0, + "grad_norm": 1.9948501445541982, + "learning_rate": 5.674215919271204e-07, + "loss": 0.7594, + "num_input_tokens_seen": 274475885, + "step": 8444 + }, + { + "epoch": 0.7615998557063625, + "flos": 18233667666240.0, + "grad_norm": 2.0443718433734777, + "learning_rate": 5.670139888079224e-07, + "loss": 0.6717, + "num_input_tokens_seen": 274503955, + "step": 8445 + }, + { + "epoch": 0.7616900392298327, + "flos": 24570994299840.0, + "grad_norm": 1.9979009341938812, + "learning_rate": 5.666065079589924e-07, + "loss": 0.8027, + "num_input_tokens_seen": 274535095, + "step": 8446 + }, + { + "epoch": 0.7617802227533029, + "flos": 25047250632000.0, + "grad_norm": 1.7691300971117392, + "learning_rate": 5.661991494150986e-07, + "loss": 0.6979, + "num_input_tokens_seen": 274564485, + "step": 8447 + }, + { + "epoch": 0.7618704062767733, + "flos": 19324809300000.0, + "grad_norm": 1.928234588130182, + "learning_rate": 5.657919132109999e-07, + "loss": 0.7917, + "num_input_tokens_seen": 274593665, + "step": 8448 + }, + { + "epoch": 0.7619605898002435, + "flos": 24679435690560.0, + "grad_norm": 1.858462326066156, + "learning_rate": 5.653847993814421e-07, + "loss": 0.7781, + "num_input_tokens_seen": 274620995, + "step": 8449 + }, + { + "epoch": 0.7620507733237137, + "flos": 22860432653760.0, + "grad_norm": 3.0729861666476332, + "learning_rate": 5.649778079611647e-07, + "loss": 0.7551, + "num_input_tokens_seen": 274650455, + "step": 8450 + }, + { + "epoch": 0.762140956847184, + "flos": 22819688863200.0, + "grad_norm": 1.807019465136331, + "learning_rate": 5.645709389848923e-07, + "loss": 0.7263, + "num_input_tokens_seen": 274678110, + "step": 8451 + }, + { + "epoch": 0.7622311403706543, + "flos": 17760533593920.0, + "grad_norm": 1.9554221720373801, + "learning_rate": 5.641641924873435e-07, + "loss": 0.7081, + "num_input_tokens_seen": 274706130, + "step": 8452 + }, + { + "epoch": 0.7623213238941245, + "flos": 21549323148480.0, + "grad_norm": 2.006473554948827, + "learning_rate": 5.637575685032217e-07, + "loss": 0.7429, + "num_input_tokens_seen": 274734175, + "step": 8453 + }, + { + "epoch": 0.7624115074175948, + "flos": 29200063812480.0, + "grad_norm": 2.246922719817437, + "learning_rate": 5.633510670672246e-07, + "loss": 0.8237, + "num_input_tokens_seen": 274763445, + "step": 8454 + }, + { + "epoch": 0.762501690941065, + "flos": 65860450376640.0, + "grad_norm": 0.6231935621400398, + "learning_rate": 5.629446882140354e-07, + "loss": 0.6044, + "num_input_tokens_seen": 274851425, + "step": 8455 + }, + { + "epoch": 0.7625918744645354, + "flos": 26101848658080.0, + "grad_norm": 1.5529482520251656, + "learning_rate": 5.625384319783295e-07, + "loss": 0.7559, + "num_input_tokens_seen": 274880260, + "step": 8456 + }, + { + "epoch": 0.7626820579880056, + "flos": 24645716984640.0, + "grad_norm": 8.289257165922226, + "learning_rate": 5.621322983947705e-07, + "loss": 0.6941, + "num_input_tokens_seen": 274907560, + "step": 8457 + }, + { + "epoch": 0.7627722415114758, + "flos": 26500111366560.0, + "grad_norm": 2.3912446459377747, + "learning_rate": 5.617262874980122e-07, + "loss": 0.7806, + "num_input_tokens_seen": 274934815, + "step": 8458 + }, + { + "epoch": 0.7628624250349462, + "flos": 27011350176480.0, + "grad_norm": 1.9390732315344406, + "learning_rate": 5.613203993226981e-07, + "loss": 0.7904, + "num_input_tokens_seen": 274968660, + "step": 8459 + }, + { + "epoch": 0.7629526085584164, + "flos": 18298837722720.0, + "grad_norm": 2.7359037648838394, + "learning_rate": 5.609146339034599e-07, + "loss": 0.7999, + "num_input_tokens_seen": 274992070, + "step": 8460 + }, + { + "epoch": 0.7630427920818866, + "flos": 17760161896320.0, + "grad_norm": 2.5750333802365115, + "learning_rate": 5.605089912749199e-07, + "loss": 0.6839, + "num_input_tokens_seen": 275020325, + "step": 8461 + }, + { + "epoch": 0.7631329756053569, + "flos": 23875625000640.0, + "grad_norm": 2.5582201502943067, + "learning_rate": 5.601034714716901e-07, + "loss": 0.6899, + "num_input_tokens_seen": 275048195, + "step": 8462 + }, + { + "epoch": 0.7632231591288272, + "flos": 24791556887520.0, + "grad_norm": 1.7776855140162346, + "learning_rate": 5.59698074528372e-07, + "loss": 0.7176, + "num_input_tokens_seen": 275080270, + "step": 8463 + }, + { + "epoch": 0.7633133426522974, + "flos": 22714592750880.0, + "grad_norm": 1.6711398495971608, + "learning_rate": 5.592928004795555e-07, + "loss": 0.7168, + "num_input_tokens_seen": 275109710, + "step": 8464 + }, + { + "epoch": 0.7634035261757677, + "flos": 17942359558080.0, + "grad_norm": 2.3430795346354425, + "learning_rate": 5.58887649359822e-07, + "loss": 0.7505, + "num_input_tokens_seen": 275135835, + "step": 8465 + }, + { + "epoch": 0.7634937096992379, + "flos": 34520116592640.0, + "grad_norm": 2.0207271613581668, + "learning_rate": 5.584826212037393e-07, + "loss": 0.5988, + "num_input_tokens_seen": 275165945, + "step": 8466 + }, + { + "epoch": 0.7635838932227083, + "flos": 39910543195680.0, + "grad_norm": 1.6529710088588758, + "learning_rate": 5.580777160458689e-07, + "loss": 0.6645, + "num_input_tokens_seen": 275199635, + "step": 8467 + }, + { + "epoch": 0.7636740767461785, + "flos": 26030842949280.0, + "grad_norm": 1.6348049291263396, + "learning_rate": 5.576729339207574e-07, + "loss": 0.7669, + "num_input_tokens_seen": 275229275, + "step": 8468 + }, + { + "epoch": 0.7637642602696487, + "flos": 26904841613280.0, + "grad_norm": 1.33797754158536, + "learning_rate": 5.572682748629449e-07, + "loss": 0.8034, + "num_input_tokens_seen": 275261900, + "step": 8469 + }, + { + "epoch": 0.763854443793119, + "flos": 29346089564160.0, + "grad_norm": 1.6325980597775525, + "learning_rate": 5.568637389069582e-07, + "loss": 0.7298, + "num_input_tokens_seen": 275292625, + "step": 8470 + }, + { + "epoch": 0.7639446273165893, + "flos": 20198176078080.0, + "grad_norm": 2.1170692551591546, + "learning_rate": 5.564593260873145e-07, + "loss": 0.755, + "num_input_tokens_seen": 275319065, + "step": 8471 + }, + { + "epoch": 0.7640348108400595, + "flos": 21547501830240.0, + "grad_norm": 1.6777500778014054, + "learning_rate": 5.560550364385206e-07, + "loss": 0.7812, + "num_input_tokens_seen": 275348345, + "step": 8472 + }, + { + "epoch": 0.7641249943635298, + "flos": 22751619565440.0, + "grad_norm": 1.897552221029454, + "learning_rate": 5.556508699950728e-07, + "loss": 0.8167, + "num_input_tokens_seen": 275374790, + "step": 8473 + }, + { + "epoch": 0.764215177887, + "flos": 17978271279840.0, + "grad_norm": 1.925680904598381, + "learning_rate": 5.552468267914577e-07, + "loss": 0.6966, + "num_input_tokens_seen": 275398680, + "step": 8474 + }, + { + "epoch": 0.7643053614104703, + "flos": 22459976929440.0, + "grad_norm": 2.3839225809788434, + "learning_rate": 5.548429068621481e-07, + "loss": 0.6581, + "num_input_tokens_seen": 275427940, + "step": 8475 + }, + { + "epoch": 0.7643955449339406, + "flos": 61970690553120.0, + "grad_norm": 0.6516548448069194, + "learning_rate": 5.544391102416115e-07, + "loss": 0.5738, + "num_input_tokens_seen": 275517440, + "step": 8476 + }, + { + "epoch": 0.7644857284574108, + "flos": 17614507842240.0, + "grad_norm": 3.1214768958739083, + "learning_rate": 5.540354369643003e-07, + "loss": 0.7997, + "num_input_tokens_seen": 275543225, + "step": 8477 + }, + { + "epoch": 0.764575911980881, + "flos": 27487123301760.0, + "grad_norm": 3.675261798278957, + "learning_rate": 5.536318870646586e-07, + "loss": 0.5956, + "num_input_tokens_seen": 275572645, + "step": 8478 + }, + { + "epoch": 0.7646660955043514, + "flos": 25629383641440.0, + "grad_norm": 2.738504136791939, + "learning_rate": 5.532284605771194e-07, + "loss": 0.6496, + "num_input_tokens_seen": 275600200, + "step": 8479 + }, + { + "epoch": 0.7647562790278216, + "flos": 32660704293120.0, + "grad_norm": 1.4329042361645494, + "learning_rate": 5.528251575361052e-07, + "loss": 0.7122, + "num_input_tokens_seen": 275631320, + "step": 8480 + }, + { + "epoch": 0.7648464625512919, + "flos": 22860878690880.0, + "grad_norm": 1.70033630803131, + "learning_rate": 5.524219779760284e-07, + "loss": 0.7409, + "num_input_tokens_seen": 275660700, + "step": 8481 + }, + { + "epoch": 0.7649366460747622, + "flos": 24609619414080.0, + "grad_norm": 1.9037285059255906, + "learning_rate": 5.520189219312907e-07, + "loss": 0.7646, + "num_input_tokens_seen": 275690625, + "step": 8482 + }, + { + "epoch": 0.7650268295982324, + "flos": 23151926610720.0, + "grad_norm": 2.2555143810871647, + "learning_rate": 5.516159894362817e-07, + "loss": 0.6302, + "num_input_tokens_seen": 275722165, + "step": 8483 + }, + { + "epoch": 0.7651170131217027, + "flos": 23330518805760.0, + "grad_norm": 1.7200999420538583, + "learning_rate": 5.512131805253839e-07, + "loss": 0.6796, + "num_input_tokens_seen": 275750625, + "step": 8484 + }, + { + "epoch": 0.7652071966451729, + "flos": 27632703016320.0, + "grad_norm": 1.6060205703928654, + "learning_rate": 5.508104952329653e-07, + "loss": 0.7581, + "num_input_tokens_seen": 275781940, + "step": 8485 + }, + { + "epoch": 0.7652973801686432, + "flos": 25775372223360.0, + "grad_norm": 2.664325510952858, + "learning_rate": 5.504079335933862e-07, + "loss": 0.7713, + "num_input_tokens_seen": 275808015, + "step": 8486 + }, + { + "epoch": 0.7653875636921135, + "flos": 16301725697760.0, + "grad_norm": 2.4539916821861913, + "learning_rate": 5.500054956409952e-07, + "loss": 0.7746, + "num_input_tokens_seen": 275834255, + "step": 8487 + }, + { + "epoch": 0.7654777472155837, + "flos": 21803976139680.0, + "grad_norm": 1.9997523869775509, + "learning_rate": 5.496031814101303e-07, + "loss": 0.7044, + "num_input_tokens_seen": 275859205, + "step": 8488 + }, + { + "epoch": 0.7655679307390539, + "flos": 20485358342880.0, + "grad_norm": 1.7610646321286865, + "learning_rate": 5.492009909351203e-07, + "loss": 0.7502, + "num_input_tokens_seen": 275887060, + "step": 8489 + }, + { + "epoch": 0.7656581142625243, + "flos": 23626027096800.0, + "grad_norm": 2.1821320201470007, + "learning_rate": 5.4879892425028e-07, + "loss": 0.6689, + "num_input_tokens_seen": 275916210, + "step": 8490 + }, + { + "epoch": 0.7657482977859945, + "flos": 21944760955200.0, + "grad_norm": 1.8412571234477864, + "learning_rate": 5.483969813899184e-07, + "loss": 0.8803, + "num_input_tokens_seen": 275944855, + "step": 8491 + }, + { + "epoch": 0.7658384813094647, + "flos": 32259988380480.0, + "grad_norm": 1.5109220864930915, + "learning_rate": 5.479951623883299e-07, + "loss": 0.8248, + "num_input_tokens_seen": 275976340, + "step": 8492 + }, + { + "epoch": 0.765928664832935, + "flos": 27336414160320.0, + "grad_norm": 2.3950286398431535, + "learning_rate": 5.475934672798004e-07, + "loss": 0.6786, + "num_input_tokens_seen": 276005245, + "step": 8493 + }, + { + "epoch": 0.7660188483564053, + "flos": 19833409056960.0, + "grad_norm": 1.904837601014749, + "learning_rate": 5.471918960986047e-07, + "loss": 0.7061, + "num_input_tokens_seen": 276032760, + "step": 8494 + }, + { + "epoch": 0.7661090318798756, + "flos": 26176013796480.0, + "grad_norm": 1.9540863112770035, + "learning_rate": 5.467904488790071e-07, + "loss": 0.8397, + "num_input_tokens_seen": 276060005, + "step": 8495 + }, + { + "epoch": 0.7661992154033458, + "flos": 26392450540800.0, + "grad_norm": 2.2372189009565244, + "learning_rate": 5.463891256552615e-07, + "loss": 0.7478, + "num_input_tokens_seen": 276089610, + "step": 8496 + }, + { + "epoch": 0.766289398926816, + "flos": 21620923573440.0, + "grad_norm": 1.9258006870612028, + "learning_rate": 5.459879264616107e-07, + "loss": 0.6511, + "num_input_tokens_seen": 276117330, + "step": 8497 + }, + { + "epoch": 0.7663795824502864, + "flos": 27706310608320.0, + "grad_norm": 2.225761189074105, + "learning_rate": 5.455868513322874e-07, + "loss": 0.6284, + "num_input_tokens_seen": 276145240, + "step": 8498 + }, + { + "epoch": 0.7664697659737566, + "flos": 24828212004480.0, + "grad_norm": 1.887899278257442, + "learning_rate": 5.451859003015143e-07, + "loss": 0.6323, + "num_input_tokens_seen": 276174005, + "step": 8499 + }, + { + "epoch": 0.7665599494972268, + "flos": 25808719231680.0, + "grad_norm": 1.8629387434769915, + "learning_rate": 5.447850734035009e-07, + "loss": 0.7805, + "num_input_tokens_seen": 276201610, + "step": 8500 + }, + { + "epoch": 0.7666501330206971, + "flos": 19509125638080.0, + "grad_norm": 9.517915797608797, + "learning_rate": 5.443843706724494e-07, + "loss": 0.7247, + "num_input_tokens_seen": 276229425, + "step": 8501 + }, + { + "epoch": 0.7667403165441674, + "flos": 22492134505440.0, + "grad_norm": 2.865417433015596, + "learning_rate": 5.439837921425494e-07, + "loss": 0.6748, + "num_input_tokens_seen": 276256905, + "step": 8502 + }, + { + "epoch": 0.7668305000676376, + "flos": 24932787740160.0, + "grad_norm": 2.3360354808046315, + "learning_rate": 5.435833378479807e-07, + "loss": 0.7351, + "num_input_tokens_seen": 276287320, + "step": 8503 + }, + { + "epoch": 0.7669206835911079, + "flos": 25884965876640.0, + "grad_norm": 1.609460328441843, + "learning_rate": 5.431830078229128e-07, + "loss": 0.6, + "num_input_tokens_seen": 276318110, + "step": 8504 + }, + { + "epoch": 0.7670108671145782, + "flos": 19978133867040.0, + "grad_norm": 3.2556045331804113, + "learning_rate": 5.427828021015022e-07, + "loss": 0.76, + "num_input_tokens_seen": 276346145, + "step": 8505 + }, + { + "epoch": 0.7671010506380485, + "flos": 29375756766240.0, + "grad_norm": 1.9045364933378321, + "learning_rate": 5.42382720717899e-07, + "loss": 0.773, + "num_input_tokens_seen": 276374490, + "step": 8506 + }, + { + "epoch": 0.7671912341615187, + "flos": 24646125852000.0, + "grad_norm": 1.6988517277161896, + "learning_rate": 5.419827637062384e-07, + "loss": 0.7741, + "num_input_tokens_seen": 276404295, + "step": 8507 + }, + { + "epoch": 0.7672814176849889, + "flos": 17572611789120.0, + "grad_norm": 2.6691178937067153, + "learning_rate": 5.415829311006487e-07, + "loss": 0.8852, + "num_input_tokens_seen": 276428910, + "step": 8508 + }, + { + "epoch": 0.7673716012084593, + "flos": 22018926093600.0, + "grad_norm": 1.9809294697059723, + "learning_rate": 5.411832229352447e-07, + "loss": 0.8256, + "num_input_tokens_seen": 276455130, + "step": 8509 + }, + { + "epoch": 0.7674617847319295, + "flos": 22059186677280.0, + "grad_norm": 1.9000626503503373, + "learning_rate": 5.407836392441319e-07, + "loss": 0.8396, + "num_input_tokens_seen": 276482650, + "step": 8510 + }, + { + "epoch": 0.7675519682553997, + "flos": 27815532564000.0, + "grad_norm": 1.6787668233271356, + "learning_rate": 5.403841800614049e-07, + "loss": 0.7612, + "num_input_tokens_seen": 276513440, + "step": 8511 + }, + { + "epoch": 0.76764215177887, + "flos": 22206699219360.0, + "grad_norm": 2.8942381675697795, + "learning_rate": 5.39984845421148e-07, + "loss": 0.7005, + "num_input_tokens_seen": 276545465, + "step": 8512 + }, + { + "epoch": 0.7677323353023403, + "flos": 25918981940640.0, + "grad_norm": 2.609806322095354, + "learning_rate": 5.395856353574344e-07, + "loss": 0.7288, + "num_input_tokens_seen": 276574460, + "step": 8513 + }, + { + "epoch": 0.7678225188258105, + "flos": 70055456968320.0, + "grad_norm": 0.5835035507172566, + "learning_rate": 5.391865499043275e-07, + "loss": 0.566, + "num_input_tokens_seen": 276677965, + "step": 8514 + }, + { + "epoch": 0.7679127023492808, + "flos": 24245707297440.0, + "grad_norm": 1.888214718739943, + "learning_rate": 5.387875890958788e-07, + "loss": 0.7483, + "num_input_tokens_seen": 276706830, + "step": 8515 + }, + { + "epoch": 0.768002885872751, + "flos": 25156769945760.0, + "grad_norm": 1.66704896674868, + "learning_rate": 5.383887529661298e-07, + "loss": 0.7688, + "num_input_tokens_seen": 276736450, + "step": 8516 + }, + { + "epoch": 0.7680930693962214, + "flos": 43335271954560.0, + "grad_norm": 2.6596760813598275, + "learning_rate": 5.379900415491116e-07, + "loss": 0.6981, + "num_input_tokens_seen": 276766860, + "step": 8517 + }, + { + "epoch": 0.7681832529196916, + "flos": 26722272253920.0, + "grad_norm": 2.0135387890974923, + "learning_rate": 5.375914548788447e-07, + "loss": 0.6958, + "num_input_tokens_seen": 276795240, + "step": 8518 + }, + { + "epoch": 0.7682734364431618, + "flos": 20635881635520.0, + "grad_norm": 2.1136664922586124, + "learning_rate": 5.371929929893384e-07, + "loss": 0.6266, + "num_input_tokens_seen": 276822565, + "step": 8519 + }, + { + "epoch": 0.768363619966632, + "flos": 38453779636320.0, + "grad_norm": 2.0384278460737044, + "learning_rate": 5.367946559145917e-07, + "loss": 0.7077, + "num_input_tokens_seen": 276852600, + "step": 8520 + }, + { + "epoch": 0.7684538034901024, + "flos": 23590041035520.0, + "grad_norm": 2.109807740080541, + "learning_rate": 5.363964436885935e-07, + "loss": 0.7616, + "num_input_tokens_seen": 276883285, + "step": 8521 + }, + { + "epoch": 0.7685439870135726, + "flos": 21658322085600.0, + "grad_norm": 1.5292317850382315, + "learning_rate": 5.359983563453199e-07, + "loss": 0.8264, + "num_input_tokens_seen": 276914830, + "step": 8522 + }, + { + "epoch": 0.7686341705370429, + "flos": 19618198914720.0, + "grad_norm": 2.074675831119102, + "learning_rate": 5.356003939187402e-07, + "loss": 0.8587, + "num_input_tokens_seen": 276943755, + "step": 8523 + }, + { + "epoch": 0.7687243540605131, + "flos": 27013766210880.0, + "grad_norm": 2.093557092690334, + "learning_rate": 5.352025564428082e-07, + "loss": 0.7215, + "num_input_tokens_seen": 276971420, + "step": 8524 + }, + { + "epoch": 0.7688145375839834, + "flos": 26028092387040.0, + "grad_norm": 1.9561939389907048, + "learning_rate": 5.348048439514723e-07, + "loss": 0.7821, + "num_input_tokens_seen": 277002200, + "step": 8525 + }, + { + "epoch": 0.7689047211074537, + "flos": 26826885159360.0, + "grad_norm": 2.2274966773148575, + "learning_rate": 5.344072564786653e-07, + "loss": 0.6233, + "num_input_tokens_seen": 277031790, + "step": 8526 + }, + { + "epoch": 0.7689949046309239, + "flos": 25445513340480.0, + "grad_norm": 1.6839604166548066, + "learning_rate": 5.340097940583123e-07, + "loss": 0.7563, + "num_input_tokens_seen": 277062610, + "step": 8527 + }, + { + "epoch": 0.7690850881543942, + "flos": 17868826305600.0, + "grad_norm": 2.8553780725278313, + "learning_rate": 5.336124567243275e-07, + "loss": 0.7063, + "num_input_tokens_seen": 277086645, + "step": 8528 + }, + { + "epoch": 0.7691752716778645, + "flos": 18416460044160.0, + "grad_norm": 2.230100156989609, + "learning_rate": 5.33215244510613e-07, + "loss": 0.7494, + "num_input_tokens_seen": 277113745, + "step": 8529 + }, + { + "epoch": 0.7692654552013347, + "flos": 14444506414080.0, + "grad_norm": 2.1239542912595515, + "learning_rate": 5.328181574510624e-07, + "loss": 0.7382, + "num_input_tokens_seen": 277139245, + "step": 8530 + }, + { + "epoch": 0.769355638724805, + "flos": 27519503896320.0, + "grad_norm": 2.339410553890561, + "learning_rate": 5.324211955795559e-07, + "loss": 0.7984, + "num_input_tokens_seen": 277167410, + "step": 8531 + }, + { + "epoch": 0.7694458222482753, + "flos": 23079396941760.0, + "grad_norm": 1.8833207833313623, + "learning_rate": 5.320243589299651e-07, + "loss": 0.6704, + "num_input_tokens_seen": 277196295, + "step": 8532 + }, + { + "epoch": 0.7695360057717455, + "flos": 23516842310880.0, + "grad_norm": 2.2597451567650166, + "learning_rate": 5.316276475361505e-07, + "loss": 0.6965, + "num_input_tokens_seen": 277224380, + "step": 8533 + }, + { + "epoch": 0.7696261892952158, + "flos": 20856741581280.0, + "grad_norm": 2.664183583122848, + "learning_rate": 5.312310614319613e-07, + "loss": 0.7811, + "num_input_tokens_seen": 277249475, + "step": 8534 + }, + { + "epoch": 0.769716372818686, + "flos": 19071457250400.0, + "grad_norm": 2.322853315598588, + "learning_rate": 5.308346006512367e-07, + "loss": 0.7134, + "num_input_tokens_seen": 277272540, + "step": 8535 + }, + { + "epoch": 0.7698065563421563, + "flos": 23476024180800.0, + "grad_norm": 1.8753431006275654, + "learning_rate": 5.30438265227805e-07, + "loss": 0.6752, + "num_input_tokens_seen": 277302260, + "step": 8536 + }, + { + "epoch": 0.7698967398656266, + "flos": 70555544850240.0, + "grad_norm": 0.602712399174197, + "learning_rate": 5.300420551954837e-07, + "loss": 0.559, + "num_input_tokens_seen": 277405985, + "step": 8537 + }, + { + "epoch": 0.7699869233890968, + "flos": 30001755826080.0, + "grad_norm": 1.5938132778396346, + "learning_rate": 5.296459705880798e-07, + "loss": 0.6689, + "num_input_tokens_seen": 277436180, + "step": 8538 + }, + { + "epoch": 0.770077106912567, + "flos": 23698779784320.0, + "grad_norm": 1.6793372920253284, + "learning_rate": 5.292500114393881e-07, + "loss": 0.7983, + "num_input_tokens_seen": 277466625, + "step": 8539 + }, + { + "epoch": 0.7701672904360374, + "flos": 20418961684320.0, + "grad_norm": 1.901392955634808, + "learning_rate": 5.288541777831963e-07, + "loss": 0.7638, + "num_input_tokens_seen": 277492220, + "step": 8540 + }, + { + "epoch": 0.7702574739595076, + "flos": 25119557282400.0, + "grad_norm": 1.7845398487698543, + "learning_rate": 5.284584696532772e-07, + "loss": 0.7981, + "num_input_tokens_seen": 277524095, + "step": 8541 + }, + { + "epoch": 0.7703476574829778, + "flos": 18524232379200.0, + "grad_norm": 1.9536663068245894, + "learning_rate": 5.280628870833954e-07, + "loss": 0.7503, + "num_input_tokens_seen": 277552445, + "step": 8542 + }, + { + "epoch": 0.7704378410064481, + "flos": 22350606294720.0, + "grad_norm": 1.944356537905685, + "learning_rate": 5.276674301073045e-07, + "loss": 0.7739, + "num_input_tokens_seen": 277580825, + "step": 8543 + }, + { + "epoch": 0.7705280245299184, + "flos": 21618804897120.0, + "grad_norm": 1.530021641896091, + "learning_rate": 5.272720987587467e-07, + "loss": 0.7576, + "num_input_tokens_seen": 277610405, + "step": 8544 + }, + { + "epoch": 0.7706182080533887, + "flos": 22023200616000.0, + "grad_norm": 2.100565961231142, + "learning_rate": 5.268768930714545e-07, + "loss": 0.6956, + "num_input_tokens_seen": 277637990, + "step": 8545 + }, + { + "epoch": 0.7707083915768589, + "flos": 25957644224640.0, + "grad_norm": 3.3647179985063618, + "learning_rate": 5.264818130791473e-07, + "loss": 0.7732, + "num_input_tokens_seen": 277668610, + "step": 8546 + }, + { + "epoch": 0.7707985751003291, + "flos": 16703928400800.0, + "grad_norm": 2.2662779820729866, + "learning_rate": 5.260868588155378e-07, + "loss": 0.7467, + "num_input_tokens_seen": 277695220, + "step": 8547 + }, + { + "epoch": 0.7708887586237995, + "flos": 23443717925760.0, + "grad_norm": 1.8686918319784116, + "learning_rate": 5.256920303143242e-07, + "loss": 0.7727, + "num_input_tokens_seen": 277725340, + "step": 8548 + }, + { + "epoch": 0.7709789421472697, + "flos": 22820878295520.0, + "grad_norm": 1.9386895742536565, + "learning_rate": 5.252973276091956e-07, + "loss": 0.8571, + "num_input_tokens_seen": 277755680, + "step": 8549 + }, + { + "epoch": 0.7710691256707399, + "flos": 21801783123840.0, + "grad_norm": 1.8705152282217323, + "learning_rate": 5.249027507338307e-07, + "loss": 0.8134, + "num_input_tokens_seen": 277784630, + "step": 8550 + }, + { + "epoch": 0.7711593091942102, + "flos": 22060376109600.0, + "grad_norm": 1.61599090092524, + "learning_rate": 5.245082997218966e-07, + "loss": 0.7164, + "num_input_tokens_seen": 277814835, + "step": 8551 + }, + { + "epoch": 0.7712494927176805, + "flos": 23475615313440.0, + "grad_norm": 2.9248522924965568, + "learning_rate": 5.241139746070499e-07, + "loss": 0.7677, + "num_input_tokens_seen": 277842840, + "step": 8552 + }, + { + "epoch": 0.7713396762411507, + "flos": 20638706537280.0, + "grad_norm": 1.72620633279317, + "learning_rate": 5.237197754229376e-07, + "loss": 0.6269, + "num_input_tokens_seen": 277872020, + "step": 8553 + }, + { + "epoch": 0.771429859764621, + "flos": 35937437303040.0, + "grad_norm": 2.1206166554541745, + "learning_rate": 5.233257022031931e-07, + "loss": 0.6541, + "num_input_tokens_seen": 277906915, + "step": 8554 + }, + { + "epoch": 0.7715200432880913, + "flos": 19472544860640.0, + "grad_norm": 1.6936808526257676, + "learning_rate": 5.229317549814432e-07, + "loss": 0.755, + "num_input_tokens_seen": 277933660, + "step": 8555 + }, + { + "epoch": 0.7716102268115616, + "flos": 20164457372160.0, + "grad_norm": 2.091070139755293, + "learning_rate": 5.225379337912998e-07, + "loss": 0.826, + "num_input_tokens_seen": 277960325, + "step": 8556 + }, + { + "epoch": 0.7717004103350318, + "flos": 23516544952800.0, + "grad_norm": 19.812244305329532, + "learning_rate": 5.221442386663663e-07, + "loss": 0.7606, + "num_input_tokens_seen": 277988860, + "step": 8557 + }, + { + "epoch": 0.771790593858502, + "flos": 26467805111520.0, + "grad_norm": 1.989347820090317, + "learning_rate": 5.217506696402354e-07, + "loss": 0.7832, + "num_input_tokens_seen": 278018745, + "step": 8558 + }, + { + "epoch": 0.7718807773819724, + "flos": 20711124696960.0, + "grad_norm": 2.7122492625991224, + "learning_rate": 5.213572267464883e-07, + "loss": 0.7589, + "num_input_tokens_seen": 278046065, + "step": 8559 + }, + { + "epoch": 0.7719709609054426, + "flos": 16700620292160.0, + "grad_norm": 1.9205783718915286, + "learning_rate": 5.209639100186965e-07, + "loss": 0.7252, + "num_input_tokens_seen": 278070655, + "step": 8560 + }, + { + "epoch": 0.7720611444289128, + "flos": 67822208226240.0, + "grad_norm": 0.6630264502739643, + "learning_rate": 5.205707194904179e-07, + "loss": 0.6401, + "num_input_tokens_seen": 278159715, + "step": 8561 + }, + { + "epoch": 0.7721513279523831, + "flos": 26026717105920.0, + "grad_norm": 2.0289396792225807, + "learning_rate": 5.201776551952042e-07, + "loss": 0.6707, + "num_input_tokens_seen": 278187320, + "step": 8562 + }, + { + "epoch": 0.7722415114758534, + "flos": 30395483823840.0, + "grad_norm": 3.111289043863545, + "learning_rate": 5.197847171665914e-07, + "loss": 0.7082, + "num_input_tokens_seen": 278213395, + "step": 8563 + }, + { + "epoch": 0.7723316949993236, + "flos": 20964253728000.0, + "grad_norm": 1.4929872935853719, + "learning_rate": 5.193919054381095e-07, + "loss": 0.7305, + "num_input_tokens_seen": 278241300, + "step": 8564 + }, + { + "epoch": 0.7724218785227939, + "flos": 20524949870880.0, + "grad_norm": 1.9678056056419977, + "learning_rate": 5.189992200432738e-07, + "loss": 0.7091, + "num_input_tokens_seen": 278269365, + "step": 8565 + }, + { + "epoch": 0.7725120620462641, + "flos": 15640595481120.0, + "grad_norm": 2.889550746595546, + "learning_rate": 5.186066610155906e-07, + "loss": 0.7583, + "num_input_tokens_seen": 278294735, + "step": 8566 + }, + { + "epoch": 0.7726022455697344, + "flos": 23297655004320.0, + "grad_norm": 2.267059608601286, + "learning_rate": 5.182142283885555e-07, + "loss": 0.7992, + "num_input_tokens_seen": 278322490, + "step": 8567 + }, + { + "epoch": 0.7726924290932047, + "flos": 62982983658720.0, + "grad_norm": 0.6539515936476762, + "learning_rate": 5.178219221956528e-07, + "loss": 0.5303, + "num_input_tokens_seen": 278404685, + "step": 8568 + }, + { + "epoch": 0.7727826126166749, + "flos": 21076040397120.0, + "grad_norm": 2.250568635929168, + "learning_rate": 5.174297424703565e-07, + "loss": 0.7648, + "num_input_tokens_seen": 278433195, + "step": 8569 + }, + { + "epoch": 0.7728727961401451, + "flos": 26828706477600.0, + "grad_norm": 1.8320939758884685, + "learning_rate": 5.170376892461299e-07, + "loss": 0.6782, + "num_input_tokens_seen": 278463925, + "step": 8570 + }, + { + "epoch": 0.7729629796636155, + "flos": 66574521798720.0, + "grad_norm": 0.969200121995601, + "learning_rate": 5.16645762556424e-07, + "loss": 0.6302, + "num_input_tokens_seen": 278549750, + "step": 8571 + }, + { + "epoch": 0.7730531631870857, + "flos": 41803153824480.0, + "grad_norm": 1.9612507968801287, + "learning_rate": 5.162539624346809e-07, + "loss": 0.7923, + "num_input_tokens_seen": 278581280, + "step": 8572 + }, + { + "epoch": 0.773143346710556, + "flos": 23914733321760.0, + "grad_norm": 1.9447551986089353, + "learning_rate": 5.158622889143309e-07, + "loss": 0.8118, + "num_input_tokens_seen": 278608115, + "step": 8573 + }, + { + "epoch": 0.7732335302340262, + "flos": 23480038514880.0, + "grad_norm": 2.4739408365266695, + "learning_rate": 5.154707420287939e-07, + "loss": 0.7047, + "num_input_tokens_seen": 278635350, + "step": 8574 + }, + { + "epoch": 0.7733237137574965, + "flos": 27377009271840.0, + "grad_norm": 1.8669595619158057, + "learning_rate": 5.150793218114793e-07, + "loss": 0.6726, + "num_input_tokens_seen": 278665945, + "step": 8575 + }, + { + "epoch": 0.7734138972809668, + "flos": 23990348080800.0, + "grad_norm": 1.7764373144701227, + "learning_rate": 5.146880282957837e-07, + "loss": 0.7323, + "num_input_tokens_seen": 278693670, + "step": 8576 + }, + { + "epoch": 0.773504080804437, + "flos": 25265434355040.0, + "grad_norm": 2.220378117192404, + "learning_rate": 5.142968615150964e-07, + "loss": 0.7324, + "num_input_tokens_seen": 278720695, + "step": 8577 + }, + { + "epoch": 0.7735942643279073, + "flos": 29017159925280.0, + "grad_norm": 2.0430567414278653, + "learning_rate": 5.139058215027921e-07, + "loss": 0.6753, + "num_input_tokens_seen": 278750695, + "step": 8578 + }, + { + "epoch": 0.7736844478513776, + "flos": 26756771524800.0, + "grad_norm": 1.4951005233344272, + "learning_rate": 5.135149082922383e-07, + "loss": 0.7335, + "num_input_tokens_seen": 278780800, + "step": 8579 + }, + { + "epoch": 0.7737746313748478, + "flos": 33746976688320.0, + "grad_norm": 1.9706841913075537, + "learning_rate": 5.131241219167879e-07, + "loss": 0.6728, + "num_input_tokens_seen": 278809940, + "step": 8580 + }, + { + "epoch": 0.773864814898318, + "flos": 59032928750880.0, + "grad_norm": 0.7776721279921592, + "learning_rate": 5.127334624097869e-07, + "loss": 0.6214, + "num_input_tokens_seen": 278897945, + "step": 8581 + }, + { + "epoch": 0.7739549984217884, + "flos": 31563950025600.0, + "grad_norm": 1.550655715435212, + "learning_rate": 5.123429298045672e-07, + "loss": 0.7994, + "num_input_tokens_seen": 278927730, + "step": 8582 + }, + { + "epoch": 0.7740451819452586, + "flos": 35794310792640.0, + "grad_norm": 2.072029409509153, + "learning_rate": 5.119525241344515e-07, + "loss": 0.6341, + "num_input_tokens_seen": 278958445, + "step": 8583 + }, + { + "epoch": 0.7741353654687289, + "flos": 25010707024320.0, + "grad_norm": 5.430220543473058, + "learning_rate": 5.115622454327515e-07, + "loss": 0.7539, + "num_input_tokens_seen": 278985950, + "step": 8584 + }, + { + "epoch": 0.7742255489921991, + "flos": 21253443159840.0, + "grad_norm": 4.132974718800626, + "learning_rate": 5.11172093732768e-07, + "loss": 0.7306, + "num_input_tokens_seen": 279011560, + "step": 8585 + }, + { + "epoch": 0.7743157325156694, + "flos": 23953284096480.0, + "grad_norm": 2.28540225538812, + "learning_rate": 5.107820690677911e-07, + "loss": 0.7722, + "num_input_tokens_seen": 279041560, + "step": 8586 + }, + { + "epoch": 0.7744059160391397, + "flos": 20747408116320.0, + "grad_norm": 1.842401884512656, + "learning_rate": 5.103921714710991e-07, + "loss": 0.6928, + "num_input_tokens_seen": 279069945, + "step": 8587 + }, + { + "epoch": 0.7744960995626099, + "flos": 19728052756320.0, + "grad_norm": 1.8679433912648344, + "learning_rate": 5.100024009759605e-07, + "loss": 0.7601, + "num_input_tokens_seen": 279097330, + "step": 8588 + }, + { + "epoch": 0.7745862830860801, + "flos": 16666790076960.0, + "grad_norm": 2.7410609873931824, + "learning_rate": 5.09612757615633e-07, + "loss": 0.7843, + "num_input_tokens_seen": 279123110, + "step": 8589 + }, + { + "epoch": 0.7746764666095505, + "flos": 21768138757440.0, + "grad_norm": 1.6504687048144535, + "learning_rate": 5.092232414233628e-07, + "loss": 0.6987, + "num_input_tokens_seen": 279151425, + "step": 8590 + }, + { + "epoch": 0.7747666501330207, + "flos": 17102265448800.0, + "grad_norm": 1.9489566114517618, + "learning_rate": 5.088338524323858e-07, + "loss": 0.7111, + "num_input_tokens_seen": 279175780, + "step": 8591 + }, + { + "epoch": 0.7748568336564909, + "flos": 19983226124160.0, + "grad_norm": 2.9668640571104743, + "learning_rate": 5.084445906759271e-07, + "loss": 0.8017, + "num_input_tokens_seen": 279203790, + "step": 8592 + }, + { + "epoch": 0.7749470171799612, + "flos": 20053674286560.0, + "grad_norm": 1.7978376380758303, + "learning_rate": 5.080554561871995e-07, + "loss": 0.8077, + "num_input_tokens_seen": 279231540, + "step": 8593 + }, + { + "epoch": 0.7750372007034315, + "flos": 22678569519840.0, + "grad_norm": 1.68927029188162, + "learning_rate": 5.076664489994078e-07, + "loss": 0.7725, + "num_input_tokens_seen": 279260830, + "step": 8594 + }, + { + "epoch": 0.7751273842269017, + "flos": 21105075713280.0, + "grad_norm": 2.1348693766602236, + "learning_rate": 5.07277569145742e-07, + "loss": 0.7135, + "num_input_tokens_seen": 279287180, + "step": 8595 + }, + { + "epoch": 0.775217567750372, + "flos": 20601977080800.0, + "grad_norm": 2.979737838375923, + "learning_rate": 5.068888166593861e-07, + "loss": 0.6599, + "num_input_tokens_seen": 279314905, + "step": 8596 + }, + { + "epoch": 0.7753077512738422, + "flos": 33717644014080.0, + "grad_norm": 2.0125114582754406, + "learning_rate": 5.065001915735087e-07, + "loss": 0.7319, + "num_input_tokens_seen": 279343385, + "step": 8597 + }, + { + "epoch": 0.7753979347973126, + "flos": 22933668548160.0, + "grad_norm": 2.1126126908260323, + "learning_rate": 5.061116939212702e-07, + "loss": 0.7122, + "num_input_tokens_seen": 279371170, + "step": 8598 + }, + { + "epoch": 0.7754881183207828, + "flos": 29855655734880.0, + "grad_norm": 2.2402415055125733, + "learning_rate": 5.05723323735819e-07, + "loss": 0.7434, + "num_input_tokens_seen": 279399370, + "step": 8599 + }, + { + "epoch": 0.775578301844253, + "flos": 61812919157280.0, + "grad_norm": 0.6272170560037615, + "learning_rate": 5.053350810502932e-07, + "loss": 0.5823, + "num_input_tokens_seen": 279487675, + "step": 8600 + }, + { + "epoch": 0.7756684853677234, + "flos": 22897868335680.0, + "grad_norm": 1.7114082281107683, + "learning_rate": 5.049469658978202e-07, + "loss": 0.7398, + "num_input_tokens_seen": 279517695, + "step": 8601 + }, + { + "epoch": 0.7757586688911936, + "flos": 68811636195840.0, + "grad_norm": 0.6280614023608057, + "learning_rate": 5.045589783115147e-07, + "loss": 0.5313, + "num_input_tokens_seen": 279617875, + "step": 8602 + }, + { + "epoch": 0.7758488524146638, + "flos": 20128173952800.0, + "grad_norm": 1.9808518981491583, + "learning_rate": 5.041711183244842e-07, + "loss": 0.7819, + "num_input_tokens_seen": 279644335, + "step": 8603 + }, + { + "epoch": 0.7759390359381341, + "flos": 21184258769280.0, + "grad_norm": 2.641199805545063, + "learning_rate": 5.037833859698211e-07, + "loss": 0.7636, + "num_input_tokens_seen": 279671670, + "step": 8604 + }, + { + "epoch": 0.7760292194616044, + "flos": 27087187954080.0, + "grad_norm": 2.2601132401647313, + "learning_rate": 5.033957812806096e-07, + "loss": 0.8107, + "num_input_tokens_seen": 279700390, + "step": 8605 + }, + { + "epoch": 0.7761194029850746, + "flos": 26685765816000.0, + "grad_norm": 2.2927622401549557, + "learning_rate": 5.030083042899223e-07, + "loss": 0.7956, + "num_input_tokens_seen": 279726670, + "step": 8606 + }, + { + "epoch": 0.7762095865085449, + "flos": 24025962444480.0, + "grad_norm": 2.3576219121452535, + "learning_rate": 5.026209550308207e-07, + "loss": 0.743, + "num_input_tokens_seen": 279754530, + "step": 8607 + }, + { + "epoch": 0.7762997700320151, + "flos": 23626361624640.0, + "grad_norm": 2.1159219372849836, + "learning_rate": 5.022337335363558e-07, + "loss": 0.8001, + "num_input_tokens_seen": 279783600, + "step": 8608 + }, + { + "epoch": 0.7763899535554855, + "flos": 22637342522400.0, + "grad_norm": 1.8249028820168307, + "learning_rate": 5.018466398395677e-07, + "loss": 0.7696, + "num_input_tokens_seen": 279809775, + "step": 8609 + }, + { + "epoch": 0.7764801370789557, + "flos": 25156435417920.0, + "grad_norm": 3.928138746239062, + "learning_rate": 5.01459673973484e-07, + "loss": 0.6774, + "num_input_tokens_seen": 279841130, + "step": 8610 + }, + { + "epoch": 0.7765703206024259, + "flos": 20237730436320.0, + "grad_norm": 2.417264919113846, + "learning_rate": 5.01072835971125e-07, + "loss": 0.8076, + "num_input_tokens_seen": 279866525, + "step": 8611 + }, + { + "epoch": 0.7766605041258962, + "flos": 30507047474400.0, + "grad_norm": 1.7666104697640288, + "learning_rate": 5.006861258654959e-07, + "loss": 0.6499, + "num_input_tokens_seen": 279898530, + "step": 8612 + }, + { + "epoch": 0.7767506876493665, + "flos": 20237841945600.0, + "grad_norm": 5.421812557333603, + "learning_rate": 5.002995436895938e-07, + "loss": 0.7729, + "num_input_tokens_seen": 279927975, + "step": 8613 + }, + { + "epoch": 0.7768408711728367, + "flos": 23625841248000.0, + "grad_norm": 1.8039636522678646, + "learning_rate": 4.999130894764039e-07, + "loss": 0.7023, + "num_input_tokens_seen": 279956700, + "step": 8614 + }, + { + "epoch": 0.776931054696307, + "flos": 23370593540640.0, + "grad_norm": 1.8036452417228566, + "learning_rate": 4.995267632589006e-07, + "loss": 0.6855, + "num_input_tokens_seen": 279985010, + "step": 8615 + }, + { + "epoch": 0.7770212382197772, + "flos": 59391116724480.0, + "grad_norm": 0.6843413706141976, + "learning_rate": 4.99140565070048e-07, + "loss": 0.5802, + "num_input_tokens_seen": 280074430, + "step": 8616 + }, + { + "epoch": 0.7771114217432475, + "flos": 18744014401920.0, + "grad_norm": 1.7840809995754858, + "learning_rate": 4.987544949427969e-07, + "loss": 0.7654, + "num_input_tokens_seen": 280101975, + "step": 8617 + }, + { + "epoch": 0.7772016052667178, + "flos": 67441978584960.0, + "grad_norm": 0.6649810048667348, + "learning_rate": 4.98368552910091e-07, + "loss": 0.5871, + "num_input_tokens_seen": 280190955, + "step": 8618 + }, + { + "epoch": 0.777291788790188, + "flos": 22782922236960.0, + "grad_norm": 2.2646255919705274, + "learning_rate": 4.979827390048596e-07, + "loss": 0.8588, + "num_input_tokens_seen": 280216875, + "step": 8619 + }, + { + "epoch": 0.7773819723136582, + "flos": 20055421265280.0, + "grad_norm": 1.5007039643483453, + "learning_rate": 4.975970532600231e-07, + "loss": 0.7085, + "num_input_tokens_seen": 280246535, + "step": 8620 + }, + { + "epoch": 0.7774721558371286, + "flos": 17797374559680.0, + "grad_norm": 1.992598547323238, + "learning_rate": 4.972114957084901e-07, + "loss": 0.7658, + "num_input_tokens_seen": 280274105, + "step": 8621 + }, + { + "epoch": 0.7775623393605988, + "flos": 22932516285600.0, + "grad_norm": 1.8580057568209418, + "learning_rate": 4.968260663831585e-07, + "loss": 0.7245, + "num_input_tokens_seen": 280304245, + "step": 8622 + }, + { + "epoch": 0.777652522884069, + "flos": 23515652878560.0, + "grad_norm": 1.6255825905729318, + "learning_rate": 4.964407653169154e-07, + "loss": 0.8549, + "num_input_tokens_seen": 280334060, + "step": 8623 + }, + { + "epoch": 0.7777427064075394, + "flos": 23917446714240.0, + "grad_norm": 1.5747226258917737, + "learning_rate": 4.960555925426366e-07, + "loss": 0.7124, + "num_input_tokens_seen": 280362475, + "step": 8624 + }, + { + "epoch": 0.7778328899310096, + "flos": 15137125151040.0, + "grad_norm": 2.1857473709280866, + "learning_rate": 4.956705480931876e-07, + "loss": 0.8058, + "num_input_tokens_seen": 280387625, + "step": 8625 + }, + { + "epoch": 0.7779230734544799, + "flos": 25953221023200.0, + "grad_norm": 1.5467354297287714, + "learning_rate": 4.952856320014225e-07, + "loss": 0.7246, + "num_input_tokens_seen": 280417930, + "step": 8626 + }, + { + "epoch": 0.7780132569779501, + "flos": 38379800346720.0, + "grad_norm": 1.5232525835539663, + "learning_rate": 4.949008443001838e-07, + "loss": 0.6803, + "num_input_tokens_seen": 280453610, + "step": 8627 + }, + { + "epoch": 0.7781034405014204, + "flos": 19834635659040.0, + "grad_norm": 2.052788034131564, + "learning_rate": 4.945161850223041e-07, + "loss": 0.758, + "num_input_tokens_seen": 280482380, + "step": 8628 + }, + { + "epoch": 0.7781936240248907, + "flos": 32988147141600.0, + "grad_norm": 2.848556815707212, + "learning_rate": 4.941316542006044e-07, + "loss": 0.6774, + "num_input_tokens_seen": 280510035, + "step": 8629 + }, + { + "epoch": 0.7782838075483609, + "flos": 18927215647200.0, + "grad_norm": 1.56363705270623, + "learning_rate": 4.937472518678956e-07, + "loss": 0.7697, + "num_input_tokens_seen": 280538695, + "step": 8630 + }, + { + "epoch": 0.7783739910718311, + "flos": 19685859345120.0, + "grad_norm": 2.072498362394452, + "learning_rate": 4.93362978056977e-07, + "loss": 0.7434, + "num_input_tokens_seen": 280566510, + "step": 8631 + }, + { + "epoch": 0.7784641745953015, + "flos": 32078831472000.0, + "grad_norm": 2.154035343124604, + "learning_rate": 4.929788328006355e-07, + "loss": 0.6957, + "num_input_tokens_seen": 280601015, + "step": 8632 + }, + { + "epoch": 0.7785543581187717, + "flos": 19618050235680.0, + "grad_norm": 2.0961399387498596, + "learning_rate": 4.925948161316506e-07, + "loss": 0.8171, + "num_input_tokens_seen": 280626440, + "step": 8633 + }, + { + "epoch": 0.778644541642242, + "flos": 23550486677280.0, + "grad_norm": 1.9329369248204848, + "learning_rate": 4.922109280827868e-07, + "loss": 0.7326, + "num_input_tokens_seen": 280655345, + "step": 8634 + }, + { + "epoch": 0.7787347251657122, + "flos": 21950076230880.0, + "grad_norm": 1.9026603670240947, + "learning_rate": 4.918271686868016e-07, + "loss": 0.7377, + "num_input_tokens_seen": 280683810, + "step": 8635 + }, + { + "epoch": 0.7788249086891825, + "flos": 20638260500160.0, + "grad_norm": 8.738558018359457, + "learning_rate": 4.914435379764379e-07, + "loss": 0.7871, + "num_input_tokens_seen": 280712225, + "step": 8636 + }, + { + "epoch": 0.7789150922126528, + "flos": 27157933474560.0, + "grad_norm": 1.9985432684067588, + "learning_rate": 4.910600359844294e-07, + "loss": 0.7361, + "num_input_tokens_seen": 280742455, + "step": 8637 + }, + { + "epoch": 0.779005275736123, + "flos": 24974572284000.0, + "grad_norm": 3.1093368144542035, + "learning_rate": 4.90676662743499e-07, + "loss": 0.6533, + "num_input_tokens_seen": 280769840, + "step": 8638 + }, + { + "epoch": 0.7790954592595932, + "flos": 22314917591520.0, + "grad_norm": 2.4937453229281896, + "learning_rate": 4.902934182863581e-07, + "loss": 0.7385, + "num_input_tokens_seen": 280797410, + "step": 8639 + }, + { + "epoch": 0.7791856427830636, + "flos": 23224567788960.0, + "grad_norm": 2.0349696214737567, + "learning_rate": 4.899103026457069e-07, + "loss": 0.6159, + "num_input_tokens_seen": 280826725, + "step": 8640 + }, + { + "epoch": 0.7792758263065338, + "flos": 66005509714560.0, + "grad_norm": 0.5866647911403038, + "learning_rate": 4.895273158542361e-07, + "loss": 0.5295, + "num_input_tokens_seen": 280918770, + "step": 8641 + }, + { + "epoch": 0.779366009830004, + "flos": 23915551056480.0, + "grad_norm": 2.024254357846669, + "learning_rate": 4.891444579446227e-07, + "loss": 0.76, + "num_input_tokens_seen": 280948760, + "step": 8642 + }, + { + "epoch": 0.7794561933534743, + "flos": 23917112186400.0, + "grad_norm": 1.973755362215302, + "learning_rate": 4.887617289495349e-07, + "loss": 0.701, + "num_input_tokens_seen": 280977120, + "step": 8643 + }, + { + "epoch": 0.7795463768769446, + "flos": 62116976493120.0, + "grad_norm": 0.5891150716701405, + "learning_rate": 4.883791289016292e-07, + "loss": 0.5101, + "num_input_tokens_seen": 281062655, + "step": 8644 + }, + { + "epoch": 0.7796365604004148, + "flos": 19728127095840.0, + "grad_norm": 2.0011029778347664, + "learning_rate": 4.879966578335514e-07, + "loss": 0.754, + "num_input_tokens_seen": 281092455, + "step": 8645 + }, + { + "epoch": 0.7797267439238851, + "flos": 23730416983680.0, + "grad_norm": 3.1167746057705665, + "learning_rate": 4.876143157779358e-07, + "loss": 0.6411, + "num_input_tokens_seen": 281120155, + "step": 8646 + }, + { + "epoch": 0.7798169274473554, + "flos": 34767632989920.0, + "grad_norm": 2.3820474365725355, + "learning_rate": 4.872321027674058e-07, + "loss": 0.631, + "num_input_tokens_seen": 281152285, + "step": 8647 + }, + { + "epoch": 0.7799071109708257, + "flos": 23625989927040.0, + "grad_norm": 1.7085764719678582, + "learning_rate": 4.868500188345748e-07, + "loss": 0.8148, + "num_input_tokens_seen": 281178110, + "step": 8648 + }, + { + "epoch": 0.7799972944942959, + "flos": 21658582273920.0, + "grad_norm": 2.007624635924747, + "learning_rate": 4.864680640120425e-07, + "loss": 0.8058, + "num_input_tokens_seen": 281207300, + "step": 8649 + }, + { + "epoch": 0.7800874780177661, + "flos": 14772469639200.0, + "grad_norm": 2.6604345234924094, + "learning_rate": 4.860862383324016e-07, + "loss": 0.7512, + "num_input_tokens_seen": 281231410, + "step": 8650 + }, + { + "epoch": 0.7801776615412365, + "flos": 22497078083520.0, + "grad_norm": 1.922851810484367, + "learning_rate": 4.857045418282295e-07, + "loss": 0.7402, + "num_input_tokens_seen": 281261000, + "step": 8651 + }, + { + "epoch": 0.7802678450647067, + "flos": 23881311973920.0, + "grad_norm": 5.713553981251922, + "learning_rate": 4.853229745320966e-07, + "loss": 0.6957, + "num_input_tokens_seen": 281288090, + "step": 8652 + }, + { + "epoch": 0.7803580285881769, + "flos": 23916889167840.0, + "grad_norm": 1.7400185122546517, + "learning_rate": 4.849415364765587e-07, + "loss": 0.728, + "num_input_tokens_seen": 281316800, + "step": 8653 + }, + { + "epoch": 0.7804482121116472, + "flos": 28507556584800.0, + "grad_norm": 2.049490502474375, + "learning_rate": 4.845602276941631e-07, + "loss": 0.808, + "num_input_tokens_seen": 281347495, + "step": 8654 + }, + { + "epoch": 0.7805383956351175, + "flos": 23552568183840.0, + "grad_norm": 1.8326481347884551, + "learning_rate": 4.841790482174449e-07, + "loss": 0.7806, + "num_input_tokens_seen": 281377695, + "step": 8655 + }, + { + "epoch": 0.7806285791585877, + "flos": 30836237301600.0, + "grad_norm": 2.0969953294955785, + "learning_rate": 4.837979980789282e-07, + "loss": 0.7334, + "num_input_tokens_seen": 281407970, + "step": 8656 + }, + { + "epoch": 0.780718762682058, + "flos": 29710856585280.0, + "grad_norm": 1.8436334217522419, + "learning_rate": 4.834170773111273e-07, + "loss": 0.6857, + "num_input_tokens_seen": 281437395, + "step": 8657 + }, + { + "epoch": 0.7808089462055282, + "flos": 20669005625280.0, + "grad_norm": 1.8519194053761598, + "learning_rate": 4.830362859465431e-07, + "loss": 0.7648, + "num_input_tokens_seen": 281465520, + "step": 8658 + }, + { + "epoch": 0.7808991297289986, + "flos": 22788237512640.0, + "grad_norm": 1.9521846491177772, + "learning_rate": 4.826556240176675e-07, + "loss": 0.8043, + "num_input_tokens_seen": 281495405, + "step": 8659 + }, + { + "epoch": 0.7809893132524688, + "flos": 20930274833760.0, + "grad_norm": 2.3514547228273566, + "learning_rate": 4.822750915569807e-07, + "loss": 0.7568, + "num_input_tokens_seen": 281521570, + "step": 8660 + }, + { + "epoch": 0.781079496775939, + "flos": 16449052391040.0, + "grad_norm": 2.082887398359692, + "learning_rate": 4.818946885969514e-07, + "loss": 0.82, + "num_input_tokens_seen": 281547925, + "step": 8661 + }, + { + "epoch": 0.7811696802994093, + "flos": 21586201284000.0, + "grad_norm": 2.092139920298343, + "learning_rate": 4.815144151700383e-07, + "loss": 0.7851, + "num_input_tokens_seen": 281578680, + "step": 8662 + }, + { + "epoch": 0.7812598638228796, + "flos": 20014454456160.0, + "grad_norm": 2.1822050042497176, + "learning_rate": 4.811342713086885e-07, + "loss": 0.7502, + "num_input_tokens_seen": 281606595, + "step": 8663 + }, + { + "epoch": 0.7813500473463498, + "flos": 15137496848640.0, + "grad_norm": 2.2639284292979314, + "learning_rate": 4.807542570453367e-07, + "loss": 0.7806, + "num_input_tokens_seen": 281632050, + "step": 8664 + }, + { + "epoch": 0.7814402308698201, + "flos": 21400769853120.0, + "grad_norm": 1.7536968284589132, + "learning_rate": 4.803743724124098e-07, + "loss": 0.7075, + "num_input_tokens_seen": 281662195, + "step": 8665 + }, + { + "epoch": 0.7815304143932903, + "flos": 36155100649440.0, + "grad_norm": 2.3502156909410066, + "learning_rate": 4.799946174423192e-07, + "loss": 0.749, + "num_input_tokens_seen": 281691400, + "step": 8666 + }, + { + "epoch": 0.7816205979167606, + "flos": 70991837956800.0, + "grad_norm": 0.7304067738398574, + "learning_rate": 4.796149921674706e-07, + "loss": 0.5316, + "num_input_tokens_seen": 281782290, + "step": 8667 + }, + { + "epoch": 0.7817107814402309, + "flos": 27597014313120.0, + "grad_norm": 1.7021173418523678, + "learning_rate": 4.792354966202534e-07, + "loss": 0.7018, + "num_input_tokens_seen": 281811145, + "step": 8668 + }, + { + "epoch": 0.7818009649637011, + "flos": 21184816315680.0, + "grad_norm": 1.9952993585188141, + "learning_rate": 4.788561308330489e-07, + "loss": 0.5712, + "num_input_tokens_seen": 281838625, + "step": 8669 + }, + { + "epoch": 0.7818911484871713, + "flos": 25265322845760.0, + "grad_norm": 2.6302199461022777, + "learning_rate": 4.784768948382272e-07, + "loss": 0.6874, + "num_input_tokens_seen": 281866270, + "step": 8670 + }, + { + "epoch": 0.7819813320106417, + "flos": 23407322997120.0, + "grad_norm": 1.8093681243080648, + "learning_rate": 4.780977886681461e-07, + "loss": 0.752, + "num_input_tokens_seen": 281895235, + "step": 8671 + }, + { + "epoch": 0.7820715155341119, + "flos": 25338484400640.0, + "grad_norm": 2.080016343896183, + "learning_rate": 4.777188123551541e-07, + "loss": 0.6906, + "num_input_tokens_seen": 281924650, + "step": 8672 + }, + { + "epoch": 0.7821616990575821, + "flos": 23590003865760.0, + "grad_norm": 2.0051412067489625, + "learning_rate": 4.773399659315856e-07, + "loss": 0.775, + "num_input_tokens_seen": 281953680, + "step": 8673 + }, + { + "epoch": 0.7822518825810525, + "flos": 16223174527680.0, + "grad_norm": 2.486303576567736, + "learning_rate": 4.769612494297681e-07, + "loss": 0.7269, + "num_input_tokens_seen": 281979620, + "step": 8674 + }, + { + "epoch": 0.7823420661045227, + "flos": 65842200024480.0, + "grad_norm": 0.6413882988335537, + "learning_rate": 4.765826628820142e-07, + "loss": 0.5544, + "num_input_tokens_seen": 282069110, + "step": 8675 + }, + { + "epoch": 0.782432249627993, + "flos": 24464002529760.0, + "grad_norm": 2.1684581857088516, + "learning_rate": 4.7620420632062775e-07, + "loss": 0.6577, + "num_input_tokens_seen": 282096370, + "step": 8676 + }, + { + "epoch": 0.7825224331514632, + "flos": 28508337149760.0, + "grad_norm": 1.9183934007725223, + "learning_rate": 4.758258797779002e-07, + "loss": 0.6995, + "num_input_tokens_seen": 282126860, + "step": 8677 + }, + { + "epoch": 0.7826126166749335, + "flos": 63906280891680.0, + "grad_norm": 0.6125996683733196, + "learning_rate": 4.7544768328611317e-07, + "loss": 0.5639, + "num_input_tokens_seen": 282213245, + "step": 8678 + }, + { + "epoch": 0.7827028001984038, + "flos": 24281693358720.0, + "grad_norm": 1.7515671388835203, + "learning_rate": 4.750696168775359e-07, + "loss": 0.7271, + "num_input_tokens_seen": 282245195, + "step": 8679 + }, + { + "epoch": 0.782792983721874, + "flos": 69944265015360.0, + "grad_norm": 0.5986500872908959, + "learning_rate": 4.746916805844279e-07, + "loss": 0.5509, + "num_input_tokens_seen": 282344220, + "step": 8680 + }, + { + "epoch": 0.7828831672453442, + "flos": 21439878174240.0, + "grad_norm": 2.6312234763940237, + "learning_rate": 4.743138744390356e-07, + "loss": 0.772, + "num_input_tokens_seen": 282373490, + "step": 8681 + }, + { + "epoch": 0.7829733507688146, + "flos": 22861101709440.0, + "grad_norm": 1.758359086569745, + "learning_rate": 4.739361984735959e-07, + "loss": 0.8165, + "num_input_tokens_seen": 282402020, + "step": 8682 + }, + { + "epoch": 0.7830635342922848, + "flos": 22824149234400.0, + "grad_norm": 2.0585885382627933, + "learning_rate": 4.7355865272033455e-07, + "loss": 0.7009, + "num_input_tokens_seen": 282430765, + "step": 8683 + }, + { + "epoch": 0.783153717815755, + "flos": 22060413279360.0, + "grad_norm": 1.8311141917414713, + "learning_rate": 4.7318123721146563e-07, + "loss": 0.7343, + "num_input_tokens_seen": 282459200, + "step": 8684 + }, + { + "epoch": 0.7832439013392253, + "flos": 25410493692960.0, + "grad_norm": 1.686396273185224, + "learning_rate": 4.728039519791924e-07, + "loss": 0.73, + "num_input_tokens_seen": 282488365, + "step": 8685 + }, + { + "epoch": 0.7833340848626956, + "flos": 67902469205280.0, + "grad_norm": 0.674695128892011, + "learning_rate": 4.72426797055707e-07, + "loss": 0.5982, + "num_input_tokens_seen": 282579205, + "step": 8686 + }, + { + "epoch": 0.7834242683861659, + "flos": 23623239364800.0, + "grad_norm": 1.8823411280247313, + "learning_rate": 4.720497724731904e-07, + "loss": 0.7115, + "num_input_tokens_seen": 282606205, + "step": 8687 + }, + { + "epoch": 0.7835144519096361, + "flos": 22460014099200.0, + "grad_norm": 2.4296726827384303, + "learning_rate": 4.7167287826381153e-07, + "loss": 0.7807, + "num_input_tokens_seen": 282634640, + "step": 8688 + }, + { + "epoch": 0.7836046354331063, + "flos": 28616778540480.0, + "grad_norm": 1.7025167796558616, + "learning_rate": 4.712961144597307e-07, + "loss": 0.7515, + "num_input_tokens_seen": 282666950, + "step": 8689 + }, + { + "epoch": 0.7836948189565767, + "flos": 23553794785920.0, + "grad_norm": 5.151499276298228, + "learning_rate": 4.7091948109309343e-07, + "loss": 0.7803, + "num_input_tokens_seen": 282694285, + "step": 8690 + }, + { + "epoch": 0.7837850024800469, + "flos": 18488543676000.0, + "grad_norm": 2.869417825062282, + "learning_rate": 4.705429781960384e-07, + "loss": 0.7536, + "num_input_tokens_seen": 282719660, + "step": 8691 + }, + { + "epoch": 0.7838751860035171, + "flos": 20748486039360.0, + "grad_norm": 2.4554021677620224, + "learning_rate": 4.7016660580068923e-07, + "loss": 0.7404, + "num_input_tokens_seen": 282748145, + "step": 8692 + }, + { + "epoch": 0.7839653695269874, + "flos": 26715581697120.0, + "grad_norm": 2.2738957418464896, + "learning_rate": 4.6979036393916093e-07, + "loss": 0.7768, + "num_input_tokens_seen": 282774365, + "step": 8693 + }, + { + "epoch": 0.7840555530504577, + "flos": 20524094966400.0, + "grad_norm": 1.6462528885082488, + "learning_rate": 4.6941425264355603e-07, + "loss": 0.8241, + "num_input_tokens_seen": 282801910, + "step": 8694 + }, + { + "epoch": 0.7841457365739279, + "flos": 20673428826720.0, + "grad_norm": 1.8052825607239313, + "learning_rate": 4.6903827194596666e-07, + "loss": 0.7549, + "num_input_tokens_seen": 282833155, + "step": 8695 + }, + { + "epoch": 0.7842359200973982, + "flos": 19800173557920.0, + "grad_norm": 1.5821950568648384, + "learning_rate": 4.686624218784743e-07, + "loss": 0.8013, + "num_input_tokens_seen": 282861290, + "step": 8696 + }, + { + "epoch": 0.7843261036208685, + "flos": 20303086341600.0, + "grad_norm": 2.2765741372841934, + "learning_rate": 4.6828670247314696e-07, + "loss": 0.7344, + "num_input_tokens_seen": 282885605, + "step": 8697 + }, + { + "epoch": 0.7844162871443388, + "flos": 31238848872000.0, + "grad_norm": 3.430222315353841, + "learning_rate": 4.679111137620442e-07, + "loss": 0.7616, + "num_input_tokens_seen": 282913905, + "step": 8698 + }, + { + "epoch": 0.784506470667809, + "flos": 27488275564320.0, + "grad_norm": 1.9595289300689336, + "learning_rate": 4.67535655777213e-07, + "loss": 0.7052, + "num_input_tokens_seen": 282943005, + "step": 8699 + }, + { + "epoch": 0.7845966541912792, + "flos": 22423656340320.0, + "grad_norm": 1.8424773268941517, + "learning_rate": 4.6716032855068956e-07, + "loss": 0.7525, + "num_input_tokens_seen": 282969545, + "step": 8700 + }, + { + "epoch": 0.7846868377147496, + "flos": 11092976379840.0, + "grad_norm": 2.0055205923586303, + "learning_rate": 4.6678513211449867e-07, + "loss": 0.7888, + "num_input_tokens_seen": 282993295, + "step": 8701 + }, + { + "epoch": 0.7847770212382198, + "flos": 17497740425280.0, + "grad_norm": 6.050534109704309, + "learning_rate": 4.6641006650065516e-07, + "loss": 0.7107, + "num_input_tokens_seen": 283012390, + "step": 8702 + }, + { + "epoch": 0.78486720476169, + "flos": 31422310305600.0, + "grad_norm": 1.9283793103938989, + "learning_rate": 4.6603513174115973e-07, + "loss": 0.7331, + "num_input_tokens_seen": 283041450, + "step": 8703 + }, + { + "epoch": 0.7849573882851603, + "flos": 24245670127680.0, + "grad_norm": 2.218205667266609, + "learning_rate": 4.6566032786800625e-07, + "loss": 0.7596, + "num_input_tokens_seen": 283069875, + "step": 8704 + }, + { + "epoch": 0.7850475718086306, + "flos": 22132422571680.0, + "grad_norm": 2.552776590244855, + "learning_rate": 4.6528565491317274e-07, + "loss": 0.7236, + "num_input_tokens_seen": 283097340, + "step": 8705 + }, + { + "epoch": 0.7851377553321008, + "flos": 31313051180160.0, + "grad_norm": 3.7369341520817074, + "learning_rate": 4.649111129086305e-07, + "loss": 0.812, + "num_input_tokens_seen": 283127900, + "step": 8706 + }, + { + "epoch": 0.7852279388555711, + "flos": 24752671584960.0, + "grad_norm": 1.8061803282925981, + "learning_rate": 4.6453670188633596e-07, + "loss": 0.7672, + "num_input_tokens_seen": 283157880, + "step": 8707 + }, + { + "epoch": 0.7853181223790413, + "flos": 24172025365920.0, + "grad_norm": 2.5946014109942963, + "learning_rate": 4.641624218782365e-07, + "loss": 0.7618, + "num_input_tokens_seen": 283185760, + "step": 8708 + }, + { + "epoch": 0.7854083059025116, + "flos": 21875167697280.0, + "grad_norm": 1.9315492861051515, + "learning_rate": 4.6378827291626765e-07, + "loss": 0.7104, + "num_input_tokens_seen": 283216455, + "step": 8709 + }, + { + "epoch": 0.7854984894259819, + "flos": 27851927492640.0, + "grad_norm": 1.3530660656082893, + "learning_rate": 4.634142550323541e-07, + "loss": 0.7967, + "num_input_tokens_seen": 283247510, + "step": 8710 + }, + { + "epoch": 0.7855886729494521, + "flos": 15208948594560.0, + "grad_norm": 2.2008752902315605, + "learning_rate": 4.6304036825840943e-07, + "loss": 0.7935, + "num_input_tokens_seen": 283273220, + "step": 8711 + }, + { + "epoch": 0.7856788564729223, + "flos": 22205249598720.0, + "grad_norm": 2.450620405313961, + "learning_rate": 4.626666126263341e-07, + "loss": 0.6514, + "num_input_tokens_seen": 283302080, + "step": 8712 + }, + { + "epoch": 0.7857690399963927, + "flos": 22055246682720.0, + "grad_norm": 2.0388343654182886, + "learning_rate": 4.622929881680213e-07, + "loss": 0.7706, + "num_input_tokens_seen": 283331010, + "step": 8713 + }, + { + "epoch": 0.7858592235198629, + "flos": 14946898821120.0, + "grad_norm": 3.200020378763159, + "learning_rate": 4.6191949491534887e-07, + "loss": 0.6646, + "num_input_tokens_seen": 283357260, + "step": 8714 + }, + { + "epoch": 0.7859494070433332, + "flos": 18525607660320.0, + "grad_norm": 1.8917502215357453, + "learning_rate": 4.6154613290018617e-07, + "loss": 0.7576, + "num_input_tokens_seen": 283385020, + "step": 8715 + }, + { + "epoch": 0.7860395905668034, + "flos": 20019249355200.0, + "grad_norm": 2.219214699665003, + "learning_rate": 4.6117290215439043e-07, + "loss": 0.6918, + "num_input_tokens_seen": 283413220, + "step": 8716 + }, + { + "epoch": 0.7861297740902737, + "flos": 21148607235840.0, + "grad_norm": 2.315476435169698, + "learning_rate": 4.6079980270980744e-07, + "loss": 0.6085, + "num_input_tokens_seen": 283441915, + "step": 8717 + }, + { + "epoch": 0.786219957613744, + "flos": 19800173557920.0, + "grad_norm": 2.0604033939974213, + "learning_rate": 4.6042683459827245e-07, + "loss": 0.7164, + "num_input_tokens_seen": 283470030, + "step": 8718 + }, + { + "epoch": 0.7863101411372142, + "flos": 70283496411360.0, + "grad_norm": 0.7025657336952429, + "learning_rate": 4.600539978516098e-07, + "loss": 0.5627, + "num_input_tokens_seen": 283564030, + "step": 8719 + }, + { + "epoch": 0.7864003246606845, + "flos": 21799590108000.0, + "grad_norm": 1.8967264887130775, + "learning_rate": 4.5968129250163004e-07, + "loss": 0.7082, + "num_input_tokens_seen": 283593790, + "step": 8720 + }, + { + "epoch": 0.7864905081841548, + "flos": 30729951756960.0, + "grad_norm": 2.1320419316847206, + "learning_rate": 4.5930871858013653e-07, + "loss": 0.6757, + "num_input_tokens_seen": 283623705, + "step": 8721 + }, + { + "epoch": 0.786580691707625, + "flos": 63967393710720.0, + "grad_norm": 0.6562919603284596, + "learning_rate": 4.589362761189182e-07, + "loss": 0.621, + "num_input_tokens_seen": 283720415, + "step": 8722 + }, + { + "epoch": 0.7866708752310952, + "flos": 24901299219840.0, + "grad_norm": 1.781773540454255, + "learning_rate": 4.585639651497539e-07, + "loss": 0.8096, + "num_input_tokens_seen": 283749805, + "step": 8723 + }, + { + "epoch": 0.7867610587545656, + "flos": 39838608242880.0, + "grad_norm": 1.9210801380446239, + "learning_rate": 4.581917857044115e-07, + "loss": 0.6972, + "num_input_tokens_seen": 283779835, + "step": 8724 + }, + { + "epoch": 0.7868512422780358, + "flos": 21841783519200.0, + "grad_norm": 1.8949356119495058, + "learning_rate": 4.5781973781464734e-07, + "loss": 0.7822, + "num_input_tokens_seen": 283811140, + "step": 8725 + }, + { + "epoch": 0.7869414258015061, + "flos": 23004414068640.0, + "grad_norm": 2.845077037338123, + "learning_rate": 4.574478215122073e-07, + "loss": 0.7858, + "num_input_tokens_seen": 283840950, + "step": 8726 + }, + { + "epoch": 0.7870316093249763, + "flos": 21658879632000.0, + "grad_norm": 5.61928742992319, + "learning_rate": 4.5707603682882357e-07, + "loss": 0.7246, + "num_input_tokens_seen": 283866915, + "step": 8727 + }, + { + "epoch": 0.7871217928484466, + "flos": 25041563658720.0, + "grad_norm": 2.8245818984950946, + "learning_rate": 4.56704383796221e-07, + "loss": 0.6101, + "num_input_tokens_seen": 283897005, + "step": 8728 + }, + { + "epoch": 0.7872119763719169, + "flos": 21003324879360.0, + "grad_norm": 2.1305091353289747, + "learning_rate": 4.5633286244610956e-07, + "loss": 0.7814, + "num_input_tokens_seen": 283926995, + "step": 8729 + }, + { + "epoch": 0.7873021598953871, + "flos": 23187652483680.0, + "grad_norm": 1.9024486473661986, + "learning_rate": 4.5596147281018993e-07, + "loss": 0.8122, + "num_input_tokens_seen": 283956775, + "step": 8730 + }, + { + "epoch": 0.7873923434188573, + "flos": 20635212579840.0, + "grad_norm": 3.576720915919493, + "learning_rate": 4.5559021492015137e-07, + "loss": 0.8443, + "num_input_tokens_seen": 283984915, + "step": 8731 + }, + { + "epoch": 0.7874825269423277, + "flos": 36484327646400.0, + "grad_norm": 1.6658351572372478, + "learning_rate": 4.552190888076712e-07, + "loss": 0.6537, + "num_input_tokens_seen": 284015290, + "step": 8732 + }, + { + "epoch": 0.7875727104657979, + "flos": 24864569763360.0, + "grad_norm": 1.9072840028112648, + "learning_rate": 4.548480945044164e-07, + "loss": 0.6928, + "num_input_tokens_seen": 284046880, + "step": 8733 + }, + { + "epoch": 0.7876628939892681, + "flos": 34295093633760.0, + "grad_norm": 1.8128375035142972, + "learning_rate": 4.54477232042042e-07, + "loss": 0.7405, + "num_input_tokens_seen": 284077550, + "step": 8734 + }, + { + "epoch": 0.7877530775127384, + "flos": 29200732868160.0, + "grad_norm": 1.9340760525382534, + "learning_rate": 4.541065014521921e-07, + "loss": 0.7364, + "num_input_tokens_seen": 284108200, + "step": 8735 + }, + { + "epoch": 0.7878432610362087, + "flos": 19399680663840.0, + "grad_norm": 2.940701789137123, + "learning_rate": 4.5373590276649996e-07, + "loss": 0.6929, + "num_input_tokens_seen": 284134005, + "step": 8736 + }, + { + "epoch": 0.787933444559679, + "flos": 35537836483200.0, + "grad_norm": 1.5817044144210382, + "learning_rate": 4.533654360165862e-07, + "loss": 0.7114, + "num_input_tokens_seen": 284165735, + "step": 8737 + }, + { + "epoch": 0.7880236280831492, + "flos": 17141225090880.0, + "grad_norm": 1.9658329257854166, + "learning_rate": 4.5299510123406115e-07, + "loss": 0.7998, + "num_input_tokens_seen": 284193100, + "step": 8738 + }, + { + "epoch": 0.7881138116066194, + "flos": 34373235936480.0, + "grad_norm": 1.6981274136612818, + "learning_rate": 4.5262489845052456e-07, + "loss": 0.7096, + "num_input_tokens_seen": 284222390, + "step": 8739 + }, + { + "epoch": 0.7882039951300898, + "flos": 16556601707520.0, + "grad_norm": 2.0341649194585143, + "learning_rate": 4.5225482769756353e-07, + "loss": 0.7123, + "num_input_tokens_seen": 284248265, + "step": 8740 + }, + { + "epoch": 0.78829417865356, + "flos": 26175493419840.0, + "grad_norm": 2.8177339807684034, + "learning_rate": 4.5188488900675545e-07, + "loss": 0.6732, + "num_input_tokens_seen": 284273170, + "step": 8741 + }, + { + "epoch": 0.7883843621770302, + "flos": 24387867394080.0, + "grad_norm": 1.99031577350948, + "learning_rate": 4.5151508240966363e-07, + "loss": 0.7731, + "num_input_tokens_seen": 284301465, + "step": 8742 + }, + { + "epoch": 0.7884745457005006, + "flos": 23371039577760.0, + "grad_norm": 1.8090662822990264, + "learning_rate": 4.511454079378445e-07, + "loss": 0.7914, + "num_input_tokens_seen": 284329645, + "step": 8743 + }, + { + "epoch": 0.7885647292239708, + "flos": 47051940707520.0, + "grad_norm": 3.3072120636899456, + "learning_rate": 4.507758656228382e-07, + "loss": 0.647, + "num_input_tokens_seen": 284360905, + "step": 8744 + }, + { + "epoch": 0.788654912747441, + "flos": 25879910789280.0, + "grad_norm": 1.9408116495054848, + "learning_rate": 4.5040645549617864e-07, + "loss": 0.8065, + "num_input_tokens_seen": 284391245, + "step": 8745 + }, + { + "epoch": 0.7887450962709113, + "flos": 25301308907040.0, + "grad_norm": 3.237580116488417, + "learning_rate": 4.5003717758938384e-07, + "loss": 0.6612, + "num_input_tokens_seen": 284422115, + "step": 8746 + }, + { + "epoch": 0.7888352797943816, + "flos": 26650002773280.0, + "grad_norm": 1.9039064603796816, + "learning_rate": 4.4966803193396365e-07, + "loss": 0.6846, + "num_input_tokens_seen": 284451600, + "step": 8747 + }, + { + "epoch": 0.7889254633178518, + "flos": 13825235080800.0, + "grad_norm": 1.8119101386661527, + "learning_rate": 4.492990185614154e-07, + "loss": 0.7502, + "num_input_tokens_seen": 284479140, + "step": 8748 + }, + { + "epoch": 0.7890156468413221, + "flos": 27229608239040.0, + "grad_norm": 2.000758384922489, + "learning_rate": 4.489301375032255e-07, + "loss": 0.7325, + "num_input_tokens_seen": 284507895, + "step": 8749 + }, + { + "epoch": 0.7891058303647923, + "flos": 26213746836480.0, + "grad_norm": 2.0372695932994414, + "learning_rate": 4.4856138879086857e-07, + "loss": 0.7493, + "num_input_tokens_seen": 284538415, + "step": 8750 + }, + { + "epoch": 0.7891960138882627, + "flos": 31858677751680.0, + "grad_norm": 1.423599371306025, + "learning_rate": 4.481927724558092e-07, + "loss": 0.6953, + "num_input_tokens_seen": 284572405, + "step": 8751 + }, + { + "epoch": 0.7892861974117329, + "flos": 20419705079520.0, + "grad_norm": 2.142212884420055, + "learning_rate": 4.478242885294985e-07, + "loss": 0.7157, + "num_input_tokens_seen": 284602750, + "step": 8752 + }, + { + "epoch": 0.7893763809352031, + "flos": 39324210003360.0, + "grad_norm": 3.117124781539315, + "learning_rate": 4.474559370433779e-07, + "loss": 0.7061, + "num_input_tokens_seen": 284634685, + "step": 8753 + }, + { + "epoch": 0.7894665644586734, + "flos": 17250632895360.0, + "grad_norm": 2.4054515554291167, + "learning_rate": 4.470877180288777e-07, + "loss": 0.6904, + "num_input_tokens_seen": 284661115, + "step": 8754 + }, + { + "epoch": 0.7895567479821437, + "flos": 26140956979200.0, + "grad_norm": 1.6127340575686664, + "learning_rate": 4.4671963151741574e-07, + "loss": 0.7124, + "num_input_tokens_seen": 284692115, + "step": 8755 + }, + { + "epoch": 0.7896469315056139, + "flos": 25667376869760.0, + "grad_norm": 5.50256825218871, + "learning_rate": 4.4635167754039973e-07, + "loss": 0.721, + "num_input_tokens_seen": 284724590, + "step": 8756 + }, + { + "epoch": 0.7897371150290842, + "flos": 23879118958080.0, + "grad_norm": 1.5693407230743384, + "learning_rate": 4.459838561292253e-07, + "loss": 0.7721, + "num_input_tokens_seen": 284754660, + "step": 8757 + }, + { + "epoch": 0.7898272985525544, + "flos": 23910161441280.0, + "grad_norm": 2.2973725694717926, + "learning_rate": 4.456161673152774e-07, + "loss": 0.6561, + "num_input_tokens_seen": 284782000, + "step": 8758 + }, + { + "epoch": 0.7899174820760247, + "flos": 21731780998560.0, + "grad_norm": 1.9405470852406572, + "learning_rate": 4.4524861112992806e-07, + "loss": 0.7334, + "num_input_tokens_seen": 284810260, + "step": 8759 + }, + { + "epoch": 0.790007665599495, + "flos": 40238692269600.0, + "grad_norm": 2.226563892029444, + "learning_rate": 4.448811876045411e-07, + "loss": 0.7422, + "num_input_tokens_seen": 284840280, + "step": 8760 + }, + { + "epoch": 0.7900978491229652, + "flos": 18998927581440.0, + "grad_norm": 2.4959514260945346, + "learning_rate": 4.445138967704647e-07, + "loss": 0.8112, + "num_input_tokens_seen": 284866960, + "step": 8761 + }, + { + "epoch": 0.7901880326464354, + "flos": 29123222451360.0, + "grad_norm": 2.0347307166687854, + "learning_rate": 4.4414673865904075e-07, + "loss": 0.7163, + "num_input_tokens_seen": 284896950, + "step": 8762 + }, + { + "epoch": 0.7902782161699058, + "flos": 24608876018880.0, + "grad_norm": 1.565169076530923, + "learning_rate": 4.437797133015955e-07, + "loss": 0.7591, + "num_input_tokens_seen": 284926915, + "step": 8763 + }, + { + "epoch": 0.790368399693376, + "flos": 22130824272000.0, + "grad_norm": 2.3170792886851563, + "learning_rate": 4.4341282072944586e-07, + "loss": 0.6575, + "num_input_tokens_seen": 284957680, + "step": 8764 + }, + { + "epoch": 0.7904585832168463, + "flos": 25520384704320.0, + "grad_norm": 1.8094669483388575, + "learning_rate": 4.430460609738973e-07, + "loss": 0.7643, + "num_input_tokens_seen": 284987310, + "step": 8765 + }, + { + "epoch": 0.7905487667403166, + "flos": 67751165347680.0, + "grad_norm": 0.7082207300403977, + "learning_rate": 4.4267943406624386e-07, + "loss": 0.6095, + "num_input_tokens_seen": 285077640, + "step": 8766 + }, + { + "epoch": 0.7906389502637868, + "flos": 15355606232160.0, + "grad_norm": 5.5411686303007155, + "learning_rate": 4.4231294003776853e-07, + "loss": 0.8533, + "num_input_tokens_seen": 285104575, + "step": 8767 + }, + { + "epoch": 0.7907291337872571, + "flos": 24864569763360.0, + "grad_norm": 2.3913216168077867, + "learning_rate": 4.419465789197416e-07, + "loss": 0.7826, + "num_input_tokens_seen": 285131055, + "step": 8768 + }, + { + "epoch": 0.7908193173107273, + "flos": 21768213096960.0, + "grad_norm": 1.9377232699804066, + "learning_rate": 4.415803507434237e-07, + "loss": 0.74, + "num_input_tokens_seen": 285158930, + "step": 8769 + }, + { + "epoch": 0.7909095008341976, + "flos": 23696066391840.0, + "grad_norm": 2.048837434440242, + "learning_rate": 4.4121425554006307e-07, + "loss": 0.766, + "num_input_tokens_seen": 285188485, + "step": 8770 + }, + { + "epoch": 0.7909996843576679, + "flos": 23509148170560.0, + "grad_norm": 2.173797415270503, + "learning_rate": 4.4084829334089744e-07, + "loss": 0.7258, + "num_input_tokens_seen": 285215310, + "step": 8771 + }, + { + "epoch": 0.7910898678811381, + "flos": 37397620480320.0, + "grad_norm": 1.8249891071778344, + "learning_rate": 4.404824641771525e-07, + "loss": 0.6823, + "num_input_tokens_seen": 285249415, + "step": 8772 + }, + { + "epoch": 0.7911800514046083, + "flos": 23729896607040.0, + "grad_norm": 1.5117138608710505, + "learning_rate": 4.4011676808004327e-07, + "loss": 0.81, + "num_input_tokens_seen": 285277925, + "step": 8773 + }, + { + "epoch": 0.7912702349280787, + "flos": 23550263658720.0, + "grad_norm": 1.7559542858368993, + "learning_rate": 4.3975120508077145e-07, + "loss": 0.7513, + "num_input_tokens_seen": 285305875, + "step": 8774 + }, + { + "epoch": 0.7913604184515489, + "flos": 26977965998400.0, + "grad_norm": 2.0663295219613103, + "learning_rate": 4.39385775210531e-07, + "loss": 0.7916, + "num_input_tokens_seen": 285334770, + "step": 8775 + }, + { + "epoch": 0.7914506019750192, + "flos": 18744237420480.0, + "grad_norm": 2.1869792863434405, + "learning_rate": 4.390204785005003e-07, + "loss": 0.7581, + "num_input_tokens_seen": 285362800, + "step": 8776 + }, + { + "epoch": 0.7915407854984894, + "flos": 24171616498560.0, + "grad_norm": 1.920206291421414, + "learning_rate": 4.386553149818504e-07, + "loss": 0.7477, + "num_input_tokens_seen": 285394720, + "step": 8777 + }, + { + "epoch": 0.7916309690219597, + "flos": 33243617867520.0, + "grad_norm": 2.86780643279178, + "learning_rate": 4.3829028468573793e-07, + "loss": 0.6106, + "num_input_tokens_seen": 285425035, + "step": 8778 + }, + { + "epoch": 0.79172115254543, + "flos": 29342446927680.0, + "grad_norm": 1.875368336420159, + "learning_rate": 4.3792538764330935e-07, + "loss": 0.6834, + "num_input_tokens_seen": 285454100, + "step": 8779 + }, + { + "epoch": 0.7918113360689002, + "flos": 17104569973920.0, + "grad_norm": 2.395060295667993, + "learning_rate": 4.3756062388569994e-07, + "loss": 0.7151, + "num_input_tokens_seen": 285478660, + "step": 8780 + }, + { + "epoch": 0.7919015195923704, + "flos": 52963524712800.0, + "grad_norm": 0.7095002958580728, + "learning_rate": 4.3719599344403346e-07, + "loss": 0.5793, + "num_input_tokens_seen": 285573365, + "step": 8781 + }, + { + "epoch": 0.7919917031158408, + "flos": 32368615620000.0, + "grad_norm": 1.5915694185410305, + "learning_rate": 4.3683149634942243e-07, + "loss": 0.7229, + "num_input_tokens_seen": 285604820, + "step": 8782 + }, + { + "epoch": 0.792081886639311, + "flos": 37249476052320.0, + "grad_norm": 1.9596033307022471, + "learning_rate": 4.364671326329663e-07, + "loss": 0.6253, + "num_input_tokens_seen": 285640620, + "step": 8783 + }, + { + "epoch": 0.7921720701627812, + "flos": 49492816960800.0, + "grad_norm": 2.405880193655866, + "learning_rate": 4.3610290232575673e-07, + "loss": 0.764, + "num_input_tokens_seen": 285673495, + "step": 8784 + }, + { + "epoch": 0.7922622536862515, + "flos": 22752400130400.0, + "grad_norm": 1.7330817134718948, + "learning_rate": 4.357388054588702e-07, + "loss": 0.7404, + "num_input_tokens_seen": 285701220, + "step": 8785 + }, + { + "epoch": 0.7923524372097218, + "flos": 22533175654080.0, + "grad_norm": 1.6783434081067194, + "learning_rate": 4.3537484206337405e-07, + "loss": 0.6877, + "num_input_tokens_seen": 285729230, + "step": 8786 + }, + { + "epoch": 0.792442620733192, + "flos": 15902236387200.0, + "grad_norm": 1.9438313433440098, + "learning_rate": 4.3501101217032366e-07, + "loss": 0.7737, + "num_input_tokens_seen": 285754470, + "step": 8787 + }, + { + "epoch": 0.7925328042566623, + "flos": 34518964330080.0, + "grad_norm": 2.088516184460318, + "learning_rate": 4.346473158107629e-07, + "loss": 0.668, + "num_input_tokens_seen": 285784305, + "step": 8788 + }, + { + "epoch": 0.7926229877801325, + "flos": 24063212277600.0, + "grad_norm": 2.1898188698124486, + "learning_rate": 4.342837530157244e-07, + "loss": 0.7463, + "num_input_tokens_seen": 285811475, + "step": 8789 + }, + { + "epoch": 0.7927131713036029, + "flos": 40456206936960.0, + "grad_norm": 2.5617483013434468, + "learning_rate": 4.3392032381622987e-07, + "loss": 0.6451, + "num_input_tokens_seen": 285841615, + "step": 8790 + }, + { + "epoch": 0.7928033548270731, + "flos": 19580911911840.0, + "grad_norm": 1.5637225899249447, + "learning_rate": 4.3355702824328765e-07, + "loss": 0.7035, + "num_input_tokens_seen": 285870190, + "step": 8791 + }, + { + "epoch": 0.7928935383505433, + "flos": 25556036237760.0, + "grad_norm": 1.8119844163725587, + "learning_rate": 4.3319386632789823e-07, + "loss": 0.7839, + "num_input_tokens_seen": 285901540, + "step": 8792 + }, + { + "epoch": 0.7929837218740137, + "flos": 20563686494400.0, + "grad_norm": 2.0713431617905527, + "learning_rate": 4.328308381010466e-07, + "loss": 0.7376, + "num_input_tokens_seen": 285929375, + "step": 8793 + }, + { + "epoch": 0.7930739053974839, + "flos": 22169412216480.0, + "grad_norm": 1.8095891532596482, + "learning_rate": 4.3246794359370933e-07, + "loss": 0.747, + "num_input_tokens_seen": 285955945, + "step": 8794 + }, + { + "epoch": 0.7931640889209541, + "flos": 21439989683520.0, + "grad_norm": 2.1244282794857905, + "learning_rate": 4.3210518283685025e-07, + "loss": 0.7173, + "num_input_tokens_seen": 285984975, + "step": 8795 + }, + { + "epoch": 0.7932542724444244, + "flos": 19763815799040.0, + "grad_norm": 2.1846528189046306, + "learning_rate": 4.317425558614225e-07, + "loss": 0.7397, + "num_input_tokens_seen": 286012175, + "step": 8796 + }, + { + "epoch": 0.7933444559678947, + "flos": 21364003226880.0, + "grad_norm": 1.4683237471828676, + "learning_rate": 4.3138006269836744e-07, + "loss": 0.7063, + "num_input_tokens_seen": 286039775, + "step": 8797 + }, + { + "epoch": 0.793434639491365, + "flos": 26026531257120.0, + "grad_norm": 2.579796543970565, + "learning_rate": 4.3101770337861376e-07, + "loss": 0.7299, + "num_input_tokens_seen": 286069990, + "step": 8798 + }, + { + "epoch": 0.7935248230148352, + "flos": 62678108588160.0, + "grad_norm": 0.6656728404945843, + "learning_rate": 4.30655477933082e-07, + "loss": 0.5928, + "num_input_tokens_seen": 286165170, + "step": 8799 + }, + { + "epoch": 0.7936150065383054, + "flos": 27887876384160.0, + "grad_norm": 1.749225240595686, + "learning_rate": 4.30293386392677e-07, + "loss": 0.7659, + "num_input_tokens_seen": 286198805, + "step": 8800 + }, + { + "epoch": 0.7937051900617758, + "flos": 25741616347680.0, + "grad_norm": 1.5809255998491203, + "learning_rate": 4.299314287882967e-07, + "loss": 0.7617, + "num_input_tokens_seen": 286232635, + "step": 8801 + }, + { + "epoch": 0.793795373585246, + "flos": 30219121814400.0, + "grad_norm": 2.1233777978853623, + "learning_rate": 4.2956960515082353e-07, + "loss": 0.7745, + "num_input_tokens_seen": 286264700, + "step": 8802 + }, + { + "epoch": 0.7938855571087162, + "flos": 16769470154880.0, + "grad_norm": 1.917401449364134, + "learning_rate": 4.29207915511131e-07, + "loss": 0.6973, + "num_input_tokens_seen": 286292240, + "step": 8803 + }, + { + "epoch": 0.7939757406321865, + "flos": 19654073466720.0, + "grad_norm": 2.0134203778749216, + "learning_rate": 4.2884635990008024e-07, + "loss": 0.8002, + "num_input_tokens_seen": 286319045, + "step": 8804 + }, + { + "epoch": 0.7940659241556568, + "flos": 21294112610880.0, + "grad_norm": 2.5511603314698266, + "learning_rate": 4.284849383485214e-07, + "loss": 0.7066, + "num_input_tokens_seen": 286348300, + "step": 8805 + }, + { + "epoch": 0.794156107679127, + "flos": 21512891050080.0, + "grad_norm": 2.9356751962763568, + "learning_rate": 4.2812365088729296e-07, + "loss": 0.7676, + "num_input_tokens_seen": 286376695, + "step": 8806 + }, + { + "epoch": 0.7942462912025973, + "flos": 21076337755200.0, + "grad_norm": 3.275064651200562, + "learning_rate": 4.2776249754722227e-07, + "loss": 0.7623, + "num_input_tokens_seen": 286406945, + "step": 8807 + }, + { + "epoch": 0.7943364747260675, + "flos": 17970725818560.0, + "grad_norm": 3.0163113302231452, + "learning_rate": 4.27401478359124e-07, + "loss": 0.8126, + "num_input_tokens_seen": 286426605, + "step": 8808 + }, + { + "epoch": 0.7944266582495378, + "flos": 23404795453440.0, + "grad_norm": 1.9822699126832164, + "learning_rate": 4.2704059335380283e-07, + "loss": 0.8367, + "num_input_tokens_seen": 286455270, + "step": 8809 + }, + { + "epoch": 0.7945168417730081, + "flos": 25922996274720.0, + "grad_norm": 1.646398069171867, + "learning_rate": 4.266798425620515e-07, + "loss": 0.8451, + "num_input_tokens_seen": 286487270, + "step": 8810 + }, + { + "epoch": 0.7946070252964783, + "flos": 50734296038400.0, + "grad_norm": 1.9736921857371572, + "learning_rate": 4.263192260146511e-07, + "loss": 0.7274, + "num_input_tokens_seen": 286523730, + "step": 8811 + }, + { + "epoch": 0.7946972088199485, + "flos": 22131976534560.0, + "grad_norm": 1.9469579761634956, + "learning_rate": 4.2595874374237216e-07, + "loss": 0.7906, + "num_input_tokens_seen": 286551105, + "step": 8812 + }, + { + "epoch": 0.7947873923434189, + "flos": 22747939759200.0, + "grad_norm": 1.4636883464178614, + "learning_rate": 4.255983957759712e-07, + "loss": 0.7376, + "num_input_tokens_seen": 286581480, + "step": 8813 + }, + { + "epoch": 0.7948775758668891, + "flos": 23698259407680.0, + "grad_norm": 1.6506933893886788, + "learning_rate": 4.2523818214619745e-07, + "loss": 0.6793, + "num_input_tokens_seen": 286612745, + "step": 8814 + }, + { + "epoch": 0.7949677593903594, + "flos": 26902685767200.0, + "grad_norm": 1.696339520289769, + "learning_rate": 4.24878102883784e-07, + "loss": 0.7882, + "num_input_tokens_seen": 286642610, + "step": 8815 + }, + { + "epoch": 0.7950579429138297, + "flos": 21984092294880.0, + "grad_norm": 1.6525496406428517, + "learning_rate": 4.24518158019457e-07, + "loss": 0.7487, + "num_input_tokens_seen": 286671235, + "step": 8816 + }, + { + "epoch": 0.7951481264372999, + "flos": 27414779481600.0, + "grad_norm": 2.756796587596648, + "learning_rate": 4.241583475839274e-07, + "loss": 0.6821, + "num_input_tokens_seen": 286699560, + "step": 8817 + }, + { + "epoch": 0.7952383099607702, + "flos": 22678718198880.0, + "grad_norm": 1.826923986581429, + "learning_rate": 4.237986716078965e-07, + "loss": 0.7865, + "num_input_tokens_seen": 286727735, + "step": 8818 + }, + { + "epoch": 0.7953284934842404, + "flos": 26066643161760.0, + "grad_norm": 1.8403171327169559, + "learning_rate": 4.2343913012205433e-07, + "loss": 0.7285, + "num_input_tokens_seen": 286757745, + "step": 8819 + }, + { + "epoch": 0.7954186770077107, + "flos": 25480718836800.0, + "grad_norm": 1.937535870127729, + "learning_rate": 4.230797231570784e-07, + "loss": 0.7168, + "num_input_tokens_seen": 286788570, + "step": 8820 + }, + { + "epoch": 0.795508860531181, + "flos": 22423358982240.0, + "grad_norm": 3.2585044510210843, + "learning_rate": 4.227204507436357e-07, + "loss": 0.7442, + "num_input_tokens_seen": 286817840, + "step": 8821 + }, + { + "epoch": 0.7955990440546512, + "flos": 74626801843680.0, + "grad_norm": 1.758516636856251, + "learning_rate": 4.223613129123811e-07, + "loss": 0.6599, + "num_input_tokens_seen": 286856415, + "step": 8822 + }, + { + "epoch": 0.7956892275781214, + "flos": 24643635478080.0, + "grad_norm": 1.7440020088569201, + "learning_rate": 4.220023096939589e-07, + "loss": 0.6472, + "num_input_tokens_seen": 286885515, + "step": 8823 + }, + { + "epoch": 0.7957794111015918, + "flos": 19107740669760.0, + "grad_norm": 2.0523383922819174, + "learning_rate": 4.21643441119e-07, + "loss": 0.7365, + "num_input_tokens_seen": 286910440, + "step": 8824 + }, + { + "epoch": 0.795869594625062, + "flos": 66915977646720.0, + "grad_norm": 0.6293727258377886, + "learning_rate": 4.212847072181256e-07, + "loss": 0.629, + "num_input_tokens_seen": 287001595, + "step": 8825 + }, + { + "epoch": 0.7959597781485322, + "flos": 21367385675040.0, + "grad_norm": 1.9387433769265436, + "learning_rate": 4.2092610802194505e-07, + "loss": 0.7159, + "num_input_tokens_seen": 287030295, + "step": 8826 + }, + { + "epoch": 0.7960499616720025, + "flos": 24243885979200.0, + "grad_norm": 2.1873791252527655, + "learning_rate": 4.2056764356105587e-07, + "loss": 0.652, + "num_input_tokens_seen": 287059545, + "step": 8827 + }, + { + "epoch": 0.7961401451954728, + "flos": 24719027218560.0, + "grad_norm": 1.8114792152621184, + "learning_rate": 4.202093138660443e-07, + "loss": 0.7878, + "num_input_tokens_seen": 287087560, + "step": 8828 + }, + { + "epoch": 0.7962303287189431, + "flos": 25044946106880.0, + "grad_norm": 1.902507673090175, + "learning_rate": 4.198511189674854e-07, + "loss": 0.7735, + "num_input_tokens_seen": 287117715, + "step": 8829 + }, + { + "epoch": 0.7963205122424133, + "flos": 28507816773120.0, + "grad_norm": 2.6354930600237965, + "learning_rate": 4.1949305889594066e-07, + "loss": 0.761, + "num_input_tokens_seen": 287146620, + "step": 8830 + }, + { + "epoch": 0.7964106957658835, + "flos": 27303476019360.0, + "grad_norm": 1.6035412729421499, + "learning_rate": 4.191351336819642e-07, + "loss": 0.7332, + "num_input_tokens_seen": 287178175, + "step": 8831 + }, + { + "epoch": 0.7965008792893539, + "flos": 34628260625280.0, + "grad_norm": 1.733207447114533, + "learning_rate": 4.187773433560939e-07, + "loss": 0.6782, + "num_input_tokens_seen": 287209490, + "step": 8832 + }, + { + "epoch": 0.7965910628128241, + "flos": 25810057343040.0, + "grad_norm": 3.043573556912645, + "learning_rate": 4.184196879488604e-07, + "loss": 0.7377, + "num_input_tokens_seen": 287237920, + "step": 8833 + }, + { + "epoch": 0.7966812463362943, + "flos": 27086556068160.0, + "grad_norm": 1.7734424742238197, + "learning_rate": 4.1806216749077936e-07, + "loss": 0.8028, + "num_input_tokens_seen": 287267705, + "step": 8834 + }, + { + "epoch": 0.7967714298597646, + "flos": 30689951361600.0, + "grad_norm": 1.586115634479, + "learning_rate": 4.177047820123569e-07, + "loss": 0.7832, + "num_input_tokens_seen": 287297645, + "step": 8835 + }, + { + "epoch": 0.7968616133832349, + "flos": 22713477658080.0, + "grad_norm": 1.8669334462240492, + "learning_rate": 4.1734753154408733e-07, + "loss": 0.6129, + "num_input_tokens_seen": 287326215, + "step": 8836 + }, + { + "epoch": 0.7969517969067051, + "flos": 20558742916320.0, + "grad_norm": 2.0834004836386666, + "learning_rate": 4.169904161164528e-07, + "loss": 0.7237, + "num_input_tokens_seen": 287353265, + "step": 8837 + }, + { + "epoch": 0.7970419804301754, + "flos": 18561147684480.0, + "grad_norm": 2.3214938563294942, + "learning_rate": 4.1663343575992526e-07, + "loss": 0.8594, + "num_input_tokens_seen": 287377675, + "step": 8838 + }, + { + "epoch": 0.7971321639536457, + "flos": 24790962171360.0, + "grad_norm": 1.9143951127019811, + "learning_rate": 4.1627659050496275e-07, + "loss": 0.7585, + "num_input_tokens_seen": 287402915, + "step": 8839 + }, + { + "epoch": 0.797222347477116, + "flos": 26358992023200.0, + "grad_norm": 1.9628767781694436, + "learning_rate": 4.1591988038201453e-07, + "loss": 0.6571, + "num_input_tokens_seen": 287433930, + "step": 8840 + }, + { + "epoch": 0.7973125310005862, + "flos": 17360300888160.0, + "grad_norm": 1.764480017017334, + "learning_rate": 4.155633054215164e-07, + "loss": 0.7859, + "num_input_tokens_seen": 287461365, + "step": 8841 + }, + { + "epoch": 0.7974027145240564, + "flos": 27050049630240.0, + "grad_norm": 2.233161436649001, + "learning_rate": 4.152068656538934e-07, + "loss": 0.6507, + "num_input_tokens_seen": 287488980, + "step": 8842 + }, + { + "epoch": 0.7974928980475268, + "flos": 34628483643840.0, + "grad_norm": 1.661742478072576, + "learning_rate": 4.148505611095594e-07, + "loss": 0.7974, + "num_input_tokens_seen": 287519625, + "step": 8843 + }, + { + "epoch": 0.797583081570997, + "flos": 26102852241600.0, + "grad_norm": 3.0375475404888697, + "learning_rate": 4.1449439181891563e-07, + "loss": 0.6155, + "num_input_tokens_seen": 287548570, + "step": 8844 + }, + { + "epoch": 0.7976732650944672, + "flos": 23188693236960.0, + "grad_norm": 2.0229651883006223, + "learning_rate": 4.14138357812353e-07, + "loss": 0.7603, + "num_input_tokens_seen": 287576195, + "step": 8845 + }, + { + "epoch": 0.7977634486179375, + "flos": 27487903866720.0, + "grad_norm": 1.639419330633088, + "learning_rate": 4.137824591202506e-07, + "loss": 0.7966, + "num_input_tokens_seen": 287608180, + "step": 8846 + }, + { + "epoch": 0.7978536321414078, + "flos": 22459605231840.0, + "grad_norm": 1.7344429758858408, + "learning_rate": 4.134266957729737e-07, + "loss": 0.7301, + "num_input_tokens_seen": 287637915, + "step": 8847 + }, + { + "epoch": 0.797943815664878, + "flos": 20165646804480.0, + "grad_norm": 1.8092072272384716, + "learning_rate": 4.1307106780088065e-07, + "loss": 0.8359, + "num_input_tokens_seen": 287664555, + "step": 8848 + }, + { + "epoch": 0.7980339991883483, + "flos": 24938028676320.0, + "grad_norm": 1.920147437256158, + "learning_rate": 4.1271557523431387e-07, + "loss": 0.751, + "num_input_tokens_seen": 287693355, + "step": 8849 + }, + { + "epoch": 0.7981241827118185, + "flos": 16521842248320.0, + "grad_norm": 2.0077796072436405, + "learning_rate": 4.1236021810360634e-07, + "loss": 0.7251, + "num_input_tokens_seen": 287719710, + "step": 8850 + }, + { + "epoch": 0.7982143662352889, + "flos": 20966112216000.0, + "grad_norm": 2.1220018891048382, + "learning_rate": 4.120049964390793e-07, + "loss": 0.6849, + "num_input_tokens_seen": 287748135, + "step": 8851 + }, + { + "epoch": 0.7983045497587591, + "flos": 23843616103680.0, + "grad_norm": 3.0715891583813324, + "learning_rate": 4.116499102710418e-07, + "loss": 0.69, + "num_input_tokens_seen": 287773740, + "step": 8852 + }, + { + "epoch": 0.7983947332822293, + "flos": 26795842676160.0, + "grad_norm": 2.7016551701816676, + "learning_rate": 4.112949596297928e-07, + "loss": 0.7132, + "num_input_tokens_seen": 287803385, + "step": 8853 + }, + { + "epoch": 0.7984849168056996, + "flos": 24714269489280.0, + "grad_norm": 2.0116026020778865, + "learning_rate": 4.1094014454561664e-07, + "loss": 0.6865, + "num_input_tokens_seen": 287832130, + "step": 8854 + }, + { + "epoch": 0.7985751003291699, + "flos": 17468742278880.0, + "grad_norm": 2.334539192686107, + "learning_rate": 4.1058546504879057e-07, + "loss": 0.6464, + "num_input_tokens_seen": 287857925, + "step": 8855 + }, + { + "epoch": 0.7986652838526401, + "flos": 24937805657760.0, + "grad_norm": 2.0068828642247643, + "learning_rate": 4.1023092116957583e-07, + "loss": 0.7974, + "num_input_tokens_seen": 287884775, + "step": 8856 + }, + { + "epoch": 0.7987554673761104, + "flos": 34155646929600.0, + "grad_norm": 1.6578644854909574, + "learning_rate": 4.098765129382249e-07, + "loss": 0.7794, + "num_input_tokens_seen": 287916690, + "step": 8857 + }, + { + "epoch": 0.7988456508995806, + "flos": 24859105808640.0, + "grad_norm": 1.922772515853172, + "learning_rate": 4.0952224038497764e-07, + "loss": 0.8124, + "num_input_tokens_seen": 287942750, + "step": 8858 + }, + { + "epoch": 0.7989358344230509, + "flos": 13236894721440.0, + "grad_norm": 2.3040058744408634, + "learning_rate": 4.091681035400627e-07, + "loss": 0.7358, + "num_input_tokens_seen": 287967825, + "step": 8859 + }, + { + "epoch": 0.7990260179465212, + "flos": 21403631924640.0, + "grad_norm": 1.9098085426319278, + "learning_rate": 4.088141024336971e-07, + "loss": 0.7767, + "num_input_tokens_seen": 287996215, + "step": 8860 + }, + { + "epoch": 0.7991162014699914, + "flos": 22748757493920.0, + "grad_norm": 1.5744458358960107, + "learning_rate": 4.0846023709608636e-07, + "loss": 0.8207, + "num_input_tokens_seen": 288025610, + "step": 8861 + }, + { + "epoch": 0.7992063849934617, + "flos": 23443606416480.0, + "grad_norm": 2.133372809621029, + "learning_rate": 4.081065075574226e-07, + "loss": 0.8155, + "num_input_tokens_seen": 288055005, + "step": 8862 + }, + { + "epoch": 0.799296568516932, + "flos": 24457311972960.0, + "grad_norm": 1.6906301642883736, + "learning_rate": 4.077529138478906e-07, + "loss": 0.6708, + "num_input_tokens_seen": 288083235, + "step": 8863 + }, + { + "epoch": 0.7993867520404022, + "flos": 67530379741440.0, + "grad_norm": 0.6803063173587, + "learning_rate": 4.073994559976588e-07, + "loss": 0.5405, + "num_input_tokens_seen": 288166660, + "step": 8864 + }, + { + "epoch": 0.7994769355638724, + "flos": 27232804838400.0, + "grad_norm": 2.02930192698932, + "learning_rate": 4.0704613403688716e-07, + "loss": 0.7051, + "num_input_tokens_seen": 288196770, + "step": 8865 + }, + { + "epoch": 0.7995671190873428, + "flos": 23477696820000.0, + "grad_norm": 2.1206419518038033, + "learning_rate": 4.0669294799572264e-07, + "loss": 0.7992, + "num_input_tokens_seen": 288225385, + "step": 8866 + }, + { + "epoch": 0.799657302610813, + "flos": 14626518227040.0, + "grad_norm": 2.3821316105815473, + "learning_rate": 4.0633989790430113e-07, + "loss": 0.8413, + "num_input_tokens_seen": 288248775, + "step": 8867 + }, + { + "epoch": 0.7997474861342833, + "flos": 22860395484000.0, + "grad_norm": 1.994444940644288, + "learning_rate": 4.059869837927477e-07, + "loss": 0.7494, + "num_input_tokens_seen": 288275510, + "step": 8868 + }, + { + "epoch": 0.7998376696577535, + "flos": 22751285037600.0, + "grad_norm": 2.55182910094044, + "learning_rate": 4.056342056911728e-07, + "loss": 0.7249, + "num_input_tokens_seen": 288303985, + "step": 8869 + }, + { + "epoch": 0.7999278531812238, + "flos": 24865052970240.0, + "grad_norm": 2.344077837585793, + "learning_rate": 4.052815636296798e-07, + "loss": 0.7613, + "num_input_tokens_seen": 288334830, + "step": 8870 + }, + { + "epoch": 0.8000180367046941, + "flos": 21730814584800.0, + "grad_norm": 2.0129175262459866, + "learning_rate": 4.0492905763835593e-07, + "loss": 0.7742, + "num_input_tokens_seen": 288363750, + "step": 8871 + }, + { + "epoch": 0.8001082202281643, + "flos": 23587922359200.0, + "grad_norm": 1.8798869977631982, + "learning_rate": 4.0457668774728115e-07, + "loss": 0.7723, + "num_input_tokens_seen": 288391060, + "step": 8872 + }, + { + "epoch": 0.8001984037516345, + "flos": 23692981301760.0, + "grad_norm": 2.478442227815172, + "learning_rate": 4.0422445398651985e-07, + "loss": 0.7601, + "num_input_tokens_seen": 288419250, + "step": 8873 + }, + { + "epoch": 0.8002885872751049, + "flos": 40566804173760.0, + "grad_norm": 1.8602947236436833, + "learning_rate": 4.0387235638612706e-07, + "loss": 0.6651, + "num_input_tokens_seen": 288453070, + "step": 8874 + }, + { + "epoch": 0.8003787707985751, + "flos": 19144432956480.0, + "grad_norm": 2.1831013733431446, + "learning_rate": 4.0352039497614586e-07, + "loss": 0.7487, + "num_input_tokens_seen": 288479835, + "step": 8875 + }, + { + "epoch": 0.8004689543220453, + "flos": 17942322388320.0, + "grad_norm": 2.352883813663689, + "learning_rate": 4.031685697866074e-07, + "loss": 0.7313, + "num_input_tokens_seen": 288507325, + "step": 8876 + }, + { + "epoch": 0.8005591378455156, + "flos": 22234953970560.0, + "grad_norm": 1.5766923652447142, + "learning_rate": 4.0281688084753165e-07, + "loss": 0.782, + "num_input_tokens_seen": 288536880, + "step": 8877 + }, + { + "epoch": 0.8006493213689859, + "flos": 21694568335200.0, + "grad_norm": 2.42961213709488, + "learning_rate": 4.0246532818892675e-07, + "loss": 0.7918, + "num_input_tokens_seen": 288563795, + "step": 8878 + }, + { + "epoch": 0.8007395048924562, + "flos": 20779193994720.0, + "grad_norm": 2.0657248283883005, + "learning_rate": 4.0211391184078814e-07, + "loss": 0.7779, + "num_input_tokens_seen": 288590765, + "step": 8879 + }, + { + "epoch": 0.8008296884159264, + "flos": 22640464782240.0, + "grad_norm": 2.1327794110871903, + "learning_rate": 4.0176263183310135e-07, + "loss": 0.7749, + "num_input_tokens_seen": 288619735, + "step": 8880 + }, + { + "epoch": 0.8009198719393966, + "flos": 23295945195360.0, + "grad_norm": 1.7377463900183272, + "learning_rate": 4.0141148819583925e-07, + "loss": 0.7976, + "num_input_tokens_seen": 288648225, + "step": 8881 + }, + { + "epoch": 0.801010055462867, + "flos": 64078845852000.0, + "grad_norm": 0.6034001607222654, + "learning_rate": 4.010604809589637e-07, + "loss": 0.569, + "num_input_tokens_seen": 288742820, + "step": 8882 + }, + { + "epoch": 0.8011002389863372, + "flos": 27778988956320.0, + "grad_norm": 1.7747644509923979, + "learning_rate": 4.0070961015242475e-07, + "loss": 0.7452, + "num_input_tokens_seen": 288770290, + "step": 8883 + }, + { + "epoch": 0.8011904225098074, + "flos": 22816566603360.0, + "grad_norm": 2.3119109150906283, + "learning_rate": 4.0035887580615933e-07, + "loss": 0.8138, + "num_input_tokens_seen": 288797900, + "step": 8884 + }, + { + "epoch": 0.8012806060332777, + "flos": 26538030255360.0, + "grad_norm": 2.323129914991564, + "learning_rate": 4.0000827795009594e-07, + "loss": 0.7914, + "num_input_tokens_seen": 288828090, + "step": 8885 + }, + { + "epoch": 0.801370789556748, + "flos": 19035025152000.0, + "grad_norm": 3.6251085199357873, + "learning_rate": 3.996578166141475e-07, + "loss": 0.7084, + "num_input_tokens_seen": 288854350, + "step": 8886 + }, + { + "epoch": 0.8014609730802182, + "flos": 19800693934560.0, + "grad_norm": 4.61327649535371, + "learning_rate": 3.9930749182821955e-07, + "loss": 0.7511, + "num_input_tokens_seen": 288882255, + "step": 8887 + }, + { + "epoch": 0.8015511566036885, + "flos": 58899800905920.0, + "grad_norm": 0.5984872513451388, + "learning_rate": 3.9895730362220116e-07, + "loss": 0.5944, + "num_input_tokens_seen": 288973515, + "step": 8888 + }, + { + "epoch": 0.8016413401271588, + "flos": 15209877838560.0, + "grad_norm": 1.9301356980446247, + "learning_rate": 3.986072520259749e-07, + "loss": 0.7843, + "num_input_tokens_seen": 288998805, + "step": 8889 + }, + { + "epoch": 0.801731523650629, + "flos": 22569384733920.0, + "grad_norm": 1.9693400997264294, + "learning_rate": 3.9825733706940736e-07, + "loss": 0.7321, + "num_input_tokens_seen": 289028125, + "step": 8890 + }, + { + "epoch": 0.8018217071740993, + "flos": 25957161017760.0, + "grad_norm": 1.9198545054105507, + "learning_rate": 3.979075587823557e-07, + "loss": 0.76, + "num_input_tokens_seen": 289057575, + "step": 8891 + }, + { + "epoch": 0.8019118906975695, + "flos": 26387246774400.0, + "grad_norm": 5.402980278728952, + "learning_rate": 3.9755791719466504e-07, + "loss": 0.6213, + "num_input_tokens_seen": 289084745, + "step": 8892 + }, + { + "epoch": 0.8020020742210399, + "flos": 23042816164320.0, + "grad_norm": 2.10978457395212, + "learning_rate": 3.9720841233616875e-07, + "loss": 0.7013, + "num_input_tokens_seen": 289113315, + "step": 8893 + }, + { + "epoch": 0.8020922577445101, + "flos": 17936821263840.0, + "grad_norm": 1.9693920360180632, + "learning_rate": 3.968590442366888e-07, + "loss": 0.7203, + "num_input_tokens_seen": 289138520, + "step": 8894 + }, + { + "epoch": 0.8021824412679803, + "flos": 22896641733600.0, + "grad_norm": 6.486740967552732, + "learning_rate": 3.9650981292603423e-07, + "loss": 0.7196, + "num_input_tokens_seen": 289167905, + "step": 8895 + }, + { + "epoch": 0.8022726247914506, + "flos": 21366642279840.0, + "grad_norm": 2.2151575152187766, + "learning_rate": 3.961607184340041e-07, + "loss": 0.7206, + "num_input_tokens_seen": 289192955, + "step": 8896 + }, + { + "epoch": 0.8023628083149209, + "flos": 38672743924320.0, + "grad_norm": 1.5800204083778484, + "learning_rate": 3.9581176079038505e-07, + "loss": 0.6414, + "num_input_tokens_seen": 289226080, + "step": 8897 + }, + { + "epoch": 0.8024529918383911, + "flos": 41404147720800.0, + "grad_norm": 1.7310734848212175, + "learning_rate": 3.954629400249516e-07, + "loss": 0.6898, + "num_input_tokens_seen": 289259925, + "step": 8898 + }, + { + "epoch": 0.8025431753618614, + "flos": 23845325912640.0, + "grad_norm": 1.7263586692112223, + "learning_rate": 3.9511425616746787e-07, + "loss": 0.8118, + "num_input_tokens_seen": 289287725, + "step": 8899 + }, + { + "epoch": 0.8026333588853316, + "flos": 20820235143360.0, + "grad_norm": 3.768061818921741, + "learning_rate": 3.947657092476853e-07, + "loss": 0.81, + "num_input_tokens_seen": 289315845, + "step": 8900 + }, + { + "epoch": 0.802723542408802, + "flos": 18084965691840.0, + "grad_norm": 3.7081164750258857, + "learning_rate": 3.944172992953425e-07, + "loss": 0.7556, + "num_input_tokens_seen": 289342360, + "step": 8901 + }, + { + "epoch": 0.8028137259322722, + "flos": 31239666606720.0, + "grad_norm": 1.7752183887292559, + "learning_rate": 3.9406902634017e-07, + "loss": 0.6973, + "num_input_tokens_seen": 289373715, + "step": 8902 + }, + { + "epoch": 0.8029039094557424, + "flos": 16630915524960.0, + "grad_norm": 1.6440721899550796, + "learning_rate": 3.9372089041188227e-07, + "loss": 0.7781, + "num_input_tokens_seen": 289400840, + "step": 8903 + }, + { + "epoch": 0.8029940929792126, + "flos": 37944324974880.0, + "grad_norm": 2.0762403159286027, + "learning_rate": 3.9337289154018593e-07, + "loss": 0.6604, + "num_input_tokens_seen": 289433075, + "step": 8904 + }, + { + "epoch": 0.803084276502683, + "flos": 27997990414080.0, + "grad_norm": 1.9951611633981172, + "learning_rate": 3.930250297547728e-07, + "loss": 0.7457, + "num_input_tokens_seen": 289462750, + "step": 8905 + }, + { + "epoch": 0.8031744600261532, + "flos": 24094143251520.0, + "grad_norm": 4.260520927564932, + "learning_rate": 3.9267730508532513e-07, + "loss": 0.6257, + "num_input_tokens_seen": 289492730, + "step": 8906 + }, + { + "epoch": 0.8032646435496235, + "flos": 20962878446880.0, + "grad_norm": 2.352055596142181, + "learning_rate": 3.923297175615121e-07, + "loss": 0.7359, + "num_input_tokens_seen": 289520135, + "step": 8907 + }, + { + "epoch": 0.8033548270730937, + "flos": 28285470036960.0, + "grad_norm": 2.270177857409516, + "learning_rate": 3.9198226721299243e-07, + "loss": 0.7856, + "num_input_tokens_seen": 289552015, + "step": 8908 + }, + { + "epoch": 0.803445010596564, + "flos": 16266185673600.0, + "grad_norm": 3.3589771123834877, + "learning_rate": 3.916349540694128e-07, + "loss": 0.7243, + "num_input_tokens_seen": 289578665, + "step": 8909 + }, + { + "epoch": 0.8035351941200343, + "flos": 28690423302240.0, + "grad_norm": 2.10542866344898, + "learning_rate": 3.912877781604063e-07, + "loss": 0.7551, + "num_input_tokens_seen": 289607535, + "step": 8910 + }, + { + "epoch": 0.8036253776435045, + "flos": 23581417651200.0, + "grad_norm": 4.074652702650472, + "learning_rate": 3.909407395155977e-07, + "loss": 0.7736, + "num_input_tokens_seen": 289633635, + "step": 8911 + }, + { + "epoch": 0.8037155611669748, + "flos": 20783617196160.0, + "grad_norm": 2.4165606859141886, + "learning_rate": 3.9059383816459725e-07, + "loss": 0.6875, + "num_input_tokens_seen": 289661145, + "step": 8912 + }, + { + "epoch": 0.8038057446904451, + "flos": 25224579055200.0, + "grad_norm": 2.465062634985788, + "learning_rate": 3.902470741370045e-07, + "loss": 0.8052, + "num_input_tokens_seen": 289690900, + "step": 8913 + }, + { + "epoch": 0.8038959282139153, + "flos": 26758667182560.0, + "grad_norm": 1.4913289261401668, + "learning_rate": 3.8990044746240746e-07, + "loss": 0.7329, + "num_input_tokens_seen": 289722585, + "step": 8914 + }, + { + "epoch": 0.8039861117373855, + "flos": 25046358557760.0, + "grad_norm": 2.327240231744983, + "learning_rate": 3.8955395817038237e-07, + "loss": 0.8422, + "num_input_tokens_seen": 289748775, + "step": 8915 + }, + { + "epoch": 0.8040762952608559, + "flos": 25192793176800.0, + "grad_norm": 1.7143836305674318, + "learning_rate": 3.892076062904934e-07, + "loss": 0.712, + "num_input_tokens_seen": 289779190, + "step": 8916 + }, + { + "epoch": 0.8041664787843261, + "flos": 15209840668800.0, + "grad_norm": 2.1936506721637277, + "learning_rate": 3.8886139185229384e-07, + "loss": 0.7041, + "num_input_tokens_seen": 289806325, + "step": 8917 + }, + { + "epoch": 0.8042566623077964, + "flos": 20086761106560.0, + "grad_norm": 1.8587873155225516, + "learning_rate": 3.8851531488532284e-07, + "loss": 0.6618, + "num_input_tokens_seen": 289833515, + "step": 8918 + }, + { + "epoch": 0.8043468458312666, + "flos": 21511069731840.0, + "grad_norm": 1.7795201652436847, + "learning_rate": 3.88169375419112e-07, + "loss": 0.7673, + "num_input_tokens_seen": 289863215, + "step": 8919 + }, + { + "epoch": 0.8044370293547369, + "flos": 43079541040320.0, + "grad_norm": 3.732278147374946, + "learning_rate": 3.8782357348317717e-07, + "loss": 0.7182, + "num_input_tokens_seen": 289895160, + "step": 8920 + }, + { + "epoch": 0.8045272128782072, + "flos": 13789434868320.0, + "grad_norm": 2.7262496987420937, + "learning_rate": 3.8747790910702437e-07, + "loss": 0.7904, + "num_input_tokens_seen": 289919850, + "step": 8921 + }, + { + "epoch": 0.8046173964016774, + "flos": 21727134778560.0, + "grad_norm": 1.987322083193213, + "learning_rate": 3.8713238232014776e-07, + "loss": 0.7444, + "num_input_tokens_seen": 289950020, + "step": 8922 + }, + { + "epoch": 0.8047075799251476, + "flos": 19363508753760.0, + "grad_norm": 1.8728362632299222, + "learning_rate": 3.867869931520296e-07, + "loss": 0.8035, + "num_input_tokens_seen": 289977200, + "step": 8923 + }, + { + "epoch": 0.804797763448618, + "flos": 27701106841920.0, + "grad_norm": 1.749899520714211, + "learning_rate": 3.864417416321406e-07, + "loss": 0.757, + "num_input_tokens_seen": 290005290, + "step": 8924 + }, + { + "epoch": 0.8048879469720882, + "flos": 23297097457920.0, + "grad_norm": 1.7547439579700015, + "learning_rate": 3.8609662778993847e-07, + "loss": 0.7267, + "num_input_tokens_seen": 290033680, + "step": 8925 + }, + { + "epoch": 0.8049781304955584, + "flos": 68111880864960.0, + "grad_norm": 0.5902273181465653, + "learning_rate": 3.85751651654872e-07, + "loss": 0.5568, + "num_input_tokens_seen": 290135140, + "step": 8926 + }, + { + "epoch": 0.8050683140190287, + "flos": 15865581270240.0, + "grad_norm": 1.978794571548644, + "learning_rate": 3.8540681325637505e-07, + "loss": 0.8564, + "num_input_tokens_seen": 290158180, + "step": 8927 + }, + { + "epoch": 0.805158497542499, + "flos": 24790925001600.0, + "grad_norm": 2.955999680742342, + "learning_rate": 3.8506211262387155e-07, + "loss": 0.7554, + "num_input_tokens_seen": 290190830, + "step": 8928 + }, + { + "epoch": 0.8052486810659693, + "flos": 26613124637760.0, + "grad_norm": 3.8983232561698533, + "learning_rate": 3.847175497867732e-07, + "loss": 0.753, + "num_input_tokens_seen": 290220255, + "step": 8929 + }, + { + "epoch": 0.8053388645894395, + "flos": 20893694056320.0, + "grad_norm": 1.8945788190709731, + "learning_rate": 3.843731247744801e-07, + "loss": 0.6943, + "num_input_tokens_seen": 290249305, + "step": 8930 + }, + { + "epoch": 0.8054290481129097, + "flos": 24277641854880.0, + "grad_norm": 1.8724736041460976, + "learning_rate": 3.8402883761638047e-07, + "loss": 0.7858, + "num_input_tokens_seen": 290276025, + "step": 8931 + }, + { + "epoch": 0.8055192316363801, + "flos": 66110085450240.0, + "grad_norm": 0.6627996666658216, + "learning_rate": 3.8368468834185076e-07, + "loss": 0.6134, + "num_input_tokens_seen": 290373585, + "step": 8932 + }, + { + "epoch": 0.8056094151598503, + "flos": 21039013582560.0, + "grad_norm": 1.8118338840710584, + "learning_rate": 3.8334067698025583e-07, + "loss": 0.7927, + "num_input_tokens_seen": 290402760, + "step": 8933 + }, + { + "epoch": 0.8056995986833205, + "flos": 21257866361280.0, + "grad_norm": 1.8109360886602028, + "learning_rate": 3.8299680356094897e-07, + "loss": 0.6779, + "num_input_tokens_seen": 290430545, + "step": 8934 + }, + { + "epoch": 0.8057897822067909, + "flos": 18525979357920.0, + "grad_norm": 3.0598221573854025, + "learning_rate": 3.8265306811327024e-07, + "loss": 0.6339, + "num_input_tokens_seen": 290459550, + "step": 8935 + }, + { + "epoch": 0.8058799657302611, + "flos": 14042340880800.0, + "grad_norm": 2.1850369346866483, + "learning_rate": 3.8230947066654994e-07, + "loss": 0.718, + "num_input_tokens_seen": 290483220, + "step": 8936 + }, + { + "epoch": 0.8059701492537313, + "flos": 20091630345120.0, + "grad_norm": 1.8675302486191345, + "learning_rate": 3.819660112501053e-07, + "loss": 0.8042, + "num_input_tokens_seen": 290512290, + "step": 8937 + }, + { + "epoch": 0.8060603327772016, + "flos": 67378258149120.0, + "grad_norm": 0.6216740566313849, + "learning_rate": 3.816226898932422e-07, + "loss": 0.5969, + "num_input_tokens_seen": 290608130, + "step": 8938 + }, + { + "epoch": 0.8061505163006719, + "flos": 24282659772480.0, + "grad_norm": 1.786206827356055, + "learning_rate": 3.812795066252557e-07, + "loss": 0.7335, + "num_input_tokens_seen": 290638980, + "step": 8939 + }, + { + "epoch": 0.8062406998241421, + "flos": 13169271460800.0, + "grad_norm": 2.295825432996317, + "learning_rate": 3.8093646147542577e-07, + "loss": 0.7604, + "num_input_tokens_seen": 290663585, + "step": 8940 + }, + { + "epoch": 0.8063308833476124, + "flos": 18954875682240.0, + "grad_norm": 2.1561400896891523, + "learning_rate": 3.805935544730259e-07, + "loss": 0.7151, + "num_input_tokens_seen": 290689675, + "step": 8941 + }, + { + "epoch": 0.8064210668710826, + "flos": 25921546654080.0, + "grad_norm": 2.5879924250319406, + "learning_rate": 3.802507856473118e-07, + "loss": 0.6837, + "num_input_tokens_seen": 290719905, + "step": 8942 + }, + { + "epoch": 0.806511250394553, + "flos": 19216999795200.0, + "grad_norm": 2.423436209249737, + "learning_rate": 3.7990815502753317e-07, + "loss": 0.7501, + "num_input_tokens_seen": 290744980, + "step": 8943 + }, + { + "epoch": 0.8066014339180232, + "flos": 19866904744320.0, + "grad_norm": 1.8049371059587298, + "learning_rate": 3.795656626429231e-07, + "loss": 0.7517, + "num_input_tokens_seen": 290772435, + "step": 8944 + }, + { + "epoch": 0.8066916174414934, + "flos": 25775483732640.0, + "grad_norm": 17.55103468059291, + "learning_rate": 3.792233085227059e-07, + "loss": 0.6649, + "num_input_tokens_seen": 290800985, + "step": 8945 + }, + { + "epoch": 0.8067818009649637, + "flos": 22125992203200.0, + "grad_norm": 2.3746230177590184, + "learning_rate": 3.788810926960928e-07, + "loss": 0.7567, + "num_input_tokens_seen": 290827075, + "step": 8946 + }, + { + "epoch": 0.806871984488434, + "flos": 24496308784800.0, + "grad_norm": 1.8625625087822137, + "learning_rate": 3.785390151922836e-07, + "loss": 0.7325, + "num_input_tokens_seen": 290855910, + "step": 8947 + }, + { + "epoch": 0.8069621680119042, + "flos": 20492643615840.0, + "grad_norm": 2.2292427929390297, + "learning_rate": 3.781970760404665e-07, + "loss": 0.7293, + "num_input_tokens_seen": 290881945, + "step": 8948 + }, + { + "epoch": 0.8070523515353745, + "flos": 29600816894880.0, + "grad_norm": 2.098781063901882, + "learning_rate": 3.778552752698176e-07, + "loss": 0.823, + "num_input_tokens_seen": 290910410, + "step": 8949 + }, + { + "epoch": 0.8071425350588447, + "flos": 39692619660960.0, + "grad_norm": 1.8222856428512644, + "learning_rate": 3.775136129095007e-07, + "loss": 0.6592, + "num_input_tokens_seen": 290943535, + "step": 8950 + }, + { + "epoch": 0.807232718582315, + "flos": 23371225426560.0, + "grad_norm": 2.567707074663474, + "learning_rate": 3.771720889886685e-07, + "loss": 0.7178, + "num_input_tokens_seen": 290973790, + "step": 8951 + }, + { + "epoch": 0.8073229021057853, + "flos": 61910804336160.0, + "grad_norm": 0.6916243430099525, + "learning_rate": 3.7683070353646194e-07, + "loss": 0.5954, + "num_input_tokens_seen": 291057440, + "step": 8952 + }, + { + "epoch": 0.8074130856292555, + "flos": 26285495940480.0, + "grad_norm": 1.9643439990227225, + "learning_rate": 3.7648945658200983e-07, + "loss": 0.7068, + "num_input_tokens_seen": 291085630, + "step": 8953 + }, + { + "epoch": 0.8075032691527257, + "flos": 39473766882240.0, + "grad_norm": 1.8746958803859781, + "learning_rate": 3.761483481544292e-07, + "loss": 0.634, + "num_input_tokens_seen": 291117695, + "step": 8954 + }, + { + "epoch": 0.8075934526761961, + "flos": 20417883761280.0, + "grad_norm": 1.9112208857292898, + "learning_rate": 3.7580737828282525e-07, + "loss": 0.7146, + "num_input_tokens_seen": 291145995, + "step": 8955 + }, + { + "epoch": 0.8076836361996663, + "flos": 20419853758560.0, + "grad_norm": 1.7081921294246858, + "learning_rate": 3.754665469962921e-07, + "loss": 0.7246, + "num_input_tokens_seen": 291174250, + "step": 8956 + }, + { + "epoch": 0.8077738197231366, + "flos": 21147752331360.0, + "grad_norm": 2.1526913232431872, + "learning_rate": 3.7512585432390973e-07, + "loss": 0.7013, + "num_input_tokens_seen": 291202510, + "step": 8957 + }, + { + "epoch": 0.8078640032466069, + "flos": 14772581148480.0, + "grad_norm": 1.8130700760484222, + "learning_rate": 3.7478530029474987e-07, + "loss": 0.6831, + "num_input_tokens_seen": 291228070, + "step": 8958 + }, + { + "epoch": 0.8079541867700771, + "flos": 70567476343200.0, + "grad_norm": 0.6191898651278103, + "learning_rate": 3.7444488493786854e-07, + "loss": 0.6273, + "num_input_tokens_seen": 291317280, + "step": 8959 + }, + { + "epoch": 0.8080443702935474, + "flos": 19035136661280.0, + "grad_norm": 2.0347063442242868, + "learning_rate": 3.7410460828231405e-07, + "loss": 0.7947, + "num_input_tokens_seen": 291344460, + "step": 8960 + }, + { + "epoch": 0.8081345538170176, + "flos": 31936002319680.0, + "grad_norm": 1.4849482985207887, + "learning_rate": 3.737644703571188e-07, + "loss": 0.6805, + "num_input_tokens_seen": 291378930, + "step": 8961 + }, + { + "epoch": 0.8082247373404879, + "flos": 23079062413920.0, + "grad_norm": 4.971991099711007, + "learning_rate": 3.734244711913059e-07, + "loss": 0.8158, + "num_input_tokens_seen": 291404700, + "step": 8962 + }, + { + "epoch": 0.8083149208639582, + "flos": 25520198855520.0, + "grad_norm": 2.1285333989367903, + "learning_rate": 3.7308461081388584e-07, + "loss": 0.7208, + "num_input_tokens_seen": 291433930, + "step": 8963 + }, + { + "epoch": 0.8084051043874284, + "flos": 24791519717760.0, + "grad_norm": 1.6093748518564102, + "learning_rate": 3.727448892538576e-07, + "loss": 0.6744, + "num_input_tokens_seen": 291465505, + "step": 8964 + }, + { + "epoch": 0.8084952879108986, + "flos": 25739237483040.0, + "grad_norm": 1.9961279700046508, + "learning_rate": 3.724053065402086e-07, + "loss": 0.8307, + "num_input_tokens_seen": 291496870, + "step": 8965 + }, + { + "epoch": 0.808585471434369, + "flos": 22751582395680.0, + "grad_norm": 1.6499276318719038, + "learning_rate": 3.7206586270191285e-07, + "loss": 0.7111, + "num_input_tokens_seen": 291528210, + "step": 8966 + }, + { + "epoch": 0.8086756549578392, + "flos": 24245261260320.0, + "grad_norm": 1.9329734963397402, + "learning_rate": 3.7172655776793385e-07, + "loss": 0.7908, + "num_input_tokens_seen": 291557155, + "step": 8967 + }, + { + "epoch": 0.8087658384813095, + "flos": 22715224636800.0, + "grad_norm": 2.1567521547161044, + "learning_rate": 3.7138739176722323e-07, + "loss": 0.6377, + "num_input_tokens_seen": 291585125, + "step": 8968 + }, + { + "epoch": 0.8088560220047797, + "flos": 25192309969920.0, + "grad_norm": 2.131006279487255, + "learning_rate": 3.710483647287206e-07, + "loss": 0.7124, + "num_input_tokens_seen": 291614090, + "step": 8969 + }, + { + "epoch": 0.80894620552825, + "flos": 17906633685120.0, + "grad_norm": 1.7117180981410642, + "learning_rate": 3.707094766813532e-07, + "loss": 0.6423, + "num_input_tokens_seen": 291640855, + "step": 8970 + }, + { + "epoch": 0.8090363890517203, + "flos": 21221731620960.0, + "grad_norm": 2.1215179187845954, + "learning_rate": 3.7037072765403754e-07, + "loss": 0.7092, + "num_input_tokens_seen": 291669665, + "step": 8971 + }, + { + "epoch": 0.8091265725751905, + "flos": 22785970157280.0, + "grad_norm": 4.269509806670762, + "learning_rate": 3.700321176756762e-07, + "loss": 0.698, + "num_input_tokens_seen": 291697925, + "step": 8972 + }, + { + "epoch": 0.8092167560986607, + "flos": 21658433594880.0, + "grad_norm": 1.8528663133053311, + "learning_rate": 3.69693646775163e-07, + "loss": 0.7879, + "num_input_tokens_seen": 291726095, + "step": 8973 + }, + { + "epoch": 0.8093069396221311, + "flos": 22824706780800.0, + "grad_norm": 2.4505488968045266, + "learning_rate": 3.693553149813764e-07, + "loss": 0.727, + "num_input_tokens_seen": 291754030, + "step": 8974 + }, + { + "epoch": 0.8093971231456013, + "flos": 60679807130880.0, + "grad_norm": 0.5775547565989227, + "learning_rate": 3.690171223231866e-07, + "loss": 0.5509, + "num_input_tokens_seen": 291847200, + "step": 8975 + }, + { + "epoch": 0.8094873066690715, + "flos": 26283154245600.0, + "grad_norm": 2.112992074752077, + "learning_rate": 3.6867906882944854e-07, + "loss": 0.8344, + "num_input_tokens_seen": 291877430, + "step": 8976 + }, + { + "epoch": 0.8095774901925418, + "flos": 21580328461920.0, + "grad_norm": 2.551536086179925, + "learning_rate": 3.6834115452900737e-07, + "loss": 0.7447, + "num_input_tokens_seen": 291906680, + "step": 8977 + }, + { + "epoch": 0.8096676737160121, + "flos": 20746999248960.0, + "grad_norm": 1.8693026130662496, + "learning_rate": 3.680033794506958e-07, + "loss": 0.6446, + "num_input_tokens_seen": 291934735, + "step": 8978 + }, + { + "epoch": 0.8097578572394823, + "flos": 15646728491520.0, + "grad_norm": 2.121978519032555, + "learning_rate": 3.676657436233346e-07, + "loss": 0.5761, + "num_input_tokens_seen": 291962690, + "step": 8979 + }, + { + "epoch": 0.8098480407629526, + "flos": 18007418105280.0, + "grad_norm": 1.9030249157475723, + "learning_rate": 3.6732824707573305e-07, + "loss": 0.7398, + "num_input_tokens_seen": 291987890, + "step": 8980 + }, + { + "epoch": 0.8099382242864229, + "flos": 24135890625600.0, + "grad_norm": 2.364458279656842, + "learning_rate": 3.6699088983668716e-07, + "loss": 0.7803, + "num_input_tokens_seen": 292015195, + "step": 8981 + }, + { + "epoch": 0.8100284078098932, + "flos": 25698010485600.0, + "grad_norm": 2.365462074061223, + "learning_rate": 3.6665367193498376e-07, + "loss": 0.6986, + "num_input_tokens_seen": 292044210, + "step": 8982 + }, + { + "epoch": 0.8101185913333634, + "flos": 18124594389600.0, + "grad_norm": 2.495364469155426, + "learning_rate": 3.663165933993948e-07, + "loss": 0.7476, + "num_input_tokens_seen": 292070805, + "step": 8983 + }, + { + "epoch": 0.8102087748568336, + "flos": 23475987011040.0, + "grad_norm": 3.370156096966418, + "learning_rate": 3.659796542586822e-07, + "loss": 0.7689, + "num_input_tokens_seen": 292098950, + "step": 8984 + }, + { + "epoch": 0.810298958380304, + "flos": 24240020324160.0, + "grad_norm": 1.7881254049098687, + "learning_rate": 3.6564285454159526e-07, + "loss": 0.702, + "num_input_tokens_seen": 292127125, + "step": 8985 + }, + { + "epoch": 0.8103891419037742, + "flos": 20963882030400.0, + "grad_norm": 1.6819083502449403, + "learning_rate": 3.653061942768718e-07, + "loss": 0.7902, + "num_input_tokens_seen": 292156695, + "step": 8986 + }, + { + "epoch": 0.8104793254272444, + "flos": 54119098741440.0, + "grad_norm": 3.4256987088741377, + "learning_rate": 3.649696734932375e-07, + "loss": 0.6975, + "num_input_tokens_seen": 292189015, + "step": 8987 + }, + { + "epoch": 0.8105695089507147, + "flos": 26319995211360.0, + "grad_norm": 2.249107077695977, + "learning_rate": 3.646332922194064e-07, + "loss": 0.6743, + "num_input_tokens_seen": 292219705, + "step": 8988 + }, + { + "epoch": 0.810659692474185, + "flos": 23587513491840.0, + "grad_norm": 1.9088211230184446, + "learning_rate": 3.6429705048407943e-07, + "loss": 0.7289, + "num_input_tokens_seen": 292245060, + "step": 8989 + }, + { + "epoch": 0.8107498759976552, + "flos": 34117765210560.0, + "grad_norm": 2.7288692302579083, + "learning_rate": 3.6396094831594804e-07, + "loss": 0.7353, + "num_input_tokens_seen": 292273760, + "step": 8990 + }, + { + "epoch": 0.8108400595211255, + "flos": 29673606752160.0, + "grad_norm": 2.0981057252018442, + "learning_rate": 3.6362498574368926e-07, + "loss": 0.696, + "num_input_tokens_seen": 292301710, + "step": 8991 + }, + { + "epoch": 0.8109302430445957, + "flos": 62241220765440.0, + "grad_norm": 0.7054061338025508, + "learning_rate": 3.6328916279596935e-07, + "loss": 0.6223, + "num_input_tokens_seen": 292389975, + "step": 8992 + }, + { + "epoch": 0.811020426568066, + "flos": 13971223662720.0, + "grad_norm": 1.9784764006361424, + "learning_rate": 3.6295347950144305e-07, + "loss": 0.7848, + "num_input_tokens_seen": 292417010, + "step": 8993 + }, + { + "epoch": 0.8111106100915363, + "flos": 31671499342080.0, + "grad_norm": 1.9855637481530872, + "learning_rate": 3.626179358887522e-07, + "loss": 0.5931, + "num_input_tokens_seen": 292446560, + "step": 8994 + }, + { + "epoch": 0.8112007936150065, + "flos": 22350383276160.0, + "grad_norm": 2.0555972561887894, + "learning_rate": 3.6228253198652816e-07, + "loss": 0.8017, + "num_input_tokens_seen": 292477230, + "step": 8995 + }, + { + "epoch": 0.8112909771384768, + "flos": 22423507661280.0, + "grad_norm": 1.8874404190461798, + "learning_rate": 3.6194726782338767e-07, + "loss": 0.6929, + "num_input_tokens_seen": 292506590, + "step": 8996 + }, + { + "epoch": 0.8113811606619471, + "flos": 17171933046240.0, + "grad_norm": 2.756166912738197, + "learning_rate": 3.6161214342793953e-07, + "loss": 0.7153, + "num_input_tokens_seen": 292534450, + "step": 8997 + }, + { + "epoch": 0.8114713441854173, + "flos": 21547873527840.0, + "grad_norm": 1.9380562506780437, + "learning_rate": 3.612771588287764e-07, + "loss": 0.7711, + "num_input_tokens_seen": 292562250, + "step": 8998 + }, + { + "epoch": 0.8115615277088876, + "flos": 15534161257440.0, + "grad_norm": 2.477920785501419, + "learning_rate": 3.609423140544827e-07, + "loss": 0.8065, + "num_input_tokens_seen": 292582425, + "step": 8999 + }, + { + "epoch": 0.8116517112323578, + "flos": 22459642401600.0, + "grad_norm": 2.701830427167021, + "learning_rate": 3.6060760913362787e-07, + "loss": 0.7588, + "num_input_tokens_seen": 292608105, + "step": 9000 + }, + { + "epoch": 0.8117418947558281, + "flos": 34263121906560.0, + "grad_norm": 2.9979703713264203, + "learning_rate": 3.6027304409477146e-07, + "loss": 0.7498, + "num_input_tokens_seen": 292638730, + "step": 9001 + }, + { + "epoch": 0.8118320782792984, + "flos": 22605296455680.0, + "grad_norm": 1.5267556075184763, + "learning_rate": 3.599386189664604e-07, + "loss": 0.7175, + "num_input_tokens_seen": 292668055, + "step": 9002 + }, + { + "epoch": 0.8119222618027686, + "flos": 17869383852000.0, + "grad_norm": 2.746185518253507, + "learning_rate": 3.5960433377722945e-07, + "loss": 0.7272, + "num_input_tokens_seen": 292694660, + "step": 9003 + }, + { + "epoch": 0.8120124453262388, + "flos": 26795024941440.0, + "grad_norm": 2.096237185671633, + "learning_rate": 3.5927018855560174e-07, + "loss": 0.8012, + "num_input_tokens_seen": 292722840, + "step": 9004 + }, + { + "epoch": 0.8121026288497092, + "flos": 34477811672160.0, + "grad_norm": 1.9865748909525867, + "learning_rate": 3.5893618333008904e-07, + "loss": 0.7356, + "num_input_tokens_seen": 292754130, + "step": 9005 + }, + { + "epoch": 0.8121928123731794, + "flos": 25331719504320.0, + "grad_norm": 2.002587313802286, + "learning_rate": 3.586023181291893e-07, + "loss": 0.8313, + "num_input_tokens_seen": 292782010, + "step": 9006 + }, + { + "epoch": 0.8122829958966497, + "flos": 69928393527840.0, + "grad_norm": 0.7490126945988479, + "learning_rate": 3.5826859298139044e-07, + "loss": 0.6096, + "num_input_tokens_seen": 292878480, + "step": 9007 + }, + { + "epoch": 0.81237317942012, + "flos": 30067817956800.0, + "grad_norm": 1.7162071143763609, + "learning_rate": 3.5793500791516773e-07, + "loss": 0.7241, + "num_input_tokens_seen": 292904895, + "step": 9008 + }, + { + "epoch": 0.8124633629435902, + "flos": 21768398945760.0, + "grad_norm": 1.7578085666188632, + "learning_rate": 3.5760156295898415e-07, + "loss": 0.7122, + "num_input_tokens_seen": 292930960, + "step": 9009 + }, + { + "epoch": 0.8125535464670605, + "flos": 28143644468160.0, + "grad_norm": 2.1708045352287426, + "learning_rate": 3.5726825814129203e-07, + "loss": 0.7382, + "num_input_tokens_seen": 292959220, + "step": 9010 + }, + { + "epoch": 0.8126437299905307, + "flos": 69362912570880.0, + "grad_norm": 0.6965447126387632, + "learning_rate": 3.5693509349052886e-07, + "loss": 0.6123, + "num_input_tokens_seen": 293044040, + "step": 9011 + }, + { + "epoch": 0.812733913514001, + "flos": 23874324059040.0, + "grad_norm": 1.6969382932882466, + "learning_rate": 3.5660206903512433e-07, + "loss": 0.8127, + "num_input_tokens_seen": 293071420, + "step": 9012 + }, + { + "epoch": 0.8128240970374713, + "flos": 17687632227360.0, + "grad_norm": 1.8459325900060146, + "learning_rate": 3.56269184803492e-07, + "loss": 0.7971, + "num_input_tokens_seen": 293097955, + "step": 9013 + }, + { + "epoch": 0.8129142805609415, + "flos": 22390681029600.0, + "grad_norm": 1.5404662880383069, + "learning_rate": 3.5593644082403727e-07, + "loss": 0.7645, + "num_input_tokens_seen": 293128770, + "step": 9014 + }, + { + "epoch": 0.8130044640844117, + "flos": 21112695514080.0, + "grad_norm": 2.149787173715796, + "learning_rate": 3.5560383712514994e-07, + "loss": 0.8457, + "num_input_tokens_seen": 293156865, + "step": 9015 + }, + { + "epoch": 0.8130946476078821, + "flos": 23476024180800.0, + "grad_norm": 1.6956669309488526, + "learning_rate": 3.5527137373521066e-07, + "loss": 0.7458, + "num_input_tokens_seen": 293185240, + "step": 9016 + }, + { + "epoch": 0.8131848311313523, + "flos": 26941236541920.0, + "grad_norm": 1.5324201200243264, + "learning_rate": 3.5493905068258645e-07, + "loss": 0.7012, + "num_input_tokens_seen": 293214045, + "step": 9017 + }, + { + "epoch": 0.8132750146548225, + "flos": 29928631440960.0, + "grad_norm": 1.6347569473382664, + "learning_rate": 3.546068679956333e-07, + "loss": 0.7575, + "num_input_tokens_seen": 293246780, + "step": 9018 + }, + { + "epoch": 0.8133651981782928, + "flos": 22532172070560.0, + "grad_norm": 1.6435913879685085, + "learning_rate": 3.5427482570269487e-07, + "loss": 0.8049, + "num_input_tokens_seen": 293277355, + "step": 9019 + }, + { + "epoch": 0.8134553817017631, + "flos": 22641914402880.0, + "grad_norm": 3.640336708802441, + "learning_rate": 3.539429238321026e-07, + "loss": 0.5895, + "num_input_tokens_seen": 293302675, + "step": 9020 + }, + { + "epoch": 0.8135455652252334, + "flos": 23480075684640.0, + "grad_norm": 1.7172676630026704, + "learning_rate": 3.536111624121769e-07, + "loss": 0.5939, + "num_input_tokens_seen": 293329965, + "step": 9021 + }, + { + "epoch": 0.8136357487487036, + "flos": 42240190326240.0, + "grad_norm": 1.9955672971445642, + "learning_rate": 3.532795414712244e-07, + "loss": 0.7065, + "num_input_tokens_seen": 293361160, + "step": 9022 + }, + { + "epoch": 0.8137259322721738, + "flos": 24865201649280.0, + "grad_norm": 1.767984012768817, + "learning_rate": 3.5294806103754124e-07, + "loss": 0.7599, + "num_input_tokens_seen": 293389685, + "step": 9023 + }, + { + "epoch": 0.8138161157956442, + "flos": 23953692963840.0, + "grad_norm": 2.4213123283182343, + "learning_rate": 3.526167211394115e-07, + "loss": 0.7563, + "num_input_tokens_seen": 293418820, + "step": 9024 + }, + { + "epoch": 0.8139062993191144, + "flos": 38237565910560.0, + "grad_norm": 1.6904337071415596, + "learning_rate": 3.522855218051066e-07, + "loss": 0.6838, + "num_input_tokens_seen": 293457085, + "step": 9025 + }, + { + "epoch": 0.8139964828425846, + "flos": 27187600676640.0, + "grad_norm": 1.8169295674866974, + "learning_rate": 3.5195446306288633e-07, + "loss": 0.7448, + "num_input_tokens_seen": 293480350, + "step": 9026 + }, + { + "epoch": 0.8140866663660549, + "flos": 34590155887680.0, + "grad_norm": 1.6141774641603763, + "learning_rate": 3.51623544940999e-07, + "loss": 0.6864, + "num_input_tokens_seen": 293514585, + "step": 9027 + }, + { + "epoch": 0.8141768498895252, + "flos": 23734951694400.0, + "grad_norm": 1.777812384559975, + "learning_rate": 3.5129276746767886e-07, + "loss": 0.7296, + "num_input_tokens_seen": 293545555, + "step": 9028 + }, + { + "epoch": 0.8142670334129954, + "flos": 22932962322720.0, + "grad_norm": 2.0047330879385745, + "learning_rate": 3.5096213067115165e-07, + "loss": 0.7167, + "num_input_tokens_seen": 293575170, + "step": 9029 + }, + { + "epoch": 0.8143572169364657, + "flos": 24639026427840.0, + "grad_norm": 2.059545539428418, + "learning_rate": 3.506316345796272e-07, + "loss": 0.7606, + "num_input_tokens_seen": 293602280, + "step": 9030 + }, + { + "epoch": 0.814447400459936, + "flos": 72160936044480.0, + "grad_norm": 0.5943065343949142, + "learning_rate": 3.5030127922130714e-07, + "loss": 0.5352, + "num_input_tokens_seen": 293701520, + "step": 9031 + }, + { + "epoch": 0.8145375839834063, + "flos": 18884687708160.0, + "grad_norm": 1.6715222280646418, + "learning_rate": 3.4997106462437784e-07, + "loss": 0.6328, + "num_input_tokens_seen": 293729055, + "step": 9032 + }, + { + "epoch": 0.8146277675068765, + "flos": 69737498142240.0, + "grad_norm": 0.7279646275279348, + "learning_rate": 3.496409908170157e-07, + "loss": 0.5606, + "num_input_tokens_seen": 293812365, + "step": 9033 + }, + { + "epoch": 0.8147179510303467, + "flos": 22641914402880.0, + "grad_norm": 1.8834342721316786, + "learning_rate": 3.493110578273839e-07, + "loss": 0.7086, + "num_input_tokens_seen": 293843230, + "step": 9034 + }, + { + "epoch": 0.8148081345538171, + "flos": 21366605110080.0, + "grad_norm": 3.0528373090751417, + "learning_rate": 3.489812656836346e-07, + "loss": 0.7976, + "num_input_tokens_seen": 293870810, + "step": 9035 + }, + { + "epoch": 0.8148983180772873, + "flos": 28690274623200.0, + "grad_norm": 1.6289684549067545, + "learning_rate": 3.486516144139078e-07, + "loss": 0.6578, + "num_input_tokens_seen": 293902085, + "step": 9036 + }, + { + "epoch": 0.8149885016007575, + "flos": 26100584886240.0, + "grad_norm": 1.870401975296844, + "learning_rate": 3.4832210404632957e-07, + "loss": 0.7134, + "num_input_tokens_seen": 293932145, + "step": 9037 + }, + { + "epoch": 0.8150786851242278, + "flos": 22857421903200.0, + "grad_norm": 1.5163706881929035, + "learning_rate": 3.479927346090179e-07, + "loss": 0.7497, + "num_input_tokens_seen": 293962735, + "step": 9038 + }, + { + "epoch": 0.8151688686476981, + "flos": 14553505351200.0, + "grad_norm": 1.9341533429126416, + "learning_rate": 3.4766350613007455e-07, + "loss": 0.7285, + "num_input_tokens_seen": 293987475, + "step": 9039 + }, + { + "epoch": 0.8152590521711683, + "flos": 24644230194240.0, + "grad_norm": 1.717462994150455, + "learning_rate": 3.4733441863759173e-07, + "loss": 0.73, + "num_input_tokens_seen": 294015565, + "step": 9040 + }, + { + "epoch": 0.8153492356946386, + "flos": 21874647320640.0, + "grad_norm": 2.170993104450199, + "learning_rate": 3.4700547215964916e-07, + "loss": 0.6323, + "num_input_tokens_seen": 294047050, + "step": 9041 + }, + { + "epoch": 0.8154394192181088, + "flos": 29417466970560.0, + "grad_norm": 1.86848589711817, + "learning_rate": 3.46676666724314e-07, + "loss": 0.7231, + "num_input_tokens_seen": 294078690, + "step": 9042 + }, + { + "epoch": 0.8155296027415792, + "flos": 19545000190080.0, + "grad_norm": 2.5131549911398157, + "learning_rate": 3.463480023596421e-07, + "loss": 0.7324, + "num_input_tokens_seen": 294107100, + "step": 9043 + }, + { + "epoch": 0.8156197862650494, + "flos": 25625555156160.0, + "grad_norm": 2.02234764647534, + "learning_rate": 3.460194790936772e-07, + "loss": 0.8142, + "num_input_tokens_seen": 294135010, + "step": 9044 + }, + { + "epoch": 0.8157099697885196, + "flos": 22494550539840.0, + "grad_norm": 1.6573471178344448, + "learning_rate": 3.456910969544495e-07, + "loss": 0.7314, + "num_input_tokens_seen": 294164590, + "step": 9045 + }, + { + "epoch": 0.8158001533119899, + "flos": 17796519655200.0, + "grad_norm": 2.199741877238566, + "learning_rate": 3.4536285596997994e-07, + "loss": 0.7737, + "num_input_tokens_seen": 294190480, + "step": 9046 + }, + { + "epoch": 0.8158903368354602, + "flos": 39801618598080.0, + "grad_norm": 1.9153390169976618, + "learning_rate": 3.450347561682747e-07, + "loss": 0.7038, + "num_input_tokens_seen": 294219685, + "step": 9047 + }, + { + "epoch": 0.8159805203589304, + "flos": 31160446380960.0, + "grad_norm": 2.0646648027787107, + "learning_rate": 3.4470679757732945e-07, + "loss": 0.686, + "num_input_tokens_seen": 294246925, + "step": 9048 + }, + { + "epoch": 0.8160707038824007, + "flos": 22605370795200.0, + "grad_norm": 13.868887989927254, + "learning_rate": 3.4437898022512735e-07, + "loss": 0.8226, + "num_input_tokens_seen": 294271970, + "step": 9049 + }, + { + "epoch": 0.8161608874058709, + "flos": 23370816559200.0, + "grad_norm": 2.0792634756704893, + "learning_rate": 3.4405130413963977e-07, + "loss": 0.7441, + "num_input_tokens_seen": 294299935, + "step": 9050 + }, + { + "epoch": 0.8162510709293412, + "flos": 21147454973280.0, + "grad_norm": 2.7027769360786564, + "learning_rate": 3.437237693488262e-07, + "loss": 0.8133, + "num_input_tokens_seen": 294329140, + "step": 9051 + }, + { + "epoch": 0.8163412544528115, + "flos": 21329838483840.0, + "grad_norm": 1.6681350517899118, + "learning_rate": 3.433963758806322e-07, + "loss": 0.758, + "num_input_tokens_seen": 294357960, + "step": 9052 + }, + { + "epoch": 0.8164314379762817, + "flos": 15789334625280.0, + "grad_norm": 1.84867904774541, + "learning_rate": 3.430691237629948e-07, + "loss": 0.6096, + "num_input_tokens_seen": 294384395, + "step": 9053 + }, + { + "epoch": 0.816521621499752, + "flos": 48762762541920.0, + "grad_norm": 1.9010031360919415, + "learning_rate": 3.427420130238354e-07, + "loss": 0.6902, + "num_input_tokens_seen": 294416805, + "step": 9054 + }, + { + "epoch": 0.8166118050232223, + "flos": 36845117503200.0, + "grad_norm": 2.214461789843473, + "learning_rate": 3.424150436910658e-07, + "loss": 0.8269, + "num_input_tokens_seen": 294447675, + "step": 9055 + }, + { + "epoch": 0.8167019885466925, + "flos": 23769042097920.0, + "grad_norm": 2.2471239994537235, + "learning_rate": 3.420882157925842e-07, + "loss": 0.7998, + "num_input_tokens_seen": 294470105, + "step": 9056 + }, + { + "epoch": 0.8167921720701627, + "flos": 18816804259200.0, + "grad_norm": 2.5680896922042677, + "learning_rate": 3.417615293562777e-07, + "loss": 0.7202, + "num_input_tokens_seen": 294498465, + "step": 9057 + }, + { + "epoch": 0.8168823555936331, + "flos": 26758109636160.0, + "grad_norm": 1.674365176705667, + "learning_rate": 3.4143498441002105e-07, + "loss": 0.6535, + "num_input_tokens_seen": 294529970, + "step": 9058 + }, + { + "epoch": 0.8169725391171033, + "flos": 25046804594880.0, + "grad_norm": 1.7302745425949244, + "learning_rate": 3.411085809816767e-07, + "loss": 0.7489, + "num_input_tokens_seen": 294558455, + "step": 9059 + }, + { + "epoch": 0.8170627226405736, + "flos": 17833397790720.0, + "grad_norm": 2.0818732167685527, + "learning_rate": 3.407823190990953e-07, + "loss": 0.8547, + "num_input_tokens_seen": 294584815, + "step": 9060 + }, + { + "epoch": 0.8171529061640438, + "flos": 22313728159200.0, + "grad_norm": 5.5912084506371755, + "learning_rate": 3.4045619879011577e-07, + "loss": 0.8239, + "num_input_tokens_seen": 294608975, + "step": 9061 + }, + { + "epoch": 0.8172430896875141, + "flos": 31636702713120.0, + "grad_norm": 1.9559730211827377, + "learning_rate": 3.4013022008256334e-07, + "loss": 0.6732, + "num_input_tokens_seen": 294636920, + "step": 9062 + }, + { + "epoch": 0.8173332732109844, + "flos": 22059669884160.0, + "grad_norm": 1.8782106884595056, + "learning_rate": 3.398043830042532e-07, + "loss": 0.6818, + "num_input_tokens_seen": 294666175, + "step": 9063 + }, + { + "epoch": 0.8174234567344546, + "flos": 19874190017280.0, + "grad_norm": 2.0432386639410205, + "learning_rate": 3.394786875829871e-07, + "loss": 0.741, + "num_input_tokens_seen": 294693820, + "step": 9064 + }, + { + "epoch": 0.8175136402579248, + "flos": 36264842981760.0, + "grad_norm": 1.8756581740733085, + "learning_rate": 3.3915313384655564e-07, + "loss": 0.6236, + "num_input_tokens_seen": 294728055, + "step": 9065 + }, + { + "epoch": 0.8176038237813952, + "flos": 26211516650880.0, + "grad_norm": 1.6951393003809223, + "learning_rate": 3.388277218227369e-07, + "loss": 0.7582, + "num_input_tokens_seen": 294758815, + "step": 9066 + }, + { + "epoch": 0.8176940073048654, + "flos": 18014963566560.0, + "grad_norm": 2.5824412887413613, + "learning_rate": 3.3850245153929557e-07, + "loss": 0.762, + "num_input_tokens_seen": 294783425, + "step": 9067 + }, + { + "epoch": 0.8177841908283356, + "flos": 22675707448320.0, + "grad_norm": 2.297810993087733, + "learning_rate": 3.381773230239875e-07, + "loss": 0.6989, + "num_input_tokens_seen": 294810895, + "step": 9068 + }, + { + "epoch": 0.8178743743518059, + "flos": 22350792143520.0, + "grad_norm": 1.686353718004735, + "learning_rate": 3.3785233630455247e-07, + "loss": 0.772, + "num_input_tokens_seen": 294839470, + "step": 9069 + }, + { + "epoch": 0.8179645578752762, + "flos": 24094514949120.0, + "grad_norm": 2.657850254694458, + "learning_rate": 3.375274914087221e-07, + "loss": 0.749, + "num_input_tokens_seen": 294865775, + "step": 9070 + }, + { + "epoch": 0.8180547413987465, + "flos": 34301821360320.0, + "grad_norm": 1.7381509453442958, + "learning_rate": 3.3720278836421234e-07, + "loss": 0.7148, + "num_input_tokens_seen": 294899165, + "step": 9071 + }, + { + "epoch": 0.8181449249222167, + "flos": 18343149810240.0, + "grad_norm": 2.752077358411665, + "learning_rate": 3.368782271987294e-07, + "loss": 0.7659, + "num_input_tokens_seen": 294926660, + "step": 9072 + }, + { + "epoch": 0.8182351084456869, + "flos": 19181125243200.0, + "grad_norm": 1.690613241383951, + "learning_rate": 3.3655380793996636e-07, + "loss": 0.7463, + "num_input_tokens_seen": 294952995, + "step": 9073 + }, + { + "epoch": 0.8183252919691573, + "flos": 21730740245280.0, + "grad_norm": 1.833497069011891, + "learning_rate": 3.362295306156047e-07, + "loss": 0.6626, + "num_input_tokens_seen": 294981255, + "step": 9074 + }, + { + "epoch": 0.8184154754926275, + "flos": 26025787861920.0, + "grad_norm": 1.8206329063717908, + "learning_rate": 3.3590539525331327e-07, + "loss": 0.7567, + "num_input_tokens_seen": 295012455, + "step": 9075 + }, + { + "epoch": 0.8185056590160977, + "flos": 20746404532800.0, + "grad_norm": 1.6631741727134677, + "learning_rate": 3.3558140188074967e-07, + "loss": 0.6926, + "num_input_tokens_seen": 295040835, + "step": 9076 + }, + { + "epoch": 0.8185958425395681, + "flos": 41368198829280.0, + "grad_norm": 2.1463744089353374, + "learning_rate": 3.3525755052555817e-07, + "loss": 0.7724, + "num_input_tokens_seen": 295070535, + "step": 9077 + }, + { + "epoch": 0.8186860260630383, + "flos": 61858792374720.0, + "grad_norm": 0.7289750437762238, + "learning_rate": 3.3493384121537147e-07, + "loss": 0.6477, + "num_input_tokens_seen": 295156840, + "step": 9078 + }, + { + "epoch": 0.8187762095865085, + "flos": 21945615859680.0, + "grad_norm": 1.8887083568816907, + "learning_rate": 3.3461027397781075e-07, + "loss": 0.7429, + "num_input_tokens_seen": 295184215, + "step": 9079 + }, + { + "epoch": 0.8188663931099788, + "flos": 21804124818720.0, + "grad_norm": 2.248192060805047, + "learning_rate": 3.3428684884048397e-07, + "loss": 0.7241, + "num_input_tokens_seen": 295212875, + "step": 9080 + }, + { + "epoch": 0.8189565766334491, + "flos": 17505137207520.0, + "grad_norm": 1.6239174391332503, + "learning_rate": 3.3396356583098826e-07, + "loss": 0.7409, + "num_input_tokens_seen": 295239660, + "step": 9081 + }, + { + "epoch": 0.8190467601569194, + "flos": 22350234597120.0, + "grad_norm": 1.8027744474741083, + "learning_rate": 3.3364042497690736e-07, + "loss": 0.775, + "num_input_tokens_seen": 295269720, + "step": 9082 + }, + { + "epoch": 0.8191369436803896, + "flos": 28911506266560.0, + "grad_norm": 1.7650111658969658, + "learning_rate": 3.3331742630581405e-07, + "loss": 0.7985, + "num_input_tokens_seen": 295303520, + "step": 9083 + }, + { + "epoch": 0.8192271272038598, + "flos": 33748909515840.0, + "grad_norm": 1.4906873901873832, + "learning_rate": 3.3299456984526717e-07, + "loss": 0.6906, + "num_input_tokens_seen": 295334395, + "step": 9084 + }, + { + "epoch": 0.8193173107273302, + "flos": 21838884277920.0, + "grad_norm": 1.9798522230353415, + "learning_rate": 3.3267185562281605e-07, + "loss": 0.777, + "num_input_tokens_seen": 295364380, + "step": 9085 + }, + { + "epoch": 0.8194074942508004, + "flos": 27706607966400.0, + "grad_norm": 2.0412890675456756, + "learning_rate": 3.3234928366599514e-07, + "loss": 0.7362, + "num_input_tokens_seen": 295395395, + "step": 9086 + }, + { + "epoch": 0.8194976777742706, + "flos": 19945716102720.0, + "grad_norm": 2.465003732365071, + "learning_rate": 3.3202685400232946e-07, + "loss": 0.7745, + "num_input_tokens_seen": 295422890, + "step": 9087 + }, + { + "epoch": 0.8195878612977409, + "flos": 18634569427680.0, + "grad_norm": 2.5778667898631413, + "learning_rate": 3.317045666593292e-07, + "loss": 0.7686, + "num_input_tokens_seen": 295449500, + "step": 9088 + }, + { + "epoch": 0.8196780448212112, + "flos": 17905927459680.0, + "grad_norm": 1.9674777809837016, + "learning_rate": 3.3138242166449426e-07, + "loss": 0.7508, + "num_input_tokens_seen": 295477565, + "step": 9089 + }, + { + "epoch": 0.8197682283446814, + "flos": 24355263780960.0, + "grad_norm": 1.5659274085212043, + "learning_rate": 3.310604190453117e-07, + "loss": 0.805, + "num_input_tokens_seen": 295506670, + "step": 9090 + }, + { + "epoch": 0.8198584118681517, + "flos": 46466239401120.0, + "grad_norm": 1.713665394437078, + "learning_rate": 3.307385588292566e-07, + "loss": 0.6921, + "num_input_tokens_seen": 295540390, + "step": 9091 + }, + { + "epoch": 0.8199485953916219, + "flos": 33134024214240.0, + "grad_norm": 2.015936893905953, + "learning_rate": 3.304168410437924e-07, + "loss": 0.7357, + "num_input_tokens_seen": 295568105, + "step": 9092 + }, + { + "epoch": 0.8200387789150922, + "flos": 69823408924800.0, + "grad_norm": 0.6252417573656257, + "learning_rate": 3.300952657163687e-07, + "loss": 0.603, + "num_input_tokens_seen": 295657610, + "step": 9093 + }, + { + "epoch": 0.8201289624385625, + "flos": 26795582487840.0, + "grad_norm": 2.833173934342247, + "learning_rate": 3.297738328744248e-07, + "loss": 0.7256, + "num_input_tokens_seen": 295686940, + "step": 9094 + }, + { + "epoch": 0.8202191459620327, + "flos": 25229150935680.0, + "grad_norm": 1.8369273013378593, + "learning_rate": 3.2945254254538714e-07, + "loss": 0.7354, + "num_input_tokens_seen": 295715135, + "step": 9095 + }, + { + "epoch": 0.820309329485503, + "flos": 26972762232000.0, + "grad_norm": 1.9187255112268824, + "learning_rate": 3.2913139475666963e-07, + "loss": 0.7102, + "num_input_tokens_seen": 295742645, + "step": 9096 + }, + { + "epoch": 0.8203995130089733, + "flos": 21949630193760.0, + "grad_norm": 2.3945767521229846, + "learning_rate": 3.288103895356749e-07, + "loss": 0.6909, + "num_input_tokens_seen": 295770935, + "step": 9097 + }, + { + "epoch": 0.8204896965324435, + "flos": 25848868306080.0, + "grad_norm": 2.046676160167003, + "learning_rate": 3.284895269097927e-07, + "loss": 0.6762, + "num_input_tokens_seen": 295800385, + "step": 9098 + }, + { + "epoch": 0.8205798800559138, + "flos": 64369447734720.0, + "grad_norm": 0.6630456431847002, + "learning_rate": 3.281688069063999e-07, + "loss": 0.5807, + "num_input_tokens_seen": 295895100, + "step": 9099 + }, + { + "epoch": 0.8206700635793841, + "flos": 23698854123840.0, + "grad_norm": 2.048041091611017, + "learning_rate": 3.2784822955286396e-07, + "loss": 0.5675, + "num_input_tokens_seen": 295922605, + "step": 9100 + }, + { + "epoch": 0.8207602471028543, + "flos": 22747753910400.0, + "grad_norm": 1.5926223954575482, + "learning_rate": 3.275277948765365e-07, + "loss": 0.8128, + "num_input_tokens_seen": 295952605, + "step": 9101 + }, + { + "epoch": 0.8208504306263246, + "flos": 29272742160480.0, + "grad_norm": 3.252984803739762, + "learning_rate": 3.2720750290475964e-07, + "loss": 0.7855, + "num_input_tokens_seen": 295982105, + "step": 9102 + }, + { + "epoch": 0.8209406141497948, + "flos": 20856704411520.0, + "grad_norm": 2.1247501226376437, + "learning_rate": 3.268873536648622e-07, + "loss": 0.7464, + "num_input_tokens_seen": 296006335, + "step": 9103 + }, + { + "epoch": 0.8210307976732651, + "flos": 31641051575040.0, + "grad_norm": 2.42873714787199, + "learning_rate": 3.265673471841612e-07, + "loss": 0.7135, + "num_input_tokens_seen": 296039730, + "step": 9104 + }, + { + "epoch": 0.8211209811967354, + "flos": 62234790396960.0, + "grad_norm": 0.6366036433920715, + "learning_rate": 3.262474834899616e-07, + "loss": 0.5399, + "num_input_tokens_seen": 296126155, + "step": 9105 + }, + { + "epoch": 0.8212111647202056, + "flos": 63442364919840.0, + "grad_norm": 0.6730266725205869, + "learning_rate": 3.2592776260955534e-07, + "loss": 0.6162, + "num_input_tokens_seen": 296225540, + "step": 9106 + }, + { + "epoch": 0.8213013482436758, + "flos": 22569273224640.0, + "grad_norm": 4.30211550105571, + "learning_rate": 3.256081845702239e-07, + "loss": 0.7809, + "num_input_tokens_seen": 296251975, + "step": 9107 + }, + { + "epoch": 0.8213915317671462, + "flos": 18451665540480.0, + "grad_norm": 2.772393946360643, + "learning_rate": 3.2528874939923335e-07, + "loss": 0.7922, + "num_input_tokens_seen": 296279395, + "step": 9108 + }, + { + "epoch": 0.8214817152906164, + "flos": 22169189197920.0, + "grad_norm": 2.0889942189415556, + "learning_rate": 3.2496945712384217e-07, + "loss": 0.6955, + "num_input_tokens_seen": 296307910, + "step": 9109 + }, + { + "epoch": 0.8215718988140867, + "flos": 26351074864320.0, + "grad_norm": 1.972328929372494, + "learning_rate": 3.246503077712923e-07, + "loss": 0.7221, + "num_input_tokens_seen": 296332530, + "step": 9110 + }, + { + "epoch": 0.8216620823375569, + "flos": 23293826519040.0, + "grad_norm": 2.477014542488835, + "learning_rate": 3.2433130136881625e-07, + "loss": 0.8329, + "num_input_tokens_seen": 296360915, + "step": 9111 + }, + { + "epoch": 0.8217522658610272, + "flos": 15822830312640.0, + "grad_norm": 2.379466524841351, + "learning_rate": 3.2401243794363287e-07, + "loss": 0.7348, + "num_input_tokens_seen": 296383060, + "step": 9112 + }, + { + "epoch": 0.8218424493844975, + "flos": 26649928433760.0, + "grad_norm": 1.7354116692381394, + "learning_rate": 3.236937175229495e-07, + "loss": 0.7961, + "num_input_tokens_seen": 296412590, + "step": 9113 + }, + { + "epoch": 0.8219326329079677, + "flos": 17282604622560.0, + "grad_norm": 1.9571587983557188, + "learning_rate": 3.233751401339615e-07, + "loss": 0.7098, + "num_input_tokens_seen": 296438920, + "step": 9114 + }, + { + "epoch": 0.8220228164314379, + "flos": 17177248321920.0, + "grad_norm": 3.1756551231873056, + "learning_rate": 3.2305670580385157e-07, + "loss": 0.8248, + "num_input_tokens_seen": 296465880, + "step": 9115 + }, + { + "epoch": 0.8221129999549083, + "flos": 27888136572480.0, + "grad_norm": 1.7577128742384376, + "learning_rate": 3.227384145597898e-07, + "loss": 0.7425, + "num_input_tokens_seen": 296495345, + "step": 9116 + }, + { + "epoch": 0.8222031834783785, + "flos": 23480596061280.0, + "grad_norm": 2.223092785490059, + "learning_rate": 3.224202664289346e-07, + "loss": 0.7293, + "num_input_tokens_seen": 296523735, + "step": 9117 + }, + { + "epoch": 0.8222933670018487, + "flos": 33134916288480.0, + "grad_norm": 1.891425034415157, + "learning_rate": 3.2210226143843257e-07, + "loss": 0.6841, + "num_input_tokens_seen": 296552440, + "step": 9118 + }, + { + "epoch": 0.822383550525319, + "flos": 59620042508160.0, + "grad_norm": 0.6550763852955619, + "learning_rate": 3.217843996154173e-07, + "loss": 0.6124, + "num_input_tokens_seen": 296640285, + "step": 9119 + }, + { + "epoch": 0.8224737340487893, + "flos": 66370648433280.0, + "grad_norm": 1.4343769314489057, + "learning_rate": 3.2146668098701055e-07, + "loss": 0.5282, + "num_input_tokens_seen": 296734555, + "step": 9120 + }, + { + "epoch": 0.8225639175722596, + "flos": 24347123603520.0, + "grad_norm": 2.0115035154884184, + "learning_rate": 3.2114910558032215e-07, + "loss": 0.8704, + "num_input_tokens_seen": 296761665, + "step": 9121 + }, + { + "epoch": 0.8226541010957298, + "flos": 17614953879360.0, + "grad_norm": 2.330266901700404, + "learning_rate": 3.2083167342244945e-07, + "loss": 0.7769, + "num_input_tokens_seen": 296789195, + "step": 9122 + }, + { + "epoch": 0.8227442846192, + "flos": 14371493538240.0, + "grad_norm": 2.085547249967774, + "learning_rate": 3.205143845404763e-07, + "loss": 0.8419, + "num_input_tokens_seen": 296813080, + "step": 9123 + }, + { + "epoch": 0.8228344681426704, + "flos": 20164606051200.0, + "grad_norm": 2.251237634839842, + "learning_rate": 3.201972389614773e-07, + "loss": 0.7171, + "num_input_tokens_seen": 296842165, + "step": 9124 + }, + { + "epoch": 0.8229246516661406, + "flos": 13351580631840.0, + "grad_norm": 2.9635816056298396, + "learning_rate": 3.198802367125115e-07, + "loss": 0.736, + "num_input_tokens_seen": 296865250, + "step": 9125 + }, + { + "epoch": 0.8230148351896108, + "flos": 25447371828480.0, + "grad_norm": 1.76819059604232, + "learning_rate": 3.195633778206288e-07, + "loss": 0.7554, + "num_input_tokens_seen": 296894275, + "step": 9126 + }, + { + "epoch": 0.8231050187130812, + "flos": 20383198641600.0, + "grad_norm": 2.1647197950841695, + "learning_rate": 3.19246662312864e-07, + "loss": 0.8188, + "num_input_tokens_seen": 296920380, + "step": 9127 + }, + { + "epoch": 0.8231952022365514, + "flos": 70075645881600.0, + "grad_norm": 0.6911355259080489, + "learning_rate": 3.189300902162417e-07, + "loss": 0.577, + "num_input_tokens_seen": 297015365, + "step": 9128 + }, + { + "epoch": 0.8232853857600216, + "flos": 19253729251680.0, + "grad_norm": 2.2952173180837248, + "learning_rate": 3.1861366155777327e-07, + "loss": 0.754, + "num_input_tokens_seen": 297043270, + "step": 9129 + }, + { + "epoch": 0.8233755692834919, + "flos": 26941087862880.0, + "grad_norm": 1.7123579982718582, + "learning_rate": 3.182973763644583e-07, + "loss": 0.753, + "num_input_tokens_seen": 297077550, + "step": 9130 + }, + { + "epoch": 0.8234657528069622, + "flos": 13606716829920.0, + "grad_norm": 2.557313928034604, + "learning_rate": 3.1798123466328463e-07, + "loss": 0.6688, + "num_input_tokens_seen": 297105610, + "step": 9131 + }, + { + "epoch": 0.8235559363304324, + "flos": 31052376687840.0, + "grad_norm": 2.596048554570343, + "learning_rate": 3.17665236481226e-07, + "loss": 0.7449, + "num_input_tokens_seen": 297134185, + "step": 9132 + }, + { + "epoch": 0.8236461198539027, + "flos": 21251807690400.0, + "grad_norm": 2.1951529556891587, + "learning_rate": 3.1734938184524576e-07, + "loss": 0.7028, + "num_input_tokens_seen": 297161250, + "step": 9133 + }, + { + "epoch": 0.8237363033773729, + "flos": 20159699642880.0, + "grad_norm": 1.796401517101489, + "learning_rate": 3.1703367078229427e-07, + "loss": 0.7506, + "num_input_tokens_seen": 297190745, + "step": 9134 + }, + { + "epoch": 0.8238264869008433, + "flos": 35175225308160.0, + "grad_norm": 2.5479450435311075, + "learning_rate": 3.167181033193096e-07, + "loss": 0.7061, + "num_input_tokens_seen": 297219795, + "step": 9135 + }, + { + "epoch": 0.8239166704243135, + "flos": 22782104502240.0, + "grad_norm": 1.8862148583477458, + "learning_rate": 3.16402679483218e-07, + "loss": 0.7557, + "num_input_tokens_seen": 297248255, + "step": 9136 + }, + { + "epoch": 0.8240068539477837, + "flos": 25160598431040.0, + "grad_norm": 1.5061888971241921, + "learning_rate": 3.1608739930093366e-07, + "loss": 0.6758, + "num_input_tokens_seen": 297282130, + "step": 9137 + }, + { + "epoch": 0.824097037471254, + "flos": 33387116075520.0, + "grad_norm": 1.9951039825913657, + "learning_rate": 3.157722627993562e-07, + "loss": 0.657, + "num_input_tokens_seen": 297314085, + "step": 9138 + }, + { + "epoch": 0.8241872209947243, + "flos": 32147681334720.0, + "grad_norm": 1.8713526536691398, + "learning_rate": 3.1545727000537727e-07, + "loss": 0.7057, + "num_input_tokens_seen": 297345250, + "step": 9139 + }, + { + "epoch": 0.8242774045181945, + "flos": 22639163840640.0, + "grad_norm": 1.9696036410618027, + "learning_rate": 3.151424209458713e-07, + "loss": 0.7419, + "num_input_tokens_seen": 297372820, + "step": 9140 + }, + { + "epoch": 0.8243675880416648, + "flos": 12987854364000.0, + "grad_norm": 2.1691048059815543, + "learning_rate": 3.148277156477053e-07, + "loss": 0.7571, + "num_input_tokens_seen": 297399935, + "step": 9141 + }, + { + "epoch": 0.824457771565135, + "flos": 17249852330400.0, + "grad_norm": 2.619245306621227, + "learning_rate": 3.145131541377299e-07, + "loss": 0.7453, + "num_input_tokens_seen": 297426140, + "step": 9142 + }, + { + "epoch": 0.8245479550886053, + "flos": 12185790652800.0, + "grad_norm": 2.092484504523836, + "learning_rate": 3.1419873644278606e-07, + "loss": 0.7101, + "num_input_tokens_seen": 297451325, + "step": 9143 + }, + { + "epoch": 0.8246381386120756, + "flos": 25225359620160.0, + "grad_norm": 1.789729098274021, + "learning_rate": 3.1388446258970147e-07, + "loss": 0.7625, + "num_input_tokens_seen": 297479955, + "step": 9144 + }, + { + "epoch": 0.8247283221355458, + "flos": 32799407602080.0, + "grad_norm": 1.8814947811827814, + "learning_rate": 3.1357033260529145e-07, + "loss": 0.7282, + "num_input_tokens_seen": 297509405, + "step": 9145 + }, + { + "epoch": 0.824818505659016, + "flos": 24460582911840.0, + "grad_norm": 2.1988289794269638, + "learning_rate": 3.1325634651636025e-07, + "loss": 0.8174, + "num_input_tokens_seen": 297539520, + "step": 9146 + }, + { + "epoch": 0.8249086891824864, + "flos": 21366530770560.0, + "grad_norm": 2.114782790625524, + "learning_rate": 3.1294250434969694e-07, + "loss": 0.7823, + "num_input_tokens_seen": 297568035, + "step": 9147 + }, + { + "epoch": 0.8249988727059566, + "flos": 17213680420320.0, + "grad_norm": 1.985514396298278, + "learning_rate": 3.1262880613208274e-07, + "loss": 0.7827, + "num_input_tokens_seen": 297591185, + "step": 9148 + }, + { + "epoch": 0.8250890562294269, + "flos": 30184399524960.0, + "grad_norm": 2.0105562573397995, + "learning_rate": 3.123152518902823e-07, + "loss": 0.7834, + "num_input_tokens_seen": 297620865, + "step": 9149 + }, + { + "epoch": 0.8251792397528972, + "flos": 26320812946080.0, + "grad_norm": 1.580047343063885, + "learning_rate": 3.1200184165105017e-07, + "loss": 0.698, + "num_input_tokens_seen": 297651030, + "step": 9150 + }, + { + "epoch": 0.8252694232763674, + "flos": 17501160043200.0, + "grad_norm": 4.7584037726556, + "learning_rate": 3.116885754411287e-07, + "loss": 0.7555, + "num_input_tokens_seen": 297676820, + "step": 9151 + }, + { + "epoch": 0.8253596067998377, + "flos": 27559095424320.0, + "grad_norm": 2.2890331279445557, + "learning_rate": 3.1137545328724703e-07, + "loss": 0.6767, + "num_input_tokens_seen": 297705655, + "step": 9152 + }, + { + "epoch": 0.8254497903233079, + "flos": 21330767727840.0, + "grad_norm": 2.089153767705577, + "learning_rate": 3.1106247521612285e-07, + "loss": 0.7455, + "num_input_tokens_seen": 297734335, + "step": 9153 + }, + { + "epoch": 0.8255399738467782, + "flos": 29921643526080.0, + "grad_norm": 2.395596987358743, + "learning_rate": 3.107496412544612e-07, + "loss": 0.7604, + "num_input_tokens_seen": 297761010, + "step": 9154 + }, + { + "epoch": 0.8256301573702485, + "flos": 19181719959360.0, + "grad_norm": 1.5546488541443637, + "learning_rate": 3.1043695142895397e-07, + "loss": 0.7706, + "num_input_tokens_seen": 297788800, + "step": 9155 + }, + { + "epoch": 0.8257203408937187, + "flos": 28579788895680.0, + "grad_norm": 1.6397005655659838, + "learning_rate": 3.101244057662828e-07, + "loss": 0.6649, + "num_input_tokens_seen": 297820505, + "step": 9156 + }, + { + "epoch": 0.8258105244171889, + "flos": 22751768244480.0, + "grad_norm": 1.9522608561966859, + "learning_rate": 3.098120042931152e-07, + "loss": 0.7309, + "num_input_tokens_seen": 297849085, + "step": 9157 + }, + { + "epoch": 0.8259007079406593, + "flos": 24937396790400.0, + "grad_norm": 2.6289551320852422, + "learning_rate": 3.0949974703610647e-07, + "loss": 0.7778, + "num_input_tokens_seen": 297877175, + "step": 9158 + }, + { + "epoch": 0.8259908914641295, + "flos": 36010450178880.0, + "grad_norm": 1.7458467244778029, + "learning_rate": 3.0918763402190107e-07, + "loss": 0.7326, + "num_input_tokens_seen": 297906040, + "step": 9159 + }, + { + "epoch": 0.8260810749875998, + "flos": 15971197759200.0, + "grad_norm": 2.963269284317474, + "learning_rate": 3.088756652771296e-07, + "loss": 0.6978, + "num_input_tokens_seen": 297930390, + "step": 9160 + }, + { + "epoch": 0.82617125851107, + "flos": 23291856521760.0, + "grad_norm": 2.6237428470421382, + "learning_rate": 3.0856384082841147e-07, + "loss": 0.8366, + "num_input_tokens_seen": 297957320, + "step": 9161 + }, + { + "epoch": 0.8262614420345403, + "flos": 29272704990720.0, + "grad_norm": 1.6948764081924208, + "learning_rate": 3.0825216070235207e-07, + "loss": 0.6913, + "num_input_tokens_seen": 297987640, + "step": 9162 + }, + { + "epoch": 0.8263516255580106, + "flos": 16885419837120.0, + "grad_norm": 2.3060364993962805, + "learning_rate": 3.0794062492554764e-07, + "loss": 0.8516, + "num_input_tokens_seen": 298013035, + "step": 9163 + }, + { + "epoch": 0.8264418090814808, + "flos": 21833234474400.0, + "grad_norm": 1.9736270744017588, + "learning_rate": 3.076292335245783e-07, + "loss": 0.7551, + "num_input_tokens_seen": 298040520, + "step": 9164 + }, + { + "epoch": 0.826531992604951, + "flos": 20820718350240.0, + "grad_norm": 2.0094515867454676, + "learning_rate": 3.073179865260145e-07, + "loss": 0.811, + "num_input_tokens_seen": 298068340, + "step": 9165 + }, + { + "epoch": 0.8266221761284214, + "flos": 24790107266880.0, + "grad_norm": 1.9729482906697935, + "learning_rate": 3.070068839564135e-07, + "loss": 0.7497, + "num_input_tokens_seen": 298096840, + "step": 9166 + }, + { + "epoch": 0.8267123596518916, + "flos": 22059372526080.0, + "grad_norm": 2.5265126107220306, + "learning_rate": 3.0669592584232006e-07, + "loss": 0.7818, + "num_input_tokens_seen": 298122855, + "step": 9167 + }, + { + "epoch": 0.8268025431753618, + "flos": 40457396369280.0, + "grad_norm": 1.886024363112628, + "learning_rate": 3.063851122102672e-07, + "loss": 0.6547, + "num_input_tokens_seen": 298155595, + "step": 9168 + }, + { + "epoch": 0.8268927266988321, + "flos": 31820981881440.0, + "grad_norm": 1.891555305037466, + "learning_rate": 3.06074443086775e-07, + "loss": 0.6618, + "num_input_tokens_seen": 298189405, + "step": 9169 + }, + { + "epoch": 0.8269829102223024, + "flos": 69936719554080.0, + "grad_norm": 0.6011022813755291, + "learning_rate": 3.057639184983514e-07, + "loss": 0.5305, + "num_input_tokens_seen": 298290565, + "step": 9170 + }, + { + "epoch": 0.8270730937457726, + "flos": 22787456947680.0, + "grad_norm": 1.495803924455632, + "learning_rate": 3.054535384714927e-07, + "loss": 0.746, + "num_input_tokens_seen": 298322060, + "step": 9171 + }, + { + "epoch": 0.8271632772692429, + "flos": 25957755733920.0, + "grad_norm": 1.8071572714607147, + "learning_rate": 3.0514330303268135e-07, + "loss": 0.9097, + "num_input_tokens_seen": 298349850, + "step": 9172 + }, + { + "epoch": 0.8272534607927132, + "flos": 18379953606240.0, + "grad_norm": 1.9194447107800994, + "learning_rate": 3.0483321220838876e-07, + "loss": 0.6973, + "num_input_tokens_seen": 298376620, + "step": 9173 + }, + { + "epoch": 0.8273436443161835, + "flos": 43954914985440.0, + "grad_norm": 6.649716031314355, + "learning_rate": 3.045232660250734e-07, + "loss": 0.6001, + "num_input_tokens_seen": 298411045, + "step": 9174 + }, + { + "epoch": 0.8274338278396537, + "flos": 24753303470880.0, + "grad_norm": 1.8498537572845708, + "learning_rate": 3.0421346450918185e-07, + "loss": 0.7693, + "num_input_tokens_seen": 298441570, + "step": 9175 + }, + { + "epoch": 0.8275240113631239, + "flos": 26905585008480.0, + "grad_norm": 1.4851235359517954, + "learning_rate": 3.039038076871485e-07, + "loss": 0.6923, + "num_input_tokens_seen": 298470335, + "step": 9176 + }, + { + "epoch": 0.8276141948865943, + "flos": 29121326793600.0, + "grad_norm": 1.8098577115734729, + "learning_rate": 3.035942955853934e-07, + "loss": 0.7608, + "num_input_tokens_seen": 298498865, + "step": 9177 + }, + { + "epoch": 0.8277043784100645, + "flos": 23152186799040.0, + "grad_norm": 2.352826865308554, + "learning_rate": 3.0328492823032804e-07, + "loss": 0.7182, + "num_input_tokens_seen": 298526755, + "step": 9178 + }, + { + "epoch": 0.8277945619335347, + "flos": 18522113702880.0, + "grad_norm": 2.286057723163449, + "learning_rate": 3.029757056483471e-07, + "loss": 0.6819, + "num_input_tokens_seen": 298555630, + "step": 9179 + }, + { + "epoch": 0.827884745457005, + "flos": 25300082304960.0, + "grad_norm": 2.0668173022388165, + "learning_rate": 3.026666278658372e-07, + "loss": 0.6333, + "num_input_tokens_seen": 298585055, + "step": 9180 + }, + { + "epoch": 0.8279749289804753, + "flos": 20235165722880.0, + "grad_norm": 1.4731650042857707, + "learning_rate": 3.023576949091691e-07, + "loss": 0.7425, + "num_input_tokens_seen": 298612965, + "step": 9181 + }, + { + "epoch": 0.8280651125039455, + "flos": 20493089652960.0, + "grad_norm": 1.9626232667345218, + "learning_rate": 3.020489068047032e-07, + "loss": 0.667, + "num_input_tokens_seen": 298639170, + "step": 9182 + }, + { + "epoch": 0.8281552960274158, + "flos": 30547270888320.0, + "grad_norm": 1.8830702355971145, + "learning_rate": 3.017402635787869e-07, + "loss": 0.7022, + "num_input_tokens_seen": 298670695, + "step": 9183 + }, + { + "epoch": 0.828245479550886, + "flos": 21659139820320.0, + "grad_norm": 1.8838256389414998, + "learning_rate": 3.0143176525775537e-07, + "loss": 0.7162, + "num_input_tokens_seen": 298699030, + "step": 9184 + }, + { + "epoch": 0.8283356630743564, + "flos": 22752251451360.0, + "grad_norm": 1.8935688523878715, + "learning_rate": 3.0112341186793155e-07, + "loss": 0.6894, + "num_input_tokens_seen": 298727895, + "step": 9185 + }, + { + "epoch": 0.8284258465978266, + "flos": 21871227702720.0, + "grad_norm": 1.7010858910397375, + "learning_rate": 3.008152034356264e-07, + "loss": 0.6815, + "num_input_tokens_seen": 298756140, + "step": 9186 + }, + { + "epoch": 0.8285160301212968, + "flos": 66407712417600.0, + "grad_norm": 0.6226044282683427, + "learning_rate": 3.005071399871366e-07, + "loss": 0.563, + "num_input_tokens_seen": 298850485, + "step": 9187 + }, + { + "epoch": 0.828606213644767, + "flos": 27596791294560.0, + "grad_norm": 2.0250629980389774, + "learning_rate": 3.0019922154874853e-07, + "loss": 0.8014, + "num_input_tokens_seen": 298879490, + "step": 9188 + }, + { + "epoch": 0.8286963971682374, + "flos": 20601828401760.0, + "grad_norm": 1.9624701965401536, + "learning_rate": 2.998914481467356e-07, + "loss": 0.6986, + "num_input_tokens_seen": 298906300, + "step": 9189 + }, + { + "epoch": 0.8287865806917076, + "flos": 19436707478400.0, + "grad_norm": 2.7089121175644024, + "learning_rate": 2.9958381980735837e-07, + "loss": 0.6903, + "num_input_tokens_seen": 298933495, + "step": 9190 + }, + { + "epoch": 0.8288767642151779, + "flos": 36231495973440.0, + "grad_norm": 1.694233564202826, + "learning_rate": 2.992763365568658e-07, + "loss": 0.6734, + "num_input_tokens_seen": 298964565, + "step": 9191 + }, + { + "epoch": 0.8289669477386481, + "flos": 23808745135200.0, + "grad_norm": 1.8369736365190081, + "learning_rate": 2.98968998421494e-07, + "loss": 0.743, + "num_input_tokens_seen": 298995190, + "step": 9192 + }, + { + "epoch": 0.8290571312621184, + "flos": 20709043190400.0, + "grad_norm": 3.512056294174703, + "learning_rate": 2.98661805427467e-07, + "loss": 0.7329, + "num_input_tokens_seen": 299020925, + "step": 9193 + }, + { + "epoch": 0.8291473147855887, + "flos": 45556217506080.0, + "grad_norm": 1.5870688025517996, + "learning_rate": 2.9835475760099483e-07, + "loss": 0.7586, + "num_input_tokens_seen": 299055560, + "step": 9194 + }, + { + "epoch": 0.8292374983090589, + "flos": 31198662627840.0, + "grad_norm": 1.9542272280463935, + "learning_rate": 2.9804785496827856e-07, + "loss": 0.7587, + "num_input_tokens_seen": 299085565, + "step": 9195 + }, + { + "epoch": 0.8293276818325293, + "flos": 15063629068320.0, + "grad_norm": 2.419622641374901, + "learning_rate": 2.977410975555028e-07, + "loss": 0.7334, + "num_input_tokens_seen": 299111850, + "step": 9196 + }, + { + "epoch": 0.8294178653559995, + "flos": 25155394664640.0, + "grad_norm": 1.9273521716950934, + "learning_rate": 2.9743448538884376e-07, + "loss": 0.8022, + "num_input_tokens_seen": 299135300, + "step": 9197 + }, + { + "epoch": 0.8295080488794697, + "flos": 21439097609280.0, + "grad_norm": 1.6665473974785723, + "learning_rate": 2.9712801849446154e-07, + "loss": 0.7304, + "num_input_tokens_seen": 299165945, + "step": 9198 + }, + { + "epoch": 0.82959823240294, + "flos": 29018312187840.0, + "grad_norm": 4.3238120503755315, + "learning_rate": 2.9682169689850665e-07, + "loss": 0.6945, + "num_input_tokens_seen": 299195865, + "step": 9199 + }, + { + "epoch": 0.8296884159264103, + "flos": 63543706886400.0, + "grad_norm": 0.6706734936788866, + "learning_rate": 2.9651552062711573e-07, + "loss": 0.5266, + "num_input_tokens_seen": 299287015, + "step": 9200 + }, + { + "epoch": 0.8297785994498805, + "flos": 20487030982080.0, + "grad_norm": 2.313996320624299, + "learning_rate": 2.9620948970641333e-07, + "loss": 0.8225, + "num_input_tokens_seen": 299317380, + "step": 9201 + }, + { + "epoch": 0.8298687829733508, + "flos": 23733985280640.0, + "grad_norm": 1.5454250437424701, + "learning_rate": 2.959036041625125e-07, + "loss": 0.7608, + "num_input_tokens_seen": 299348025, + "step": 9202 + }, + { + "epoch": 0.829958966496821, + "flos": 28070445743520.0, + "grad_norm": 1.8233493473086644, + "learning_rate": 2.95597864021512e-07, + "loss": 0.6525, + "num_input_tokens_seen": 299378925, + "step": 9203 + }, + { + "epoch": 0.8300491500202913, + "flos": 16079639149920.0, + "grad_norm": 2.4170403246469094, + "learning_rate": 2.9529226930949966e-07, + "loss": 0.7936, + "num_input_tokens_seen": 299401080, + "step": 9204 + }, + { + "epoch": 0.8301393335437616, + "flos": 22710132379680.0, + "grad_norm": 1.8371617557326532, + "learning_rate": 2.949868200525505e-07, + "loss": 0.7541, + "num_input_tokens_seen": 299429245, + "step": 9205 + }, + { + "epoch": 0.8302295170672318, + "flos": 23407880543520.0, + "grad_norm": 2.4087505476376396, + "learning_rate": 2.9468151627672734e-07, + "loss": 0.7364, + "num_input_tokens_seen": 299459400, + "step": 9206 + }, + { + "epoch": 0.830319700590702, + "flos": 27485599341600.0, + "grad_norm": 1.8414304334357765, + "learning_rate": 2.9437635800808026e-07, + "loss": 0.7003, + "num_input_tokens_seen": 299488620, + "step": 9207 + }, + { + "epoch": 0.8304098841141724, + "flos": 22059409695840.0, + "grad_norm": 6.331104622590556, + "learning_rate": 2.940713452726473e-07, + "loss": 0.7322, + "num_input_tokens_seen": 299514460, + "step": 9208 + }, + { + "epoch": 0.8305000676376426, + "flos": 19217780360160.0, + "grad_norm": 2.0025701697696072, + "learning_rate": 2.937664780964526e-07, + "loss": 0.7378, + "num_input_tokens_seen": 299542295, + "step": 9209 + }, + { + "epoch": 0.8305902511611128, + "flos": 27887950723680.0, + "grad_norm": 2.0573365283050418, + "learning_rate": 2.9346175650551133e-07, + "loss": 0.7483, + "num_input_tokens_seen": 299569090, + "step": 9210 + }, + { + "epoch": 0.8306804346845831, + "flos": 23218546287840.0, + "grad_norm": 2.0037860797149176, + "learning_rate": 2.931571805258215e-07, + "loss": 0.7628, + "num_input_tokens_seen": 299598230, + "step": 9211 + }, + { + "epoch": 0.8307706182080534, + "flos": 25119482942880.0, + "grad_norm": 5.939656398095887, + "learning_rate": 2.9285275018337353e-07, + "loss": 0.792, + "num_input_tokens_seen": 299626045, + "step": 9212 + }, + { + "epoch": 0.8308608017315237, + "flos": 23217617043840.0, + "grad_norm": 2.614589939387716, + "learning_rate": 2.9254846550414146e-07, + "loss": 0.8431, + "num_input_tokens_seen": 299652045, + "step": 9213 + }, + { + "epoch": 0.8309509852549939, + "flos": 19945567423680.0, + "grad_norm": 2.67747971584925, + "learning_rate": 2.922443265140893e-07, + "loss": 0.7916, + "num_input_tokens_seen": 299678750, + "step": 9214 + }, + { + "epoch": 0.8310411687784641, + "flos": 24281990716800.0, + "grad_norm": 1.9421682416677684, + "learning_rate": 2.919403332391674e-07, + "loss": 0.7425, + "num_input_tokens_seen": 299706265, + "step": 9215 + }, + { + "epoch": 0.8311313523019345, + "flos": 31457887499520.0, + "grad_norm": 1.844284174203942, + "learning_rate": 2.9163648570531464e-07, + "loss": 0.7653, + "num_input_tokens_seen": 299740365, + "step": 9216 + }, + { + "epoch": 0.8312215358254047, + "flos": 38563187440800.0, + "grad_norm": 1.9324696365478033, + "learning_rate": 2.9133278393845717e-07, + "loss": 0.681, + "num_input_tokens_seen": 299774075, + "step": 9217 + }, + { + "epoch": 0.8313117193488749, + "flos": 24718432502400.0, + "grad_norm": 2.3026309390533246, + "learning_rate": 2.9102922796450703e-07, + "loss": 0.8117, + "num_input_tokens_seen": 299803520, + "step": 9218 + }, + { + "epoch": 0.8314019028723453, + "flos": 41843860445280.0, + "grad_norm": 1.6731884511915218, + "learning_rate": 2.907258178093672e-07, + "loss": 0.6999, + "num_input_tokens_seen": 299838155, + "step": 9219 + }, + { + "epoch": 0.8314920863958155, + "flos": 24135667607040.0, + "grad_norm": 1.6989233702947126, + "learning_rate": 2.904225534989251e-07, + "loss": 0.6664, + "num_input_tokens_seen": 299867475, + "step": 9220 + }, + { + "epoch": 0.8315822699192857, + "flos": 52367904814080.0, + "grad_norm": 2.1072272775655705, + "learning_rate": 2.901194350590572e-07, + "loss": 0.6439, + "num_input_tokens_seen": 299900450, + "step": 9221 + }, + { + "epoch": 0.831672453442756, + "flos": 34044752334720.0, + "grad_norm": 1.705615261632978, + "learning_rate": 2.898164625156274e-07, + "loss": 0.6203, + "num_input_tokens_seen": 299933665, + "step": 9222 + }, + { + "epoch": 0.8317626369662263, + "flos": 23261706112800.0, + "grad_norm": 6.247896069217865, + "learning_rate": 2.8951363589448676e-07, + "loss": 0.6984, + "num_input_tokens_seen": 299963810, + "step": 9223 + }, + { + "epoch": 0.8318528204896966, + "flos": 18671187374880.0, + "grad_norm": 3.111716457541366, + "learning_rate": 2.8921095522147434e-07, + "loss": 0.7371, + "num_input_tokens_seen": 299991515, + "step": 9224 + }, + { + "epoch": 0.8319430040131668, + "flos": 16412583122880.0, + "grad_norm": 1.9942540837963765, + "learning_rate": 2.8890842052241683e-07, + "loss": 0.7752, + "num_input_tokens_seen": 300017090, + "step": 9225 + }, + { + "epoch": 0.832033187536637, + "flos": 65738516363040.0, + "grad_norm": 0.5908165545383456, + "learning_rate": 2.886060318231267e-07, + "loss": 0.5257, + "num_input_tokens_seen": 300108605, + "step": 9226 + }, + { + "epoch": 0.8321233710601074, + "flos": 25811953000800.0, + "grad_norm": 1.694588572551331, + "learning_rate": 2.8830378914940755e-07, + "loss": 0.7654, + "num_input_tokens_seen": 300138775, + "step": 9227 + }, + { + "epoch": 0.8322135545835776, + "flos": 17287325182080.0, + "grad_norm": 2.0823385788361075, + "learning_rate": 2.8800169252704675e-07, + "loss": 0.7294, + "num_input_tokens_seen": 300163990, + "step": 9228 + }, + { + "epoch": 0.8323037381070478, + "flos": 22714890108960.0, + "grad_norm": 1.6878577665499985, + "learning_rate": 2.8769974198182143e-07, + "loss": 0.7633, + "num_input_tokens_seen": 300192535, + "step": 9229 + }, + { + "epoch": 0.8323939216305181, + "flos": 30876460715520.0, + "grad_norm": 1.7761987372456707, + "learning_rate": 2.873979375394955e-07, + "loss": 0.7412, + "num_input_tokens_seen": 300222255, + "step": 9230 + }, + { + "epoch": 0.8324841051539884, + "flos": 24756016863360.0, + "grad_norm": 1.6386090199506629, + "learning_rate": 2.870962792258209e-07, + "loss": 0.7756, + "num_input_tokens_seen": 300249985, + "step": 9231 + }, + { + "epoch": 0.8325742886774586, + "flos": 22531056977760.0, + "grad_norm": 1.6459977594965245, + "learning_rate": 2.8679476706653716e-07, + "loss": 0.7533, + "num_input_tokens_seen": 300279185, + "step": 9232 + }, + { + "epoch": 0.8326644722009289, + "flos": 24500434628160.0, + "grad_norm": 2.861459274671409, + "learning_rate": 2.864934010873692e-07, + "loss": 0.7616, + "num_input_tokens_seen": 300307225, + "step": 9233 + }, + { + "epoch": 0.8327546557243991, + "flos": 25880431165920.0, + "grad_norm": 1.514870052492933, + "learning_rate": 2.8619218131403357e-07, + "loss": 0.7165, + "num_input_tokens_seen": 300338130, + "step": 9234 + }, + { + "epoch": 0.8328448392478695, + "flos": 18561965419200.0, + "grad_norm": 1.9363466336964663, + "learning_rate": 2.858911077722299e-07, + "loss": 0.7676, + "num_input_tokens_seen": 300365840, + "step": 9235 + }, + { + "epoch": 0.8329350227713397, + "flos": 28906042311840.0, + "grad_norm": 2.0907120941677895, + "learning_rate": 2.855901804876493e-07, + "loss": 0.5915, + "num_input_tokens_seen": 300397255, + "step": 9236 + }, + { + "epoch": 0.8330252062948099, + "flos": 26391632806080.0, + "grad_norm": 1.8326852115226149, + "learning_rate": 2.852893994859673e-07, + "loss": 0.7773, + "num_input_tokens_seen": 300428270, + "step": 9237 + }, + { + "epoch": 0.8331153898182801, + "flos": 32110431501600.0, + "grad_norm": 1.9682192616862466, + "learning_rate": 2.849887647928484e-07, + "loss": 0.6866, + "num_input_tokens_seen": 300460660, + "step": 9238 + }, + { + "epoch": 0.8332055733417505, + "flos": 22314322875360.0, + "grad_norm": 3.57048865620374, + "learning_rate": 2.8468827643394465e-07, + "loss": 0.6423, + "num_input_tokens_seen": 300490450, + "step": 9239 + }, + { + "epoch": 0.8332957568652207, + "flos": 26641862595840.0, + "grad_norm": 1.9869981983984895, + "learning_rate": 2.843879344348954e-07, + "loss": 0.7536, + "num_input_tokens_seen": 300518220, + "step": 9240 + }, + { + "epoch": 0.833385940388691, + "flos": 60385382496480.0, + "grad_norm": 12.50011352487511, + "learning_rate": 2.840877388213272e-07, + "loss": 0.7142, + "num_input_tokens_seen": 300554720, + "step": 9241 + }, + { + "epoch": 0.8334761239121612, + "flos": 25119445773120.0, + "grad_norm": 2.415009050687965, + "learning_rate": 2.8378768961885515e-07, + "loss": 0.7456, + "num_input_tokens_seen": 300579880, + "step": 9242 + }, + { + "epoch": 0.8335663074356315, + "flos": 16411505199840.0, + "grad_norm": 1.8057195181451766, + "learning_rate": 2.8348778685307983e-07, + "loss": 0.7757, + "num_input_tokens_seen": 300606015, + "step": 9243 + }, + { + "epoch": 0.8336564909591018, + "flos": 24063546805440.0, + "grad_norm": 3.5439841971809303, + "learning_rate": 2.831880305495915e-07, + "loss": 0.8195, + "num_input_tokens_seen": 300634930, + "step": 9244 + }, + { + "epoch": 0.833746674482572, + "flos": 64792471236960.0, + "grad_norm": 0.6506707673552671, + "learning_rate": 2.828884207339668e-07, + "loss": 0.6111, + "num_input_tokens_seen": 300728600, + "step": 9245 + }, + { + "epoch": 0.8338368580060423, + "flos": 17212044950880.0, + "grad_norm": 2.5103414268351365, + "learning_rate": 2.8258895743177014e-07, + "loss": 0.8196, + "num_input_tokens_seen": 300753655, + "step": 9246 + }, + { + "epoch": 0.8339270415295126, + "flos": 22532952635520.0, + "grad_norm": 1.8691457272004552, + "learning_rate": 2.8228964066855356e-07, + "loss": 0.8158, + "num_input_tokens_seen": 300779730, + "step": 9247 + }, + { + "epoch": 0.8340172250529828, + "flos": 27885869217120.0, + "grad_norm": 1.6852326172001648, + "learning_rate": 2.819904704698555e-07, + "loss": 0.7202, + "num_input_tokens_seen": 300809345, + "step": 9248 + }, + { + "epoch": 0.834107408576453, + "flos": 59527063737600.0, + "grad_norm": 0.633798772564112, + "learning_rate": 2.8169144686120437e-07, + "loss": 0.5268, + "num_input_tokens_seen": 300899160, + "step": 9249 + }, + { + "epoch": 0.8341975920999234, + "flos": 28508411489280.0, + "grad_norm": 2.202553316438758, + "learning_rate": 2.8139256986811254e-07, + "loss": 0.6546, + "num_input_tokens_seen": 300929165, + "step": 9250 + }, + { + "epoch": 0.8342877756233936, + "flos": 20601642552960.0, + "grad_norm": 2.089778378613965, + "learning_rate": 2.8109383951608424e-07, + "loss": 0.7945, + "num_input_tokens_seen": 300954885, + "step": 9251 + }, + { + "epoch": 0.8343779591468639, + "flos": 29527246472640.0, + "grad_norm": 1.9288512359968795, + "learning_rate": 2.8079525583060683e-07, + "loss": 0.8079, + "num_input_tokens_seen": 300983620, + "step": 9252 + }, + { + "epoch": 0.8344681426703341, + "flos": 25265248506240.0, + "grad_norm": 1.5682927137675922, + "learning_rate": 2.804968188371577e-07, + "loss": 0.7488, + "num_input_tokens_seen": 301013955, + "step": 9253 + }, + { + "epoch": 0.8345583261938044, + "flos": 30182912734560.0, + "grad_norm": 1.9129346366120812, + "learning_rate": 2.801985285612014e-07, + "loss": 0.7444, + "num_input_tokens_seen": 301045445, + "step": 9254 + }, + { + "epoch": 0.8346485097172747, + "flos": 25119557282400.0, + "grad_norm": 1.7863228156142053, + "learning_rate": 2.7990038502818934e-07, + "loss": 0.7158, + "num_input_tokens_seen": 301074160, + "step": 9255 + }, + { + "epoch": 0.8347386932407449, + "flos": 22309267788000.0, + "grad_norm": 1.7077776152319009, + "learning_rate": 2.796023882635612e-07, + "loss": 0.7184, + "num_input_tokens_seen": 301103150, + "step": 9256 + }, + { + "epoch": 0.8348288767642151, + "flos": 20196912306240.0, + "grad_norm": 2.055814020844191, + "learning_rate": 2.7930453829274323e-07, + "loss": 0.7912, + "num_input_tokens_seen": 301131580, + "step": 9257 + }, + { + "epoch": 0.8349190602876855, + "flos": 21767729890080.0, + "grad_norm": 2.5186989525859027, + "learning_rate": 2.7900683514115054e-07, + "loss": 0.7848, + "num_input_tokens_seen": 301160295, + "step": 9258 + }, + { + "epoch": 0.8350092438111557, + "flos": 22314174196320.0, + "grad_norm": 3.057164741407227, + "learning_rate": 2.787092788341836e-07, + "loss": 0.6786, + "num_input_tokens_seen": 301188105, + "step": 9259 + }, + { + "epoch": 0.8350994273346259, + "flos": 21658768122720.0, + "grad_norm": 3.0784036732891633, + "learning_rate": 2.7841186939723195e-07, + "loss": 0.7419, + "num_input_tokens_seen": 301213950, + "step": 9260 + }, + { + "epoch": 0.8351896108580962, + "flos": 18489324240960.0, + "grad_norm": 1.8573594268345182, + "learning_rate": 2.7811460685567255e-07, + "loss": 0.7189, + "num_input_tokens_seen": 301240875, + "step": 9261 + }, + { + "epoch": 0.8352797943815665, + "flos": 27774156887520.0, + "grad_norm": 2.1570231944938545, + "learning_rate": 2.778174912348692e-07, + "loss": 0.7377, + "num_input_tokens_seen": 301267990, + "step": 9262 + }, + { + "epoch": 0.8353699779050368, + "flos": 24244963902240.0, + "grad_norm": 1.9235080510826972, + "learning_rate": 2.7752052256017354e-07, + "loss": 0.7045, + "num_input_tokens_seen": 301294935, + "step": 9263 + }, + { + "epoch": 0.835460161428507, + "flos": 17468221902240.0, + "grad_norm": 2.0316817778394896, + "learning_rate": 2.7722370085692493e-07, + "loss": 0.7123, + "num_input_tokens_seen": 301322265, + "step": 9264 + }, + { + "epoch": 0.8355503449519772, + "flos": 22533473012160.0, + "grad_norm": 2.676215678123675, + "learning_rate": 2.769270261504486e-07, + "loss": 0.7145, + "num_input_tokens_seen": 301349490, + "step": 9265 + }, + { + "epoch": 0.8356405284754476, + "flos": 29596207844640.0, + "grad_norm": 1.663177254768724, + "learning_rate": 2.7663049846606015e-07, + "loss": 0.702, + "num_input_tokens_seen": 301381955, + "step": 9266 + }, + { + "epoch": 0.8357307119989178, + "flos": 34810198098720.0, + "grad_norm": 1.66174914964884, + "learning_rate": 2.763341178290592e-07, + "loss": 0.7538, + "num_input_tokens_seen": 301412500, + "step": 9267 + }, + { + "epoch": 0.835820895522388, + "flos": 21476570460960.0, + "grad_norm": 2.3472921435032124, + "learning_rate": 2.7603788426473663e-07, + "loss": 0.7666, + "num_input_tokens_seen": 301442830, + "step": 9268 + }, + { + "epoch": 0.8359110790458584, + "flos": 24167527824960.0, + "grad_norm": 3.412410519722788, + "learning_rate": 2.7574179779836695e-07, + "loss": 0.6751, + "num_input_tokens_seen": 301473255, + "step": 9269 + }, + { + "epoch": 0.8360012625693286, + "flos": 42934481702400.0, + "grad_norm": 1.8252775399273014, + "learning_rate": 2.754458584552146e-07, + "loss": 0.6997, + "num_input_tokens_seen": 301504645, + "step": 9270 + }, + { + "epoch": 0.8360914460927988, + "flos": 28616704200960.0, + "grad_norm": 2.5909564129524796, + "learning_rate": 2.751500662605308e-07, + "loss": 0.7262, + "num_input_tokens_seen": 301532115, + "step": 9271 + }, + { + "epoch": 0.8361816296162691, + "flos": 24425340245760.0, + "grad_norm": 6.789188923772543, + "learning_rate": 2.7485442123955383e-07, + "loss": 0.6955, + "num_input_tokens_seen": 301561935, + "step": 9272 + }, + { + "epoch": 0.8362718131397394, + "flos": 28580197763040.0, + "grad_norm": 3.530427279886169, + "learning_rate": 2.7455892341751075e-07, + "loss": 0.6262, + "num_input_tokens_seen": 301591550, + "step": 9273 + }, + { + "epoch": 0.8363619966632097, + "flos": 23079582790560.0, + "grad_norm": 1.996712545792281, + "learning_rate": 2.7426357281961365e-07, + "loss": 0.6166, + "num_input_tokens_seen": 301617570, + "step": 9274 + }, + { + "epoch": 0.8364521801866799, + "flos": 25154130892800.0, + "grad_norm": 2.5108232815205036, + "learning_rate": 2.7396836947106416e-07, + "loss": 0.7503, + "num_input_tokens_seen": 301641845, + "step": 9275 + }, + { + "epoch": 0.8365423637101501, + "flos": 24641665480800.0, + "grad_norm": 2.4772203215591078, + "learning_rate": 2.736733133970506e-07, + "loss": 0.839, + "num_input_tokens_seen": 301670495, + "step": 9276 + }, + { + "epoch": 0.8366325472336205, + "flos": 24680327764800.0, + "grad_norm": 1.4681766283645186, + "learning_rate": 2.7337840462274896e-07, + "loss": 0.7816, + "num_input_tokens_seen": 301702420, + "step": 9277 + }, + { + "epoch": 0.8367227307570907, + "flos": 24097005323040.0, + "grad_norm": 2.3672494408085822, + "learning_rate": 2.730836431733221e-07, + "loss": 0.7095, + "num_input_tokens_seen": 301731280, + "step": 9278 + }, + { + "epoch": 0.8368129142805609, + "flos": 22087478598240.0, + "grad_norm": 1.8269173961813634, + "learning_rate": 2.727890290739212e-07, + "loss": 0.7894, + "num_input_tokens_seen": 301757855, + "step": 9279 + }, + { + "epoch": 0.8369030978040312, + "flos": 24244703713920.0, + "grad_norm": 1.718366406724005, + "learning_rate": 2.7249456234968395e-07, + "loss": 0.7705, + "num_input_tokens_seen": 301787890, + "step": 9280 + }, + { + "epoch": 0.8369932813275015, + "flos": 21257977870560.0, + "grad_norm": 1.6843242432372727, + "learning_rate": 2.722002430257364e-07, + "loss": 0.8095, + "num_input_tokens_seen": 301815865, + "step": 9281 + }, + { + "epoch": 0.8370834648509717, + "flos": 27742854216000.0, + "grad_norm": 1.9661551435304887, + "learning_rate": 2.7190607112719035e-07, + "loss": 0.7055, + "num_input_tokens_seen": 301846265, + "step": 9282 + }, + { + "epoch": 0.837173648374442, + "flos": 29090804687040.0, + "grad_norm": 2.4664561564169554, + "learning_rate": 2.716120466791476e-07, + "loss": 0.6616, + "num_input_tokens_seen": 301875700, + "step": 9283 + }, + { + "epoch": 0.8372638318979122, + "flos": 23771569641600.0, + "grad_norm": 2.1045487186070266, + "learning_rate": 2.7131816970669483e-07, + "loss": 0.7193, + "num_input_tokens_seen": 301903920, + "step": 9284 + }, + { + "epoch": 0.8373540154213825, + "flos": 48145015168800.0, + "grad_norm": 3.2717432266996562, + "learning_rate": 2.7102444023490777e-07, + "loss": 0.7478, + "num_input_tokens_seen": 301934565, + "step": 9285 + }, + { + "epoch": 0.8374441989448528, + "flos": 23950533534240.0, + "grad_norm": 1.6822915406911072, + "learning_rate": 2.70730858288849e-07, + "loss": 0.7961, + "num_input_tokens_seen": 301964985, + "step": 9286 + }, + { + "epoch": 0.837534382468323, + "flos": 35353594484640.0, + "grad_norm": 1.9196592754789834, + "learning_rate": 2.704374238935685e-07, + "loss": 0.6769, + "num_input_tokens_seen": 301995605, + "step": 9287 + }, + { + "epoch": 0.8376245659917932, + "flos": 51796136434080.0, + "grad_norm": 0.6721696872736321, + "learning_rate": 2.70144137074104e-07, + "loss": 0.5735, + "num_input_tokens_seen": 302082475, + "step": 9288 + }, + { + "epoch": 0.8377147495152636, + "flos": 26796400222560.0, + "grad_norm": 1.8687064404976632, + "learning_rate": 2.6985099785547926e-07, + "loss": 0.7611, + "num_input_tokens_seen": 302113320, + "step": 9289 + }, + { + "epoch": 0.8378049330387338, + "flos": 23079136753440.0, + "grad_norm": 1.9000191712667087, + "learning_rate": 2.695580062627083e-07, + "loss": 0.7674, + "num_input_tokens_seen": 302144065, + "step": 9290 + }, + { + "epoch": 0.8378951165622041, + "flos": 21512705201280.0, + "grad_norm": 3.0436847906940923, + "learning_rate": 2.692651623207891e-07, + "loss": 0.7382, + "num_input_tokens_seen": 302169640, + "step": 9291 + }, + { + "epoch": 0.8379853000856744, + "flos": 34226057922240.0, + "grad_norm": 1.7602695501428627, + "learning_rate": 2.689724660547097e-07, + "loss": 0.7268, + "num_input_tokens_seen": 302199855, + "step": 9292 + }, + { + "epoch": 0.8380754836091446, + "flos": 65227723590240.0, + "grad_norm": 0.6165622489607735, + "learning_rate": 2.686799174894441e-07, + "loss": 0.5702, + "num_input_tokens_seen": 302290655, + "step": 9293 + }, + { + "epoch": 0.8381656671326149, + "flos": 65708886330720.0, + "grad_norm": 0.6399388942381049, + "learning_rate": 2.683875166499545e-07, + "loss": 0.5705, + "num_input_tokens_seen": 302388260, + "step": 9294 + }, + { + "epoch": 0.8382558506560851, + "flos": 15828665964960.0, + "grad_norm": 1.8832084904143063, + "learning_rate": 2.680952635611899e-07, + "loss": 0.7478, + "num_input_tokens_seen": 302411475, + "step": 9295 + }, + { + "epoch": 0.8383460341795554, + "flos": 18707507964000.0, + "grad_norm": 2.1916742008016876, + "learning_rate": 2.678031582480875e-07, + "loss": 0.6684, + "num_input_tokens_seen": 302438715, + "step": 9296 + }, + { + "epoch": 0.8384362177030257, + "flos": 24753377810400.0, + "grad_norm": 1.8142944529358336, + "learning_rate": 2.6751120073557e-07, + "loss": 0.7811, + "num_input_tokens_seen": 302467890, + "step": 9297 + }, + { + "epoch": 0.8385264012264959, + "flos": 17869123663680.0, + "grad_norm": 3.751273384059488, + "learning_rate": 2.672193910485505e-07, + "loss": 0.7081, + "num_input_tokens_seen": 302495495, + "step": 9298 + }, + { + "epoch": 0.8386165847499661, + "flos": 21804199158240.0, + "grad_norm": 1.8667920703430991, + "learning_rate": 2.669277292119265e-07, + "loss": 0.7401, + "num_input_tokens_seen": 302524770, + "step": 9299 + }, + { + "epoch": 0.8387067682734365, + "flos": 24136522511520.0, + "grad_norm": 3.1215679395725457, + "learning_rate": 2.666362152505848e-07, + "loss": 0.7427, + "num_input_tokens_seen": 302550880, + "step": 9300 + }, + { + "epoch": 0.8387969517969067, + "flos": 23036943342240.0, + "grad_norm": 2.438515671617695, + "learning_rate": 2.663448491893989e-07, + "loss": 0.6746, + "num_input_tokens_seen": 302580670, + "step": 9301 + }, + { + "epoch": 0.838887135320377, + "flos": 24128977050240.0, + "grad_norm": 1.7448331306047595, + "learning_rate": 2.6605363105322974e-07, + "loss": 0.7146, + "num_input_tokens_seen": 302609130, + "step": 9302 + }, + { + "epoch": 0.8389773188438472, + "flos": 28143904656480.0, + "grad_norm": 1.8806874799761035, + "learning_rate": 2.657625608669263e-07, + "loss": 0.7383, + "num_input_tokens_seen": 302636080, + "step": 9303 + }, + { + "epoch": 0.8390675023673175, + "flos": 43444270891680.0, + "grad_norm": 1.6279001033042202, + "learning_rate": 2.654716386553224e-07, + "loss": 0.6535, + "num_input_tokens_seen": 302668215, + "step": 9304 + }, + { + "epoch": 0.8391576858907878, + "flos": 34628260625280.0, + "grad_norm": 2.3277140004968677, + "learning_rate": 2.651808644432436e-07, + "loss": 0.7371, + "num_input_tokens_seen": 302700575, + "step": 9305 + }, + { + "epoch": 0.839247869414258, + "flos": 22934114585280.0, + "grad_norm": 1.990579592008715, + "learning_rate": 2.6489023825549807e-07, + "loss": 0.8717, + "num_input_tokens_seen": 302724650, + "step": 9306 + }, + { + "epoch": 0.8393380529377282, + "flos": 27305222998080.0, + "grad_norm": 2.6277341681319135, + "learning_rate": 2.6459976011688547e-07, + "loss": 0.6612, + "num_input_tokens_seen": 302750200, + "step": 9307 + }, + { + "epoch": 0.8394282364611986, + "flos": 21476942158560.0, + "grad_norm": 3.5136184052178105, + "learning_rate": 2.6430943005219e-07, + "loss": 0.7371, + "num_input_tokens_seen": 302778305, + "step": 9308 + }, + { + "epoch": 0.8395184199846688, + "flos": 25774294300320.0, + "grad_norm": 4.173098495465503, + "learning_rate": 2.6401924808618447e-07, + "loss": 0.7674, + "num_input_tokens_seen": 302806705, + "step": 9309 + }, + { + "epoch": 0.839608603508139, + "flos": 20671719017760.0, + "grad_norm": 2.0639508086486074, + "learning_rate": 2.637292142436287e-07, + "loss": 0.7347, + "num_input_tokens_seen": 302835315, + "step": 9310 + }, + { + "epoch": 0.8396987870316093, + "flos": 21585309209760.0, + "grad_norm": 2.4747099581225784, + "learning_rate": 2.6343932854927e-07, + "loss": 0.7095, + "num_input_tokens_seen": 302863365, + "step": 9311 + }, + { + "epoch": 0.8397889705550796, + "flos": 69888430302240.0, + "grad_norm": 0.6835868399134565, + "learning_rate": 2.6314959102784316e-07, + "loss": 0.5692, + "num_input_tokens_seen": 302953995, + "step": 9312 + }, + { + "epoch": 0.8398791540785498, + "flos": 16558051328160.0, + "grad_norm": 3.462799329248103, + "learning_rate": 2.6286000170407074e-07, + "loss": 0.7312, + "num_input_tokens_seen": 302979230, + "step": 9313 + }, + { + "epoch": 0.8399693376020201, + "flos": 23735323392000.0, + "grad_norm": 2.4060633203001838, + "learning_rate": 2.625705606026607e-07, + "loss": 0.7144, + "num_input_tokens_seen": 303008280, + "step": 9314 + }, + { + "epoch": 0.8400595211254904, + "flos": 23254792537440.0, + "grad_norm": 1.5485059059037833, + "learning_rate": 2.622812677483106e-07, + "loss": 0.8201, + "num_input_tokens_seen": 303035220, + "step": 9315 + }, + { + "epoch": 0.8401497046489607, + "flos": 14444655093120.0, + "grad_norm": 3.9747679732638708, + "learning_rate": 2.6199212316570453e-07, + "loss": 0.7921, + "num_input_tokens_seen": 303059495, + "step": 9316 + }, + { + "epoch": 0.8402398881724309, + "flos": 21111840609600.0, + "grad_norm": 2.4414368162149067, + "learning_rate": 2.617031268795138e-07, + "loss": 0.7199, + "num_input_tokens_seen": 303090055, + "step": 9317 + }, + { + "epoch": 0.8403300716959011, + "flos": 24570734111520.0, + "grad_norm": 1.4902660521494022, + "learning_rate": 2.614142789143976e-07, + "loss": 0.7681, + "num_input_tokens_seen": 303121020, + "step": 9318 + }, + { + "epoch": 0.8404202552193715, + "flos": 19108595574240.0, + "grad_norm": 1.7221611851298209, + "learning_rate": 2.6112557929500047e-07, + "loss": 0.7231, + "num_input_tokens_seen": 303149210, + "step": 9319 + }, + { + "epoch": 0.8405104387428417, + "flos": 21076226245920.0, + "grad_norm": 1.8628167107145257, + "learning_rate": 2.6083702804595817e-07, + "loss": 0.7626, + "num_input_tokens_seen": 303177675, + "step": 9320 + }, + { + "epoch": 0.8406006222663119, + "flos": 22897533807840.0, + "grad_norm": 1.742121947556333, + "learning_rate": 2.6054862519188915e-07, + "loss": 0.7238, + "num_input_tokens_seen": 303207150, + "step": 9321 + }, + { + "epoch": 0.8406908057897822, + "flos": 18634383578880.0, + "grad_norm": 1.7973906067520093, + "learning_rate": 2.6026037075740357e-07, + "loss": 0.6741, + "num_input_tokens_seen": 303234590, + "step": 9322 + }, + { + "epoch": 0.8407809893132525, + "flos": 30365036056800.0, + "grad_norm": 1.995591789181919, + "learning_rate": 2.5997226476709524e-07, + "loss": 0.7067, + "num_input_tokens_seen": 303262470, + "step": 9323 + }, + { + "epoch": 0.8408711728367227, + "flos": 18087790593600.0, + "grad_norm": 1.9702239195887614, + "learning_rate": 2.5968430724554856e-07, + "loss": 0.7418, + "num_input_tokens_seen": 303288800, + "step": 9324 + }, + { + "epoch": 0.840961356360193, + "flos": 36850804476480.0, + "grad_norm": 2.033337151674948, + "learning_rate": 2.5939649821733225e-07, + "loss": 0.6731, + "num_input_tokens_seen": 303318540, + "step": 9325 + }, + { + "epoch": 0.8410515398836632, + "flos": 20562980268960.0, + "grad_norm": 6.422432756914613, + "learning_rate": 2.5910883770700433e-07, + "loss": 0.7015, + "num_input_tokens_seen": 303347935, + "step": 9326 + }, + { + "epoch": 0.8411417234071336, + "flos": 37870866061920.0, + "grad_norm": 3.103851626664703, + "learning_rate": 2.5882132573910965e-07, + "loss": 0.5819, + "num_input_tokens_seen": 303381380, + "step": 9327 + }, + { + "epoch": 0.8412319069306038, + "flos": 24353851330080.0, + "grad_norm": 1.3774678884595395, + "learning_rate": 2.585339623381801e-07, + "loss": 0.7783, + "num_input_tokens_seen": 303413075, + "step": 9328 + }, + { + "epoch": 0.841322090454074, + "flos": 27305780544480.0, + "grad_norm": 1.9331686766432816, + "learning_rate": 2.582467475287358e-07, + "loss": 0.7173, + "num_input_tokens_seen": 303444145, + "step": 9329 + }, + { + "epoch": 0.8414122739775443, + "flos": 24427867789440.0, + "grad_norm": 1.9914128611332553, + "learning_rate": 2.5795968133528224e-07, + "loss": 0.7487, + "num_input_tokens_seen": 303471530, + "step": 9330 + }, + { + "epoch": 0.8415024575010146, + "flos": 22642286100480.0, + "grad_norm": 1.790297835521863, + "learning_rate": 2.576727637823144e-07, + "loss": 0.7409, + "num_input_tokens_seen": 303499350, + "step": 9331 + }, + { + "epoch": 0.8415926410244848, + "flos": 21031170763200.0, + "grad_norm": 2.9608033484513427, + "learning_rate": 2.5738599489431335e-07, + "loss": 0.7796, + "num_input_tokens_seen": 303525980, + "step": 9332 + }, + { + "epoch": 0.8416828245479551, + "flos": 27045180391680.0, + "grad_norm": 1.9877163247475347, + "learning_rate": 2.5709937469574794e-07, + "loss": 0.8006, + "num_input_tokens_seen": 303555145, + "step": 9333 + }, + { + "epoch": 0.8417730080714253, + "flos": 25156100890080.0, + "grad_norm": 2.80874994132802, + "learning_rate": 2.568129032110742e-07, + "loss": 0.7778, + "num_input_tokens_seen": 303584480, + "step": 9334 + }, + { + "epoch": 0.8418631915948956, + "flos": 19910064569280.0, + "grad_norm": 1.638334291806463, + "learning_rate": 2.5652658046473565e-07, + "loss": 0.8154, + "num_input_tokens_seen": 303612575, + "step": 9335 + }, + { + "epoch": 0.8419533751183659, + "flos": 28140670887360.0, + "grad_norm": 1.7363476918408536, + "learning_rate": 2.5624040648116184e-07, + "loss": 0.7391, + "num_input_tokens_seen": 303641135, + "step": 9336 + }, + { + "epoch": 0.8420435586418361, + "flos": 15829372190400.0, + "grad_norm": 2.0988347953493016, + "learning_rate": 2.5595438128477245e-07, + "loss": 0.6771, + "num_input_tokens_seen": 303668570, + "step": 9337 + }, + { + "epoch": 0.8421337421653065, + "flos": 22532469428640.0, + "grad_norm": 1.700218660653333, + "learning_rate": 2.5566850489997096e-07, + "loss": 0.7457, + "num_input_tokens_seen": 303694825, + "step": 9338 + }, + { + "epoch": 0.8422239256887767, + "flos": 22861027369920.0, + "grad_norm": 2.210908168422926, + "learning_rate": 2.5538277735115166e-07, + "loss": 0.7632, + "num_input_tokens_seen": 303725180, + "step": 9339 + }, + { + "epoch": 0.8423141092122469, + "flos": 28688936511840.0, + "grad_norm": 1.5591206436799792, + "learning_rate": 2.5509719866269306e-07, + "loss": 0.7922, + "num_input_tokens_seen": 303754960, + "step": 9340 + }, + { + "epoch": 0.8424042927357172, + "flos": 18744460439040.0, + "grad_norm": 2.0511939223932933, + "learning_rate": 2.548117688589628e-07, + "loss": 0.8015, + "num_input_tokens_seen": 303781680, + "step": 9341 + }, + { + "epoch": 0.8424944762591875, + "flos": 24682483610880.0, + "grad_norm": 2.0652213205697327, + "learning_rate": 2.545264879643152e-07, + "loss": 0.7237, + "num_input_tokens_seen": 303811835, + "step": 9342 + }, + { + "epoch": 0.8425846597826577, + "flos": 23407099978560.0, + "grad_norm": 1.817842502251318, + "learning_rate": 2.542413560030923e-07, + "loss": 0.697, + "num_input_tokens_seen": 303838970, + "step": 9343 + }, + { + "epoch": 0.842674843306128, + "flos": 23658928068000.0, + "grad_norm": 1.7814041204868902, + "learning_rate": 2.53956372999623e-07, + "loss": 0.6332, + "num_input_tokens_seen": 303868435, + "step": 9344 + }, + { + "epoch": 0.8427650268295982, + "flos": 69899729909280.0, + "grad_norm": 0.6372181858533366, + "learning_rate": 2.5367153897822293e-07, + "loss": 0.6082, + "num_input_tokens_seen": 303964635, + "step": 9345 + }, + { + "epoch": 0.8428552103530685, + "flos": 25264244922720.0, + "grad_norm": 2.1267211899698895, + "learning_rate": 2.5338685396319715e-07, + "loss": 0.7323, + "num_input_tokens_seen": 303994945, + "step": 9346 + }, + { + "epoch": 0.8429453938765388, + "flos": 28471161656160.0, + "grad_norm": 2.5211007367850873, + "learning_rate": 2.531023179788352e-07, + "loss": 0.6762, + "num_input_tokens_seen": 304023190, + "step": 9347 + }, + { + "epoch": 0.843035577400009, + "flos": 26613161807520.0, + "grad_norm": 2.926057647158064, + "learning_rate": 2.528179310494158e-07, + "loss": 0.7113, + "num_input_tokens_seen": 304053525, + "step": 9348 + }, + { + "epoch": 0.8431257609234792, + "flos": 38271210276960.0, + "grad_norm": 1.7248776876236653, + "learning_rate": 2.5253369319920436e-07, + "loss": 0.7594, + "num_input_tokens_seen": 304087310, + "step": 9349 + }, + { + "epoch": 0.8432159444469496, + "flos": 21731892507840.0, + "grad_norm": 2.3475909485835342, + "learning_rate": 2.522496044524538e-07, + "loss": 0.7404, + "num_input_tokens_seen": 304114745, + "step": 9350 + }, + { + "epoch": 0.8433061279704198, + "flos": 12586841093280.0, + "grad_norm": 2.4899844773509594, + "learning_rate": 2.5196566483340386e-07, + "loss": 0.8292, + "num_input_tokens_seen": 304139595, + "step": 9351 + }, + { + "epoch": 0.84339631149389, + "flos": 26756920203840.0, + "grad_norm": 5.777721090935307, + "learning_rate": 2.516818743662825e-07, + "loss": 0.7546, + "num_input_tokens_seen": 304169000, + "step": 9352 + }, + { + "epoch": 0.8434864950173603, + "flos": 12871161286560.0, + "grad_norm": 2.136466002640396, + "learning_rate": 2.5139823307530285e-07, + "loss": 0.7374, + "num_input_tokens_seen": 304194060, + "step": 9353 + }, + { + "epoch": 0.8435766785408306, + "flos": 25629680999520.0, + "grad_norm": 2.3220236277597945, + "learning_rate": 2.5111474098466836e-07, + "loss": 0.7762, + "num_input_tokens_seen": 304221610, + "step": 9354 + }, + { + "epoch": 0.8436668620643009, + "flos": 22273393236000.0, + "grad_norm": 1.7229814633320464, + "learning_rate": 2.50831398118567e-07, + "loss": 0.7849, + "num_input_tokens_seen": 304250130, + "step": 9355 + }, + { + "epoch": 0.8437570455877711, + "flos": 17577815555520.0, + "grad_norm": 1.969867933852127, + "learning_rate": 2.5054820450117576e-07, + "loss": 0.6855, + "num_input_tokens_seen": 304276250, + "step": 9356 + }, + { + "epoch": 0.8438472291112413, + "flos": 25627339304640.0, + "grad_norm": 2.517476034646998, + "learning_rate": 2.502651601566579e-07, + "loss": 0.7031, + "num_input_tokens_seen": 304303825, + "step": 9357 + }, + { + "epoch": 0.8439374126347117, + "flos": 25593100222080.0, + "grad_norm": 2.332910011128941, + "learning_rate": 2.499822651091645e-07, + "loss": 0.6907, + "num_input_tokens_seen": 304331575, + "step": 9358 + }, + { + "epoch": 0.8440275961581819, + "flos": 17793248716320.0, + "grad_norm": 2.303295581561615, + "learning_rate": 2.496995193828344e-07, + "loss": 0.7595, + "num_input_tokens_seen": 304358295, + "step": 9359 + }, + { + "epoch": 0.8441177796816521, + "flos": 24893493570240.0, + "grad_norm": 2.0309169789733352, + "learning_rate": 2.494169230017913e-07, + "loss": 0.738, + "num_input_tokens_seen": 304388865, + "step": 9360 + }, + { + "epoch": 0.8442079632051224, + "flos": 24132136479840.0, + "grad_norm": 1.7068402425718465, + "learning_rate": 2.491344759901499e-07, + "loss": 0.6964, + "num_input_tokens_seen": 304417655, + "step": 9361 + }, + { + "epoch": 0.8442981467285927, + "flos": 27086853426240.0, + "grad_norm": 2.1183178126108957, + "learning_rate": 2.488521783720088e-07, + "loss": 0.7314, + "num_input_tokens_seen": 304451125, + "step": 9362 + }, + { + "epoch": 0.844388330252063, + "flos": 30548311641600.0, + "grad_norm": 2.0957274706618274, + "learning_rate": 2.4857003017145526e-07, + "loss": 0.7742, + "num_input_tokens_seen": 304479625, + "step": 9363 + }, + { + "epoch": 0.8444785137755332, + "flos": 18343186980000.0, + "grad_norm": 1.984530825766934, + "learning_rate": 2.482880314125644e-07, + "loss": 0.7782, + "num_input_tokens_seen": 304506190, + "step": 9364 + }, + { + "epoch": 0.8445686972990035, + "flos": 30110494574880.0, + "grad_norm": 1.7433244220051232, + "learning_rate": 2.4800618211939726e-07, + "loss": 0.7312, + "num_input_tokens_seen": 304535010, + "step": 9365 + }, + { + "epoch": 0.8446588808224738, + "flos": 18671224544640.0, + "grad_norm": 1.9756862220823985, + "learning_rate": 2.477244823160034e-07, + "loss": 0.7487, + "num_input_tokens_seen": 304561820, + "step": 9366 + }, + { + "epoch": 0.844749064345944, + "flos": 23953507115040.0, + "grad_norm": 2.2651505598338977, + "learning_rate": 2.474429320264184e-07, + "loss": 0.7561, + "num_input_tokens_seen": 304591235, + "step": 9367 + }, + { + "epoch": 0.8448392478694142, + "flos": 17687706566880.0, + "grad_norm": 1.6248005964739187, + "learning_rate": 2.47161531274666e-07, + "loss": 0.6972, + "num_input_tokens_seen": 304619455, + "step": 9368 + }, + { + "epoch": 0.8449294313928846, + "flos": 43950528953760.0, + "grad_norm": 1.8446778371261536, + "learning_rate": 2.4688028008475714e-07, + "loss": 0.6714, + "num_input_tokens_seen": 304651760, + "step": 9369 + }, + { + "epoch": 0.8450196149163548, + "flos": 23438216801280.0, + "grad_norm": 2.2249696622344697, + "learning_rate": 2.465991784806891e-07, + "loss": 0.7317, + "num_input_tokens_seen": 304677195, + "step": 9370 + }, + { + "epoch": 0.845109798439825, + "flos": 16228452633600.0, + "grad_norm": 2.1095797969158845, + "learning_rate": 2.463182264864472e-07, + "loss": 0.8167, + "num_input_tokens_seen": 304703555, + "step": 9371 + }, + { + "epoch": 0.8451999819632953, + "flos": 21622038666240.0, + "grad_norm": 2.245268615421217, + "learning_rate": 2.460374241260039e-07, + "loss": 0.6666, + "num_input_tokens_seen": 304731125, + "step": 9372 + }, + { + "epoch": 0.8452901654867656, + "flos": 17687372039040.0, + "grad_norm": 2.309719372832508, + "learning_rate": 2.4575677142331884e-07, + "loss": 0.8054, + "num_input_tokens_seen": 304759685, + "step": 9373 + }, + { + "epoch": 0.8453803490102358, + "flos": 23735583580320.0, + "grad_norm": 2.0519642932529094, + "learning_rate": 2.454762684023395e-07, + "loss": 0.6807, + "num_input_tokens_seen": 304790925, + "step": 9374 + }, + { + "epoch": 0.8454705325337061, + "flos": 25630015527360.0, + "grad_norm": 2.492792133899907, + "learning_rate": 2.4519591508699823e-07, + "loss": 0.6862, + "num_input_tokens_seen": 304819450, + "step": 9375 + }, + { + "epoch": 0.8455607160571763, + "flos": 17323125394560.0, + "grad_norm": 2.069508061579698, + "learning_rate": 2.4491571150121815e-07, + "loss": 0.785, + "num_input_tokens_seen": 304845560, + "step": 9376 + }, + { + "epoch": 0.8456508995806467, + "flos": 27085626824160.0, + "grad_norm": 2.0966314728678523, + "learning_rate": 2.446356576689062e-07, + "loss": 0.737, + "num_input_tokens_seen": 304875475, + "step": 9377 + }, + { + "epoch": 0.8457410831041169, + "flos": 26900269732800.0, + "grad_norm": 2.988168272986906, + "learning_rate": 2.4435575361395976e-07, + "loss": 0.7461, + "num_input_tokens_seen": 304903895, + "step": 9378 + }, + { + "epoch": 0.8458312666275871, + "flos": 21731743828800.0, + "grad_norm": 1.9935472223847137, + "learning_rate": 2.440759993602606e-07, + "loss": 0.7154, + "num_input_tokens_seen": 304933990, + "step": 9379 + }, + { + "epoch": 0.8459214501510574, + "flos": 21732524393760.0, + "grad_norm": 2.6289258392193635, + "learning_rate": 2.437963949316793e-07, + "loss": 0.6802, + "num_input_tokens_seen": 304963665, + "step": 9380 + }, + { + "epoch": 0.8460116336745277, + "flos": 23366727885600.0, + "grad_norm": 2.097182463661346, + "learning_rate": 2.435169403520729e-07, + "loss": 0.7326, + "num_input_tokens_seen": 304989890, + "step": 9381 + }, + { + "epoch": 0.8461018171979979, + "flos": 25587227400000.0, + "grad_norm": 1.9916036311220502, + "learning_rate": 2.4323763564528653e-07, + "loss": 0.7847, + "num_input_tokens_seen": 305017660, + "step": 9382 + }, + { + "epoch": 0.8461920007214682, + "flos": 24825164084160.0, + "grad_norm": 2.131566057234389, + "learning_rate": 2.429584808351517e-07, + "loss": 0.7568, + "num_input_tokens_seen": 305046885, + "step": 9383 + }, + { + "epoch": 0.8462821842449384, + "flos": 17388667148640.0, + "grad_norm": 2.6174776286541834, + "learning_rate": 2.42679475945488e-07, + "loss": 0.862, + "num_input_tokens_seen": 305069860, + "step": 9384 + }, + { + "epoch": 0.8463723677684087, + "flos": 39947086803360.0, + "grad_norm": 1.5914007337546685, + "learning_rate": 2.424006210001008e-07, + "loss": 0.7367, + "num_input_tokens_seen": 305104275, + "step": 9385 + }, + { + "epoch": 0.846462551291879, + "flos": 30801849540000.0, + "grad_norm": 1.8172148186687591, + "learning_rate": 2.421219160227839e-07, + "loss": 0.7631, + "num_input_tokens_seen": 305135350, + "step": 9386 + }, + { + "epoch": 0.8465527348153492, + "flos": 24317642250240.0, + "grad_norm": 1.7596752554227497, + "learning_rate": 2.4184336103731785e-07, + "loss": 0.7323, + "num_input_tokens_seen": 305167265, + "step": 9387 + }, + { + "epoch": 0.8466429183388196, + "flos": 19181942977920.0, + "grad_norm": 2.4624402498848434, + "learning_rate": 2.4156495606747065e-07, + "loss": 0.7597, + "num_input_tokens_seen": 305193625, + "step": 9388 + }, + { + "epoch": 0.8467331018622898, + "flos": 23370259012800.0, + "grad_norm": 2.268842740435938, + "learning_rate": 2.412867011369972e-07, + "loss": 0.8566, + "num_input_tokens_seen": 305221190, + "step": 9389 + }, + { + "epoch": 0.84682328538576, + "flos": 26503568154240.0, + "grad_norm": 1.8290498138618683, + "learning_rate": 2.4100859626963997e-07, + "loss": 0.7464, + "num_input_tokens_seen": 305250140, + "step": 9390 + }, + { + "epoch": 0.8469134689092302, + "flos": 26941794088320.0, + "grad_norm": 6.128427908655841, + "learning_rate": 2.407306414891288e-07, + "loss": 0.7376, + "num_input_tokens_seen": 305282330, + "step": 9391 + }, + { + "epoch": 0.8470036524327006, + "flos": 31311564389760.0, + "grad_norm": 1.5375485792471475, + "learning_rate": 2.4045283681917893e-07, + "loss": 0.8406, + "num_input_tokens_seen": 305313215, + "step": 9392 + }, + { + "epoch": 0.8470938359561708, + "flos": 25957086678240.0, + "grad_norm": 1.8794372189025477, + "learning_rate": 2.4017518228349586e-07, + "loss": 0.8529, + "num_input_tokens_seen": 305342265, + "step": 9393 + }, + { + "epoch": 0.8471840194796411, + "flos": 66797314572000.0, + "grad_norm": 0.622417335192046, + "learning_rate": 2.3989767790576887e-07, + "loss": 0.5908, + "num_input_tokens_seen": 305437885, + "step": 9394 + }, + { + "epoch": 0.8472742030031113, + "flos": 22860767181600.0, + "grad_norm": 2.106314692649856, + "learning_rate": 2.396203237096781e-07, + "loss": 0.7081, + "num_input_tokens_seen": 305466850, + "step": 9395 + }, + { + "epoch": 0.8473643865265816, + "flos": 30329384523360.0, + "grad_norm": 2.0734725155780738, + "learning_rate": 2.393431197188873e-07, + "loss": 0.7321, + "num_input_tokens_seen": 305496725, + "step": 9396 + }, + { + "epoch": 0.8474545700500519, + "flos": 18413597972640.0, + "grad_norm": 2.598018213738313, + "learning_rate": 2.3906606595705004e-07, + "loss": 0.8117, + "num_input_tokens_seen": 305522050, + "step": 9397 + }, + { + "epoch": 0.8475447535735221, + "flos": 22383284247360.0, + "grad_norm": 1.9812661395934819, + "learning_rate": 2.387891624478056e-07, + "loss": 0.7933, + "num_input_tokens_seen": 305550325, + "step": 9398 + }, + { + "epoch": 0.8476349370969923, + "flos": 22059335356320.0, + "grad_norm": 2.0020787641596844, + "learning_rate": 2.3851240921478075e-07, + "loss": 0.7589, + "num_input_tokens_seen": 305577070, + "step": 9399 + }, + { + "epoch": 0.8477251206204627, + "flos": 26795693997120.0, + "grad_norm": 2.1382472065176232, + "learning_rate": 2.3823580628159057e-07, + "loss": 0.7193, + "num_input_tokens_seen": 305606425, + "step": 9400 + }, + { + "epoch": 0.8478153041439329, + "flos": 21622744891680.0, + "grad_norm": 2.2469313146248155, + "learning_rate": 2.3795935367183517e-07, + "loss": 0.7212, + "num_input_tokens_seen": 305632790, + "step": 9401 + }, + { + "epoch": 0.8479054876674031, + "flos": 25921658163360.0, + "grad_norm": 1.9545287923929309, + "learning_rate": 2.376830514091035e-07, + "loss": 0.7604, + "num_input_tokens_seen": 305663415, + "step": 9402 + }, + { + "epoch": 0.8479956711908734, + "flos": 25375139517600.0, + "grad_norm": 1.8037270636630227, + "learning_rate": 2.3740689951697135e-07, + "loss": 0.8222, + "num_input_tokens_seen": 305694070, + "step": 9403 + }, + { + "epoch": 0.8480858547143437, + "flos": 21870670156320.0, + "grad_norm": 1.6816479265185202, + "learning_rate": 2.371308980190012e-07, + "loss": 0.7384, + "num_input_tokens_seen": 305721820, + "step": 9404 + }, + { + "epoch": 0.848176038237814, + "flos": 61731803273760.0, + "grad_norm": 0.622340432631577, + "learning_rate": 2.3685504693874337e-07, + "loss": 0.5971, + "num_input_tokens_seen": 305814660, + "step": 9405 + }, + { + "epoch": 0.8482662217612842, + "flos": 27414667972320.0, + "grad_norm": 2.1919462521221598, + "learning_rate": 2.3657934629973497e-07, + "loss": 0.7748, + "num_input_tokens_seen": 305842200, + "step": 9406 + }, + { + "epoch": 0.8483564052847544, + "flos": 12441930434400.0, + "grad_norm": 2.0534780445545975, + "learning_rate": 2.3630379612549944e-07, + "loss": 0.7464, + "num_input_tokens_seen": 305868990, + "step": 9407 + }, + { + "epoch": 0.8484465888082248, + "flos": 69642698053440.0, + "grad_norm": 0.7352139976271016, + "learning_rate": 2.3602839643954997e-07, + "loss": 0.6201, + "num_input_tokens_seen": 305964645, + "step": 9408 + }, + { + "epoch": 0.848536772331695, + "flos": 22387893297600.0, + "grad_norm": 1.9395531266506891, + "learning_rate": 2.3575314726538308e-07, + "loss": 0.7428, + "num_input_tokens_seen": 305994720, + "step": 9409 + }, + { + "epoch": 0.8486269558551652, + "flos": 23549520263520.0, + "grad_norm": 2.273819933838697, + "learning_rate": 2.3547804862648645e-07, + "loss": 0.7754, + "num_input_tokens_seen": 306023655, + "step": 9410 + }, + { + "epoch": 0.8487171393786356, + "flos": 28107138030240.0, + "grad_norm": 1.9471019802374765, + "learning_rate": 2.3520310054633174e-07, + "loss": 0.7859, + "num_input_tokens_seen": 306055575, + "step": 9411 + }, + { + "epoch": 0.8488073229021058, + "flos": 25046841764640.0, + "grad_norm": 1.8735567260168633, + "learning_rate": 2.3492830304837973e-07, + "loss": 0.7017, + "num_input_tokens_seen": 306083300, + "step": 9412 + }, + { + "epoch": 0.848897506425576, + "flos": 23915142189120.0, + "grad_norm": 2.0324981376010673, + "learning_rate": 2.3465365615607723e-07, + "loss": 0.6855, + "num_input_tokens_seen": 306115485, + "step": 9413 + }, + { + "epoch": 0.8489876899490463, + "flos": 22350569124960.0, + "grad_norm": 2.121937733145619, + "learning_rate": 2.3437915989285884e-07, + "loss": 0.7325, + "num_input_tokens_seen": 306144845, + "step": 9414 + }, + { + "epoch": 0.8490778734725166, + "flos": 17177025303360.0, + "grad_norm": 2.0571723098576524, + "learning_rate": 2.3410481428214647e-07, + "loss": 0.8027, + "num_input_tokens_seen": 306171520, + "step": 9415 + }, + { + "epoch": 0.8491680569959869, + "flos": 25441759194720.0, + "grad_norm": 1.7421159943416766, + "learning_rate": 2.338306193473476e-07, + "loss": 0.7662, + "num_input_tokens_seen": 306199260, + "step": 9416 + }, + { + "epoch": 0.8492582405194571, + "flos": 39729386287200.0, + "grad_norm": 1.8400491328521134, + "learning_rate": 2.3355657511185957e-07, + "loss": 0.7416, + "num_input_tokens_seen": 306231345, + "step": 9417 + }, + { + "epoch": 0.8493484240429273, + "flos": 63884159150880.0, + "grad_norm": 0.659718159334042, + "learning_rate": 2.3328268159906428e-07, + "loss": 0.5621, + "num_input_tokens_seen": 306318585, + "step": 9418 + }, + { + "epoch": 0.8494386075663977, + "flos": 20310260105280.0, + "grad_norm": 22.73205797900704, + "learning_rate": 2.330089388323322e-07, + "loss": 0.7891, + "num_input_tokens_seen": 306346760, + "step": 9419 + }, + { + "epoch": 0.8495287910898679, + "flos": 23620191444480.0, + "grad_norm": 1.9103165697715587, + "learning_rate": 2.327353468350204e-07, + "loss": 0.7794, + "num_input_tokens_seen": 306374730, + "step": 9420 + }, + { + "epoch": 0.8496189746133381, + "flos": 33608459228160.0, + "grad_norm": 1.900110708324107, + "learning_rate": 2.3246190563047352e-07, + "loss": 0.7319, + "num_input_tokens_seen": 306404675, + "step": 9421 + }, + { + "epoch": 0.8497091581368084, + "flos": 23407360166880.0, + "grad_norm": 1.782744867912569, + "learning_rate": 2.3218861524202293e-07, + "loss": 0.726, + "num_input_tokens_seen": 306433035, + "step": 9422 + }, + { + "epoch": 0.8497993416602787, + "flos": 26682420537600.0, + "grad_norm": 2.5247771778557104, + "learning_rate": 2.3191547569298775e-07, + "loss": 0.7507, + "num_input_tokens_seen": 306462090, + "step": 9423 + }, + { + "epoch": 0.8498895251837489, + "flos": 22308896090400.0, + "grad_norm": 4.21550937095291, + "learning_rate": 2.3164248700667245e-07, + "loss": 0.7265, + "num_input_tokens_seen": 306488655, + "step": 9424 + }, + { + "epoch": 0.8499797087072192, + "flos": 23077538453760.0, + "grad_norm": 1.807611449988362, + "learning_rate": 2.313696492063717e-07, + "loss": 0.7745, + "num_input_tokens_seen": 306519385, + "step": 9425 + }, + { + "epoch": 0.8500698922306894, + "flos": 18780595179360.0, + "grad_norm": 2.388765796427495, + "learning_rate": 2.3109696231536401e-07, + "loss": 0.7292, + "num_input_tokens_seen": 306546880, + "step": 9426 + }, + { + "epoch": 0.8501600757541597, + "flos": 24682409271360.0, + "grad_norm": 1.9693440343419288, + "learning_rate": 2.3082442635691722e-07, + "loss": 0.7, + "num_input_tokens_seen": 306576240, + "step": 9427 + }, + { + "epoch": 0.85025025927763, + "flos": 19213654516800.0, + "grad_norm": 2.965472530388911, + "learning_rate": 2.305520413542854e-07, + "loss": 0.7544, + "num_input_tokens_seen": 306604315, + "step": 9428 + }, + { + "epoch": 0.8503404428011002, + "flos": 27159866302080.0, + "grad_norm": 1.6281436443176573, + "learning_rate": 2.3027980733071018e-07, + "loss": 0.6645, + "num_input_tokens_seen": 306635890, + "step": 9429 + }, + { + "epoch": 0.8504306263245704, + "flos": 65965657998240.0, + "grad_norm": 0.5947750178299054, + "learning_rate": 2.3000772430942027e-07, + "loss": 0.5444, + "num_input_tokens_seen": 306737310, + "step": 9430 + }, + { + "epoch": 0.8505208098480408, + "flos": 21913458283680.0, + "grad_norm": 2.7390022580213036, + "learning_rate": 2.2973579231363028e-07, + "loss": 0.7546, + "num_input_tokens_seen": 306764905, + "step": 9431 + }, + { + "epoch": 0.850610993371511, + "flos": 17505546074880.0, + "grad_norm": 2.0052225351862814, + "learning_rate": 2.2946401136654446e-07, + "loss": 0.7651, + "num_input_tokens_seen": 306793630, + "step": 9432 + }, + { + "epoch": 0.8507011768949813, + "flos": 21258238058880.0, + "grad_norm": 2.5640442079848174, + "learning_rate": 2.2919238149135077e-07, + "loss": 0.7681, + "num_input_tokens_seen": 306821800, + "step": 9433 + }, + { + "epoch": 0.8507913604184516, + "flos": 24937508299680.0, + "grad_norm": 1.7696458387851548, + "learning_rate": 2.289209027112282e-07, + "loss": 0.7432, + "num_input_tokens_seen": 306852595, + "step": 9434 + }, + { + "epoch": 0.8508815439419218, + "flos": 25119334263840.0, + "grad_norm": 4.375418288806953, + "learning_rate": 2.2864957504933934e-07, + "loss": 0.7992, + "num_input_tokens_seen": 306878155, + "step": 9435 + }, + { + "epoch": 0.8509717274653921, + "flos": 23876591414400.0, + "grad_norm": 2.328252084834958, + "learning_rate": 2.2837839852883589e-07, + "loss": 0.7221, + "num_input_tokens_seen": 306908040, + "step": 9436 + }, + { + "epoch": 0.8510619109888623, + "flos": 24571365997440.0, + "grad_norm": 1.820279951991571, + "learning_rate": 2.2810737317285623e-07, + "loss": 0.736, + "num_input_tokens_seen": 306936080, + "step": 9437 + }, + { + "epoch": 0.8511520945123326, + "flos": 37506619417440.0, + "grad_norm": 2.4048621398686625, + "learning_rate": 2.278364990045254e-07, + "loss": 0.6803, + "num_input_tokens_seen": 306967040, + "step": 9438 + }, + { + "epoch": 0.8512422780358029, + "flos": 22642360440000.0, + "grad_norm": 2.2320463410355695, + "learning_rate": 2.2756577604695625e-07, + "loss": 0.7303, + "num_input_tokens_seen": 306994890, + "step": 9439 + }, + { + "epoch": 0.8513324615592731, + "flos": 22821844709280.0, + "grad_norm": 4.212290247554016, + "learning_rate": 2.2729520432324855e-07, + "loss": 0.7574, + "num_input_tokens_seen": 307024890, + "step": 9440 + }, + { + "epoch": 0.8514226450827433, + "flos": 28071263478240.0, + "grad_norm": 2.3562565082490523, + "learning_rate": 2.2702478385648826e-07, + "loss": 0.6554, + "num_input_tokens_seen": 307053765, + "step": 9441 + }, + { + "epoch": 0.8515128286062137, + "flos": 25593880787040.0, + "grad_norm": 1.8916434075134152, + "learning_rate": 2.2675451466974938e-07, + "loss": 0.7355, + "num_input_tokens_seen": 307082875, + "step": 9442 + }, + { + "epoch": 0.8516030121296839, + "flos": 33135064967520.0, + "grad_norm": 2.2130687133560496, + "learning_rate": 2.26484396786093e-07, + "loss": 0.6724, + "num_input_tokens_seen": 307112205, + "step": 9443 + }, + { + "epoch": 0.8516931956531542, + "flos": 57359282410080.0, + "grad_norm": 0.6470181891775076, + "learning_rate": 2.2621443022856667e-07, + "loss": 0.6019, + "num_input_tokens_seen": 307195395, + "step": 9444 + }, + { + "epoch": 0.8517833791766244, + "flos": 24499356705120.0, + "grad_norm": 1.8523586830879049, + "learning_rate": 2.2594461502020646e-07, + "loss": 0.7818, + "num_input_tokens_seen": 307225650, + "step": 9445 + }, + { + "epoch": 0.8518735627000947, + "flos": 30470652545760.0, + "grad_norm": 2.008263563260625, + "learning_rate": 2.2567495118403278e-07, + "loss": 0.619, + "num_input_tokens_seen": 307255160, + "step": 9446 + }, + { + "epoch": 0.851963746223565, + "flos": 19363434414240.0, + "grad_norm": 1.6499243083122905, + "learning_rate": 2.254054387430566e-07, + "loss": 0.7531, + "num_input_tokens_seen": 307283590, + "step": 9447 + }, + { + "epoch": 0.8520539297470352, + "flos": 21986396820000.0, + "grad_norm": 2.1601868225472733, + "learning_rate": 2.2513607772027243e-07, + "loss": 0.7257, + "num_input_tokens_seen": 307312420, + "step": 9448 + }, + { + "epoch": 0.8521441132705054, + "flos": 21140727246720.0, + "grad_norm": 2.4707374781455607, + "learning_rate": 2.2486686813866562e-07, + "loss": 0.7015, + "num_input_tokens_seen": 307338110, + "step": 9449 + }, + { + "epoch": 0.8522342967939758, + "flos": 21076263415680.0, + "grad_norm": 1.9637842070710132, + "learning_rate": 2.2459781002120514e-07, + "loss": 0.7912, + "num_input_tokens_seen": 307365920, + "step": 9450 + }, + { + "epoch": 0.852324480317446, + "flos": 28981657070880.0, + "grad_norm": 1.5995103295372195, + "learning_rate": 2.243289033908491e-07, + "loss": 0.8081, + "num_input_tokens_seen": 307396130, + "step": 9451 + }, + { + "epoch": 0.8524146638409162, + "flos": 20420522814240.0, + "grad_norm": 2.192637704420007, + "learning_rate": 2.2406014827054176e-07, + "loss": 0.7522, + "num_input_tokens_seen": 307423515, + "step": 9452 + }, + { + "epoch": 0.8525048473643865, + "flos": 22642100251680.0, + "grad_norm": 4.139137778312237, + "learning_rate": 2.2379154468321525e-07, + "loss": 0.8042, + "num_input_tokens_seen": 307450195, + "step": 9453 + }, + { + "epoch": 0.8525950308878568, + "flos": 26941310881440.0, + "grad_norm": 1.6148476743165987, + "learning_rate": 2.2352309265178793e-07, + "loss": 0.7743, + "num_input_tokens_seen": 307481615, + "step": 9454 + }, + { + "epoch": 0.852685214411327, + "flos": 22751508056160.0, + "grad_norm": 2.3059740516545926, + "learning_rate": 2.2325479219916565e-07, + "loss": 0.8117, + "num_input_tokens_seen": 307501815, + "step": 9455 + }, + { + "epoch": 0.8527753979347973, + "flos": 27597646199040.0, + "grad_norm": 1.861397612836747, + "learning_rate": 2.229866433482419e-07, + "loss": 0.7735, + "num_input_tokens_seen": 307529790, + "step": 9456 + }, + { + "epoch": 0.8528655814582676, + "flos": 19472805048960.0, + "grad_norm": 1.909806233109124, + "learning_rate": 2.2271864612189552e-07, + "loss": 0.7914, + "num_input_tokens_seen": 307558105, + "step": 9457 + }, + { + "epoch": 0.8529557649817379, + "flos": 26098837907520.0, + "grad_norm": 1.81492005941929, + "learning_rate": 2.2245080054299415e-07, + "loss": 0.7718, + "num_input_tokens_seen": 307586720, + "step": 9458 + }, + { + "epoch": 0.8530459485052081, + "flos": 42455697826560.0, + "grad_norm": 1.9450303451406783, + "learning_rate": 2.2218310663439198e-07, + "loss": 0.6904, + "num_input_tokens_seen": 307616510, + "step": 9459 + }, + { + "epoch": 0.8531361320286783, + "flos": 45193717840320.0, + "grad_norm": 1.948932145019393, + "learning_rate": 2.2191556441892968e-07, + "loss": 0.7648, + "num_input_tokens_seen": 307647890, + "step": 9460 + }, + { + "epoch": 0.8532263155521487, + "flos": 24424373832000.0, + "grad_norm": 1.7684262764630687, + "learning_rate": 2.216481739194358e-07, + "loss": 0.7448, + "num_input_tokens_seen": 307679275, + "step": 9461 + }, + { + "epoch": 0.8533164990756189, + "flos": 25002455337600.0, + "grad_norm": 2.218685202928438, + "learning_rate": 2.2138093515872592e-07, + "loss": 0.6912, + "num_input_tokens_seen": 307707260, + "step": 9462 + }, + { + "epoch": 0.8534066825990891, + "flos": 24024847351680.0, + "grad_norm": 1.5675670099612948, + "learning_rate": 2.2111384815960087e-07, + "loss": 0.8211, + "num_input_tokens_seen": 307737550, + "step": 9463 + }, + { + "epoch": 0.8534968661225594, + "flos": 18561928249440.0, + "grad_norm": 1.8630395888515885, + "learning_rate": 2.208469129448518e-07, + "loss": 0.6662, + "num_input_tokens_seen": 307765640, + "step": 9464 + }, + { + "epoch": 0.8535870496460297, + "flos": 65636133643200.0, + "grad_norm": 0.6223126093768662, + "learning_rate": 2.2058012953725357e-07, + "loss": 0.5676, + "num_input_tokens_seen": 307856635, + "step": 9465 + }, + { + "epoch": 0.8536772331695, + "flos": 21363668699040.0, + "grad_norm": 2.222438593768131, + "learning_rate": 2.203134979595711e-07, + "loss": 0.7099, + "num_input_tokens_seen": 307882960, + "step": 9466 + }, + { + "epoch": 0.8537674166929702, + "flos": 26390369034240.0, + "grad_norm": 1.7708801113756873, + "learning_rate": 2.2004701823455374e-07, + "loss": 0.6901, + "num_input_tokens_seen": 307913460, + "step": 9467 + }, + { + "epoch": 0.8538576002164404, + "flos": 22860841521120.0, + "grad_norm": 1.9251359396161059, + "learning_rate": 2.1978069038493906e-07, + "loss": 0.7718, + "num_input_tokens_seen": 307942215, + "step": 9468 + }, + { + "epoch": 0.8539477837399108, + "flos": 20638520688480.0, + "grad_norm": 1.4742535496553042, + "learning_rate": 2.1951451443345225e-07, + "loss": 0.7427, + "num_input_tokens_seen": 307970425, + "step": 9469 + }, + { + "epoch": 0.854037967263381, + "flos": 24463853850720.0, + "grad_norm": 2.5020547052588418, + "learning_rate": 2.1924849040280425e-07, + "loss": 0.7279, + "num_input_tokens_seen": 307998510, + "step": 9470 + }, + { + "epoch": 0.8541281507868512, + "flos": 42861914863680.0, + "grad_norm": 3.0484114465426266, + "learning_rate": 2.1898261831569465e-07, + "loss": 0.8254, + "num_input_tokens_seen": 308024590, + "step": 9471 + }, + { + "epoch": 0.8542183343103215, + "flos": 24974312095680.0, + "grad_norm": 1.7504412231078825, + "learning_rate": 2.1871689819480798e-07, + "loss": 0.7518, + "num_input_tokens_seen": 308055205, + "step": 9472 + }, + { + "epoch": 0.8543085178337918, + "flos": 18670927186560.0, + "grad_norm": 8.163553176397345, + "learning_rate": 2.1845133006281745e-07, + "loss": 0.7268, + "num_input_tokens_seen": 308084380, + "step": 9473 + }, + { + "epoch": 0.854398701357262, + "flos": 59181673628640.0, + "grad_norm": 1.5028147514074146, + "learning_rate": 2.1818591394238294e-07, + "loss": 0.6019, + "num_input_tokens_seen": 308120465, + "step": 9474 + }, + { + "epoch": 0.8544888848807323, + "flos": 28285023999840.0, + "grad_norm": 3.3928995545683542, + "learning_rate": 2.1792064985615076e-07, + "loss": 0.7238, + "num_input_tokens_seen": 308148695, + "step": 9475 + }, + { + "epoch": 0.8545790684042025, + "flos": 27411917410080.0, + "grad_norm": 1.6017170156979867, + "learning_rate": 2.1765553782675528e-07, + "loss": 0.7117, + "num_input_tokens_seen": 308178010, + "step": 9476 + }, + { + "epoch": 0.8546692519276728, + "flos": 24824234840160.0, + "grad_norm": 2.018563545993806, + "learning_rate": 2.1739057787681703e-07, + "loss": 0.8154, + "num_input_tokens_seen": 308208070, + "step": 9477 + }, + { + "epoch": 0.8547594354511431, + "flos": 20419370551680.0, + "grad_norm": 1.889751579367781, + "learning_rate": 2.1712577002894372e-07, + "loss": 0.7624, + "num_input_tokens_seen": 308235360, + "step": 9478 + }, + { + "epoch": 0.8548496189746133, + "flos": 30469574622720.0, + "grad_norm": 1.7536332697503747, + "learning_rate": 2.1686111430573105e-07, + "loss": 0.7151, + "num_input_tokens_seen": 308266215, + "step": 9479 + }, + { + "epoch": 0.8549398024980835, + "flos": 19035694207680.0, + "grad_norm": 1.9126928917842143, + "learning_rate": 2.165966107297592e-07, + "loss": 0.794, + "num_input_tokens_seen": 308295685, + "step": 9480 + }, + { + "epoch": 0.8550299860215539, + "flos": 21914127339360.0, + "grad_norm": 2.6269827121358906, + "learning_rate": 2.16332259323599e-07, + "loss": 0.7669, + "num_input_tokens_seen": 308324115, + "step": 9481 + }, + { + "epoch": 0.8551201695450241, + "flos": 16885531346400.0, + "grad_norm": 3.8551001715602244, + "learning_rate": 2.1606806010980504e-07, + "loss": 0.6891, + "num_input_tokens_seen": 308348355, + "step": 9482 + }, + { + "epoch": 0.8552103530684944, + "flos": 21475827065760.0, + "grad_norm": 1.9607016128976218, + "learning_rate": 2.1580401311092067e-07, + "loss": 0.74, + "num_input_tokens_seen": 308376230, + "step": 9483 + }, + { + "epoch": 0.8553005365919647, + "flos": 15100618713120.0, + "grad_norm": 1.6949667863770626, + "learning_rate": 2.1554011834947604e-07, + "loss": 0.7295, + "num_input_tokens_seen": 308402945, + "step": 9484 + }, + { + "epoch": 0.8553907201154349, + "flos": 27524298795360.0, + "grad_norm": 2.3072121450757956, + "learning_rate": 2.1527637584798764e-07, + "loss": 0.771, + "num_input_tokens_seen": 308433155, + "step": 9485 + }, + { + "epoch": 0.8554809036389052, + "flos": 24605456400960.0, + "grad_norm": 1.840751407861592, + "learning_rate": 2.150127856289603e-07, + "loss": 0.8067, + "num_input_tokens_seen": 308459035, + "step": 9486 + }, + { + "epoch": 0.8555710871623754, + "flos": 67923698871840.0, + "grad_norm": 0.6143778576013151, + "learning_rate": 2.1474934771488363e-07, + "loss": 0.5562, + "num_input_tokens_seen": 308559220, + "step": 9487 + }, + { + "epoch": 0.8556612706858457, + "flos": 22970769702240.0, + "grad_norm": 1.733327050686723, + "learning_rate": 2.1448606212823715e-07, + "loss": 0.7845, + "num_input_tokens_seen": 308586590, + "step": 9488 + }, + { + "epoch": 0.855751454209316, + "flos": 24426938545440.0, + "grad_norm": 1.9718429853013781, + "learning_rate": 2.1422292889148452e-07, + "loss": 0.7247, + "num_input_tokens_seen": 308616995, + "step": 9489 + }, + { + "epoch": 0.8558416377327862, + "flos": 27123434203680.0, + "grad_norm": 1.7186168272270945, + "learning_rate": 2.139599480270784e-07, + "loss": 0.761, + "num_input_tokens_seen": 308648060, + "step": 9490 + }, + { + "epoch": 0.8559318212562564, + "flos": 23079731469600.0, + "grad_norm": 2.580245746766865, + "learning_rate": 2.1369711955745773e-07, + "loss": 0.7302, + "num_input_tokens_seen": 308674025, + "step": 9491 + }, + { + "epoch": 0.8560220047797268, + "flos": 32733159622560.0, + "grad_norm": 3.500066756565514, + "learning_rate": 2.1343444350504813e-07, + "loss": 0.6932, + "num_input_tokens_seen": 308704825, + "step": 9492 + }, + { + "epoch": 0.856112188303197, + "flos": 15683941154880.0, + "grad_norm": 1.9201705321436473, + "learning_rate": 2.1317191989226302e-07, + "loss": 0.7427, + "num_input_tokens_seen": 308732405, + "step": 9493 + }, + { + "epoch": 0.8562023718266673, + "flos": 22059744223680.0, + "grad_norm": 2.104777319902448, + "learning_rate": 2.129095487415027e-07, + "loss": 0.7248, + "num_input_tokens_seen": 308761390, + "step": 9494 + }, + { + "epoch": 0.8562925553501375, + "flos": 20127876594720.0, + "grad_norm": 2.1439378261305135, + "learning_rate": 2.1264733007515257e-07, + "loss": 0.7506, + "num_input_tokens_seen": 308789325, + "step": 9495 + }, + { + "epoch": 0.8563827388736078, + "flos": 40421075780160.0, + "grad_norm": 3.144790695576088, + "learning_rate": 2.1238526391558852e-07, + "loss": 0.7255, + "num_input_tokens_seen": 308820645, + "step": 9496 + }, + { + "epoch": 0.8564729223970781, + "flos": 17867599703520.0, + "grad_norm": 2.048615509143426, + "learning_rate": 2.1212335028517003e-07, + "loss": 0.8798, + "num_input_tokens_seen": 308844005, + "step": 9497 + }, + { + "epoch": 0.8565631059205483, + "flos": 21840073710240.0, + "grad_norm": 2.0743141511576475, + "learning_rate": 2.1186158920624563e-07, + "loss": 0.7503, + "num_input_tokens_seen": 308872025, + "step": 9498 + }, + { + "epoch": 0.8566532894440185, + "flos": 21765685553280.0, + "grad_norm": 1.84109053108292, + "learning_rate": 2.1159998070115015e-07, + "loss": 0.7406, + "num_input_tokens_seen": 308899225, + "step": 9499 + }, + { + "epoch": 0.8567434729674889, + "flos": 19470723542400.0, + "grad_norm": 2.254446510341987, + "learning_rate": 2.113385247922055e-07, + "loss": 0.7273, + "num_input_tokens_seen": 308927860, + "step": 9500 + }, + { + "epoch": 0.8568336564909591, + "flos": 16332061955520.0, + "grad_norm": 1.7534530495916314, + "learning_rate": 2.1107722150172068e-07, + "loss": 0.7485, + "num_input_tokens_seen": 308951375, + "step": 9501 + }, + { + "epoch": 0.8569238400144293, + "flos": 30911777721120.0, + "grad_norm": 2.5315367924734433, + "learning_rate": 2.108160708519906e-07, + "loss": 0.7789, + "num_input_tokens_seen": 308983115, + "step": 9502 + }, + { + "epoch": 0.8570140235378996, + "flos": 23330593145280.0, + "grad_norm": 2.3607492044874308, + "learning_rate": 2.1055507286529984e-07, + "loss": 0.7197, + "num_input_tokens_seen": 309010445, + "step": 9503 + }, + { + "epoch": 0.8571042070613699, + "flos": 26280998399520.0, + "grad_norm": 2.123676112047653, + "learning_rate": 2.1029422756391612e-07, + "loss": 0.7626, + "num_input_tokens_seen": 309040280, + "step": 9504 + }, + { + "epoch": 0.8571943905848401, + "flos": 40637140826880.0, + "grad_norm": 1.7700673484920375, + "learning_rate": 2.1003353497009812e-07, + "loss": 0.6542, + "num_input_tokens_seen": 309072365, + "step": 9505 + }, + { + "epoch": 0.8572845741083104, + "flos": 24062766240480.0, + "grad_norm": 2.0302732026364705, + "learning_rate": 2.0977299510608825e-07, + "loss": 0.6635, + "num_input_tokens_seen": 309101115, + "step": 9506 + }, + { + "epoch": 0.8573747576317807, + "flos": 23443940944320.0, + "grad_norm": 1.718261094349311, + "learning_rate": 2.0951260799411784e-07, + "loss": 0.7813, + "num_input_tokens_seen": 309130110, + "step": 9507 + }, + { + "epoch": 0.857464941155251, + "flos": 21840779935680.0, + "grad_norm": 1.804255247988263, + "learning_rate": 2.0925237365640424e-07, + "loss": 0.7333, + "num_input_tokens_seen": 309159310, + "step": 9508 + }, + { + "epoch": 0.8575551246787212, + "flos": 70466920675200.0, + "grad_norm": 0.6383063293908788, + "learning_rate": 2.0899229211515211e-07, + "loss": 0.575, + "num_input_tokens_seen": 309252775, + "step": 9509 + }, + { + "epoch": 0.8576453082021914, + "flos": 22708459740480.0, + "grad_norm": 4.642484247512674, + "learning_rate": 2.0873236339255306e-07, + "loss": 0.6679, + "num_input_tokens_seen": 309278095, + "step": 9510 + }, + { + "epoch": 0.8577354917256618, + "flos": 19252205291520.0, + "grad_norm": 2.0941342812505925, + "learning_rate": 2.0847258751078644e-07, + "loss": 0.6901, + "num_input_tokens_seen": 309306915, + "step": 9511 + }, + { + "epoch": 0.857825675249132, + "flos": 19327225334400.0, + "grad_norm": 2.555570863825333, + "learning_rate": 2.082129644920163e-07, + "loss": 0.7186, + "num_input_tokens_seen": 309334955, + "step": 9512 + }, + { + "epoch": 0.8579158587726022, + "flos": 24281581849440.0, + "grad_norm": 1.6149290965669287, + "learning_rate": 2.0795349435839605e-07, + "loss": 0.7518, + "num_input_tokens_seen": 309365380, + "step": 9513 + }, + { + "epoch": 0.8580060422960725, + "flos": 11821135140960.0, + "grad_norm": 2.0021662900994137, + "learning_rate": 2.0769417713206484e-07, + "loss": 0.714, + "num_input_tokens_seen": 309390020, + "step": 9514 + }, + { + "epoch": 0.8580962258195428, + "flos": 24025516407360.0, + "grad_norm": 1.4251022603335652, + "learning_rate": 2.074350128351492e-07, + "loss": 0.8043, + "num_input_tokens_seen": 309421015, + "step": 9515 + }, + { + "epoch": 0.858186409343013, + "flos": 30074843041440.0, + "grad_norm": 1.6351108710585, + "learning_rate": 2.0717600148976256e-07, + "loss": 0.6609, + "num_input_tokens_seen": 309448820, + "step": 9516 + }, + { + "epoch": 0.8582765928664833, + "flos": 24244666544160.0, + "grad_norm": 2.362721879823456, + "learning_rate": 2.0691714311800436e-07, + "loss": 0.6723, + "num_input_tokens_seen": 309476985, + "step": 9517 + }, + { + "epoch": 0.8583667763899535, + "flos": 20091890533440.0, + "grad_norm": 1.9763761892452067, + "learning_rate": 2.066584377419631e-07, + "loss": 0.8048, + "num_input_tokens_seen": 309504930, + "step": 9518 + }, + { + "epoch": 0.8584569599134239, + "flos": 21002692993440.0, + "grad_norm": 1.6825712068178789, + "learning_rate": 2.0639988538371167e-07, + "loss": 0.7643, + "num_input_tokens_seen": 309531875, + "step": 9519 + }, + { + "epoch": 0.8585471434368941, + "flos": 32224225337760.0, + "grad_norm": 1.9930644180981008, + "learning_rate": 2.0614148606531258e-07, + "loss": 0.707, + "num_input_tokens_seen": 309563080, + "step": 9520 + }, + { + "epoch": 0.8586373269603643, + "flos": 32733940187520.0, + "grad_norm": 2.475114368917362, + "learning_rate": 2.0588323980881285e-07, + "loss": 0.7721, + "num_input_tokens_seen": 309590640, + "step": 9521 + }, + { + "epoch": 0.8587275104838346, + "flos": 20817447411360.0, + "grad_norm": 1.8407407869924166, + "learning_rate": 2.0562514663624752e-07, + "loss": 0.7301, + "num_input_tokens_seen": 309618210, + "step": 9522 + }, + { + "epoch": 0.8588176940073049, + "flos": 25994410850880.0, + "grad_norm": 1.81285157575679, + "learning_rate": 2.0536720656963902e-07, + "loss": 0.8071, + "num_input_tokens_seen": 309649675, + "step": 9523 + }, + { + "epoch": 0.8589078775307751, + "flos": 34659042920160.0, + "grad_norm": 2.7010032838408113, + "learning_rate": 2.051094196309957e-07, + "loss": 0.7673, + "num_input_tokens_seen": 309675950, + "step": 9524 + }, + { + "epoch": 0.8589980610542454, + "flos": 22496855064960.0, + "grad_norm": 2.8481518558778824, + "learning_rate": 2.0485178584231378e-07, + "loss": 0.7018, + "num_input_tokens_seen": 309706280, + "step": 9525 + }, + { + "epoch": 0.8590882445777156, + "flos": 63474373816800.0, + "grad_norm": 0.6734855112543644, + "learning_rate": 2.0459430522557587e-07, + "loss": 0.5768, + "num_input_tokens_seen": 309789660, + "step": 9526 + }, + { + "epoch": 0.8591784281011859, + "flos": 24318571494240.0, + "grad_norm": 2.857822684743846, + "learning_rate": 2.0433697780275195e-07, + "loss": 0.7427, + "num_input_tokens_seen": 309818715, + "step": 9527 + }, + { + "epoch": 0.8592686116246562, + "flos": 25487112035520.0, + "grad_norm": 1.8088030766079233, + "learning_rate": 2.040798035957978e-07, + "loss": 0.7579, + "num_input_tokens_seen": 309850115, + "step": 9528 + }, + { + "epoch": 0.8593587951481264, + "flos": 70776367626240.0, + "grad_norm": 0.7240020180747848, + "learning_rate": 2.038227826266574e-07, + "loss": 0.5926, + "num_input_tokens_seen": 309947390, + "step": 9529 + }, + { + "epoch": 0.8594489786715968, + "flos": 22278225304800.0, + "grad_norm": 2.014304623133406, + "learning_rate": 2.0356591491726126e-07, + "loss": 0.7506, + "num_input_tokens_seen": 309974585, + "step": 9530 + }, + { + "epoch": 0.859539162195067, + "flos": 27779137635360.0, + "grad_norm": 1.8029757882564426, + "learning_rate": 2.033092004895267e-07, + "loss": 0.7831, + "num_input_tokens_seen": 310006380, + "step": 9531 + }, + { + "epoch": 0.8596293457185372, + "flos": 20201595696000.0, + "grad_norm": 1.8618368952534423, + "learning_rate": 2.03052639365358e-07, + "loss": 0.7569, + "num_input_tokens_seen": 310034415, + "step": 9532 + }, + { + "epoch": 0.8597195292420075, + "flos": 26869004231040.0, + "grad_norm": 1.9612994220990891, + "learning_rate": 2.0279623156664694e-07, + "loss": 0.7614, + "num_input_tokens_seen": 310064760, + "step": 9533 + }, + { + "epoch": 0.8598097127654778, + "flos": 23042964843360.0, + "grad_norm": 2.214803920703936, + "learning_rate": 2.0253997711527005e-07, + "loss": 0.8004, + "num_input_tokens_seen": 310094200, + "step": 9534 + }, + { + "epoch": 0.859899896288948, + "flos": 18670890016800.0, + "grad_norm": 1.7936189424174676, + "learning_rate": 2.0228387603309428e-07, + "loss": 0.7913, + "num_input_tokens_seen": 310122610, + "step": 9535 + }, + { + "epoch": 0.8599900798124183, + "flos": 24536532198720.0, + "grad_norm": 1.793457762408815, + "learning_rate": 2.0202792834197035e-07, + "loss": 0.7629, + "num_input_tokens_seen": 310150915, + "step": 9536 + }, + { + "epoch": 0.8600802633358885, + "flos": 18375641914080.0, + "grad_norm": 2.2177712874274036, + "learning_rate": 2.017721340637375e-07, + "loss": 0.6727, + "num_input_tokens_seen": 310176730, + "step": 9537 + }, + { + "epoch": 0.8601704468593588, + "flos": 21476347442400.0, + "grad_norm": 1.7538554772587374, + "learning_rate": 2.0151649322022134e-07, + "loss": 0.7265, + "num_input_tokens_seen": 310204750, + "step": 9538 + }, + { + "epoch": 0.8602606303828291, + "flos": 27851146927680.0, + "grad_norm": 2.131144000239275, + "learning_rate": 2.012610058332349e-07, + "loss": 0.5506, + "num_input_tokens_seen": 310232500, + "step": 9539 + }, + { + "epoch": 0.8603508139062993, + "flos": 71460437318400.0, + "grad_norm": 0.6848923038831697, + "learning_rate": 2.010056719245774e-07, + "loss": 0.573, + "num_input_tokens_seen": 310320190, + "step": 9540 + }, + { + "epoch": 0.8604409974297695, + "flos": 35718324336000.0, + "grad_norm": 2.0364284529449157, + "learning_rate": 2.0075049151603563e-07, + "loss": 0.7664, + "num_input_tokens_seen": 310350355, + "step": 9541 + }, + { + "epoch": 0.8605311809532399, + "flos": 17031148230720.0, + "grad_norm": 1.9098830076407183, + "learning_rate": 2.0049546462938326e-07, + "loss": 0.748, + "num_input_tokens_seen": 310376575, + "step": 9542 + }, + { + "epoch": 0.8606213644767101, + "flos": 30547085039520.0, + "grad_norm": 1.732762317257595, + "learning_rate": 2.0024059128637961e-07, + "loss": 0.7642, + "num_input_tokens_seen": 310407600, + "step": 9543 + }, + { + "epoch": 0.8607115480001803, + "flos": 29745355856160.0, + "grad_norm": 1.6087235548903236, + "learning_rate": 1.99985871508773e-07, + "loss": 0.6076, + "num_input_tokens_seen": 310439075, + "step": 9544 + }, + { + "epoch": 0.8608017315236506, + "flos": 27160275169440.0, + "grad_norm": 2.320727070465116, + "learning_rate": 1.9973130531829674e-07, + "loss": 0.7686, + "num_input_tokens_seen": 310471205, + "step": 9545 + }, + { + "epoch": 0.8608919150471209, + "flos": 19618793630880.0, + "grad_norm": 2.1629872870069446, + "learning_rate": 1.994768927366721e-07, + "loss": 0.6913, + "num_input_tokens_seen": 310499575, + "step": 9546 + }, + { + "epoch": 0.8609820985705912, + "flos": 15536019745440.0, + "grad_norm": 2.7312054383856736, + "learning_rate": 1.992226337856069e-07, + "loss": 0.7116, + "num_input_tokens_seen": 310526315, + "step": 9547 + }, + { + "epoch": 0.8610722820940614, + "flos": 23625766908480.0, + "grad_norm": 1.7542974974470409, + "learning_rate": 1.9896852848679592e-07, + "loss": 0.698, + "num_input_tokens_seen": 310555370, + "step": 9548 + }, + { + "epoch": 0.8611624656175316, + "flos": 23735026033920.0, + "grad_norm": 1.8008985501892467, + "learning_rate": 1.9871457686192094e-07, + "loss": 0.751, + "num_input_tokens_seen": 310585330, + "step": 9549 + }, + { + "epoch": 0.861252649141002, + "flos": 20303978415840.0, + "grad_norm": 2.8702288304460213, + "learning_rate": 1.984607789326509e-07, + "loss": 0.735, + "num_input_tokens_seen": 310612540, + "step": 9550 + }, + { + "epoch": 0.8613428326644722, + "flos": 25010223817440.0, + "grad_norm": 1.640177275101431, + "learning_rate": 1.982071347206402e-07, + "loss": 0.6894, + "num_input_tokens_seen": 310641920, + "step": 9551 + }, + { + "epoch": 0.8614330161879424, + "flos": 43878333812640.0, + "grad_norm": 1.7425217927858865, + "learning_rate": 1.9795364424753202e-07, + "loss": 0.7208, + "num_input_tokens_seen": 310674445, + "step": 9552 + }, + { + "epoch": 0.8615231997114128, + "flos": 20889419533920.0, + "grad_norm": 10.047327959875739, + "learning_rate": 1.9770030753495505e-07, + "loss": 0.7117, + "num_input_tokens_seen": 310699740, + "step": 9553 + }, + { + "epoch": 0.861613383234883, + "flos": 20745698307360.0, + "grad_norm": 1.833264356818348, + "learning_rate": 1.9744712460452573e-07, + "loss": 0.7022, + "num_input_tokens_seen": 310727765, + "step": 9554 + }, + { + "epoch": 0.8617035667583532, + "flos": 24573744862080.0, + "grad_norm": 2.0901546068820283, + "learning_rate": 1.9719409547784703e-07, + "loss": 0.723, + "num_input_tokens_seen": 310756550, + "step": 9555 + }, + { + "epoch": 0.8617937502818235, + "flos": 23074862231040.0, + "grad_norm": 2.754088916537168, + "learning_rate": 1.9694122017650837e-07, + "loss": 0.7652, + "num_input_tokens_seen": 310783510, + "step": 9556 + }, + { + "epoch": 0.8618839338052938, + "flos": 30217486344960.0, + "grad_norm": 1.7926312962200626, + "learning_rate": 1.9668849872208738e-07, + "loss": 0.7522, + "num_input_tokens_seen": 310812705, + "step": 9557 + }, + { + "epoch": 0.8619741173287641, + "flos": 19907537025600.0, + "grad_norm": 1.8128970796436887, + "learning_rate": 1.9643593113614632e-07, + "loss": 0.5897, + "num_input_tokens_seen": 310840905, + "step": 9558 + }, + { + "epoch": 0.8620643008522343, + "flos": 25082864995680.0, + "grad_norm": 1.5962701208030945, + "learning_rate": 1.961835174402371e-07, + "loss": 0.7838, + "num_input_tokens_seen": 310870250, + "step": 9559 + }, + { + "epoch": 0.8621544843757045, + "flos": 23626027096800.0, + "grad_norm": 3.5989880411087705, + "learning_rate": 1.9593125765589535e-07, + "loss": 0.7072, + "num_input_tokens_seen": 310897715, + "step": 9560 + }, + { + "epoch": 0.8622446678991749, + "flos": 23115643191360.0, + "grad_norm": 1.8323260560169066, + "learning_rate": 1.9567915180464721e-07, + "loss": 0.6867, + "num_input_tokens_seen": 310926105, + "step": 9561 + }, + { + "epoch": 0.8623348514226451, + "flos": 26140027735200.0, + "grad_norm": 2.3826775184295994, + "learning_rate": 1.9542719990800217e-07, + "loss": 0.7259, + "num_input_tokens_seen": 310954555, + "step": 9562 + }, + { + "epoch": 0.8624250349461153, + "flos": 17936152208160.0, + "grad_norm": 1.9060919023835212, + "learning_rate": 1.9517540198745896e-07, + "loss": 0.8109, + "num_input_tokens_seen": 310981460, + "step": 9563 + }, + { + "epoch": 0.8625152184695856, + "flos": 21767506871520.0, + "grad_norm": 2.570940213529472, + "learning_rate": 1.94923758064502e-07, + "loss": 0.7389, + "num_input_tokens_seen": 311008710, + "step": 9564 + }, + { + "epoch": 0.8626054019930559, + "flos": 20416545649920.0, + "grad_norm": 1.7643054994143366, + "learning_rate": 1.9467226816060322e-07, + "loss": 0.708, + "num_input_tokens_seen": 311036950, + "step": 9565 + }, + { + "epoch": 0.8626955855165261, + "flos": 21622373194080.0, + "grad_norm": 1.8600766436875245, + "learning_rate": 1.9442093229722122e-07, + "loss": 0.6838, + "num_input_tokens_seen": 311065675, + "step": 9566 + }, + { + "epoch": 0.8627857690399964, + "flos": 14845333836000.0, + "grad_norm": 2.820326532978852, + "learning_rate": 1.9416975049580085e-07, + "loss": 0.8151, + "num_input_tokens_seen": 311089250, + "step": 9567 + }, + { + "epoch": 0.8628759525634666, + "flos": 28325247413760.0, + "grad_norm": 2.088786832985785, + "learning_rate": 1.9391872277777456e-07, + "loss": 0.7242, + "num_input_tokens_seen": 311116430, + "step": 9568 + }, + { + "epoch": 0.862966136086937, + "flos": 22459419383040.0, + "grad_norm": 2.5371599317157765, + "learning_rate": 1.9366784916456158e-07, + "loss": 0.7302, + "num_input_tokens_seen": 311141785, + "step": 9569 + }, + { + "epoch": 0.8630563196104072, + "flos": 24536829556800.0, + "grad_norm": 2.027029494550423, + "learning_rate": 1.9341712967756774e-07, + "loss": 0.6942, + "num_input_tokens_seen": 311172240, + "step": 9570 + }, + { + "epoch": 0.8631465031338774, + "flos": 23115977719200.0, + "grad_norm": 3.4248840479470726, + "learning_rate": 1.9316656433818566e-07, + "loss": 0.6508, + "num_input_tokens_seen": 311202570, + "step": 9571 + }, + { + "epoch": 0.8632366866573477, + "flos": 29673866940480.0, + "grad_norm": 2.787694686454468, + "learning_rate": 1.929161531677954e-07, + "loss": 0.6414, + "num_input_tokens_seen": 311231375, + "step": 9572 + }, + { + "epoch": 0.863326870180818, + "flos": 23547067059360.0, + "grad_norm": 1.7961657243787343, + "learning_rate": 1.9266589618776251e-07, + "loss": 0.6972, + "num_input_tokens_seen": 311259010, + "step": 9573 + }, + { + "epoch": 0.8634170537042882, + "flos": 48723468372000.0, + "grad_norm": 1.9058056923291165, + "learning_rate": 1.924157934194417e-07, + "loss": 0.792, + "num_input_tokens_seen": 311290865, + "step": 9574 + }, + { + "epoch": 0.8635072372277585, + "flos": 26358062779200.0, + "grad_norm": 1.9494245951790026, + "learning_rate": 1.9216584488417142e-07, + "loss": 0.6324, + "num_input_tokens_seen": 311321090, + "step": 9575 + }, + { + "epoch": 0.8635974207512288, + "flos": 14553393841920.0, + "grad_norm": 2.275658849534106, + "learning_rate": 1.919160506032802e-07, + "loss": 0.7976, + "num_input_tokens_seen": 311347185, + "step": 9576 + }, + { + "epoch": 0.863687604274699, + "flos": 20128434141120.0, + "grad_norm": 2.615422989166755, + "learning_rate": 1.916664105980812e-07, + "loss": 0.8022, + "num_input_tokens_seen": 311373060, + "step": 9577 + }, + { + "epoch": 0.8637777877981693, + "flos": 35211322878720.0, + "grad_norm": 4.939901805936523, + "learning_rate": 1.914169248898747e-07, + "loss": 0.7742, + "num_input_tokens_seen": 311402050, + "step": 9578 + }, + { + "epoch": 0.8638679713216395, + "flos": 21767246683200.0, + "grad_norm": 1.6561856527244412, + "learning_rate": 1.9116759349994882e-07, + "loss": 0.6551, + "num_input_tokens_seen": 311431480, + "step": 9579 + }, + { + "epoch": 0.8639581548451098, + "flos": 20197990229280.0, + "grad_norm": 1.8976642933561592, + "learning_rate": 1.9091841644957763e-07, + "loss": 0.7746, + "num_input_tokens_seen": 311459695, + "step": 9580 + }, + { + "epoch": 0.8640483383685801, + "flos": 23336354458080.0, + "grad_norm": 1.4436147254884084, + "learning_rate": 1.9066939376002278e-07, + "loss": 0.8248, + "num_input_tokens_seen": 311491845, + "step": 9581 + }, + { + "epoch": 0.8641385218920503, + "flos": 19873186433760.0, + "grad_norm": 2.4992416185225323, + "learning_rate": 1.9042052545253085e-07, + "loss": 0.6561, + "num_input_tokens_seen": 311516985, + "step": 9582 + }, + { + "epoch": 0.8642287054155205, + "flos": 21548691262560.0, + "grad_norm": 2.560739051712766, + "learning_rate": 1.901718115483384e-07, + "loss": 0.7179, + "num_input_tokens_seen": 311544585, + "step": 9583 + }, + { + "epoch": 0.8643188889389909, + "flos": 22235585856480.0, + "grad_norm": 2.0709478590962194, + "learning_rate": 1.8992325206866598e-07, + "loss": 0.7259, + "num_input_tokens_seen": 311573170, + "step": 9584 + }, + { + "epoch": 0.8644090724624611, + "flos": 27632851695360.0, + "grad_norm": 1.4803750202230812, + "learning_rate": 1.8967484703472225e-07, + "loss": 0.714, + "num_input_tokens_seen": 311606095, + "step": 9585 + }, + { + "epoch": 0.8644992559859314, + "flos": 24170464236000.0, + "grad_norm": 1.5152149235577583, + "learning_rate": 1.8942659646770288e-07, + "loss": 0.7877, + "num_input_tokens_seen": 311636380, + "step": 9586 + }, + { + "epoch": 0.8645894395094016, + "flos": 24791928585120.0, + "grad_norm": 2.0759137083726227, + "learning_rate": 1.8917850038878936e-07, + "loss": 0.8058, + "num_input_tokens_seen": 311663360, + "step": 9587 + }, + { + "epoch": 0.8646796230328719, + "flos": 29637509181600.0, + "grad_norm": 1.8377920320924934, + "learning_rate": 1.8893055881915121e-07, + "loss": 0.6525, + "num_input_tokens_seen": 311693685, + "step": 9588 + }, + { + "epoch": 0.8647698065563422, + "flos": 19290347198880.0, + "grad_norm": 2.5768208297302113, + "learning_rate": 1.886827717799442e-07, + "loss": 0.6874, + "num_input_tokens_seen": 311719545, + "step": 9589 + }, + { + "epoch": 0.8648599900798124, + "flos": 17759827368480.0, + "grad_norm": 2.1353688552590286, + "learning_rate": 1.884351392923096e-07, + "loss": 0.7557, + "num_input_tokens_seen": 311747000, + "step": 9590 + }, + { + "epoch": 0.8649501736032826, + "flos": 22234507933440.0, + "grad_norm": 2.50953249312405, + "learning_rate": 1.8818766137737896e-07, + "loss": 0.7812, + "num_input_tokens_seen": 311771720, + "step": 9591 + }, + { + "epoch": 0.865040357126753, + "flos": 18298466025120.0, + "grad_norm": 2.2620465662014064, + "learning_rate": 1.8794033805626653e-07, + "loss": 0.7156, + "num_input_tokens_seen": 311795215, + "step": 9592 + }, + { + "epoch": 0.8651305406502232, + "flos": 38342178816000.0, + "grad_norm": 2.3430328289146582, + "learning_rate": 1.876931693500763e-07, + "loss": 0.735, + "num_input_tokens_seen": 311829100, + "step": 9593 + }, + { + "epoch": 0.8652207241736934, + "flos": 36516485222400.0, + "grad_norm": 2.3688123307293325, + "learning_rate": 1.8744615527989783e-07, + "loss": 0.588, + "num_input_tokens_seen": 311859925, + "step": 9594 + }, + { + "epoch": 0.8653109076971637, + "flos": 21585532228320.0, + "grad_norm": 2.333556630031237, + "learning_rate": 1.871992958668076e-07, + "loss": 0.8277, + "num_input_tokens_seen": 311888075, + "step": 9595 + }, + { + "epoch": 0.865401091220634, + "flos": 18306680542080.0, + "grad_norm": 2.3328931484091875, + "learning_rate": 1.8695259113186944e-07, + "loss": 0.7828, + "num_input_tokens_seen": 311912215, + "step": 9596 + }, + { + "epoch": 0.8654912747441043, + "flos": 18999039090720.0, + "grad_norm": 1.6332627547189547, + "learning_rate": 1.8670604109613252e-07, + "loss": 0.6997, + "num_input_tokens_seen": 311940110, + "step": 9597 + }, + { + "epoch": 0.8655814582675745, + "flos": 26210959104480.0, + "grad_norm": 2.020378127068463, + "learning_rate": 1.8645964578063533e-07, + "loss": 0.7841, + "num_input_tokens_seen": 311970825, + "step": 9598 + }, + { + "epoch": 0.8656716417910447, + "flos": 24646237361280.0, + "grad_norm": 1.5502912770660062, + "learning_rate": 1.862134052064006e-07, + "loss": 0.7987, + "num_input_tokens_seen": 312000040, + "step": 9599 + }, + { + "epoch": 0.8657618253145151, + "flos": 18707507964000.0, + "grad_norm": 7.693639769024009, + "learning_rate": 1.8596731939443932e-07, + "loss": 0.784, + "num_input_tokens_seen": 312025020, + "step": 9600 + }, + { + "epoch": 0.8658520088379853, + "flos": 16483923359520.0, + "grad_norm": 1.7031840042952442, + "learning_rate": 1.857213883657487e-07, + "loss": 0.7694, + "num_input_tokens_seen": 312052025, + "step": 9601 + }, + { + "epoch": 0.8659421923614555, + "flos": 56332716116640.0, + "grad_norm": 0.6366782146258713, + "learning_rate": 1.8547561214131303e-07, + "loss": 0.5497, + "num_input_tokens_seen": 312140245, + "step": 9602 + }, + { + "epoch": 0.8660323758849259, + "flos": 23407546015680.0, + "grad_norm": 2.0797948796661645, + "learning_rate": 1.8522999074210355e-07, + "loss": 0.8178, + "num_input_tokens_seen": 312167235, + "step": 9603 + }, + { + "epoch": 0.8661225594083961, + "flos": 19546078113120.0, + "grad_norm": 2.1903495756917226, + "learning_rate": 1.849845241890775e-07, + "loss": 0.7815, + "num_input_tokens_seen": 312194230, + "step": 9604 + }, + { + "epoch": 0.8662127429318663, + "flos": 25192941855840.0, + "grad_norm": 2.3717580050235543, + "learning_rate": 1.8473921250317992e-07, + "loss": 0.7456, + "num_input_tokens_seen": 312224620, + "step": 9605 + }, + { + "epoch": 0.8663029264553366, + "flos": 31783880727360.0, + "grad_norm": 1.7938226399350916, + "learning_rate": 1.8449405570534225e-07, + "loss": 0.6708, + "num_input_tokens_seen": 312251220, + "step": 9606 + }, + { + "epoch": 0.8663931099788069, + "flos": 22314174196320.0, + "grad_norm": 1.8378160321688686, + "learning_rate": 1.8424905381648204e-07, + "loss": 0.7865, + "num_input_tokens_seen": 312280305, + "step": 9607 + }, + { + "epoch": 0.8664832935022772, + "flos": 25920617410080.0, + "grad_norm": 1.9396883388767607, + "learning_rate": 1.8400420685750452e-07, + "loss": 0.696, + "num_input_tokens_seen": 312313220, + "step": 9608 + }, + { + "epoch": 0.8665734770257474, + "flos": 23807778721440.0, + "grad_norm": 1.9209625435135775, + "learning_rate": 1.8375951484930142e-07, + "loss": 0.6758, + "num_input_tokens_seen": 312341665, + "step": 9609 + }, + { + "epoch": 0.8666636605492176, + "flos": 22202536206240.0, + "grad_norm": 1.8098038318404188, + "learning_rate": 1.8351497781275094e-07, + "loss": 0.7448, + "num_input_tokens_seen": 312370455, + "step": 9610 + }, + { + "epoch": 0.866753844072688, + "flos": 19399940852160.0, + "grad_norm": 1.9079151072859892, + "learning_rate": 1.8327059576871907e-07, + "loss": 0.7282, + "num_input_tokens_seen": 312397165, + "step": 9611 + }, + { + "epoch": 0.8668440275961582, + "flos": 22895489471040.0, + "grad_norm": 1.5845428498271141, + "learning_rate": 1.8302636873805665e-07, + "loss": 0.7906, + "num_input_tokens_seen": 312427530, + "step": 9612 + }, + { + "epoch": 0.8669342111196284, + "flos": 25338261382080.0, + "grad_norm": 2.039379109908372, + "learning_rate": 1.8278229674160373e-07, + "loss": 0.8174, + "num_input_tokens_seen": 312453770, + "step": 9613 + }, + { + "epoch": 0.8670243946430987, + "flos": 43115229743520.0, + "grad_norm": 1.8756277305128084, + "learning_rate": 1.825383798001845e-07, + "loss": 0.6736, + "num_input_tokens_seen": 312486905, + "step": 9614 + }, + { + "epoch": 0.867114578166569, + "flos": 22386555186240.0, + "grad_norm": 1.7024046048005288, + "learning_rate": 1.8229461793461297e-07, + "loss": 0.684, + "num_input_tokens_seen": 312517965, + "step": 9615 + }, + { + "epoch": 0.8672047616900392, + "flos": 46287163999200.0, + "grad_norm": 1.8712932822726662, + "learning_rate": 1.8205101116568698e-07, + "loss": 0.7535, + "num_input_tokens_seen": 312548555, + "step": 9616 + }, + { + "epoch": 0.8672949452135095, + "flos": 18889556946720.0, + "grad_norm": 2.2004237287114647, + "learning_rate": 1.818075595141928e-07, + "loss": 0.7914, + "num_input_tokens_seen": 312575310, + "step": 9617 + }, + { + "epoch": 0.8673851287369797, + "flos": 22640799310080.0, + "grad_norm": 3.0569221927522956, + "learning_rate": 1.8156426300090288e-07, + "loss": 0.7382, + "num_input_tokens_seen": 312603405, + "step": 9618 + }, + { + "epoch": 0.86747531226045, + "flos": 21731706659040.0, + "grad_norm": 1.7870077802619995, + "learning_rate": 1.8132112164657686e-07, + "loss": 0.732, + "num_input_tokens_seen": 312633110, + "step": 9619 + }, + { + "epoch": 0.8675654957839203, + "flos": 28289112673440.0, + "grad_norm": 2.839993185653033, + "learning_rate": 1.8107813547196106e-07, + "loss": 0.6338, + "num_input_tokens_seen": 312661925, + "step": 9620 + }, + { + "epoch": 0.8676556793073905, + "flos": 16922520991200.0, + "grad_norm": 1.9365645580764035, + "learning_rate": 1.8083530449778817e-07, + "loss": 0.7788, + "num_input_tokens_seen": 312689310, + "step": 9621 + }, + { + "epoch": 0.8677458628308607, + "flos": 36923408484960.0, + "grad_norm": 1.7426315220764885, + "learning_rate": 1.8059262874477787e-07, + "loss": 0.6616, + "num_input_tokens_seen": 312718495, + "step": 9622 + }, + { + "epoch": 0.8678360463543311, + "flos": 26759484917280.0, + "grad_norm": 1.753013468575841, + "learning_rate": 1.8035010823363627e-07, + "loss": 0.7248, + "num_input_tokens_seen": 312751030, + "step": 9623 + }, + { + "epoch": 0.8679262298778013, + "flos": 25732918623840.0, + "grad_norm": 1.9959287929867597, + "learning_rate": 1.8010774298505705e-07, + "loss": 0.7787, + "num_input_tokens_seen": 312781065, + "step": 9624 + }, + { + "epoch": 0.8680164134012716, + "flos": 16630915524960.0, + "grad_norm": 1.9711758034433267, + "learning_rate": 1.7986553301972007e-07, + "loss": 0.8323, + "num_input_tokens_seen": 312802770, + "step": 9625 + }, + { + "epoch": 0.8681065969247419, + "flos": 20269293296160.0, + "grad_norm": 1.7703258086307796, + "learning_rate": 1.7962347835829171e-07, + "loss": 0.7305, + "num_input_tokens_seen": 312830415, + "step": 9626 + }, + { + "epoch": 0.8681967804482121, + "flos": 22164394298880.0, + "grad_norm": 2.1343581638817124, + "learning_rate": 1.793815790214257e-07, + "loss": 0.7462, + "num_input_tokens_seen": 312857765, + "step": 9627 + }, + { + "epoch": 0.8682869639716824, + "flos": 16011904380000.0, + "grad_norm": 2.3603563099546276, + "learning_rate": 1.791398350297626e-07, + "loss": 0.7018, + "num_input_tokens_seen": 312883385, + "step": 9628 + }, + { + "epoch": 0.8683771474951526, + "flos": 25950990837600.0, + "grad_norm": 1.569715775560369, + "learning_rate": 1.7889824640392813e-07, + "loss": 0.7465, + "num_input_tokens_seen": 312912795, + "step": 9629 + }, + { + "epoch": 0.868467331018623, + "flos": 19654519503840.0, + "grad_norm": 1.8963209370550853, + "learning_rate": 1.7865681316453741e-07, + "loss": 0.7216, + "num_input_tokens_seen": 312939145, + "step": 9630 + }, + { + "epoch": 0.8685575145420932, + "flos": 26285607449760.0, + "grad_norm": 1.9547094403323138, + "learning_rate": 1.7841553533218968e-07, + "loss": 0.738, + "num_input_tokens_seen": 312970260, + "step": 9631 + }, + { + "epoch": 0.8686476980655634, + "flos": 18306420353760.0, + "grad_norm": 2.183861021769092, + "learning_rate": 1.7817441292747292e-07, + "loss": 0.843, + "num_input_tokens_seen": 312996980, + "step": 9632 + }, + { + "epoch": 0.8687378815890336, + "flos": 24714789865920.0, + "grad_norm": 2.321047990643844, + "learning_rate": 1.779334459709607e-07, + "loss": 0.7377, + "num_input_tokens_seen": 313025905, + "step": 9633 + }, + { + "epoch": 0.868828065112504, + "flos": 22421760682560.0, + "grad_norm": 3.342919077084071, + "learning_rate": 1.7769263448321347e-07, + "loss": 0.7561, + "num_input_tokens_seen": 313050795, + "step": 9634 + }, + { + "epoch": 0.8689182486359742, + "flos": 21876580148160.0, + "grad_norm": 2.1205349180750677, + "learning_rate": 1.7745197848477879e-07, + "loss": 0.7323, + "num_input_tokens_seen": 313079350, + "step": 9635 + }, + { + "epoch": 0.8690084321594445, + "flos": 20492086069440.0, + "grad_norm": 2.3033414368917047, + "learning_rate": 1.7721147799619063e-07, + "loss": 0.842, + "num_input_tokens_seen": 313103775, + "step": 9636 + }, + { + "epoch": 0.8690986156829147, + "flos": 70013077708320.0, + "grad_norm": 0.592849600862142, + "learning_rate": 1.769711330379704e-07, + "loss": 0.5469, + "num_input_tokens_seen": 313208110, + "step": 9637 + }, + { + "epoch": 0.869188799206385, + "flos": 24317270552640.0, + "grad_norm": 1.992548950915034, + "learning_rate": 1.767309436306248e-07, + "loss": 0.7926, + "num_input_tokens_seen": 313234500, + "step": 9638 + }, + { + "epoch": 0.8692789827298553, + "flos": 23371708633440.0, + "grad_norm": 2.0594003678700035, + "learning_rate": 1.764909097946483e-07, + "loss": 0.8046, + "num_input_tokens_seen": 313262915, + "step": 9639 + }, + { + "epoch": 0.8693691662533255, + "flos": 21616351692960.0, + "grad_norm": 2.023518800022725, + "learning_rate": 1.7625103155052236e-07, + "loss": 0.8072, + "num_input_tokens_seen": 313289695, + "step": 9640 + }, + { + "epoch": 0.8694593497767957, + "flos": 20121929433120.0, + "grad_norm": 2.6760144208682024, + "learning_rate": 1.760113089187143e-07, + "loss": 0.7534, + "num_input_tokens_seen": 313316280, + "step": 9641 + }, + { + "epoch": 0.8695495333002661, + "flos": 27412437786720.0, + "grad_norm": 1.8545108932829308, + "learning_rate": 1.7577174191967868e-07, + "loss": 0.7676, + "num_input_tokens_seen": 313345655, + "step": 9642 + }, + { + "epoch": 0.8696397168237363, + "flos": 23480373042720.0, + "grad_norm": 2.027081246553022, + "learning_rate": 1.755323305738574e-07, + "loss": 0.7515, + "num_input_tokens_seen": 313373460, + "step": 9643 + }, + { + "epoch": 0.8697299003472065, + "flos": 22351052331840.0, + "grad_norm": 1.884224896122403, + "learning_rate": 1.7529307490167677e-07, + "loss": 0.7728, + "num_input_tokens_seen": 313404835, + "step": 9644 + }, + { + "epoch": 0.8698200838706768, + "flos": 25006469671680.0, + "grad_norm": 2.733299959568256, + "learning_rate": 1.7505397492355288e-07, + "loss": 0.7231, + "num_input_tokens_seen": 313432950, + "step": 9645 + }, + { + "epoch": 0.8699102673941471, + "flos": 22059855732960.0, + "grad_norm": 1.9029300107488756, + "learning_rate": 1.7481503065988589e-07, + "loss": 0.7784, + "num_input_tokens_seen": 313461760, + "step": 9646 + }, + { + "epoch": 0.8700004509176174, + "flos": 28072155552480.0, + "grad_norm": 1.5331105198197734, + "learning_rate": 1.7457624213106526e-07, + "loss": 0.7083, + "num_input_tokens_seen": 313495505, + "step": 9647 + }, + { + "epoch": 0.8700906344410876, + "flos": 19363174225920.0, + "grad_norm": 2.971683699009369, + "learning_rate": 1.7433760935746465e-07, + "loss": 0.7406, + "num_input_tokens_seen": 313520625, + "step": 9648 + }, + { + "epoch": 0.8701808179645579, + "flos": 20091927703200.0, + "grad_norm": 1.962779543945706, + "learning_rate": 1.740991323594456e-07, + "loss": 0.6849, + "num_input_tokens_seen": 313550825, + "step": 9649 + }, + { + "epoch": 0.8702710014880282, + "flos": 22386629525760.0, + "grad_norm": 1.9632119897844695, + "learning_rate": 1.7386081115735651e-07, + "loss": 0.7143, + "num_input_tokens_seen": 313578955, + "step": 9650 + }, + { + "epoch": 0.8703611850114984, + "flos": 25411088409120.0, + "grad_norm": 2.950995267661943, + "learning_rate": 1.736226457715324e-07, + "loss": 0.7006, + "num_input_tokens_seen": 313608285, + "step": 9651 + }, + { + "epoch": 0.8704513685349686, + "flos": 24901931105760.0, + "grad_norm": 1.5217910868796423, + "learning_rate": 1.7338463622229505e-07, + "loss": 0.6936, + "num_input_tokens_seen": 313641765, + "step": 9652 + }, + { + "epoch": 0.870541552058439, + "flos": 35205004019520.0, + "grad_norm": 2.118678282477624, + "learning_rate": 1.7314678252995152e-07, + "loss": 0.6923, + "num_input_tokens_seen": 313669900, + "step": 9653 + }, + { + "epoch": 0.8706317355819092, + "flos": 29707362627840.0, + "grad_norm": 1.758547209738594, + "learning_rate": 1.7290908471479847e-07, + "loss": 0.6747, + "num_input_tokens_seen": 313699690, + "step": 9654 + }, + { + "epoch": 0.8707219191053794, + "flos": 25338707419200.0, + "grad_norm": 1.8980992019703895, + "learning_rate": 1.7267154279711637e-07, + "loss": 0.7394, + "num_input_tokens_seen": 313729625, + "step": 9655 + }, + { + "epoch": 0.8708121026288497, + "flos": 22714518411360.0, + "grad_norm": 1.5659708002858874, + "learning_rate": 1.724341567971741e-07, + "loss": 0.7854, + "num_input_tokens_seen": 313759435, + "step": 9656 + }, + { + "epoch": 0.87090228615232, + "flos": 21836282394720.0, + "grad_norm": 1.7467019474795513, + "learning_rate": 1.7219692673522657e-07, + "loss": 0.8401, + "num_input_tokens_seen": 313786380, + "step": 9657 + }, + { + "epoch": 0.8709924696757902, + "flos": 24683450024640.0, + "grad_norm": 1.668693152898847, + "learning_rate": 1.7195985263151558e-07, + "loss": 0.7604, + "num_input_tokens_seen": 313815920, + "step": 9658 + }, + { + "epoch": 0.8710826531992605, + "flos": 27813228038880.0, + "grad_norm": 2.0432974731426703, + "learning_rate": 1.7172293450626985e-07, + "loss": 0.7899, + "num_input_tokens_seen": 313846440, + "step": 9659 + }, + { + "epoch": 0.8711728367227307, + "flos": 24749214797280.0, + "grad_norm": 1.9312976497956815, + "learning_rate": 1.7148617237970475e-07, + "loss": 0.7424, + "num_input_tokens_seen": 313875245, + "step": 9660 + }, + { + "epoch": 0.8712630202462011, + "flos": 27999737392800.0, + "grad_norm": 1.5761970603093562, + "learning_rate": 1.7124956627202102e-07, + "loss": 0.7065, + "num_input_tokens_seen": 313910080, + "step": 9661 + }, + { + "epoch": 0.8713532037696713, + "flos": 22128742765440.0, + "grad_norm": 2.240052486043189, + "learning_rate": 1.7101311620340852e-07, + "loss": 0.755, + "num_input_tokens_seen": 313931100, + "step": 9662 + }, + { + "epoch": 0.8714433872931415, + "flos": 11706003193440.0, + "grad_norm": 2.0352759737907156, + "learning_rate": 1.7077682219404155e-07, + "loss": 0.6833, + "num_input_tokens_seen": 313955405, + "step": 9663 + }, + { + "epoch": 0.8715335708166118, + "flos": 14363241851520.0, + "grad_norm": 2.383428075103311, + "learning_rate": 1.705406842640824e-07, + "loss": 0.8707, + "num_input_tokens_seen": 313973275, + "step": 9664 + }, + { + "epoch": 0.8716237543400821, + "flos": 21841225972800.0, + "grad_norm": 3.41939493125053, + "learning_rate": 1.7030470243367946e-07, + "loss": 0.7317, + "num_input_tokens_seen": 314003435, + "step": 9665 + }, + { + "epoch": 0.8717139378635523, + "flos": 17031780116640.0, + "grad_norm": 1.5601810031201342, + "learning_rate": 1.7006887672296834e-07, + "loss": 0.6677, + "num_input_tokens_seen": 314029420, + "step": 9666 + }, + { + "epoch": 0.8718041213870226, + "flos": 24427050054720.0, + "grad_norm": 2.188102530195444, + "learning_rate": 1.6983320715207094e-07, + "loss": 0.6857, + "num_input_tokens_seen": 314058920, + "step": 9667 + }, + { + "epoch": 0.8718943049104928, + "flos": 27704266271520.0, + "grad_norm": 1.8622299203331694, + "learning_rate": 1.6959769374109523e-07, + "loss": 0.7583, + "num_input_tokens_seen": 314088730, + "step": 9668 + }, + { + "epoch": 0.8719844884339631, + "flos": 22893073436640.0, + "grad_norm": 1.6963796465135954, + "learning_rate": 1.6936233651013754e-07, + "loss": 0.6752, + "num_input_tokens_seen": 314116045, + "step": 9669 + }, + { + "epoch": 0.8720746719574334, + "flos": 24097191171840.0, + "grad_norm": 1.8490918894784172, + "learning_rate": 1.691271354792787e-07, + "loss": 0.7702, + "num_input_tokens_seen": 314144835, + "step": 9670 + }, + { + "epoch": 0.8721648554809036, + "flos": 28361642342400.0, + "grad_norm": 2.0534536624098805, + "learning_rate": 1.6889209066858866e-07, + "loss": 0.6216, + "num_input_tokens_seen": 314176685, + "step": 9671 + }, + { + "epoch": 0.872255039004374, + "flos": 21220988225760.0, + "grad_norm": 2.120162506532852, + "learning_rate": 1.6865720209812185e-07, + "loss": 0.7833, + "num_input_tokens_seen": 314205515, + "step": 9672 + }, + { + "epoch": 0.8723452225278442, + "flos": 22606114190400.0, + "grad_norm": 2.787238627954007, + "learning_rate": 1.684224697879204e-07, + "loss": 0.7063, + "num_input_tokens_seen": 314232390, + "step": 9673 + }, + { + "epoch": 0.8724354060513144, + "flos": 21402405322560.0, + "grad_norm": 2.0109450546869674, + "learning_rate": 1.6818789375801302e-07, + "loss": 0.8157, + "num_input_tokens_seen": 314258620, + "step": 9674 + }, + { + "epoch": 0.8725255895747847, + "flos": 20784286251840.0, + "grad_norm": 1.6066890635980478, + "learning_rate": 1.679534740284152e-07, + "loss": 0.7003, + "num_input_tokens_seen": 314288420, + "step": 9675 + }, + { + "epoch": 0.872615773098255, + "flos": 19692066695040.0, + "grad_norm": 1.9929385097337329, + "learning_rate": 1.6771921061912853e-07, + "loss": 0.6829, + "num_input_tokens_seen": 314315295, + "step": 9676 + }, + { + "epoch": 0.8727059566217252, + "flos": 21038976412800.0, + "grad_norm": 2.0994322236312115, + "learning_rate": 1.6748510355014234e-07, + "loss": 0.6805, + "num_input_tokens_seen": 314342510, + "step": 9677 + }, + { + "epoch": 0.8727961401451955, + "flos": 23805808724160.0, + "grad_norm": 1.6291290789444606, + "learning_rate": 1.6725115284143132e-07, + "loss": 0.7549, + "num_input_tokens_seen": 314372900, + "step": 9678 + }, + { + "epoch": 0.8728863236686657, + "flos": 41434037941440.0, + "grad_norm": 4.131931838471375, + "learning_rate": 1.670173585129575e-07, + "loss": 0.7144, + "num_input_tokens_seen": 314399965, + "step": 9679 + }, + { + "epoch": 0.872976507192136, + "flos": 21730703075520.0, + "grad_norm": 1.956521399931026, + "learning_rate": 1.667837205846696e-07, + "loss": 0.8002, + "num_input_tokens_seen": 314430050, + "step": 9680 + }, + { + "epoch": 0.8730666907156063, + "flos": 20602646136480.0, + "grad_norm": 2.166740880820016, + "learning_rate": 1.6655023907650278e-07, + "loss": 0.7499, + "num_input_tokens_seen": 314458105, + "step": 9681 + }, + { + "epoch": 0.8731568742390765, + "flos": 24754307054400.0, + "grad_norm": 2.150862445400264, + "learning_rate": 1.6631691400837954e-07, + "loss": 0.7818, + "num_input_tokens_seen": 314485805, + "step": 9682 + }, + { + "epoch": 0.8732470577625467, + "flos": 19691174620800.0, + "grad_norm": 1.9404790187606684, + "learning_rate": 1.6608374540020752e-07, + "loss": 0.7883, + "num_input_tokens_seen": 314512390, + "step": 9683 + }, + { + "epoch": 0.8733372412860171, + "flos": 66561357970080.0, + "grad_norm": 0.692920986795941, + "learning_rate": 1.658507332718828e-07, + "loss": 0.5265, + "num_input_tokens_seen": 314602645, + "step": 9684 + }, + { + "epoch": 0.8734274248094873, + "flos": 29124486223200.0, + "grad_norm": 1.7779165832982007, + "learning_rate": 1.656178776432864e-07, + "loss": 0.676, + "num_input_tokens_seen": 314635135, + "step": 9685 + }, + { + "epoch": 0.8735176083329576, + "flos": 19983040275360.0, + "grad_norm": 2.638975361751303, + "learning_rate": 1.6538517853428814e-07, + "loss": 0.7178, + "num_input_tokens_seen": 314661085, + "step": 9686 + }, + { + "epoch": 0.8736077918564278, + "flos": 25196807510880.0, + "grad_norm": 1.526306658793657, + "learning_rate": 1.6515263596474194e-07, + "loss": 0.7233, + "num_input_tokens_seen": 314693515, + "step": 9687 + }, + { + "epoch": 0.8736979753798981, + "flos": 22168743160800.0, + "grad_norm": 1.9655506841684118, + "learning_rate": 1.6492024995449017e-07, + "loss": 0.7227, + "num_input_tokens_seen": 314719585, + "step": 9688 + }, + { + "epoch": 0.8737881589033684, + "flos": 21221136904800.0, + "grad_norm": 1.8963211885088962, + "learning_rate": 1.6468802052336116e-07, + "loss": 0.7081, + "num_input_tokens_seen": 314747855, + "step": 9689 + }, + { + "epoch": 0.8738783424268386, + "flos": 37725360686880.0, + "grad_norm": 2.2257578684207027, + "learning_rate": 1.6445594769116998e-07, + "loss": 0.7616, + "num_input_tokens_seen": 314780075, + "step": 9690 + }, + { + "epoch": 0.8739685259503088, + "flos": 24718952879040.0, + "grad_norm": 1.8541609160373391, + "learning_rate": 1.6422403147771836e-07, + "loss": 0.8089, + "num_input_tokens_seen": 314809765, + "step": 9691 + }, + { + "epoch": 0.8740587094737792, + "flos": 22752214281600.0, + "grad_norm": 2.1333768328563094, + "learning_rate": 1.6399227190279485e-07, + "loss": 0.7036, + "num_input_tokens_seen": 314838750, + "step": 9692 + }, + { + "epoch": 0.8741488929972494, + "flos": 27854046168960.0, + "grad_norm": 1.9508420495521992, + "learning_rate": 1.637606689861748e-07, + "loss": 0.6695, + "num_input_tokens_seen": 314872100, + "step": 9693 + }, + { + "epoch": 0.8742390765207196, + "flos": 21330284520960.0, + "grad_norm": 2.8658440792159214, + "learning_rate": 1.6352922274761883e-07, + "loss": 0.6991, + "num_input_tokens_seen": 314899205, + "step": 9694 + }, + { + "epoch": 0.87432926004419, + "flos": 34446174472800.0, + "grad_norm": 1.6841312975270606, + "learning_rate": 1.6329793320687602e-07, + "loss": 0.6073, + "num_input_tokens_seen": 314933380, + "step": 9695 + }, + { + "epoch": 0.8744194435676602, + "flos": 24496606142880.0, + "grad_norm": 1.8855377355100775, + "learning_rate": 1.630668003836808e-07, + "loss": 0.7758, + "num_input_tokens_seen": 314963765, + "step": 9696 + }, + { + "epoch": 0.8745096270911304, + "flos": 19942854031200.0, + "grad_norm": 1.9307075923938535, + "learning_rate": 1.62835824297755e-07, + "loss": 0.7472, + "num_input_tokens_seen": 314992850, + "step": 9697 + }, + { + "epoch": 0.8745998106146007, + "flos": 22059112337760.0, + "grad_norm": 1.6103138083525366, + "learning_rate": 1.626050049688066e-07, + "loss": 0.7487, + "num_input_tokens_seen": 315020905, + "step": 9698 + }, + { + "epoch": 0.874689994138071, + "flos": 20162821902720.0, + "grad_norm": 1.7502187183485083, + "learning_rate": 1.623743424165309e-07, + "loss": 0.7173, + "num_input_tokens_seen": 315050795, + "step": 9699 + }, + { + "epoch": 0.8747801776615413, + "flos": 16886051723040.0, + "grad_norm": 1.883244888231882, + "learning_rate": 1.6214383666060826e-07, + "loss": 0.688, + "num_input_tokens_seen": 315076095, + "step": 9700 + }, + { + "epoch": 0.8748703611850115, + "flos": 20092336570560.0, + "grad_norm": 1.7227558963526923, + "learning_rate": 1.619134877207078e-07, + "loss": 0.7602, + "num_input_tokens_seen": 315104700, + "step": 9701 + }, + { + "epoch": 0.8749605447084817, + "flos": 23661009574560.0, + "grad_norm": 1.7458925408616792, + "learning_rate": 1.616832956164831e-07, + "loss": 0.7394, + "num_input_tokens_seen": 315133510, + "step": 9702 + }, + { + "epoch": 0.8750507282319521, + "flos": 31494951483840.0, + "grad_norm": 1.7187579761666842, + "learning_rate": 1.6145326036757667e-07, + "loss": 0.7124, + "num_input_tokens_seen": 315166285, + "step": 9703 + }, + { + "epoch": 0.8751409117554223, + "flos": 25374470461920.0, + "grad_norm": 1.8641634278261472, + "learning_rate": 1.612233819936155e-07, + "loss": 0.7066, + "num_input_tokens_seen": 315195015, + "step": 9704 + }, + { + "epoch": 0.8752310952788925, + "flos": 32220025154880.0, + "grad_norm": 1.8530262501026087, + "learning_rate": 1.6099366051421414e-07, + "loss": 0.7332, + "num_input_tokens_seen": 315223405, + "step": 9705 + }, + { + "epoch": 0.8753212788023628, + "flos": 23877000281760.0, + "grad_norm": 1.839111182282327, + "learning_rate": 1.6076409594897378e-07, + "loss": 0.7997, + "num_input_tokens_seen": 315251770, + "step": 9706 + }, + { + "epoch": 0.8754114623258331, + "flos": 69743928510720.0, + "grad_norm": 0.6584454777926344, + "learning_rate": 1.605346883174823e-07, + "loss": 0.5269, + "num_input_tokens_seen": 315341515, + "step": 9707 + }, + { + "epoch": 0.8755016458493033, + "flos": 32005372559040.0, + "grad_norm": 2.0058007756395937, + "learning_rate": 1.6030543763931427e-07, + "loss": 0.7112, + "num_input_tokens_seen": 315370715, + "step": 9708 + }, + { + "epoch": 0.8755918293727736, + "flos": 14881617255360.0, + "grad_norm": 1.95856414273453, + "learning_rate": 1.600763439340298e-07, + "loss": 0.7607, + "num_input_tokens_seen": 315394900, + "step": 9709 + }, + { + "epoch": 0.8756820128962438, + "flos": 27809027856000.0, + "grad_norm": 1.904602902593607, + "learning_rate": 1.5984740722117707e-07, + "loss": 0.7949, + "num_input_tokens_seen": 315424785, + "step": 9710 + }, + { + "epoch": 0.8757721964197142, + "flos": 24608318472480.0, + "grad_norm": 1.4311984889993212, + "learning_rate": 1.5961862752028998e-07, + "loss": 0.7541, + "num_input_tokens_seen": 315454655, + "step": 9711 + }, + { + "epoch": 0.8758623799431844, + "flos": 24499839912000.0, + "grad_norm": 2.426235678550436, + "learning_rate": 1.5939000485088937e-07, + "loss": 0.6482, + "num_input_tokens_seen": 315486400, + "step": 9712 + }, + { + "epoch": 0.8759525634666546, + "flos": 21359468516160.0, + "grad_norm": 1.8497676600368393, + "learning_rate": 1.5916153923248254e-07, + "loss": 0.8343, + "num_input_tokens_seen": 315513530, + "step": 9713 + }, + { + "epoch": 0.8760427469901249, + "flos": 21147975349920.0, + "grad_norm": 1.7059750289967748, + "learning_rate": 1.5893323068456342e-07, + "loss": 0.787, + "num_input_tokens_seen": 315541105, + "step": 9714 + }, + { + "epoch": 0.8761329305135952, + "flos": 25302163811520.0, + "grad_norm": 1.600462223878735, + "learning_rate": 1.5870507922661248e-07, + "loss": 0.7866, + "num_input_tokens_seen": 315569385, + "step": 9715 + }, + { + "epoch": 0.8762231140370654, + "flos": 69382692616800.0, + "grad_norm": 0.6188690119256984, + "learning_rate": 1.5847708487809763e-07, + "loss": 0.5659, + "num_input_tokens_seen": 315667190, + "step": 9716 + }, + { + "epoch": 0.8763132975605357, + "flos": 22459828250400.0, + "grad_norm": 2.230520912535851, + "learning_rate": 1.5824924765847113e-07, + "loss": 0.8163, + "num_input_tokens_seen": 315695850, + "step": 9717 + }, + { + "epoch": 0.8764034810840059, + "flos": 23331150691680.0, + "grad_norm": 1.530463191709215, + "learning_rate": 1.5802156758717478e-07, + "loss": 0.7173, + "num_input_tokens_seen": 315725555, + "step": 9718 + }, + { + "epoch": 0.8764936646074762, + "flos": 23079508451040.0, + "grad_norm": 1.7453889725886764, + "learning_rate": 1.5779404468363433e-07, + "loss": 0.7398, + "num_input_tokens_seen": 315753755, + "step": 9719 + }, + { + "epoch": 0.8765838481309465, + "flos": 23079768639360.0, + "grad_norm": 2.0151748746046922, + "learning_rate": 1.5756667896726405e-07, + "loss": 0.7973, + "num_input_tokens_seen": 315782445, + "step": 9720 + }, + { + "epoch": 0.8766740316544167, + "flos": 28903589107680.0, + "grad_norm": 2.3770074892997313, + "learning_rate": 1.5733947045746377e-07, + "loss": 0.7591, + "num_input_tokens_seen": 315810070, + "step": 9721 + }, + { + "epoch": 0.876764215177887, + "flos": 25702916893920.0, + "grad_norm": 2.0169399023828696, + "learning_rate": 1.5711241917362018e-07, + "loss": 0.8087, + "num_input_tokens_seen": 315840330, + "step": 9722 + }, + { + "epoch": 0.8768543987013573, + "flos": 21768213096960.0, + "grad_norm": 2.3811047046618903, + "learning_rate": 1.5688552513510688e-07, + "loss": 0.6802, + "num_input_tokens_seen": 315867970, + "step": 9723 + }, + { + "epoch": 0.8769445822248275, + "flos": 26502192873120.0, + "grad_norm": 2.4691601183217924, + "learning_rate": 1.5665878836128266e-07, + "loss": 0.7657, + "num_input_tokens_seen": 315896180, + "step": 9724 + }, + { + "epoch": 0.8770347657482978, + "flos": 34117690871040.0, + "grad_norm": 12.045232084389113, + "learning_rate": 1.5643220887149554e-07, + "loss": 0.7898, + "num_input_tokens_seen": 315929955, + "step": 9725 + }, + { + "epoch": 0.8771249492717681, + "flos": 69536747036640.0, + "grad_norm": 0.7003248406156578, + "learning_rate": 1.562057866850772e-07, + "loss": 0.5881, + "num_input_tokens_seen": 316015265, + "step": 9726 + }, + { + "epoch": 0.8772151327952383, + "flos": 38927396915520.0, + "grad_norm": 2.742981450706898, + "learning_rate": 1.5597952182134777e-07, + "loss": 0.6163, + "num_input_tokens_seen": 316044405, + "step": 9727 + }, + { + "epoch": 0.8773053163187086, + "flos": 25007621934240.0, + "grad_norm": 1.6238262632578184, + "learning_rate": 1.557534142996133e-07, + "loss": 0.7955, + "num_input_tokens_seen": 316074455, + "step": 9728 + }, + { + "epoch": 0.8773954998421788, + "flos": 23042741824800.0, + "grad_norm": 2.3502764559521667, + "learning_rate": 1.5552746413916662e-07, + "loss": 0.7743, + "num_input_tokens_seen": 316102825, + "step": 9729 + }, + { + "epoch": 0.8774856833656491, + "flos": 23115791870400.0, + "grad_norm": 1.9319708284462174, + "learning_rate": 1.5530167135928697e-07, + "loss": 0.7336, + "num_input_tokens_seen": 316126930, + "step": 9730 + }, + { + "epoch": 0.8775758668891194, + "flos": 27159829132320.0, + "grad_norm": 2.500164026601477, + "learning_rate": 1.5507603597924068e-07, + "loss": 0.6803, + "num_input_tokens_seen": 316156770, + "step": 9731 + }, + { + "epoch": 0.8776660504125896, + "flos": 25484212794240.0, + "grad_norm": 1.8756277305128084, + "learning_rate": 1.548505580182793e-07, + "loss": 0.6999, + "num_input_tokens_seen": 316185460, + "step": 9732 + }, + { + "epoch": 0.8777562339360598, + "flos": 23262226489440.0, + "grad_norm": 1.7802462510352188, + "learning_rate": 1.5462523749564271e-07, + "loss": 0.7384, + "num_input_tokens_seen": 316214940, + "step": 9733 + }, + { + "epoch": 0.8778464174595302, + "flos": 24714752696160.0, + "grad_norm": 2.300620832825493, + "learning_rate": 1.5440007443055602e-07, + "loss": 0.7389, + "num_input_tokens_seen": 316243395, + "step": 9734 + }, + { + "epoch": 0.8779366009830004, + "flos": 26172408329760.0, + "grad_norm": 2.056407484151729, + "learning_rate": 1.541750688422314e-07, + "loss": 0.6739, + "num_input_tokens_seen": 316272415, + "step": 9735 + }, + { + "epoch": 0.8780267845064706, + "flos": 24534450692160.0, + "grad_norm": 2.308230195570231, + "learning_rate": 1.5395022074986797e-07, + "loss": 0.7412, + "num_input_tokens_seen": 316301155, + "step": 9736 + }, + { + "epoch": 0.8781169680299409, + "flos": 24828174834720.0, + "grad_norm": 1.6654505425390316, + "learning_rate": 1.5372553017265033e-07, + "loss": 0.6845, + "num_input_tokens_seen": 316332475, + "step": 9737 + }, + { + "epoch": 0.8782071515534112, + "flos": 21940932469920.0, + "grad_norm": 3.1310491379446015, + "learning_rate": 1.5350099712975116e-07, + "loss": 0.7973, + "num_input_tokens_seen": 316354610, + "step": 9738 + }, + { + "epoch": 0.8782973350768815, + "flos": 19363248565440.0, + "grad_norm": 1.601597334320106, + "learning_rate": 1.5327662164032785e-07, + "loss": 0.7313, + "num_input_tokens_seen": 316382395, + "step": 9739 + }, + { + "epoch": 0.8783875186003517, + "flos": 27961335297120.0, + "grad_norm": 1.9206660740560253, + "learning_rate": 1.5305240372352656e-07, + "loss": 0.5957, + "num_input_tokens_seen": 316413445, + "step": 9740 + }, + { + "epoch": 0.8784777021238219, + "flos": 23917186525920.0, + "grad_norm": 1.7783577175071488, + "learning_rate": 1.5282834339847738e-07, + "loss": 0.7512, + "num_input_tokens_seen": 316442375, + "step": 9741 + }, + { + "epoch": 0.8785678856472923, + "flos": 31713915771840.0, + "grad_norm": 1.7564988038112999, + "learning_rate": 1.526044406842999e-07, + "loss": 0.7201, + "num_input_tokens_seen": 316474675, + "step": 9742 + }, + { + "epoch": 0.8786580691707625, + "flos": 28727375777280.0, + "grad_norm": 1.9700996979789644, + "learning_rate": 1.523806956000977e-07, + "loss": 0.6683, + "num_input_tokens_seen": 316506585, + "step": 9743 + }, + { + "epoch": 0.8787482526942327, + "flos": 29675651088960.0, + "grad_norm": 1.5327723973245742, + "learning_rate": 1.5215710816496197e-07, + "loss": 0.7918, + "num_input_tokens_seen": 316538140, + "step": 9744 + }, + { + "epoch": 0.8788384362177031, + "flos": 29418842251680.0, + "grad_norm": 1.9784548299172329, + "learning_rate": 1.5193367839797077e-07, + "loss": 0.687, + "num_input_tokens_seen": 316566965, + "step": 9745 + }, + { + "epoch": 0.8789286197411733, + "flos": 20565396303360.0, + "grad_norm": 1.7282490497786809, + "learning_rate": 1.5171040631818842e-07, + "loss": 0.7397, + "num_input_tokens_seen": 316594780, + "step": 9746 + }, + { + "epoch": 0.8790188032646435, + "flos": 36267630713760.0, + "grad_norm": 1.8898979239171987, + "learning_rate": 1.5148729194466547e-07, + "loss": 0.6743, + "num_input_tokens_seen": 316625265, + "step": 9747 + }, + { + "epoch": 0.8791089867881138, + "flos": 19727012003040.0, + "grad_norm": 1.7568430622119637, + "learning_rate": 1.5126433529643956e-07, + "loss": 0.8104, + "num_input_tokens_seen": 316652325, + "step": 9748 + }, + { + "epoch": 0.8791991703115841, + "flos": 25812250358880.0, + "grad_norm": 1.747041244648835, + "learning_rate": 1.5104153639253436e-07, + "loss": 0.7407, + "num_input_tokens_seen": 316681220, + "step": 9749 + }, + { + "epoch": 0.8792893538350544, + "flos": 57799329662400.0, + "grad_norm": 0.6126381727965053, + "learning_rate": 1.5081889525196002e-07, + "loss": 0.6291, + "num_input_tokens_seen": 316772465, + "step": 9750 + }, + { + "epoch": 0.8793795373585246, + "flos": 21912863567520.0, + "grad_norm": 2.1031361187078397, + "learning_rate": 1.5059641189371398e-07, + "loss": 0.799, + "num_input_tokens_seen": 316800460, + "step": 9751 + }, + { + "epoch": 0.8794697208819948, + "flos": 28617075898560.0, + "grad_norm": 1.7431248764634388, + "learning_rate": 1.503740863367795e-07, + "loss": 0.756, + "num_input_tokens_seen": 316830805, + "step": 9752 + }, + { + "epoch": 0.8795599044054652, + "flos": 20455876989600.0, + "grad_norm": 2.2495543780379608, + "learning_rate": 1.50151918600127e-07, + "loss": 0.7692, + "num_input_tokens_seen": 316856765, + "step": 9753 + }, + { + "epoch": 0.8796500879289354, + "flos": 28395621236640.0, + "grad_norm": 3.9326993954034846, + "learning_rate": 1.4992990870271217e-07, + "loss": 0.7967, + "num_input_tokens_seen": 316886240, + "step": 9754 + }, + { + "epoch": 0.8797402714524056, + "flos": 18816098033760.0, + "grad_norm": 2.236960136027686, + "learning_rate": 1.497080566634794e-07, + "loss": 0.7448, + "num_input_tokens_seen": 316911935, + "step": 9755 + }, + { + "epoch": 0.8798304549758759, + "flos": 20420374135200.0, + "grad_norm": 2.063625144393299, + "learning_rate": 1.4948636250135693e-07, + "loss": 0.7502, + "num_input_tokens_seen": 316938410, + "step": 9756 + }, + { + "epoch": 0.8799206384993462, + "flos": 19618124575200.0, + "grad_norm": 2.2368752955404267, + "learning_rate": 1.4926482623526249e-07, + "loss": 0.8023, + "num_input_tokens_seen": 316966190, + "step": 9757 + }, + { + "epoch": 0.8800108220228164, + "flos": 22350420445920.0, + "grad_norm": 2.170185996667529, + "learning_rate": 1.4904344788409694e-07, + "loss": 0.7431, + "num_input_tokens_seen": 316991845, + "step": 9758 + }, + { + "epoch": 0.8801010055462867, + "flos": 20238473831520.0, + "grad_norm": 2.440802562081622, + "learning_rate": 1.4882222746675143e-07, + "loss": 0.7032, + "num_input_tokens_seen": 317019760, + "step": 9759 + }, + { + "epoch": 0.8801911890697569, + "flos": 22678681029120.0, + "grad_norm": 2.4067924185399043, + "learning_rate": 1.4860116500210018e-07, + "loss": 0.7852, + "num_input_tokens_seen": 317046295, + "step": 9760 + }, + { + "epoch": 0.8802813725932273, + "flos": 18926323572960.0, + "grad_norm": 2.4064724311573085, + "learning_rate": 1.4838026050900632e-07, + "loss": 0.8135, + "num_input_tokens_seen": 317072405, + "step": 9761 + }, + { + "epoch": 0.8803715561166975, + "flos": 22533101314560.0, + "grad_norm": 1.61125621288395, + "learning_rate": 1.481595140063181e-07, + "loss": 0.79, + "num_input_tokens_seen": 317101165, + "step": 9762 + }, + { + "epoch": 0.8804617396401677, + "flos": 23953544284800.0, + "grad_norm": 2.2787561240979723, + "learning_rate": 1.4793892551287136e-07, + "loss": 0.8318, + "num_input_tokens_seen": 317128395, + "step": 9763 + }, + { + "epoch": 0.880551923163638, + "flos": 26573421600480.0, + "grad_norm": 2.334996981404414, + "learning_rate": 1.4771849504748768e-07, + "loss": 0.6754, + "num_input_tokens_seen": 317157785, + "step": 9764 + }, + { + "epoch": 0.8806421066871083, + "flos": 24355152271680.0, + "grad_norm": 1.618873123140759, + "learning_rate": 1.4749822262897517e-07, + "loss": 0.8198, + "num_input_tokens_seen": 317186735, + "step": 9765 + }, + { + "epoch": 0.8807322902105785, + "flos": 53499976087200.0, + "grad_norm": 1.9839362913562524, + "learning_rate": 1.4727810827612895e-07, + "loss": 0.6839, + "num_input_tokens_seen": 317219445, + "step": 9766 + }, + { + "epoch": 0.8808224737340488, + "flos": 25738456918080.0, + "grad_norm": 2.1480149841993588, + "learning_rate": 1.470581520077303e-07, + "loss": 0.7517, + "num_input_tokens_seen": 317247485, + "step": 9767 + }, + { + "epoch": 0.8809126572575191, + "flos": 13206038087040.0, + "grad_norm": 5.354623361943079, + "learning_rate": 1.4683835384254705e-07, + "loss": 0.7373, + "num_input_tokens_seen": 317272390, + "step": 9768 + }, + { + "epoch": 0.8810028407809893, + "flos": 21653043979680.0, + "grad_norm": 1.9070001595731607, + "learning_rate": 1.4661871379933376e-07, + "loss": 0.8229, + "num_input_tokens_seen": 317299625, + "step": 9769 + }, + { + "epoch": 0.8810930243044596, + "flos": 17030590684320.0, + "grad_norm": 2.034326541634068, + "learning_rate": 1.4639923189683169e-07, + "loss": 0.7922, + "num_input_tokens_seen": 317326270, + "step": 9770 + }, + { + "epoch": 0.8811832078279298, + "flos": 21221248414080.0, + "grad_norm": 1.642808460319883, + "learning_rate": 1.461799081537669e-07, + "loss": 0.7928, + "num_input_tokens_seen": 317357795, + "step": 9771 + }, + { + "epoch": 0.8812733913514001, + "flos": 21512853880320.0, + "grad_norm": 2.4158861500270614, + "learning_rate": 1.4596074258885514e-07, + "loss": 0.8086, + "num_input_tokens_seen": 317386680, + "step": 9772 + }, + { + "epoch": 0.8813635748748704, + "flos": 25116472192320.0, + "grad_norm": 1.6888653212221894, + "learning_rate": 1.4574173522079502e-07, + "loss": 0.7159, + "num_input_tokens_seen": 317415130, + "step": 9773 + }, + { + "epoch": 0.8814537583983406, + "flos": 23438811517440.0, + "grad_norm": 4.625234030911111, + "learning_rate": 1.4552288606827513e-07, + "loss": 0.7193, + "num_input_tokens_seen": 317444395, + "step": 9774 + }, + { + "epoch": 0.8815439419218108, + "flos": 20201372677440.0, + "grad_norm": 4.340061735074492, + "learning_rate": 1.4530419514996761e-07, + "loss": 0.8202, + "num_input_tokens_seen": 317472345, + "step": 9775 + }, + { + "epoch": 0.8816341254452812, + "flos": 71353296869280.0, + "grad_norm": 0.6964347868771269, + "learning_rate": 1.4508566248453291e-07, + "loss": 0.5878, + "num_input_tokens_seen": 317559230, + "step": 9776 + }, + { + "epoch": 0.8817243089687514, + "flos": 34045570069440.0, + "grad_norm": 3.2634775502254025, + "learning_rate": 1.448672880906172e-07, + "loss": 0.7382, + "num_input_tokens_seen": 317586960, + "step": 9777 + }, + { + "epoch": 0.8818144924922217, + "flos": 41914234268160.0, + "grad_norm": 2.5459978898512374, + "learning_rate": 1.4464907198685382e-07, + "loss": 0.6982, + "num_input_tokens_seen": 317619555, + "step": 9778 + }, + { + "epoch": 0.8819046760156919, + "flos": 35536721390400.0, + "grad_norm": 2.0767840885052435, + "learning_rate": 1.444310141918621e-07, + "loss": 0.7467, + "num_input_tokens_seen": 317650030, + "step": 9779 + }, + { + "epoch": 0.8819948595391622, + "flos": 21505048230720.0, + "grad_norm": 3.268689191847085, + "learning_rate": 1.4421311472424735e-07, + "loss": 0.7728, + "num_input_tokens_seen": 317677370, + "step": 9780 + }, + { + "epoch": 0.8820850430626325, + "flos": 18707545133760.0, + "grad_norm": 1.857229004161392, + "learning_rate": 1.4399537360260273e-07, + "loss": 0.7798, + "num_input_tokens_seen": 317702835, + "step": 9781 + }, + { + "epoch": 0.8821752265861027, + "flos": 24244592204640.0, + "grad_norm": 1.9797364335007757, + "learning_rate": 1.4377779084550645e-07, + "loss": 0.7702, + "num_input_tokens_seen": 317728320, + "step": 9782 + }, + { + "epoch": 0.8822654101095729, + "flos": 23188916255520.0, + "grad_norm": 2.673441199205917, + "learning_rate": 1.4356036647152413e-07, + "loss": 0.7364, + "num_input_tokens_seen": 317757870, + "step": 9783 + }, + { + "epoch": 0.8823555936330433, + "flos": 34699749540960.0, + "grad_norm": 1.5969604618621898, + "learning_rate": 1.4334310049920785e-07, + "loss": 0.7204, + "num_input_tokens_seen": 317787470, + "step": 9784 + }, + { + "epoch": 0.8824457771565135, + "flos": 70088952655680.0, + "grad_norm": 0.607752571976122, + "learning_rate": 1.431259929470956e-07, + "loss": 0.5284, + "num_input_tokens_seen": 317878585, + "step": 9785 + }, + { + "epoch": 0.8825359606799837, + "flos": 32073776384640.0, + "grad_norm": 2.2111668552015127, + "learning_rate": 1.4290904383371237e-07, + "loss": 0.8192, + "num_input_tokens_seen": 317909840, + "step": 9786 + }, + { + "epoch": 0.882626144203454, + "flos": 31458779573760.0, + "grad_norm": 2.0770733689273535, + "learning_rate": 1.4269225317756961e-07, + "loss": 0.7272, + "num_input_tokens_seen": 317940370, + "step": 9787 + }, + { + "epoch": 0.8827163277269243, + "flos": 18634792446240.0, + "grad_norm": 2.3005113945363544, + "learning_rate": 1.424756209971645e-07, + "loss": 0.7065, + "num_input_tokens_seen": 317967000, + "step": 9788 + }, + { + "epoch": 0.8828065112503946, + "flos": 17322642187680.0, + "grad_norm": 2.2695265226593944, + "learning_rate": 1.4225914731098199e-07, + "loss": 0.7413, + "num_input_tokens_seen": 317993520, + "step": 9789 + }, + { + "epoch": 0.8828966947738648, + "flos": 23516544952800.0, + "grad_norm": 1.997439056154289, + "learning_rate": 1.4204283213749248e-07, + "loss": 0.6742, + "num_input_tokens_seen": 318023975, + "step": 9790 + }, + { + "epoch": 0.8829868782973351, + "flos": 21913495453440.0, + "grad_norm": 1.8087896978296651, + "learning_rate": 1.4182667549515315e-07, + "loss": 0.8025, + "num_input_tokens_seen": 318052145, + "step": 9791 + }, + { + "epoch": 0.8830770618208054, + "flos": 25811767152000.0, + "grad_norm": 2.1453743477704617, + "learning_rate": 1.4161067740240752e-07, + "loss": 0.7263, + "num_input_tokens_seen": 318080610, + "step": 9792 + }, + { + "epoch": 0.8831672453442756, + "flos": 28398260289600.0, + "grad_norm": 1.7126118031781195, + "learning_rate": 1.4139483787768614e-07, + "loss": 0.736, + "num_input_tokens_seen": 318110210, + "step": 9793 + }, + { + "epoch": 0.8832574288677458, + "flos": 29417281121760.0, + "grad_norm": 2.4809785093816656, + "learning_rate": 1.4117915693940584e-07, + "loss": 0.6531, + "num_input_tokens_seen": 318141680, + "step": 9794 + }, + { + "epoch": 0.8833476123912162, + "flos": 36377781913440.0, + "grad_norm": 1.735080815942287, + "learning_rate": 1.409636346059684e-07, + "loss": 0.7399, + "num_input_tokens_seen": 318174045, + "step": 9795 + }, + { + "epoch": 0.8834377959146864, + "flos": 21723640821120.0, + "grad_norm": 2.745952575440775, + "learning_rate": 1.4074827089576501e-07, + "loss": 0.7753, + "num_input_tokens_seen": 318200565, + "step": 9796 + }, + { + "epoch": 0.8835279794381566, + "flos": 19873000584960.0, + "grad_norm": 7.336636088525789, + "learning_rate": 1.4053306582717085e-07, + "loss": 0.7588, + "num_input_tokens_seen": 318228380, + "step": 9797 + }, + { + "epoch": 0.8836181629616269, + "flos": 23042853334080.0, + "grad_norm": 1.9882188948350976, + "learning_rate": 1.4031801941854827e-07, + "loss": 0.7701, + "num_input_tokens_seen": 318258130, + "step": 9798 + }, + { + "epoch": 0.8837083464850972, + "flos": 26505017774880.0, + "grad_norm": 1.91077754676105, + "learning_rate": 1.401031316882466e-07, + "loss": 0.7109, + "num_input_tokens_seen": 318290685, + "step": 9799 + }, + { + "epoch": 0.8837985300085675, + "flos": 22308450053280.0, + "grad_norm": 2.6276831764664874, + "learning_rate": 1.39888402654601e-07, + "loss": 0.7582, + "num_input_tokens_seen": 318317585, + "step": 9800 + }, + { + "epoch": 0.8838887135320377, + "flos": 60906316880160.0, + "grad_norm": 0.6078664740988802, + "learning_rate": 1.3967383233593344e-07, + "loss": 0.5866, + "num_input_tokens_seen": 318415355, + "step": 9801 + }, + { + "epoch": 0.8839788970555079, + "flos": 22934337603840.0, + "grad_norm": 2.0218682408404094, + "learning_rate": 1.3945942075055218e-07, + "loss": 0.704, + "num_input_tokens_seen": 318443110, + "step": 9802 + }, + { + "epoch": 0.8840690805789783, + "flos": 18958778507040.0, + "grad_norm": 2.4933733854119824, + "learning_rate": 1.3924516791675212e-07, + "loss": 0.769, + "num_input_tokens_seen": 318470955, + "step": 9803 + }, + { + "epoch": 0.8841592641024485, + "flos": 22933036662240.0, + "grad_norm": 1.6759498351335083, + "learning_rate": 1.3903107385281487e-07, + "loss": 0.654, + "num_input_tokens_seen": 318503655, + "step": 9804 + }, + { + "epoch": 0.8842494476259187, + "flos": 19724633138400.0, + "grad_norm": 1.9679053780595879, + "learning_rate": 1.3881713857700717e-07, + "loss": 0.7533, + "num_input_tokens_seen": 318533115, + "step": 9805 + }, + { + "epoch": 0.884339631149389, + "flos": 28726558042560.0, + "grad_norm": 2.4144680830599086, + "learning_rate": 1.3860336210758372e-07, + "loss": 0.7223, + "num_input_tokens_seen": 318560690, + "step": 9806 + }, + { + "epoch": 0.8844298146728593, + "flos": 24240280512480.0, + "grad_norm": 1.6639641387954078, + "learning_rate": 1.3838974446278506e-07, + "loss": 0.7024, + "num_input_tokens_seen": 318588935, + "step": 9807 + }, + { + "epoch": 0.8845199981963295, + "flos": 46724720877600.0, + "grad_norm": 1.821521406719652, + "learning_rate": 1.3817628566083817e-07, + "loss": 0.7203, + "num_input_tokens_seen": 318621115, + "step": 9808 + }, + { + "epoch": 0.8846101817197998, + "flos": 21513076898880.0, + "grad_norm": 1.6664838770130241, + "learning_rate": 1.3796298571995712e-07, + "loss": 0.754, + "num_input_tokens_seen": 318649350, + "step": 9809 + }, + { + "epoch": 0.88470036524327, + "flos": 19545929434080.0, + "grad_norm": 2.315624909394503, + "learning_rate": 1.377498446583405e-07, + "loss": 0.642, + "num_input_tokens_seen": 318677290, + "step": 9810 + }, + { + "epoch": 0.8847905487667403, + "flos": 25337778175200.0, + "grad_norm": 1.910709729972267, + "learning_rate": 1.3753686249417596e-07, + "loss": 0.7168, + "num_input_tokens_seen": 318703815, + "step": 9811 + }, + { + "epoch": 0.8848807322902106, + "flos": 19068149141760.0, + "grad_norm": 2.415381901242807, + "learning_rate": 1.373240392456354e-07, + "loss": 0.7011, + "num_input_tokens_seen": 318731240, + "step": 9812 + }, + { + "epoch": 0.8849709158136808, + "flos": 23225013826080.0, + "grad_norm": 1.8004621442286124, + "learning_rate": 1.37111374930879e-07, + "loss": 0.6786, + "num_input_tokens_seen": 318758670, + "step": 9813 + }, + { + "epoch": 0.885061099337151, + "flos": 27415522876800.0, + "grad_norm": 1.6587626992424418, + "learning_rate": 1.3689886956805176e-07, + "loss": 0.6734, + "num_input_tokens_seen": 318787900, + "step": 9814 + }, + { + "epoch": 0.8851512828606214, + "flos": 22310828917920.0, + "grad_norm": 1.7064915552430504, + "learning_rate": 1.3668652317528585e-07, + "loss": 0.7343, + "num_input_tokens_seen": 318815990, + "step": 9815 + }, + { + "epoch": 0.8852414663840916, + "flos": 21549285978720.0, + "grad_norm": 1.7637302731195492, + "learning_rate": 1.3647433577070012e-07, + "loss": 0.7587, + "num_input_tokens_seen": 318846135, + "step": 9816 + }, + { + "epoch": 0.8853316499075619, + "flos": 67963476248640.0, + "grad_norm": 0.6508471287312576, + "learning_rate": 1.3626230737239942e-07, + "loss": 0.5486, + "num_input_tokens_seen": 318930970, + "step": 9817 + }, + { + "epoch": 0.8854218334310322, + "flos": 69961034310720.0, + "grad_norm": 0.7043890928129621, + "learning_rate": 1.3605043799847527e-07, + "loss": 0.5553, + "num_input_tokens_seen": 319026065, + "step": 9818 + }, + { + "epoch": 0.8855120169545024, + "flos": 46979634057120.0, + "grad_norm": 1.54486014415229, + "learning_rate": 1.3583872766700567e-07, + "loss": 0.6798, + "num_input_tokens_seen": 319059655, + "step": 9819 + }, + { + "epoch": 0.8856022004779727, + "flos": 41912264270880.0, + "grad_norm": 3.267325656620584, + "learning_rate": 1.3562717639605437e-07, + "loss": 0.7227, + "num_input_tokens_seen": 319092065, + "step": 9820 + }, + { + "epoch": 0.8856923840014429, + "flos": 21435863840160.0, + "grad_norm": 2.6490497612835338, + "learning_rate": 1.3541578420367229e-07, + "loss": 0.7127, + "num_input_tokens_seen": 319120930, + "step": 9821 + }, + { + "epoch": 0.8857825675249132, + "flos": 20419482060960.0, + "grad_norm": 3.4122555487972837, + "learning_rate": 1.3520455110789697e-07, + "loss": 0.7578, + "num_input_tokens_seen": 319149110, + "step": 9822 + }, + { + "epoch": 0.8858727510483835, + "flos": 18233333138400.0, + "grad_norm": 2.1185597832288674, + "learning_rate": 1.3499347712675158e-07, + "loss": 0.7203, + "num_input_tokens_seen": 319176860, + "step": 9823 + }, + { + "epoch": 0.8859629345718537, + "flos": 31636219506240.0, + "grad_norm": 3.9161047566228038, + "learning_rate": 1.3478256227824635e-07, + "loss": 0.7039, + "num_input_tokens_seen": 319206315, + "step": 9824 + }, + { + "epoch": 0.8860531180953239, + "flos": 30183581790240.0, + "grad_norm": 1.7614335804137573, + "learning_rate": 1.3457180658037759e-07, + "loss": 0.6521, + "num_input_tokens_seen": 319237595, + "step": 9825 + }, + { + "epoch": 0.8861433016187943, + "flos": 20203528523520.0, + "grad_norm": 1.7721916225320151, + "learning_rate": 1.3436121005112843e-07, + "loss": 0.7745, + "num_input_tokens_seen": 319267520, + "step": 9826 + }, + { + "epoch": 0.8862334851422645, + "flos": 26540780817600.0, + "grad_norm": 2.9895387085741967, + "learning_rate": 1.3415077270846719e-07, + "loss": 0.7133, + "num_input_tokens_seen": 319296875, + "step": 9827 + }, + { + "epoch": 0.8863236686657348, + "flos": 15173371400640.0, + "grad_norm": 3.4543914543851852, + "learning_rate": 1.3394049457035105e-07, + "loss": 0.8519, + "num_input_tokens_seen": 319321705, + "step": 9828 + }, + { + "epoch": 0.886413852189205, + "flos": 20237693266560.0, + "grad_norm": 1.8291070370813693, + "learning_rate": 1.3373037565472034e-07, + "loss": 0.7507, + "num_input_tokens_seen": 319349200, + "step": 9829 + }, + { + "epoch": 0.8865040357126753, + "flos": 17104421294880.0, + "grad_norm": 2.739484536211038, + "learning_rate": 1.3352041597950537e-07, + "loss": 0.7189, + "num_input_tokens_seen": 319376155, + "step": 9830 + }, + { + "epoch": 0.8865942192361456, + "flos": 21214111820160.0, + "grad_norm": 2.011236218108285, + "learning_rate": 1.333106155626196e-07, + "loss": 0.7321, + "num_input_tokens_seen": 319403310, + "step": 9831 + }, + { + "epoch": 0.8866844027596158, + "flos": 23840902711200.0, + "grad_norm": 1.6729089979832876, + "learning_rate": 1.331009744219651e-07, + "loss": 0.6846, + "num_input_tokens_seen": 319432305, + "step": 9832 + }, + { + "epoch": 0.886774586283086, + "flos": 69530130819360.0, + "grad_norm": 0.6682199819939071, + "learning_rate": 1.3289149257542943e-07, + "loss": 0.6104, + "num_input_tokens_seen": 319519790, + "step": 9833 + }, + { + "epoch": 0.8868647698065564, + "flos": 31969089139680.0, + "grad_norm": 3.0766207509800956, + "learning_rate": 1.3268217004088666e-07, + "loss": 0.6473, + "num_input_tokens_seen": 319550980, + "step": 9834 + }, + { + "epoch": 0.8869549533300266, + "flos": 29309880484320.0, + "grad_norm": 2.037619127406794, + "learning_rate": 1.3247300683619788e-07, + "loss": 0.6307, + "num_input_tokens_seen": 319581370, + "step": 9835 + }, + { + "epoch": 0.8870451368534968, + "flos": 24062952089280.0, + "grad_norm": 3.202841546147196, + "learning_rate": 1.3226400297920903e-07, + "loss": 0.7187, + "num_input_tokens_seen": 319610245, + "step": 9836 + }, + { + "epoch": 0.8871353203769671, + "flos": 26504014191360.0, + "grad_norm": 1.866094357029614, + "learning_rate": 1.3205515848775428e-07, + "loss": 0.7225, + "num_input_tokens_seen": 319637355, + "step": 9837 + }, + { + "epoch": 0.8872255039004374, + "flos": 16631212883040.0, + "grad_norm": 2.0863177349415385, + "learning_rate": 1.3184647337965316e-07, + "loss": 0.683, + "num_input_tokens_seen": 319664925, + "step": 9838 + }, + { + "epoch": 0.8873156874239077, + "flos": 36887013556320.0, + "grad_norm": 1.5888736398820276, + "learning_rate": 1.3163794767271163e-07, + "loss": 0.6899, + "num_input_tokens_seen": 319698770, + "step": 9839 + }, + { + "epoch": 0.8874058709473779, + "flos": 42747823669440.0, + "grad_norm": 1.7212171706848263, + "learning_rate": 1.314295813847226e-07, + "loss": 0.7131, + "num_input_tokens_seen": 319731450, + "step": 9840 + }, + { + "epoch": 0.8874960544708482, + "flos": 24062952089280.0, + "grad_norm": 2.0842096265991676, + "learning_rate": 1.3122137453346515e-07, + "loss": 0.8443, + "num_input_tokens_seen": 319760915, + "step": 9841 + }, + { + "epoch": 0.8875862379943185, + "flos": 20125757918400.0, + "grad_norm": 1.9396982335831836, + "learning_rate": 1.3101332713670376e-07, + "loss": 0.798, + "num_input_tokens_seen": 319790160, + "step": 9842 + }, + { + "epoch": 0.8876764215177887, + "flos": 62957373694080.0, + "grad_norm": 0.7024855567085415, + "learning_rate": 1.3080543921219133e-07, + "loss": 0.6061, + "num_input_tokens_seen": 319873960, + "step": 9843 + }, + { + "epoch": 0.8877666050412589, + "flos": 18780260651520.0, + "grad_norm": 7.221661661678821, + "learning_rate": 1.3059771077766478e-07, + "loss": 0.7344, + "num_input_tokens_seen": 319899795, + "step": 9844 + }, + { + "epoch": 0.8878567885647293, + "flos": 20381600341920.0, + "grad_norm": 2.1757922078101535, + "learning_rate": 1.3039014185085018e-07, + "loss": 0.7822, + "num_input_tokens_seen": 319922625, + "step": 9845 + }, + { + "epoch": 0.8879469720881995, + "flos": 26832423453600.0, + "grad_norm": 2.0000972724147927, + "learning_rate": 1.301827324494571e-07, + "loss": 0.7492, + "num_input_tokens_seen": 319949780, + "step": 9846 + }, + { + "epoch": 0.8880371556116697, + "flos": 35932493724960.0, + "grad_norm": 2.556403187682961, + "learning_rate": 1.2997548259118342e-07, + "loss": 0.7804, + "num_input_tokens_seen": 319975855, + "step": 9847 + }, + { + "epoch": 0.88812733913514, + "flos": 24754604412480.0, + "grad_norm": 2.257545006557903, + "learning_rate": 1.2976839229371272e-07, + "loss": 0.7692, + "num_input_tokens_seen": 320005090, + "step": 9848 + }, + { + "epoch": 0.8882175226586103, + "flos": 60760960184160.0, + "grad_norm": 0.6200972906810456, + "learning_rate": 1.2956146157471515e-07, + "loss": 0.5947, + "num_input_tokens_seen": 320100520, + "step": 9849 + }, + { + "epoch": 0.8883077061820805, + "flos": 28143495789120.0, + "grad_norm": 1.539307628029273, + "learning_rate": 1.2935469045184745e-07, + "loss": 0.7619, + "num_input_tokens_seen": 320129860, + "step": 9850 + }, + { + "epoch": 0.8883978897055508, + "flos": 21290469974400.0, + "grad_norm": 2.4970324069216723, + "learning_rate": 1.291480789427517e-07, + "loss": 0.7216, + "num_input_tokens_seen": 320156020, + "step": 9851 + }, + { + "epoch": 0.888488073229021, + "flos": 27157821965280.0, + "grad_norm": 2.0552355829421605, + "learning_rate": 1.2894162706505807e-07, + "loss": 0.7842, + "num_input_tokens_seen": 320186075, + "step": 9852 + }, + { + "epoch": 0.8885782567524914, + "flos": 20528852695680.0, + "grad_norm": 1.6983940841895209, + "learning_rate": 1.2873533483638155e-07, + "loss": 0.7342, + "num_input_tokens_seen": 320215180, + "step": 9853 + }, + { + "epoch": 0.8886684402759616, + "flos": 23227950237120.0, + "grad_norm": 2.3124039088724433, + "learning_rate": 1.285292022743243e-07, + "loss": 0.6911, + "num_input_tokens_seen": 320246860, + "step": 9854 + }, + { + "epoch": 0.8887586237994318, + "flos": 26937816924000.0, + "grad_norm": 1.8806820921568972, + "learning_rate": 1.2832322939647467e-07, + "loss": 0.6886, + "num_input_tokens_seen": 320276200, + "step": 9855 + }, + { + "epoch": 0.8888488073229021, + "flos": 22023051936960.0, + "grad_norm": 1.9193998696278969, + "learning_rate": 1.281174162204075e-07, + "loss": 0.7276, + "num_input_tokens_seen": 320306920, + "step": 9856 + }, + { + "epoch": 0.8889389908463724, + "flos": 24569581848960.0, + "grad_norm": 1.5914420826566582, + "learning_rate": 1.2791176276368366e-07, + "loss": 0.6209, + "num_input_tokens_seen": 320337865, + "step": 9857 + }, + { + "epoch": 0.8890291743698426, + "flos": 29999116773120.0, + "grad_norm": 1.594502589235928, + "learning_rate": 1.2770626904385128e-07, + "loss": 0.7551, + "num_input_tokens_seen": 320370850, + "step": 9858 + }, + { + "epoch": 0.8891193578933129, + "flos": 33899395638720.0, + "grad_norm": 2.002423725175689, + "learning_rate": 1.2750093507844306e-07, + "loss": 0.6047, + "num_input_tokens_seen": 320398260, + "step": 9859 + }, + { + "epoch": 0.8892095414167831, + "flos": 25263873225120.0, + "grad_norm": 1.526418185953251, + "learning_rate": 1.272957608849805e-07, + "loss": 0.7798, + "num_input_tokens_seen": 320428890, + "step": 9860 + }, + { + "epoch": 0.8892997249402534, + "flos": 22095024059520.0, + "grad_norm": 1.694482908003426, + "learning_rate": 1.270907464809694e-07, + "loss": 0.7163, + "num_input_tokens_seen": 320459150, + "step": 9861 + }, + { + "epoch": 0.8893899084637237, + "flos": 24571997883360.0, + "grad_norm": 1.751984628927831, + "learning_rate": 1.2688589188390285e-07, + "loss": 0.7453, + "num_input_tokens_seen": 320488775, + "step": 9862 + }, + { + "epoch": 0.8894800919871939, + "flos": 45594731111040.0, + "grad_norm": 2.0221505214686246, + "learning_rate": 1.2668119711126023e-07, + "loss": 0.7506, + "num_input_tokens_seen": 320521965, + "step": 9863 + }, + { + "epoch": 0.8895702755106643, + "flos": 28507482245280.0, + "grad_norm": 1.7108795295544446, + "learning_rate": 1.2647666218050735e-07, + "loss": 0.7448, + "num_input_tokens_seen": 320552175, + "step": 9864 + }, + { + "epoch": 0.8896604590341345, + "flos": 14663619381120.0, + "grad_norm": 2.163956622346487, + "learning_rate": 1.2627228710909643e-07, + "loss": 0.7059, + "num_input_tokens_seen": 320579340, + "step": 9865 + }, + { + "epoch": 0.8897506425576047, + "flos": 26686063174080.0, + "grad_norm": 1.7608238366312179, + "learning_rate": 1.260680719144649e-07, + "loss": 0.7082, + "num_input_tokens_seen": 320609340, + "step": 9866 + }, + { + "epoch": 0.889840826081075, + "flos": 13453071277440.0, + "grad_norm": 3.137510814876079, + "learning_rate": 1.2586401661403877e-07, + "loss": 0.6759, + "num_input_tokens_seen": 320631655, + "step": 9867 + }, + { + "epoch": 0.8899310096045453, + "flos": 26577621783360.0, + "grad_norm": 2.786214105949048, + "learning_rate": 1.2566012122522817e-07, + "loss": 0.7134, + "num_input_tokens_seen": 320660960, + "step": 9868 + }, + { + "epoch": 0.8900211931280155, + "flos": 25220973588480.0, + "grad_norm": 2.7276737604640644, + "learning_rate": 1.254563857654316e-07, + "loss": 0.7141, + "num_input_tokens_seen": 320685865, + "step": 9869 + }, + { + "epoch": 0.8901113766514858, + "flos": 18774610848000.0, + "grad_norm": 3.6183469821339243, + "learning_rate": 1.2525281025203205e-07, + "loss": 0.6309, + "num_input_tokens_seen": 320712825, + "step": 9870 + }, + { + "epoch": 0.890201560174956, + "flos": 26795954185440.0, + "grad_norm": 2.6358661996981527, + "learning_rate": 1.2504939470240006e-07, + "loss": 0.7222, + "num_input_tokens_seen": 320741840, + "step": 9871 + }, + { + "epoch": 0.8902917436984263, + "flos": 26394755065920.0, + "grad_norm": 1.6447695742913804, + "learning_rate": 1.2484613913389196e-07, + "loss": 0.8076, + "num_input_tokens_seen": 320772665, + "step": 9872 + }, + { + "epoch": 0.8903819272218966, + "flos": 28216917532320.0, + "grad_norm": 2.0989498646252143, + "learning_rate": 1.2464304356385057e-07, + "loss": 0.6544, + "num_input_tokens_seen": 320803485, + "step": 9873 + }, + { + "epoch": 0.8904721107453668, + "flos": 27046815861120.0, + "grad_norm": 2.083969845493223, + "learning_rate": 1.2444010800960558e-07, + "loss": 0.7998, + "num_input_tokens_seen": 320831070, + "step": 9874 + }, + { + "epoch": 0.890562294268837, + "flos": 35096859986880.0, + "grad_norm": 2.4924559254322167, + "learning_rate": 1.2423733248847267e-07, + "loss": 0.7269, + "num_input_tokens_seen": 320861005, + "step": 9875 + }, + { + "epoch": 0.8906524777923074, + "flos": 32253186314400.0, + "grad_norm": 3.481145893505838, + "learning_rate": 1.2403471701775293e-07, + "loss": 0.6806, + "num_input_tokens_seen": 320886670, + "step": 9876 + }, + { + "epoch": 0.8907426613157776, + "flos": 17286879144960.0, + "grad_norm": 2.13223314625827, + "learning_rate": 1.2383226161473515e-07, + "loss": 0.7777, + "num_input_tokens_seen": 320913650, + "step": 9877 + }, + { + "epoch": 0.8908328448392479, + "flos": 25848719627040.0, + "grad_norm": 1.9857317273901987, + "learning_rate": 1.2362996629669376e-07, + "loss": 0.8164, + "num_input_tokens_seen": 320941640, + "step": 9878 + }, + { + "epoch": 0.8909230283627181, + "flos": 23662347685920.0, + "grad_norm": 1.902575844895823, + "learning_rate": 1.2342783108089007e-07, + "loss": 0.7544, + "num_input_tokens_seen": 320972070, + "step": 9879 + }, + { + "epoch": 0.8910132118861884, + "flos": 25484361473280.0, + "grad_norm": 1.9722441289995438, + "learning_rate": 1.2322585598457135e-07, + "loss": 0.7985, + "num_input_tokens_seen": 321001025, + "step": 9880 + }, + { + "epoch": 0.8911033954096587, + "flos": 25410270674400.0, + "grad_norm": 1.3895456647366533, + "learning_rate": 1.2302404102497034e-07, + "loss": 0.7139, + "num_input_tokens_seen": 321032755, + "step": 9881 + }, + { + "epoch": 0.8911935789331289, + "flos": 16157892961920.0, + "grad_norm": 2.64662364860698, + "learning_rate": 1.228223862193083e-07, + "loss": 0.6678, + "num_input_tokens_seen": 321059295, + "step": 9882 + }, + { + "epoch": 0.8912837624565991, + "flos": 13422958038240.0, + "grad_norm": 2.496155739553271, + "learning_rate": 1.2262089158479038e-07, + "loss": 0.7508, + "num_input_tokens_seen": 321082995, + "step": 9883 + }, + { + "epoch": 0.8913739459800695, + "flos": 18416460044160.0, + "grad_norm": 2.01764725366355, + "learning_rate": 1.2241955713861042e-07, + "loss": 0.7851, + "num_input_tokens_seen": 321110370, + "step": 9884 + }, + { + "epoch": 0.8914641295035397, + "flos": 25592505505920.0, + "grad_norm": 1.5164893920054514, + "learning_rate": 1.222183828979464e-07, + "loss": 0.6806, + "num_input_tokens_seen": 321142525, + "step": 9885 + }, + { + "epoch": 0.8915543130270099, + "flos": 74632302968160.0, + "grad_norm": 1.6259324259681573, + "learning_rate": 1.2201736887996372e-07, + "loss": 0.7392, + "num_input_tokens_seen": 321180835, + "step": 9886 + }, + { + "epoch": 0.8916444965504803, + "flos": 20747556795360.0, + "grad_norm": 2.4625150922128825, + "learning_rate": 1.2181651510181444e-07, + "loss": 0.8199, + "num_input_tokens_seen": 321206290, + "step": 9887 + }, + { + "epoch": 0.8917346800739505, + "flos": 18379879266720.0, + "grad_norm": 1.8777607620127912, + "learning_rate": 1.2161582158063622e-07, + "loss": 0.7695, + "num_input_tokens_seen": 321233835, + "step": 9888 + }, + { + "epoch": 0.8918248635974207, + "flos": 24281172982080.0, + "grad_norm": 1.8970190329755676, + "learning_rate": 1.214152883335533e-07, + "loss": 0.6474, + "num_input_tokens_seen": 321265095, + "step": 9889 + }, + { + "epoch": 0.891915047120891, + "flos": 18416422874400.0, + "grad_norm": 2.3722445917476915, + "learning_rate": 1.2121491537767648e-07, + "loss": 0.6961, + "num_input_tokens_seen": 321292625, + "step": 9890 + }, + { + "epoch": 0.8920052306443613, + "flos": 19762477687680.0, + "grad_norm": 1.6380500568477796, + "learning_rate": 1.2101470273010294e-07, + "loss": 0.7768, + "num_input_tokens_seen": 321319935, + "step": 9891 + }, + { + "epoch": 0.8920954141678316, + "flos": 15465088376160.0, + "grad_norm": 2.2877943672927867, + "learning_rate": 1.2081465040791528e-07, + "loss": 0.7108, + "num_input_tokens_seen": 321344880, + "step": 9892 + }, + { + "epoch": 0.8921855976913018, + "flos": 30832483155840.0, + "grad_norm": 1.9619803524258477, + "learning_rate": 1.2061475842818335e-07, + "loss": 0.7046, + "num_input_tokens_seen": 321376315, + "step": 9893 + }, + { + "epoch": 0.892275781214772, + "flos": 29670261473760.0, + "grad_norm": 1.57747878639237, + "learning_rate": 1.2041502680796313e-07, + "loss": 0.7182, + "num_input_tokens_seen": 321406340, + "step": 9894 + }, + { + "epoch": 0.8923659647382424, + "flos": 26500222875840.0, + "grad_norm": 1.5381235269464277, + "learning_rate": 1.2021545556429648e-07, + "loss": 0.7468, + "num_input_tokens_seen": 321433925, + "step": 9895 + }, + { + "epoch": 0.8924561482617126, + "flos": 23808001740000.0, + "grad_norm": 1.487021887099313, + "learning_rate": 1.2001604471421245e-07, + "loss": 0.7403, + "num_input_tokens_seen": 321465320, + "step": 9896 + }, + { + "epoch": 0.8925463317851828, + "flos": 24538948233120.0, + "grad_norm": 2.2033069075568976, + "learning_rate": 1.1981679427472567e-07, + "loss": 0.7922, + "num_input_tokens_seen": 321495845, + "step": 9897 + }, + { + "epoch": 0.8926365153086531, + "flos": 25042084035360.0, + "grad_norm": 1.8656120517096073, + "learning_rate": 1.196177042628368e-07, + "loss": 0.7938, + "num_input_tokens_seen": 321525025, + "step": 9898 + }, + { + "epoch": 0.8927266988321234, + "flos": 15720224574240.0, + "grad_norm": 3.090352823058216, + "learning_rate": 1.194187746955344e-07, + "loss": 0.7461, + "num_input_tokens_seen": 321548225, + "step": 9899 + }, + { + "epoch": 0.8928168823555936, + "flos": 29339510516640.0, + "grad_norm": 1.6444995723038347, + "learning_rate": 1.1922000558979094e-07, + "loss": 0.7163, + "num_input_tokens_seen": 321575030, + "step": 9900 + }, + { + "epoch": 0.8929070658790639, + "flos": 22526708115840.0, + "grad_norm": 1.4754976338309715, + "learning_rate": 1.1902139696256752e-07, + "loss": 0.7724, + "num_input_tokens_seen": 321602865, + "step": 9901 + }, + { + "epoch": 0.8929972494025341, + "flos": 20602200099360.0, + "grad_norm": 3.9633343609379965, + "learning_rate": 1.188229488308099e-07, + "loss": 0.6794, + "num_input_tokens_seen": 321630935, + "step": 9902 + }, + { + "epoch": 0.8930874329260045, + "flos": 57560622498240.0, + "grad_norm": 0.5601170184284627, + "learning_rate": 1.1862466121145098e-07, + "loss": 0.5564, + "num_input_tokens_seen": 321730300, + "step": 9903 + }, + { + "epoch": 0.8931776164494747, + "flos": 16193804683680.0, + "grad_norm": 2.2437419753077252, + "learning_rate": 1.184265341214099e-07, + "loss": 0.6835, + "num_input_tokens_seen": 321757210, + "step": 9904 + }, + { + "epoch": 0.8932677999729449, + "flos": 31815555096480.0, + "grad_norm": 1.8714315790362301, + "learning_rate": 1.182285675775918e-07, + "loss": 0.7655, + "num_input_tokens_seen": 321788425, + "step": 9905 + }, + { + "epoch": 0.8933579834964152, + "flos": 18270025425120.0, + "grad_norm": 2.4093723598916004, + "learning_rate": 1.1803076159688851e-07, + "loss": 0.7238, + "num_input_tokens_seen": 321814170, + "step": 9906 + }, + { + "epoch": 0.8934481670198855, + "flos": 25811432624160.0, + "grad_norm": 1.8889666764130593, + "learning_rate": 1.1783311619617741e-07, + "loss": 0.7757, + "num_input_tokens_seen": 321843460, + "step": 9907 + }, + { + "epoch": 0.8935383505433557, + "flos": 30729803077920.0, + "grad_norm": 2.316856404866094, + "learning_rate": 1.1763563139232302e-07, + "loss": 0.708, + "num_input_tokens_seen": 321872235, + "step": 9908 + }, + { + "epoch": 0.893628534066826, + "flos": 24682334931840.0, + "grad_norm": 1.8517262833020123, + "learning_rate": 1.1743830720217562e-07, + "loss": 0.6768, + "num_input_tokens_seen": 321897880, + "step": 9909 + }, + { + "epoch": 0.8937187175902963, + "flos": 26357765421120.0, + "grad_norm": 2.561495048897921, + "learning_rate": 1.1724114364257243e-07, + "loss": 0.7173, + "num_input_tokens_seen": 321926400, + "step": 9910 + }, + { + "epoch": 0.8938089011137665, + "flos": 27924940368480.0, + "grad_norm": 5.544837426983302, + "learning_rate": 1.1704414073033619e-07, + "loss": 0.7479, + "num_input_tokens_seen": 321956700, + "step": 9911 + }, + { + "epoch": 0.8938990846372368, + "flos": 19618867970400.0, + "grad_norm": 2.033062405072683, + "learning_rate": 1.1684729848227636e-07, + "loss": 0.7817, + "num_input_tokens_seen": 321985615, + "step": 9912 + }, + { + "epoch": 0.893989268160707, + "flos": 26680896577440.0, + "grad_norm": 2.367805727560954, + "learning_rate": 1.1665061691518884e-07, + "loss": 0.7976, + "num_input_tokens_seen": 322014020, + "step": 9913 + }, + { + "epoch": 0.8940794516841774, + "flos": 23254495179360.0, + "grad_norm": 2.167013421799129, + "learning_rate": 1.1645409604585532e-07, + "loss": 0.7409, + "num_input_tokens_seen": 322042165, + "step": 9914 + }, + { + "epoch": 0.8941696352076476, + "flos": 23149696425120.0, + "grad_norm": 1.8031852381542874, + "learning_rate": 1.162577358910437e-07, + "loss": 0.7496, + "num_input_tokens_seen": 322070795, + "step": 9915 + }, + { + "epoch": 0.8942598187311178, + "flos": 30001644316800.0, + "grad_norm": 2.120309140140174, + "learning_rate": 1.160615364675095e-07, + "loss": 0.723, + "num_input_tokens_seen": 322100215, + "step": 9916 + }, + { + "epoch": 0.894350002254588, + "flos": 65824352806080.0, + "grad_norm": 0.6099896387877467, + "learning_rate": 1.1586549779199262e-07, + "loss": 0.5583, + "num_input_tokens_seen": 322190040, + "step": 9917 + }, + { + "epoch": 0.8944401857780584, + "flos": 21731000433600.0, + "grad_norm": 1.6455330957117336, + "learning_rate": 1.1566961988122037e-07, + "loss": 0.799, + "num_input_tokens_seen": 322220735, + "step": 9918 + }, + { + "epoch": 0.8945303693015286, + "flos": 21546609756000.0, + "grad_norm": 2.0975617968216325, + "learning_rate": 1.1547390275190627e-07, + "loss": 0.705, + "num_input_tokens_seen": 322249295, + "step": 9919 + }, + { + "epoch": 0.8946205528249989, + "flos": 12768964415520.0, + "grad_norm": 1.762514260416981, + "learning_rate": 1.1527834642075007e-07, + "loss": 0.6802, + "num_input_tokens_seen": 322275880, + "step": 9920 + }, + { + "epoch": 0.8947107363484691, + "flos": 27450988561440.0, + "grad_norm": 2.4365516186997516, + "learning_rate": 1.1508295090443797e-07, + "loss": 0.7241, + "num_input_tokens_seen": 322304005, + "step": 9921 + }, + { + "epoch": 0.8948009198719394, + "flos": 24642408876000.0, + "grad_norm": 2.0764398841450387, + "learning_rate": 1.148877162196411e-07, + "loss": 0.6914, + "num_input_tokens_seen": 322334390, + "step": 9922 + }, + { + "epoch": 0.8948911033954097, + "flos": 24680959650720.0, + "grad_norm": 2.4276852714010397, + "learning_rate": 1.1469264238301924e-07, + "loss": 0.6705, + "num_input_tokens_seen": 322358275, + "step": 9923 + }, + { + "epoch": 0.8949812869188799, + "flos": 16667459132640.0, + "grad_norm": 1.9488052204718298, + "learning_rate": 1.1449772941121638e-07, + "loss": 0.7667, + "num_input_tokens_seen": 322385830, + "step": 9924 + }, + { + "epoch": 0.8950714704423501, + "flos": 17068323724320.0, + "grad_norm": 11.741559832276199, + "learning_rate": 1.1430297732086369e-07, + "loss": 0.7102, + "num_input_tokens_seen": 322413725, + "step": 9925 + }, + { + "epoch": 0.8951616539658205, + "flos": 28180150906080.0, + "grad_norm": 1.7489854051298364, + "learning_rate": 1.1410838612857876e-07, + "loss": 0.7305, + "num_input_tokens_seen": 322445130, + "step": 9926 + }, + { + "epoch": 0.8952518374892907, + "flos": 24208866331680.0, + "grad_norm": 2.9933050474035423, + "learning_rate": 1.1391395585096497e-07, + "loss": 0.8403, + "num_input_tokens_seen": 322472290, + "step": 9927 + }, + { + "epoch": 0.895342021012761, + "flos": 29740895484960.0, + "grad_norm": 1.8099149487993094, + "learning_rate": 1.1371968650461216e-07, + "loss": 0.6691, + "num_input_tokens_seen": 322503350, + "step": 9928 + }, + { + "epoch": 0.8954322045362312, + "flos": 66492662520000.0, + "grad_norm": 0.691155555233186, + "learning_rate": 1.1352557810609687e-07, + "loss": 0.5719, + "num_input_tokens_seen": 322593240, + "step": 9929 + }, + { + "epoch": 0.8955223880597015, + "flos": 22970063476800.0, + "grad_norm": 2.806452713677258, + "learning_rate": 1.1333163067198048e-07, + "loss": 0.7245, + "num_input_tokens_seen": 322620185, + "step": 9930 + }, + { + "epoch": 0.8956125715831718, + "flos": 25626335721120.0, + "grad_norm": 1.57893011121354, + "learning_rate": 1.1313784421881311e-07, + "loss": 0.8037, + "num_input_tokens_seen": 322647600, + "step": 9931 + }, + { + "epoch": 0.895702755106642, + "flos": 20346766543200.0, + "grad_norm": 1.7985917224927286, + "learning_rate": 1.1294421876312865e-07, + "loss": 0.7763, + "num_input_tokens_seen": 322675460, + "step": 9932 + }, + { + "epoch": 0.8957929386301122, + "flos": 24172211214720.0, + "grad_norm": 2.2903955113048853, + "learning_rate": 1.1275075432144831e-07, + "loss": 0.7168, + "num_input_tokens_seen": 322702050, + "step": 9933 + }, + { + "epoch": 0.8958831221535826, + "flos": 22241904715680.0, + "grad_norm": 2.5320083989068856, + "learning_rate": 1.1255745091028002e-07, + "loss": 0.7535, + "num_input_tokens_seen": 322728315, + "step": 9934 + }, + { + "epoch": 0.8959733056770528, + "flos": 24536718047520.0, + "grad_norm": 2.1497699836292785, + "learning_rate": 1.1236430854611723e-07, + "loss": 0.7512, + "num_input_tokens_seen": 322755600, + "step": 9935 + }, + { + "epoch": 0.896063489200523, + "flos": 26905138971360.0, + "grad_norm": 2.3244769297583403, + "learning_rate": 1.1217132724544032e-07, + "loss": 0.7043, + "num_input_tokens_seen": 322783075, + "step": 9936 + }, + { + "epoch": 0.8961536727239934, + "flos": 16084471218720.0, + "grad_norm": 2.3922097930766433, + "learning_rate": 1.1197850702471434e-07, + "loss": 0.7893, + "num_input_tokens_seen": 322808670, + "step": 9937 + }, + { + "epoch": 0.8962438562474636, + "flos": 23735137543200.0, + "grad_norm": 1.8964193158132963, + "learning_rate": 1.1178584790039348e-07, + "loss": 0.8059, + "num_input_tokens_seen": 322837160, + "step": 9938 + }, + { + "epoch": 0.8963340397709338, + "flos": 27633409241760.0, + "grad_norm": 2.02902463641751, + "learning_rate": 1.1159334988891478e-07, + "loss": 0.6844, + "num_input_tokens_seen": 322866030, + "step": 9939 + }, + { + "epoch": 0.8964242232944041, + "flos": 24864606933120.0, + "grad_norm": 2.1569327779949616, + "learning_rate": 1.1140101300670446e-07, + "loss": 0.719, + "num_input_tokens_seen": 322896440, + "step": 9940 + }, + { + "epoch": 0.8965144068178744, + "flos": 22387187072160.0, + "grad_norm": 1.7610994252693981, + "learning_rate": 1.1120883727017338e-07, + "loss": 0.6076, + "num_input_tokens_seen": 322922700, + "step": 9941 + }, + { + "epoch": 0.8966045903413447, + "flos": 20529075714240.0, + "grad_norm": 2.4907608014348948, + "learning_rate": 1.1101682269571889e-07, + "loss": 0.794, + "num_input_tokens_seen": 322951000, + "step": 9942 + }, + { + "epoch": 0.8966947738648149, + "flos": 25226400373440.0, + "grad_norm": 2.3393836484520394, + "learning_rate": 1.1082496929972473e-07, + "loss": 0.7192, + "num_input_tokens_seen": 322980255, + "step": 9943 + }, + { + "epoch": 0.8967849573882851, + "flos": 26905287650400.0, + "grad_norm": 2.0344615490107145, + "learning_rate": 1.1063327709856096e-07, + "loss": 0.7092, + "num_input_tokens_seen": 323010675, + "step": 9944 + }, + { + "epoch": 0.8968751409117555, + "flos": 39109334388960.0, + "grad_norm": 2.0991912283475718, + "learning_rate": 1.1044174610858403e-07, + "loss": 0.6247, + "num_input_tokens_seen": 323038880, + "step": 9945 + }, + { + "epoch": 0.8969653244352257, + "flos": 25301346076800.0, + "grad_norm": 2.024939020439634, + "learning_rate": 1.1025037634613643e-07, + "loss": 0.6962, + "num_input_tokens_seen": 323067705, + "step": 9946 + }, + { + "epoch": 0.8970555079586959, + "flos": 24172471403040.0, + "grad_norm": 2.0008239241059314, + "learning_rate": 1.1005916782754643e-07, + "loss": 0.7665, + "num_input_tokens_seen": 323098610, + "step": 9947 + }, + { + "epoch": 0.8971456914821662, + "flos": 22714741429920.0, + "grad_norm": 2.033163020813053, + "learning_rate": 1.0986812056912919e-07, + "loss": 0.7265, + "num_input_tokens_seen": 323125730, + "step": 9948 + }, + { + "epoch": 0.8972358750056365, + "flos": 33893931684000.0, + "grad_norm": 2.284559488217871, + "learning_rate": 1.0967723458718613e-07, + "loss": 0.62, + "num_input_tokens_seen": 323155440, + "step": 9949 + }, + { + "epoch": 0.8973260585291067, + "flos": 20374129220160.0, + "grad_norm": 2.6636584022987937, + "learning_rate": 1.0948650989800445e-07, + "loss": 0.7082, + "num_input_tokens_seen": 323176980, + "step": 9950 + }, + { + "epoch": 0.897416242052577, + "flos": 25336923270720.0, + "grad_norm": 1.5962430117699848, + "learning_rate": 1.0929594651785823e-07, + "loss": 0.7205, + "num_input_tokens_seen": 323208325, + "step": 9951 + }, + { + "epoch": 0.8975064255760472, + "flos": 23037203530560.0, + "grad_norm": 1.8774290245649907, + "learning_rate": 1.0910554446300646e-07, + "loss": 0.7917, + "num_input_tokens_seen": 323236355, + "step": 9952 + }, + { + "epoch": 0.8975966090995176, + "flos": 24971933231040.0, + "grad_norm": 2.0266280187928905, + "learning_rate": 1.089153037496966e-07, + "loss": 0.638, + "num_input_tokens_seen": 323266260, + "step": 9953 + }, + { + "epoch": 0.8976867926229878, + "flos": 14225950993440.0, + "grad_norm": 4.137939503540492, + "learning_rate": 1.0872522439415966e-07, + "loss": 0.7603, + "num_input_tokens_seen": 323290780, + "step": 9954 + }, + { + "epoch": 0.897776976146458, + "flos": 20125237541760.0, + "grad_norm": 2.242529440934284, + "learning_rate": 1.0853530641261554e-07, + "loss": 0.6874, + "num_input_tokens_seen": 323318275, + "step": 9955 + }, + { + "epoch": 0.8978671596699282, + "flos": 22569347564160.0, + "grad_norm": 1.6803585397643008, + "learning_rate": 1.083455498212682e-07, + "loss": 0.7521, + "num_input_tokens_seen": 323346415, + "step": 9956 + }, + { + "epoch": 0.8979573431933986, + "flos": 24609879602400.0, + "grad_norm": 2.010353588109528, + "learning_rate": 1.0815595463630911e-07, + "loss": 0.777, + "num_input_tokens_seen": 323373700, + "step": 9957 + }, + { + "epoch": 0.8980475267168688, + "flos": 45157620269760.0, + "grad_norm": 2.1382934792962995, + "learning_rate": 1.0796652087391556e-07, + "loss": 0.6108, + "num_input_tokens_seen": 323405525, + "step": 9958 + }, + { + "epoch": 0.8981377102403391, + "flos": 27958324546560.0, + "grad_norm": 1.6694061732264756, + "learning_rate": 1.0777724855025083e-07, + "loss": 0.8347, + "num_input_tokens_seen": 323436945, + "step": 9959 + }, + { + "epoch": 0.8982278937638094, + "flos": 24937545469440.0, + "grad_norm": 1.7977337982546824, + "learning_rate": 1.075881376814649e-07, + "loss": 0.7401, + "num_input_tokens_seen": 323465375, + "step": 9960 + }, + { + "epoch": 0.8983180772872796, + "flos": 21549546167040.0, + "grad_norm": 1.906464486484424, + "learning_rate": 1.0739918828369377e-07, + "loss": 0.7956, + "num_input_tokens_seen": 323493870, + "step": 9961 + }, + { + "epoch": 0.8984082608107499, + "flos": 19691546318400.0, + "grad_norm": 2.3350705990951943, + "learning_rate": 1.0721040037305983e-07, + "loss": 0.7472, + "num_input_tokens_seen": 323521380, + "step": 9962 + }, + { + "epoch": 0.8984984443342201, + "flos": 18233370308160.0, + "grad_norm": 1.809102788026165, + "learning_rate": 1.0702177396567114e-07, + "loss": 0.6881, + "num_input_tokens_seen": 323548475, + "step": 9963 + }, + { + "epoch": 0.8985886278576904, + "flos": 69710693011680.0, + "grad_norm": 0.6221573318149374, + "learning_rate": 1.0683330907762233e-07, + "loss": 0.5618, + "num_input_tokens_seen": 323642675, + "step": 9964 + }, + { + "epoch": 0.8986788113811607, + "flos": 41186781732480.0, + "grad_norm": 1.7362210870454247, + "learning_rate": 1.0664500572499435e-07, + "loss": 0.6689, + "num_input_tokens_seen": 323674355, + "step": 9965 + }, + { + "epoch": 0.8987689949046309, + "flos": 33860956373280.0, + "grad_norm": 2.257877546911192, + "learning_rate": 1.0645686392385455e-07, + "loss": 0.6802, + "num_input_tokens_seen": 323704160, + "step": 9966 + }, + { + "epoch": 0.8988591784281011, + "flos": 27488535752640.0, + "grad_norm": 1.8283623435339422, + "learning_rate": 1.0626888369025588e-07, + "loss": 0.6699, + "num_input_tokens_seen": 323736940, + "step": 9967 + }, + { + "epoch": 0.8989493619515715, + "flos": 24534636540960.0, + "grad_norm": 1.7663777864939143, + "learning_rate": 1.0608106504023817e-07, + "loss": 0.7699, + "num_input_tokens_seen": 323767310, + "step": 9968 + }, + { + "epoch": 0.8990395454750417, + "flos": 22750541642400.0, + "grad_norm": 1.6556877045441383, + "learning_rate": 1.0589340798982637e-07, + "loss": 0.727, + "num_input_tokens_seen": 323797480, + "step": 9969 + }, + { + "epoch": 0.899129728998512, + "flos": 21877286373600.0, + "grad_norm": 1.7905284089937519, + "learning_rate": 1.057059125550337e-07, + "loss": 0.6816, + "num_input_tokens_seen": 323827415, + "step": 9970 + }, + { + "epoch": 0.8992199125219822, + "flos": 26353230710400.0, + "grad_norm": 1.7356134752626535, + "learning_rate": 1.0551857875185732e-07, + "loss": 0.7512, + "num_input_tokens_seen": 323856900, + "step": 9971 + }, + { + "epoch": 0.8993100960454525, + "flos": 25447223149440.0, + "grad_norm": 1.6816415465526293, + "learning_rate": 1.0533140659628181e-07, + "loss": 0.6272, + "num_input_tokens_seen": 323888565, + "step": 9972 + }, + { + "epoch": 0.8994002795689228, + "flos": 21688286645760.0, + "grad_norm": 2.503986137175043, + "learning_rate": 1.0514439610427772e-07, + "loss": 0.7968, + "num_input_tokens_seen": 323915890, + "step": 9973 + }, + { + "epoch": 0.899490463092393, + "flos": 21258386737920.0, + "grad_norm": 2.1416433098059926, + "learning_rate": 1.0495754729180206e-07, + "loss": 0.7677, + "num_input_tokens_seen": 323944770, + "step": 9974 + }, + { + "epoch": 0.8995806466158632, + "flos": 19250606991840.0, + "grad_norm": 1.94405903705386, + "learning_rate": 1.0477086017479741e-07, + "loss": 0.8381, + "num_input_tokens_seen": 323974625, + "step": 9975 + }, + { + "epoch": 0.8996708301393336, + "flos": 18343893205440.0, + "grad_norm": 2.721830617759614, + "learning_rate": 1.0458433476919327e-07, + "loss": 0.6969, + "num_input_tokens_seen": 324001850, + "step": 9976 + }, + { + "epoch": 0.8997610136628038, + "flos": 26758072466400.0, + "grad_norm": 1.4925359349102088, + "learning_rate": 1.0439797109090509e-07, + "loss": 0.768, + "num_input_tokens_seen": 324033640, + "step": 9977 + }, + { + "epoch": 0.899851197186274, + "flos": 22059744223680.0, + "grad_norm": 2.385729096010182, + "learning_rate": 1.0421176915583396e-07, + "loss": 0.7116, + "num_input_tokens_seen": 324061695, + "step": 9978 + }, + { + "epoch": 0.8999413807097443, + "flos": 19361724605280.0, + "grad_norm": 2.1015657899965534, + "learning_rate": 1.0402572897986828e-07, + "loss": 0.7696, + "num_input_tokens_seen": 324089055, + "step": 9979 + }, + { + "epoch": 0.9000315642332146, + "flos": 65095710838080.0, + "grad_norm": 0.5852510372952765, + "learning_rate": 1.0383985057888134e-07, + "loss": 0.5673, + "num_input_tokens_seen": 324193250, + "step": 9980 + }, + { + "epoch": 0.9001217477566849, + "flos": 33168114617760.0, + "grad_norm": 2.0360136074967943, + "learning_rate": 1.036541339687338e-07, + "loss": 0.6436, + "num_input_tokens_seen": 324223265, + "step": 9981 + }, + { + "epoch": 0.9002119312801551, + "flos": 70325578313280.0, + "grad_norm": 0.6406340714719528, + "learning_rate": 1.0346857916527186e-07, + "loss": 0.5373, + "num_input_tokens_seen": 324313750, + "step": 9982 + }, + { + "epoch": 0.9003021148036254, + "flos": 16959436296480.0, + "grad_norm": 1.9299168624167349, + "learning_rate": 1.0328318618432819e-07, + "loss": 0.7356, + "num_input_tokens_seen": 324342500, + "step": 9983 + }, + { + "epoch": 0.9003922983270957, + "flos": 26175530589600.0, + "grad_norm": 2.05647020639197, + "learning_rate": 1.0309795504172148e-07, + "loss": 0.7993, + "num_input_tokens_seen": 324370595, + "step": 9984 + }, + { + "epoch": 0.9004824818505659, + "flos": 15902533745280.0, + "grad_norm": 2.064856714445384, + "learning_rate": 1.0291288575325685e-07, + "loss": 0.7153, + "num_input_tokens_seen": 324397910, + "step": 9985 + }, + { + "epoch": 0.9005726653740361, + "flos": 31022114769600.0, + "grad_norm": 2.003133345897505, + "learning_rate": 1.0272797833472502e-07, + "loss": 0.6994, + "num_input_tokens_seen": 324429265, + "step": 9986 + }, + { + "epoch": 0.9006628488975065, + "flos": 70342899421440.0, + "grad_norm": 0.6835972159161579, + "learning_rate": 1.0254323280190335e-07, + "loss": 0.5469, + "num_input_tokens_seen": 324522280, + "step": 9987 + }, + { + "epoch": 0.9007530324209767, + "flos": 17869904228640.0, + "grad_norm": 2.6042215773833237, + "learning_rate": 1.023586491705557e-07, + "loss": 0.7143, + "num_input_tokens_seen": 324549070, + "step": 9988 + }, + { + "epoch": 0.9008432159444469, + "flos": 22707865024320.0, + "grad_norm": 2.298941895449636, + "learning_rate": 1.0217422745643145e-07, + "loss": 0.6746, + "num_input_tokens_seen": 324578175, + "step": 9989 + }, + { + "epoch": 0.9009333994679172, + "flos": 22278411153600.0, + "grad_norm": 1.7189459949063381, + "learning_rate": 1.0198996767526691e-07, + "loss": 0.6613, + "num_input_tokens_seen": 324607105, + "step": 9990 + }, + { + "epoch": 0.9010235829913875, + "flos": 37907260990560.0, + "grad_norm": 3.1668875433201182, + "learning_rate": 1.018058698427835e-07, + "loss": 0.7209, + "num_input_tokens_seen": 324637060, + "step": 9991 + }, + { + "epoch": 0.9011137665148578, + "flos": 23479666817280.0, + "grad_norm": 1.7346966419449814, + "learning_rate": 1.0162193397469021e-07, + "loss": 0.7875, + "num_input_tokens_seen": 324666330, + "step": 9992 + }, + { + "epoch": 0.901203950038328, + "flos": 24935649811680.0, + "grad_norm": 2.0046958631785556, + "learning_rate": 1.0143816008668049e-07, + "loss": 0.6757, + "num_input_tokens_seen": 324692980, + "step": 9993 + }, + { + "epoch": 0.9012941335617982, + "flos": 60182506980960.0, + "grad_norm": 0.6366522818666257, + "learning_rate": 1.0125454819443624e-07, + "loss": 0.5853, + "num_input_tokens_seen": 324785610, + "step": 9994 + }, + { + "epoch": 0.9013843170852686, + "flos": 40564127951040.0, + "grad_norm": 1.7521043795101388, + "learning_rate": 1.0107109831362315e-07, + "loss": 0.6442, + "num_input_tokens_seen": 324819570, + "step": 9995 + }, + { + "epoch": 0.9014745006087388, + "flos": 26393751482400.0, + "grad_norm": 1.823673338295255, + "learning_rate": 1.0088781045989447e-07, + "loss": 0.7457, + "num_input_tokens_seen": 324850015, + "step": 9996 + }, + { + "epoch": 0.901564684132209, + "flos": 19287819655200.0, + "grad_norm": 2.073664416937974, + "learning_rate": 1.0070468464888926e-07, + "loss": 0.7882, + "num_input_tokens_seen": 324876650, + "step": 9997 + }, + { + "epoch": 0.9016548676556793, + "flos": 25587599097600.0, + "grad_norm": 2.0082082395519993, + "learning_rate": 1.0052172089623324e-07, + "loss": 0.7565, + "num_input_tokens_seen": 324904355, + "step": 9998 + }, + { + "epoch": 0.9017450511791496, + "flos": 23845028554560.0, + "grad_norm": 2.3259332568104183, + "learning_rate": 1.0033891921753746e-07, + "loss": 0.7782, + "num_input_tokens_seen": 324935105, + "step": 9999 + }, + { + "epoch": 0.9018352347026198, + "flos": 13242581694720.0, + "grad_norm": 1.903720490695931, + "learning_rate": 1.0015627962839968e-07, + "loss": 0.6466, + "num_input_tokens_seen": 324961375, + "step": 10000 + }, + { + "epoch": 0.9019254182260901, + "flos": 26571674621760.0, + "grad_norm": 2.3799443227201866, + "learning_rate": 9.99738021444041e-08, + "loss": 0.7125, + "num_input_tokens_seen": 324990770, + "step": 10001 + }, + { + "epoch": 0.9020156017495603, + "flos": 32916286528320.0, + "grad_norm": 1.7101264320168454, + "learning_rate": 9.979148678112003e-08, + "loss": 0.6673, + "num_input_tokens_seen": 325022040, + "step": 10002 + }, + { + "epoch": 0.9021057852730306, + "flos": 16083653484000.0, + "grad_norm": 2.453382733649309, + "learning_rate": 9.960933355410417e-08, + "loss": 0.8368, + "num_input_tokens_seen": 325046800, + "step": 10003 + }, + { + "epoch": 0.9021959687965009, + "flos": 18888181665600.0, + "grad_norm": 4.4049743537286155, + "learning_rate": 9.942734247889828e-08, + "loss": 0.6526, + "num_input_tokens_seen": 325075770, + "step": 10004 + }, + { + "epoch": 0.9022861523199711, + "flos": 19909953060000.0, + "grad_norm": 1.742145554931407, + "learning_rate": 9.92455135710315e-08, + "loss": 0.7423, + "num_input_tokens_seen": 325103715, + "step": 10005 + }, + { + "epoch": 0.9023763358434415, + "flos": 25411051239360.0, + "grad_norm": 2.5193624277257785, + "learning_rate": 9.906384684601787e-08, + "loss": 0.6694, + "num_input_tokens_seen": 325133285, + "step": 10006 + }, + { + "epoch": 0.9024665193669117, + "flos": 23621826913920.0, + "grad_norm": 1.8216312199547064, + "learning_rate": 9.8882342319359e-08, + "loss": 0.7842, + "num_input_tokens_seen": 325162835, + "step": 10007 + }, + { + "epoch": 0.9025567028903819, + "flos": 28033047231360.0, + "grad_norm": 1.6742854344931768, + "learning_rate": 9.870100000654048e-08, + "loss": 0.6876, + "num_input_tokens_seen": 325194685, + "step": 10008 + }, + { + "epoch": 0.9026468864138522, + "flos": 28031709120000.0, + "grad_norm": 1.5297763506419486, + "learning_rate": 9.851981992303704e-08, + "loss": 0.7083, + "num_input_tokens_seen": 325225120, + "step": 10009 + }, + { + "epoch": 0.9027370699373225, + "flos": 25336440063840.0, + "grad_norm": 1.5900062271811943, + "learning_rate": 9.833880208430678e-08, + "loss": 0.7396, + "num_input_tokens_seen": 325256725, + "step": 10010 + }, + { + "epoch": 0.9028272534607927, + "flos": 22023423634560.0, + "grad_norm": 2.7311625765638343, + "learning_rate": 9.815794650579601e-08, + "loss": 0.7306, + "num_input_tokens_seen": 325284320, + "step": 10011 + }, + { + "epoch": 0.902917436984263, + "flos": 21579956764320.0, + "grad_norm": 2.7170610114342217, + "learning_rate": 9.797725320293548e-08, + "loss": 0.7153, + "num_input_tokens_seen": 325311750, + "step": 10012 + }, + { + "epoch": 0.9030076205077332, + "flos": 19982891596320.0, + "grad_norm": 2.296829949800015, + "learning_rate": 9.779672219114354e-08, + "loss": 0.8148, + "num_input_tokens_seen": 325335970, + "step": 10013 + }, + { + "epoch": 0.9030978040312035, + "flos": 31200929983200.0, + "grad_norm": 1.980840584122875, + "learning_rate": 9.761635348582386e-08, + "loss": 0.715, + "num_input_tokens_seen": 325369825, + "step": 10014 + }, + { + "epoch": 0.9031879875546738, + "flos": 18015558282720.0, + "grad_norm": 1.8776815947829704, + "learning_rate": 9.743614710236658e-08, + "loss": 0.7676, + "num_input_tokens_seen": 325397540, + "step": 10015 + }, + { + "epoch": 0.903278171078144, + "flos": 15064558312320.0, + "grad_norm": 1.832192289842919, + "learning_rate": 9.725610305614806e-08, + "loss": 0.6994, + "num_input_tokens_seen": 325425110, + "step": 10016 + }, + { + "epoch": 0.9033683546016142, + "flos": 35064888259680.0, + "grad_norm": 2.4108045006074064, + "learning_rate": 9.707622136253002e-08, + "loss": 0.7324, + "num_input_tokens_seen": 325456410, + "step": 10017 + }, + { + "epoch": 0.9034585381250846, + "flos": 23261483094240.0, + "grad_norm": 1.7803214480506835, + "learning_rate": 9.689650203686173e-08, + "loss": 0.7601, + "num_input_tokens_seen": 325485055, + "step": 10018 + }, + { + "epoch": 0.9035487216485548, + "flos": 22496706385920.0, + "grad_norm": 2.184626845002207, + "learning_rate": 9.671694509447715e-08, + "loss": 0.7118, + "num_input_tokens_seen": 325514215, + "step": 10019 + }, + { + "epoch": 0.903638905172025, + "flos": 25039519321920.0, + "grad_norm": 3.1558077662899366, + "learning_rate": 9.653755055069757e-08, + "loss": 0.7513, + "num_input_tokens_seen": 325542000, + "step": 10020 + }, + { + "epoch": 0.9037290886954953, + "flos": 39514399163520.0, + "grad_norm": 1.432368032558176, + "learning_rate": 9.635831842082987e-08, + "loss": 0.7146, + "num_input_tokens_seen": 325577720, + "step": 10021 + }, + { + "epoch": 0.9038192722189656, + "flos": 23625952757280.0, + "grad_norm": 1.8074639830951424, + "learning_rate": 9.617924872016691e-08, + "loss": 0.755, + "num_input_tokens_seen": 325603400, + "step": 10022 + }, + { + "epoch": 0.9039094557424359, + "flos": 21767432532000.0, + "grad_norm": 1.9263202152908998, + "learning_rate": 9.600034146398806e-08, + "loss": 0.7753, + "num_input_tokens_seen": 325632105, + "step": 10023 + }, + { + "epoch": 0.9039996392659061, + "flos": 71615606831040.0, + "grad_norm": 0.696580352417519, + "learning_rate": 9.582159666755863e-08, + "loss": 0.5799, + "num_input_tokens_seen": 325733400, + "step": 10024 + }, + { + "epoch": 0.9040898227893763, + "flos": 33059115680640.0, + "grad_norm": 1.6178075366563947, + "learning_rate": 9.564301434612976e-08, + "loss": 0.7078, + "num_input_tokens_seen": 325761750, + "step": 10025 + }, + { + "epoch": 0.9041800063128467, + "flos": 20050812215040.0, + "grad_norm": 1.7272505964111677, + "learning_rate": 9.546459451494015e-08, + "loss": 0.7794, + "num_input_tokens_seen": 325785560, + "step": 10026 + }, + { + "epoch": 0.9042701898363169, + "flos": 67044124743840.0, + "grad_norm": 0.6320357323712757, + "learning_rate": 9.528633718921231e-08, + "loss": 0.6046, + "num_input_tokens_seen": 325872395, + "step": 10027 + }, + { + "epoch": 0.9043603733597871, + "flos": 24024921691200.0, + "grad_norm": 1.9161085131543931, + "learning_rate": 9.510824238415672e-08, + "loss": 0.7863, + "num_input_tokens_seen": 325901845, + "step": 10028 + }, + { + "epoch": 0.9044505568832575, + "flos": 26974286192160.0, + "grad_norm": 1.8311658169401943, + "learning_rate": 9.493031011496944e-08, + "loss": 0.7706, + "num_input_tokens_seen": 325931770, + "step": 10029 + }, + { + "epoch": 0.9045407404067277, + "flos": 23953730133600.0, + "grad_norm": 15.978158825354134, + "learning_rate": 9.475254039683234e-08, + "loss": 0.5934, + "num_input_tokens_seen": 325960420, + "step": 10030 + }, + { + "epoch": 0.904630923930198, + "flos": 28508634507840.0, + "grad_norm": 1.6756196199734126, + "learning_rate": 9.45749332449144e-08, + "loss": 0.7018, + "num_input_tokens_seen": 325992380, + "step": 10031 + }, + { + "epoch": 0.9047211074536682, + "flos": 62354828752800.0, + "grad_norm": 0.6431258397889285, + "learning_rate": 9.439748867436903e-08, + "loss": 0.5886, + "num_input_tokens_seen": 326082890, + "step": 10032 + }, + { + "epoch": 0.9048112909771385, + "flos": 20783096819520.0, + "grad_norm": 1.7280800480713023, + "learning_rate": 9.42202067003377e-08, + "loss": 0.698, + "num_input_tokens_seen": 326112385, + "step": 10033 + }, + { + "epoch": 0.9049014745006088, + "flos": 21694940032800.0, + "grad_norm": 1.963276414299811, + "learning_rate": 9.404308733794652e-08, + "loss": 0.8746, + "num_input_tokens_seen": 326137730, + "step": 10034 + }, + { + "epoch": 0.904991658024079, + "flos": 24606051117120.0, + "grad_norm": 1.8297490787975474, + "learning_rate": 9.38661306023083e-08, + "loss": 0.7558, + "num_input_tokens_seen": 326168145, + "step": 10035 + }, + { + "epoch": 0.9050818415475492, + "flos": 18525533320800.0, + "grad_norm": 3.3007917957024584, + "learning_rate": 9.368933650852229e-08, + "loss": 0.729, + "num_input_tokens_seen": 326194605, + "step": 10036 + }, + { + "epoch": 0.9051720250710196, + "flos": 66296148767040.0, + "grad_norm": 0.6902151030617133, + "learning_rate": 9.351270507167352e-08, + "loss": 0.5928, + "num_input_tokens_seen": 326290925, + "step": 10037 + }, + { + "epoch": 0.9052622085944898, + "flos": 37579520784000.0, + "grad_norm": 2.089174309579477, + "learning_rate": 9.333623630683285e-08, + "loss": 0.672, + "num_input_tokens_seen": 326321615, + "step": 10038 + }, + { + "epoch": 0.90535239211796, + "flos": 27919810941600.0, + "grad_norm": 2.381486667676629, + "learning_rate": 9.315993022905799e-08, + "loss": 0.588, + "num_input_tokens_seen": 326350495, + "step": 10039 + }, + { + "epoch": 0.9054425756414303, + "flos": 22168557312000.0, + "grad_norm": 1.914628237203606, + "learning_rate": 9.298378685339158e-08, + "loss": 0.7161, + "num_input_tokens_seen": 326377975, + "step": 10040 + }, + { + "epoch": 0.9055327591649006, + "flos": 22204357524480.0, + "grad_norm": 2.556824236775945, + "learning_rate": 9.280780619486406e-08, + "loss": 0.8014, + "num_input_tokens_seen": 326405785, + "step": 10041 + }, + { + "epoch": 0.9056229426883708, + "flos": 25229373954240.0, + "grad_norm": 1.9795771591399633, + "learning_rate": 9.26319882684905e-08, + "loss": 0.818, + "num_input_tokens_seen": 326433290, + "step": 10042 + }, + { + "epoch": 0.9057131262118411, + "flos": 32368652789760.0, + "grad_norm": 1.7814146266954782, + "learning_rate": 9.245633308927293e-08, + "loss": 0.7082, + "num_input_tokens_seen": 326464955, + "step": 10043 + }, + { + "epoch": 0.9058033097353113, + "flos": 29782568519520.0, + "grad_norm": 1.8830034467814498, + "learning_rate": 9.228084067219888e-08, + "loss": 0.7178, + "num_input_tokens_seen": 326495220, + "step": 10044 + }, + { + "epoch": 0.9058934932587817, + "flos": 22897496638080.0, + "grad_norm": 1.9394502361233834, + "learning_rate": 9.210551103224284e-08, + "loss": 0.6825, + "num_input_tokens_seen": 326523740, + "step": 10045 + }, + { + "epoch": 0.9059836767822519, + "flos": 25445513340480.0, + "grad_norm": 1.664312925882266, + "learning_rate": 9.193034418436463e-08, + "loss": 0.7889, + "num_input_tokens_seen": 326553810, + "step": 10046 + }, + { + "epoch": 0.9060738603057221, + "flos": 63983828478240.0, + "grad_norm": 0.653668936413597, + "learning_rate": 9.175534014351005e-08, + "loss": 0.6126, + "num_input_tokens_seen": 326649375, + "step": 10047 + }, + { + "epoch": 0.9061640438291924, + "flos": 19472098823520.0, + "grad_norm": 3.1161994611097823, + "learning_rate": 9.158049892461228e-08, + "loss": 0.6798, + "num_input_tokens_seen": 326675665, + "step": 10048 + }, + { + "epoch": 0.9062542273526627, + "flos": 44428606604160.0, + "grad_norm": 2.2881810701320586, + "learning_rate": 9.140582054258871e-08, + "loss": 0.6756, + "num_input_tokens_seen": 326709535, + "step": 10049 + }, + { + "epoch": 0.9063444108761329, + "flos": 33494145015360.0, + "grad_norm": 1.7135055986529457, + "learning_rate": 9.123130501234499e-08, + "loss": 0.7458, + "num_input_tokens_seen": 326741040, + "step": 10050 + }, + { + "epoch": 0.9064345943996032, + "flos": 22963335750240.0, + "grad_norm": 2.367228491903483, + "learning_rate": 9.105695234877098e-08, + "loss": 0.7857, + "num_input_tokens_seen": 326769540, + "step": 10051 + }, + { + "epoch": 0.9065247779230734, + "flos": 21291213369600.0, + "grad_norm": 3.444868063428013, + "learning_rate": 9.088276256674344e-08, + "loss": 0.7638, + "num_input_tokens_seen": 326798465, + "step": 10052 + }, + { + "epoch": 0.9066149614465437, + "flos": 23880903106560.0, + "grad_norm": 1.693093671709531, + "learning_rate": 9.070873568112536e-08, + "loss": 0.7234, + "num_input_tokens_seen": 326826600, + "step": 10053 + }, + { + "epoch": 0.906705144970014, + "flos": 25994039153280.0, + "grad_norm": 1.8704503492196403, + "learning_rate": 9.053487170676577e-08, + "loss": 0.7961, + "num_input_tokens_seen": 326857005, + "step": 10054 + }, + { + "epoch": 0.9067953284934842, + "flos": 20053562777280.0, + "grad_norm": 1.9338153750632279, + "learning_rate": 9.036117065849968e-08, + "loss": 0.7934, + "num_input_tokens_seen": 326885730, + "step": 10055 + }, + { + "epoch": 0.9068855120169546, + "flos": 24858213734400.0, + "grad_norm": 1.8453204691793152, + "learning_rate": 9.018763255114837e-08, + "loss": 0.721, + "num_input_tokens_seen": 326915275, + "step": 10056 + }, + { + "epoch": 0.9069756955404248, + "flos": 20199068152320.0, + "grad_norm": 1.7362274037724648, + "learning_rate": 9.00142573995184e-08, + "loss": 0.7055, + "num_input_tokens_seen": 326942455, + "step": 10057 + }, + { + "epoch": 0.907065879063895, + "flos": 32187533051040.0, + "grad_norm": 2.688701893169046, + "learning_rate": 8.984104521840375e-08, + "loss": 0.8164, + "num_input_tokens_seen": 326972680, + "step": 10058 + }, + { + "epoch": 0.9071560625873653, + "flos": 22569979450080.0, + "grad_norm": 2.1564846187602886, + "learning_rate": 8.966799602258346e-08, + "loss": 0.7582, + "num_input_tokens_seen": 326999890, + "step": 10059 + }, + { + "epoch": 0.9072462461108356, + "flos": 16704077079840.0, + "grad_norm": 1.896059658528373, + "learning_rate": 8.949510982682329e-08, + "loss": 0.7789, + "num_input_tokens_seen": 327026230, + "step": 10060 + }, + { + "epoch": 0.9073364296343058, + "flos": 32288168792160.0, + "grad_norm": 2.66898488567785, + "learning_rate": 8.932238664587499e-08, + "loss": 0.8055, + "num_input_tokens_seen": 327050915, + "step": 10061 + }, + { + "epoch": 0.9074266131577761, + "flos": 22707530496480.0, + "grad_norm": 2.0600872376867643, + "learning_rate": 8.914982649447567e-08, + "loss": 0.829, + "num_input_tokens_seen": 327077835, + "step": 10062 + }, + { + "epoch": 0.9075167966812463, + "flos": 22496632046400.0, + "grad_norm": 1.6739752581282197, + "learning_rate": 8.897742938734975e-08, + "loss": 0.8034, + "num_input_tokens_seen": 327106400, + "step": 10063 + }, + { + "epoch": 0.9076069802047166, + "flos": 21364932470880.0, + "grad_norm": 1.9134020853261908, + "learning_rate": 8.880519533920661e-08, + "loss": 0.7645, + "num_input_tokens_seen": 327134030, + "step": 10064 + }, + { + "epoch": 0.9076971637281869, + "flos": 23259327248160.0, + "grad_norm": 1.5442751991077315, + "learning_rate": 8.863312436474268e-08, + "loss": 0.7384, + "num_input_tokens_seen": 327166240, + "step": 10065 + }, + { + "epoch": 0.9077873472516571, + "flos": 18372408144960.0, + "grad_norm": 2.3365245068777334, + "learning_rate": 8.846121647863936e-08, + "loss": 0.7433, + "num_input_tokens_seen": 327193190, + "step": 10066 + }, + { + "epoch": 0.9078775307751273, + "flos": 25375176687360.0, + "grad_norm": 1.6369158277921962, + "learning_rate": 8.828947169556555e-08, + "loss": 0.7399, + "num_input_tokens_seen": 327221900, + "step": 10067 + }, + { + "epoch": 0.9079677142985977, + "flos": 24750887436480.0, + "grad_norm": 1.721385253379312, + "learning_rate": 8.81178900301749e-08, + "loss": 0.7067, + "num_input_tokens_seen": 327252060, + "step": 10068 + }, + { + "epoch": 0.9080578978220679, + "flos": 37792240552320.0, + "grad_norm": 1.593660838307107, + "learning_rate": 8.794647149710787e-08, + "loss": 0.6551, + "num_input_tokens_seen": 327283880, + "step": 10069 + }, + { + "epoch": 0.9081480813455381, + "flos": 24899589410880.0, + "grad_norm": 1.922105046592118, + "learning_rate": 8.777521611099081e-08, + "loss": 0.7304, + "num_input_tokens_seen": 327308650, + "step": 10070 + }, + { + "epoch": 0.9082382648690084, + "flos": 19290384368640.0, + "grad_norm": 1.8506541539001884, + "learning_rate": 8.760412388643624e-08, + "loss": 0.6771, + "num_input_tokens_seen": 327338245, + "step": 10071 + }, + { + "epoch": 0.9083284483924787, + "flos": 19870398701760.0, + "grad_norm": 4.22413009626721, + "learning_rate": 8.74331948380429e-08, + "loss": 0.7525, + "num_input_tokens_seen": 327365420, + "step": 10072 + }, + { + "epoch": 0.908418631915949, + "flos": 18051990381120.0, + "grad_norm": 1.8617420036602141, + "learning_rate": 8.726242898039516e-08, + "loss": 0.7387, + "num_input_tokens_seen": 327392520, + "step": 10073 + }, + { + "epoch": 0.9085088154394192, + "flos": 16521619229760.0, + "grad_norm": 1.9620861929257463, + "learning_rate": 8.709182632806334e-08, + "loss": 0.7472, + "num_input_tokens_seen": 327420180, + "step": 10074 + }, + { + "epoch": 0.9085989989628894, + "flos": 27377938515840.0, + "grad_norm": 2.060064322569215, + "learning_rate": 8.692138689560469e-08, + "loss": 0.7462, + "num_input_tokens_seen": 327448880, + "step": 10075 + }, + { + "epoch": 0.9086891824863598, + "flos": 23553869125440.0, + "grad_norm": 1.9973057242056922, + "learning_rate": 8.675111069756203e-08, + "loss": 0.719, + "num_input_tokens_seen": 327480835, + "step": 10076 + }, + { + "epoch": 0.90877936600983, + "flos": 36044837940480.0, + "grad_norm": 1.7392987991292894, + "learning_rate": 8.658099774846395e-08, + "loss": 0.7087, + "num_input_tokens_seen": 327512315, + "step": 10077 + }, + { + "epoch": 0.9088695495333002, + "flos": 32078199586080.0, + "grad_norm": 1.482937729333916, + "learning_rate": 8.641104806282595e-08, + "loss": 0.6816, + "num_input_tokens_seen": 327545885, + "step": 10078 + }, + { + "epoch": 0.9089597330567706, + "flos": 20856964599840.0, + "grad_norm": 2.088928935104483, + "learning_rate": 8.624126165514845e-08, + "loss": 0.7476, + "num_input_tokens_seen": 327573545, + "step": 10079 + }, + { + "epoch": 0.9090499165802408, + "flos": 29491446260160.0, + "grad_norm": 1.9762785334400208, + "learning_rate": 8.607163853991917e-08, + "loss": 0.7012, + "num_input_tokens_seen": 327603870, + "step": 10080 + }, + { + "epoch": 0.909140100103711, + "flos": 36960063601920.0, + "grad_norm": 1.9929674603896566, + "learning_rate": 8.590217873161054e-08, + "loss": 0.5849, + "num_input_tokens_seen": 327637385, + "step": 10081 + }, + { + "epoch": 0.9092302836271813, + "flos": 26937854093760.0, + "grad_norm": 1.7548854617702248, + "learning_rate": 8.573288224468255e-08, + "loss": 0.755, + "num_input_tokens_seen": 327666420, + "step": 10082 + }, + { + "epoch": 0.9093204671506516, + "flos": 20927672950560.0, + "grad_norm": 2.3408848861426095, + "learning_rate": 8.556374909358011e-08, + "loss": 0.6668, + "num_input_tokens_seen": 327696195, + "step": 10083 + }, + { + "epoch": 0.9094106506741219, + "flos": 53278558595040.0, + "grad_norm": 2.1092569953971556, + "learning_rate": 8.539477929273476e-08, + "loss": 0.716, + "num_input_tokens_seen": 327729975, + "step": 10084 + }, + { + "epoch": 0.9095008341975921, + "flos": 27592999979040.0, + "grad_norm": 1.7278640467267519, + "learning_rate": 8.522597285656386e-08, + "loss": 0.7745, + "num_input_tokens_seen": 327758455, + "step": 10085 + }, + { + "epoch": 0.9095910177210623, + "flos": 71581813785600.0, + "grad_norm": 3.19751721267899, + "learning_rate": 8.505732979947078e-08, + "loss": 0.5888, + "num_input_tokens_seen": 327843120, + "step": 10086 + }, + { + "epoch": 0.9096812012445327, + "flos": 22207070916960.0, + "grad_norm": 1.3283598299546773, + "learning_rate": 8.488885013584557e-08, + "loss": 0.7535, + "num_input_tokens_seen": 327873185, + "step": 10087 + }, + { + "epoch": 0.9097713847680029, + "flos": 24971598703200.0, + "grad_norm": 1.8126057561078182, + "learning_rate": 8.472053388006295e-08, + "loss": 0.6785, + "num_input_tokens_seen": 327904275, + "step": 10088 + }, + { + "epoch": 0.9098615682914731, + "flos": 23115791870400.0, + "grad_norm": 2.6802718968322012, + "learning_rate": 8.455238104648565e-08, + "loss": 0.8259, + "num_input_tokens_seen": 327932705, + "step": 10089 + }, + { + "epoch": 0.9099517518149434, + "flos": 47889470103360.0, + "grad_norm": 3.2928673851910704, + "learning_rate": 8.438439164946043e-08, + "loss": 0.6113, + "num_input_tokens_seen": 327965300, + "step": 10090 + }, + { + "epoch": 0.9100419353384137, + "flos": 24317382061920.0, + "grad_norm": 1.9556926481763077, + "learning_rate": 8.42165657033218e-08, + "loss": 0.7512, + "num_input_tokens_seen": 327995225, + "step": 10091 + }, + { + "epoch": 0.9101321188618839, + "flos": 24570882790560.0, + "grad_norm": 2.0501217317491744, + "learning_rate": 8.4048903222389e-08, + "loss": 0.7717, + "num_input_tokens_seen": 328023530, + "step": 10092 + }, + { + "epoch": 0.9102223023853542, + "flos": 24099123999360.0, + "grad_norm": 2.0703513087827785, + "learning_rate": 8.388140422096856e-08, + "loss": 0.7883, + "num_input_tokens_seen": 328052795, + "step": 10093 + }, + { + "epoch": 0.9103124859088244, + "flos": 24317939608320.0, + "grad_norm": 2.2083635747986143, + "learning_rate": 8.371406871335173e-08, + "loss": 0.751, + "num_input_tokens_seen": 328082375, + "step": 10094 + }, + { + "epoch": 0.9104026694322948, + "flos": 22093277080800.0, + "grad_norm": 1.886860591382322, + "learning_rate": 8.354689671381732e-08, + "loss": 0.7401, + "num_input_tokens_seen": 328112000, + "step": 10095 + }, + { + "epoch": 0.910492852955765, + "flos": 38052803535360.0, + "grad_norm": 1.638704392811856, + "learning_rate": 8.337988823662834e-08, + "loss": 0.7339, + "num_input_tokens_seen": 328145685, + "step": 10096 + }, + { + "epoch": 0.9105830364792352, + "flos": 19290607387200.0, + "grad_norm": 1.7264288941209507, + "learning_rate": 8.321304329603607e-08, + "loss": 0.6926, + "num_input_tokens_seen": 328172765, + "step": 10097 + }, + { + "epoch": 0.9106732200027055, + "flos": 70995146065440.0, + "grad_norm": 0.6275951152692261, + "learning_rate": 8.304636190627557e-08, + "loss": 0.586, + "num_input_tokens_seen": 328270340, + "step": 10098 + }, + { + "epoch": 0.9107634035261758, + "flos": 20273902346400.0, + "grad_norm": 2.7867743672391865, + "learning_rate": 8.287984408156945e-08, + "loss": 0.7536, + "num_input_tokens_seen": 328296230, + "step": 10099 + }, + { + "epoch": 0.910853587049646, + "flos": 23734877354880.0, + "grad_norm": 11.955047333167746, + "learning_rate": 8.271348983612591e-08, + "loss": 0.64, + "num_input_tokens_seen": 328323645, + "step": 10100 + }, + { + "epoch": 0.9109437705731163, + "flos": 32988184311360.0, + "grad_norm": 1.9688197532285998, + "learning_rate": 8.254729918413938e-08, + "loss": 0.6986, + "num_input_tokens_seen": 328356440, + "step": 10101 + }, + { + "epoch": 0.9110339540965866, + "flos": 21695311730400.0, + "grad_norm": 1.7599504143058775, + "learning_rate": 8.238127213979006e-08, + "loss": 0.7055, + "num_input_tokens_seen": 328385890, + "step": 10102 + }, + { + "epoch": 0.9111241376200568, + "flos": 32582599160160.0, + "grad_norm": 2.0423493184756567, + "learning_rate": 8.221540871724398e-08, + "loss": 0.81, + "num_input_tokens_seen": 328414370, + "step": 10103 + }, + { + "epoch": 0.9112143211435271, + "flos": 27414853821120.0, + "grad_norm": 1.7267537421040562, + "learning_rate": 8.2049708930654e-08, + "loss": 0.6257, + "num_input_tokens_seen": 328447020, + "step": 10104 + }, + { + "epoch": 0.9113045046669973, + "flos": 29891790475200.0, + "grad_norm": 2.2302887371262377, + "learning_rate": 8.188417279415793e-08, + "loss": 0.7024, + "num_input_tokens_seen": 328477685, + "step": 10105 + }, + { + "epoch": 0.9113946881904677, + "flos": 21731037603360.0, + "grad_norm": 2.224708508031834, + "learning_rate": 8.171880032188117e-08, + "loss": 0.8311, + "num_input_tokens_seen": 328504310, + "step": 10106 + }, + { + "epoch": 0.9114848717139379, + "flos": 24864792781920.0, + "grad_norm": 1.5724246874606096, + "learning_rate": 8.155359152793351e-08, + "loss": 0.7184, + "num_input_tokens_seen": 328534410, + "step": 10107 + }, + { + "epoch": 0.9115750552374081, + "flos": 18523488984000.0, + "grad_norm": 2.1299697879206723, + "learning_rate": 8.138854642641147e-08, + "loss": 0.7262, + "num_input_tokens_seen": 328564075, + "step": 10108 + }, + { + "epoch": 0.9116652387608783, + "flos": 34846370008800.0, + "grad_norm": 2.118851911856381, + "learning_rate": 8.122366503139777e-08, + "loss": 0.6637, + "num_input_tokens_seen": 328595330, + "step": 10109 + }, + { + "epoch": 0.9117554222843487, + "flos": 23875104624000.0, + "grad_norm": 13.447102650837035, + "learning_rate": 8.105894735696117e-08, + "loss": 0.6783, + "num_input_tokens_seen": 328625020, + "step": 10110 + }, + { + "epoch": 0.9118456058078189, + "flos": 20675027126400.0, + "grad_norm": 2.0543524755456444, + "learning_rate": 8.089439341715576e-08, + "loss": 0.719, + "num_input_tokens_seen": 328652365, + "step": 10111 + }, + { + "epoch": 0.9119357893312892, + "flos": 18087902102880.0, + "grad_norm": 2.0742123131373758, + "learning_rate": 8.073000322602319e-08, + "loss": 0.7691, + "num_input_tokens_seen": 328680240, + "step": 10112 + }, + { + "epoch": 0.9120259728547594, + "flos": 21836951450400.0, + "grad_norm": 1.6571881498291998, + "learning_rate": 8.056577679758891e-08, + "loss": 0.7529, + "num_input_tokens_seen": 328709700, + "step": 10113 + }, + { + "epoch": 0.9121161563782297, + "flos": 34373310276000.0, + "grad_norm": 2.0120090904713033, + "learning_rate": 8.040171414586638e-08, + "loss": 0.6835, + "num_input_tokens_seen": 328736670, + "step": 10114 + }, + { + "epoch": 0.9122063399017, + "flos": 27196595758560.0, + "grad_norm": 1.6306229654155422, + "learning_rate": 8.023781528485419e-08, + "loss": 0.7009, + "num_input_tokens_seen": 328767585, + "step": 10115 + }, + { + "epoch": 0.9122965234251702, + "flos": 21366270582240.0, + "grad_norm": 2.3721997668205796, + "learning_rate": 8.00740802285369e-08, + "loss": 0.7635, + "num_input_tokens_seen": 328795480, + "step": 10116 + }, + { + "epoch": 0.9123867069486404, + "flos": 24318497154720.0, + "grad_norm": 1.7164922971344283, + "learning_rate": 7.99105089908858e-08, + "loss": 0.6028, + "num_input_tokens_seen": 328823910, + "step": 10117 + }, + { + "epoch": 0.9124768904721108, + "flos": 22932033078720.0, + "grad_norm": 1.749794130477627, + "learning_rate": 7.974710158585685e-08, + "loss": 0.7104, + "num_input_tokens_seen": 328854125, + "step": 10118 + }, + { + "epoch": 0.912567073995581, + "flos": 23626175775840.0, + "grad_norm": 1.83496464053395, + "learning_rate": 7.958385802739375e-08, + "loss": 0.7629, + "num_input_tokens_seen": 328883020, + "step": 10119 + }, + { + "epoch": 0.9126572575190512, + "flos": 22206030163680.0, + "grad_norm": 1.860804673330223, + "learning_rate": 7.942077832942452e-08, + "loss": 0.799, + "num_input_tokens_seen": 328913150, + "step": 10120 + }, + { + "epoch": 0.9127474410425215, + "flos": 68367500269920.0, + "grad_norm": 0.7874255054298216, + "learning_rate": 7.925786250586508e-08, + "loss": 0.58, + "num_input_tokens_seen": 329004735, + "step": 10121 + }, + { + "epoch": 0.9128376245659918, + "flos": 62324604004320.0, + "grad_norm": 0.6527559725408916, + "learning_rate": 7.909511057061524e-08, + "loss": 0.5795, + "num_input_tokens_seen": 329089945, + "step": 10122 + }, + { + "epoch": 0.9129278080894621, + "flos": 20420299795680.0, + "grad_norm": 2.083277688872772, + "learning_rate": 7.893252253756234e-08, + "loss": 0.7257, + "num_input_tokens_seen": 329118680, + "step": 10123 + }, + { + "epoch": 0.9130179916129323, + "flos": 23989976383200.0, + "grad_norm": 2.1780319215871553, + "learning_rate": 7.877009842057925e-08, + "loss": 0.6654, + "num_input_tokens_seen": 329148240, + "step": 10124 + }, + { + "epoch": 0.9131081751364026, + "flos": 19289975501280.0, + "grad_norm": 3.08859486182957, + "learning_rate": 7.860783823352512e-08, + "loss": 0.7655, + "num_input_tokens_seen": 329174340, + "step": 10125 + }, + { + "epoch": 0.9131983586598729, + "flos": 28576555126560.0, + "grad_norm": 1.9004568027684403, + "learning_rate": 7.844574199024445e-08, + "loss": 0.7451, + "num_input_tokens_seen": 329203925, + "step": 10126 + }, + { + "epoch": 0.9132885421833431, + "flos": 23006607084480.0, + "grad_norm": 1.8786039049676342, + "learning_rate": 7.82838097045686e-08, + "loss": 0.7794, + "num_input_tokens_seen": 329233225, + "step": 10127 + }, + { + "epoch": 0.9133787257068133, + "flos": 21257792021760.0, + "grad_norm": 2.2260272537167083, + "learning_rate": 7.812204139031454e-08, + "loss": 0.7242, + "num_input_tokens_seen": 329260020, + "step": 10128 + }, + { + "epoch": 0.9134689092302837, + "flos": 16193247137280.0, + "grad_norm": 4.960505909806538, + "learning_rate": 7.796043706128474e-08, + "loss": 0.7547, + "num_input_tokens_seen": 329285740, + "step": 10129 + }, + { + "epoch": 0.9135590927537539, + "flos": 19327002315840.0, + "grad_norm": 1.9882823292067922, + "learning_rate": 7.779899673126844e-08, + "loss": 0.7653, + "num_input_tokens_seen": 329312790, + "step": 10130 + }, + { + "epoch": 0.9136492762772241, + "flos": 33640282276320.0, + "grad_norm": 1.6909420794916894, + "learning_rate": 7.76377204140406e-08, + "loss": 0.7819, + "num_input_tokens_seen": 329342170, + "step": 10131 + }, + { + "epoch": 0.9137394598006944, + "flos": 23188433048640.0, + "grad_norm": 1.7947561625465052, + "learning_rate": 7.74766081233622e-08, + "loss": 0.7297, + "num_input_tokens_seen": 329371195, + "step": 10132 + }, + { + "epoch": 0.9138296433241647, + "flos": 24682260592320.0, + "grad_norm": 1.6760461415605887, + "learning_rate": 7.73156598729805e-08, + "loss": 0.8039, + "num_input_tokens_seen": 329398630, + "step": 10133 + }, + { + "epoch": 0.913919826847635, + "flos": 24209275199040.0, + "grad_norm": 2.1893541380040507, + "learning_rate": 7.715487567662849e-08, + "loss": 0.7899, + "num_input_tokens_seen": 329428160, + "step": 10134 + }, + { + "epoch": 0.9140100103711052, + "flos": 21148718745120.0, + "grad_norm": 1.9544060130109193, + "learning_rate": 7.69942555480243e-08, + "loss": 0.7075, + "num_input_tokens_seen": 329454895, + "step": 10135 + }, + { + "epoch": 0.9141001938945754, + "flos": 17141113581600.0, + "grad_norm": 2.920838197567128, + "learning_rate": 7.68337995008741e-08, + "loss": 0.8411, + "num_input_tokens_seen": 329482595, + "step": 10136 + }, + { + "epoch": 0.9141903774180458, + "flos": 28362460077120.0, + "grad_norm": 1.8899748133439405, + "learning_rate": 7.667350754886803e-08, + "loss": 0.596, + "num_input_tokens_seen": 329511680, + "step": 10137 + }, + { + "epoch": 0.914280560941516, + "flos": 22016695908000.0, + "grad_norm": 2.410752283028971, + "learning_rate": 7.651337970568361e-08, + "loss": 0.7299, + "num_input_tokens_seen": 329539680, + "step": 10138 + }, + { + "epoch": 0.9143707444649862, + "flos": 27670659074880.0, + "grad_norm": 2.0280258413094243, + "learning_rate": 7.635341598498368e-08, + "loss": 0.7012, + "num_input_tokens_seen": 329569840, + "step": 10139 + }, + { + "epoch": 0.9144609279884565, + "flos": 24282102226080.0, + "grad_norm": 1.818337084902988, + "learning_rate": 7.61936164004171e-08, + "loss": 0.7375, + "num_input_tokens_seen": 329597600, + "step": 10140 + }, + { + "epoch": 0.9145511115119268, + "flos": 20926260499680.0, + "grad_norm": 1.5886386362298435, + "learning_rate": 7.603398096561875e-08, + "loss": 0.7803, + "num_input_tokens_seen": 329626490, + "step": 10141 + }, + { + "epoch": 0.914641295035397, + "flos": 21804161988480.0, + "grad_norm": 2.3753947883548143, + "learning_rate": 7.587450969420994e-08, + "loss": 0.7404, + "num_input_tokens_seen": 329656060, + "step": 10142 + }, + { + "epoch": 0.9147314785588673, + "flos": 67836704432640.0, + "grad_norm": 0.6774469035270504, + "learning_rate": 7.571520259979757e-08, + "loss": 0.5604, + "num_input_tokens_seen": 329743580, + "step": 10143 + }, + { + "epoch": 0.9148216620823375, + "flos": 20309962747200.0, + "grad_norm": 2.145162631896509, + "learning_rate": 7.555605969597455e-08, + "loss": 0.5779, + "num_input_tokens_seen": 329770240, + "step": 10144 + }, + { + "epoch": 0.9149118456058078, + "flos": 30033876232320.0, + "grad_norm": 2.48415253800814, + "learning_rate": 7.539708099631959e-08, + "loss": 0.8216, + "num_input_tokens_seen": 329802220, + "step": 10145 + }, + { + "epoch": 0.9150020291292781, + "flos": 26575763295360.0, + "grad_norm": 2.4476568422338096, + "learning_rate": 7.52382665143978e-08, + "loss": 0.7447, + "num_input_tokens_seen": 329832315, + "step": 10146 + }, + { + "epoch": 0.9150922126527483, + "flos": 21985541915520.0, + "grad_norm": 2.004830130708541, + "learning_rate": 7.507961626376014e-08, + "loss": 0.7927, + "num_input_tokens_seen": 329860185, + "step": 10147 + }, + { + "epoch": 0.9151823961762187, + "flos": 28690125944160.0, + "grad_norm": 2.213315729316349, + "learning_rate": 7.492113025794378e-08, + "loss": 0.7755, + "num_input_tokens_seen": 329890335, + "step": 10148 + }, + { + "epoch": 0.9152725796996889, + "flos": 19873483791840.0, + "grad_norm": 2.310860645597298, + "learning_rate": 7.476280851047101e-08, + "loss": 0.684, + "num_input_tokens_seen": 329919535, + "step": 10149 + }, + { + "epoch": 0.9153627632231591, + "flos": 24281247321600.0, + "grad_norm": 1.5018902630294322, + "learning_rate": 7.460465103485125e-08, + "loss": 0.7613, + "num_input_tokens_seen": 329949160, + "step": 10150 + }, + { + "epoch": 0.9154529467466294, + "flos": 23152744345440.0, + "grad_norm": 2.0963718453050837, + "learning_rate": 7.444665784457948e-08, + "loss": 0.7268, + "num_input_tokens_seen": 329974995, + "step": 10151 + }, + { + "epoch": 0.9155431302700997, + "flos": 24427124394240.0, + "grad_norm": 16.45599322844919, + "learning_rate": 7.42888289531356e-08, + "loss": 0.6996, + "num_input_tokens_seen": 330002095, + "step": 10152 + }, + { + "epoch": 0.9156333137935699, + "flos": 33200420872800.0, + "grad_norm": 1.8304860445150706, + "learning_rate": 7.41311643739877e-08, + "loss": 0.661, + "num_input_tokens_seen": 330030535, + "step": 10153 + }, + { + "epoch": 0.9157234973170402, + "flos": 23152112459520.0, + "grad_norm": 1.7153465559225294, + "learning_rate": 7.39736641205877e-08, + "loss": 0.7758, + "num_input_tokens_seen": 330058210, + "step": 10154 + }, + { + "epoch": 0.9158136808405104, + "flos": 17797300220160.0, + "grad_norm": 1.8413855051806982, + "learning_rate": 7.381632820637462e-08, + "loss": 0.7644, + "num_input_tokens_seen": 330085515, + "step": 10155 + }, + { + "epoch": 0.9159038643639807, + "flos": 20746776230400.0, + "grad_norm": 2.0080800156551732, + "learning_rate": 7.365915664477352e-08, + "loss": 0.7255, + "num_input_tokens_seen": 330114920, + "step": 10156 + }, + { + "epoch": 0.915994047887451, + "flos": 30906053578080.0, + "grad_norm": 1.955931516322174, + "learning_rate": 7.350214944919474e-08, + "loss": 0.7424, + "num_input_tokens_seen": 330143610, + "step": 10157 + }, + { + "epoch": 0.9160842314109212, + "flos": 26972799401760.0, + "grad_norm": 11.546234071775714, + "learning_rate": 7.334530663303539e-08, + "loss": 0.7675, + "num_input_tokens_seen": 330171910, + "step": 10158 + }, + { + "epoch": 0.9161744149343914, + "flos": 24354929253120.0, + "grad_norm": 2.022478855998605, + "learning_rate": 7.318862820967742e-08, + "loss": 0.7438, + "num_input_tokens_seen": 330200250, + "step": 10159 + }, + { + "epoch": 0.9162645984578618, + "flos": 21513039729120.0, + "grad_norm": 2.800124853620248, + "learning_rate": 7.303211419249056e-08, + "loss": 0.7722, + "num_input_tokens_seen": 330227980, + "step": 10160 + }, + { + "epoch": 0.916354781981332, + "flos": 30441580059840.0, + "grad_norm": 1.8516800597651013, + "learning_rate": 7.287576459482858e-08, + "loss": 0.7551, + "num_input_tokens_seen": 330261820, + "step": 10161 + }, + { + "epoch": 0.9164449655048023, + "flos": 21110651177280.0, + "grad_norm": 4.587324566306132, + "learning_rate": 7.271957943003259e-08, + "loss": 0.7786, + "num_input_tokens_seen": 330290875, + "step": 10162 + }, + { + "epoch": 0.9165351490282725, + "flos": 17833249111680.0, + "grad_norm": 2.0257749275670838, + "learning_rate": 7.256355871142883e-08, + "loss": 0.7362, + "num_input_tokens_seen": 330318965, + "step": 10163 + }, + { + "epoch": 0.9166253325517428, + "flos": 25957681394400.0, + "grad_norm": 1.8818130689734962, + "learning_rate": 7.240770245233019e-08, + "loss": 0.7887, + "num_input_tokens_seen": 330346930, + "step": 10164 + }, + { + "epoch": 0.9167155160752131, + "flos": 26754243981120.0, + "grad_norm": 2.1579945459164342, + "learning_rate": 7.225201066603492e-08, + "loss": 0.655, + "num_input_tokens_seen": 330375945, + "step": 10165 + }, + { + "epoch": 0.9168056995986833, + "flos": 24752114038560.0, + "grad_norm": 3.0176433864807803, + "learning_rate": 7.209648336582774e-08, + "loss": 0.8482, + "num_input_tokens_seen": 330405570, + "step": 10166 + }, + { + "epoch": 0.9168958831221535, + "flos": 23407694694720.0, + "grad_norm": 1.956330438489663, + "learning_rate": 7.19411205649787e-08, + "loss": 0.7264, + "num_input_tokens_seen": 330435765, + "step": 10167 + }, + { + "epoch": 0.9169860666456239, + "flos": 23219549871360.0, + "grad_norm": 2.018539214570898, + "learning_rate": 7.178592227674474e-08, + "loss": 0.7605, + "num_input_tokens_seen": 330464760, + "step": 10168 + }, + { + "epoch": 0.9170762501690941, + "flos": 21476719140000.0, + "grad_norm": 1.9880809870443819, + "learning_rate": 7.163088851436771e-08, + "loss": 0.6565, + "num_input_tokens_seen": 330493340, + "step": 10169 + }, + { + "epoch": 0.9171664336925643, + "flos": 24391547200320.0, + "grad_norm": 1.8427256308592472, + "learning_rate": 7.147601929107639e-08, + "loss": 0.7068, + "num_input_tokens_seen": 330524175, + "step": 10170 + }, + { + "epoch": 0.9172566172160346, + "flos": 21731111942880.0, + "grad_norm": 1.6124646796564719, + "learning_rate": 7.132131462008461e-08, + "loss": 0.7831, + "num_input_tokens_seen": 330552975, + "step": 10171 + }, + { + "epoch": 0.9173468007395049, + "flos": 22896678903360.0, + "grad_norm": 2.0729271842140204, + "learning_rate": 7.116677451459297e-08, + "loss": 0.7455, + "num_input_tokens_seen": 330581860, + "step": 10172 + }, + { + "epoch": 0.9174369842629752, + "flos": 25045429313760.0, + "grad_norm": 1.356202380491417, + "learning_rate": 7.101239898778799e-08, + "loss": 0.7629, + "num_input_tokens_seen": 330614365, + "step": 10173 + }, + { + "epoch": 0.9175271677864454, + "flos": 27997284188640.0, + "grad_norm": 2.129461989354406, + "learning_rate": 7.085818805284094e-08, + "loss": 0.7747, + "num_input_tokens_seen": 330644720, + "step": 10174 + }, + { + "epoch": 0.9176173513099157, + "flos": 18307200918720.0, + "grad_norm": 6.132663836772996, + "learning_rate": 7.070414172291083e-08, + "loss": 0.8175, + "num_input_tokens_seen": 330669755, + "step": 10175 + }, + { + "epoch": 0.917707534833386, + "flos": 63804238433280.0, + "grad_norm": 1.51133033040414, + "learning_rate": 7.055026001114095e-08, + "loss": 0.6801, + "num_input_tokens_seen": 330704540, + "step": 10176 + }, + { + "epoch": 0.9177977183568562, + "flos": 57816433485600.0, + "grad_norm": 0.6508064887933156, + "learning_rate": 7.039654293066211e-08, + "loss": 0.5647, + "num_input_tokens_seen": 330783235, + "step": 10177 + }, + { + "epoch": 0.9178879018803264, + "flos": 21220988225760.0, + "grad_norm": 2.6047279871933013, + "learning_rate": 7.024299049459003e-08, + "loss": 0.6578, + "num_input_tokens_seen": 330811985, + "step": 10178 + }, + { + "epoch": 0.9179780854037968, + "flos": 24354743404320.0, + "grad_norm": 3.3636875593896027, + "learning_rate": 7.008960271602627e-08, + "loss": 0.6494, + "num_input_tokens_seen": 330838610, + "step": 10179 + }, + { + "epoch": 0.918068268927267, + "flos": 29417652819360.0, + "grad_norm": 1.8499738098558882, + "learning_rate": 6.993637960805921e-08, + "loss": 0.7021, + "num_input_tokens_seen": 330869155, + "step": 10180 + }, + { + "epoch": 0.9181584524507372, + "flos": 20959124301120.0, + "grad_norm": 1.6332458213146597, + "learning_rate": 6.97833211837624e-08, + "loss": 0.7657, + "num_input_tokens_seen": 330896495, + "step": 10181 + }, + { + "epoch": 0.9182486359742075, + "flos": 17322642187680.0, + "grad_norm": 2.018532009578706, + "learning_rate": 6.963042745619562e-08, + "loss": 0.6877, + "num_input_tokens_seen": 330921930, + "step": 10182 + }, + { + "epoch": 0.9183388194976778, + "flos": 16267003408320.0, + "grad_norm": 1.77417245835553, + "learning_rate": 6.947769843840511e-08, + "loss": 0.7071, + "num_input_tokens_seen": 330949540, + "step": 10183 + }, + { + "epoch": 0.918429003021148, + "flos": 26791307965440.0, + "grad_norm": 3.5022248962099525, + "learning_rate": 6.9325134143422e-08, + "loss": 0.7692, + "num_input_tokens_seen": 330977490, + "step": 10184 + }, + { + "epoch": 0.9185191865446183, + "flos": 47489348906880.0, + "grad_norm": 4.072794854699509, + "learning_rate": 6.917273458426387e-08, + "loss": 0.7177, + "num_input_tokens_seen": 331009945, + "step": 10185 + }, + { + "epoch": 0.9186093700680885, + "flos": 62861086814880.0, + "grad_norm": 0.6838085599994458, + "learning_rate": 6.902049977393476e-08, + "loss": 0.5529, + "num_input_tokens_seen": 331100860, + "step": 10186 + }, + { + "epoch": 0.9186995535915589, + "flos": 28872509454720.0, + "grad_norm": 1.7107141078922714, + "learning_rate": 6.886842972542362e-08, + "loss": 0.745, + "num_input_tokens_seen": 331131430, + "step": 10187 + }, + { + "epoch": 0.9187897371150291, + "flos": 37360668005280.0, + "grad_norm": 1.7546288715734812, + "learning_rate": 6.871652445170672e-08, + "loss": 0.6284, + "num_input_tokens_seen": 331162910, + "step": 10188 + }, + { + "epoch": 0.9188799206384993, + "flos": 30401839852800.0, + "grad_norm": 2.8864752598384036, + "learning_rate": 6.856478396574416e-08, + "loss": 0.821, + "num_input_tokens_seen": 331193315, + "step": 10189 + }, + { + "epoch": 0.9189701041619696, + "flos": 20857373467200.0, + "grad_norm": 1.9374193974618343, + "learning_rate": 6.841320828048491e-08, + "loss": 0.7999, + "num_input_tokens_seen": 331220390, + "step": 10190 + }, + { + "epoch": 0.9190602876854399, + "flos": 21949444344960.0, + "grad_norm": 1.659386956985412, + "learning_rate": 6.826179740886062e-08, + "loss": 0.7396, + "num_input_tokens_seen": 331248555, + "step": 10191 + }, + { + "epoch": 0.9191504712089101, + "flos": 18958741337280.0, + "grad_norm": 2.5681243208742486, + "learning_rate": 6.811055136379184e-08, + "loss": 0.8002, + "num_input_tokens_seen": 331275835, + "step": 10192 + }, + { + "epoch": 0.9192406547323804, + "flos": 22350717804000.0, + "grad_norm": 1.9669753053372685, + "learning_rate": 6.79594701581827e-08, + "loss": 0.7174, + "num_input_tokens_seen": 331303125, + "step": 10193 + }, + { + "epoch": 0.9193308382558506, + "flos": 19362876867840.0, + "grad_norm": 2.0488381326465546, + "learning_rate": 6.780855380492511e-08, + "loss": 0.7462, + "num_input_tokens_seen": 331330375, + "step": 10194 + }, + { + "epoch": 0.919421021779321, + "flos": 24239648626560.0, + "grad_norm": 1.8981861842961258, + "learning_rate": 6.765780231689544e-08, + "loss": 0.7548, + "num_input_tokens_seen": 331360150, + "step": 10195 + }, + { + "epoch": 0.9195112053027912, + "flos": 14043864840960.0, + "grad_norm": 2.1900847287096, + "learning_rate": 6.750721570695695e-08, + "loss": 0.796, + "num_input_tokens_seen": 331383840, + "step": 10196 + }, + { + "epoch": 0.9196013888262614, + "flos": 25190971858560.0, + "grad_norm": 2.7893344949546033, + "learning_rate": 6.735679398795868e-08, + "loss": 0.7127, + "num_input_tokens_seen": 331408060, + "step": 10197 + }, + { + "epoch": 0.9196915723497318, + "flos": 27633149053440.0, + "grad_norm": 2.5218720200909828, + "learning_rate": 6.720653717273506e-08, + "loss": 0.6743, + "num_input_tokens_seen": 331438255, + "step": 10198 + }, + { + "epoch": 0.919781755873202, + "flos": 21947474347680.0, + "grad_norm": 2.8284137220166086, + "learning_rate": 6.705644527410714e-08, + "loss": 0.7401, + "num_input_tokens_seen": 331466495, + "step": 10199 + }, + { + "epoch": 0.9198719393966722, + "flos": 15792679903680.0, + "grad_norm": 2.301079256353104, + "learning_rate": 6.690651830488136e-08, + "loss": 0.7058, + "num_input_tokens_seen": 331492985, + "step": 10200 + }, + { + "epoch": 0.9199621229201425, + "flos": 12259063716960.0, + "grad_norm": 3.6236667154247972, + "learning_rate": 6.675675627785037e-08, + "loss": 0.736, + "num_input_tokens_seen": 331519705, + "step": 10201 + }, + { + "epoch": 0.9200523064436128, + "flos": 22166364296160.0, + "grad_norm": 1.952068073398977, + "learning_rate": 6.660715920579263e-08, + "loss": 0.7574, + "num_input_tokens_seen": 331548265, + "step": 10202 + }, + { + "epoch": 0.920142489967083, + "flos": 68083328755680.0, + "grad_norm": 0.5902961612780808, + "learning_rate": 6.645772710147279e-08, + "loss": 0.5769, + "num_input_tokens_seen": 331644795, + "step": 10203 + }, + { + "epoch": 0.9202326734905533, + "flos": 30984902106240.0, + "grad_norm": 1.4387094137000327, + "learning_rate": 6.630845997764112e-08, + "loss": 0.7106, + "num_input_tokens_seen": 331679450, + "step": 10204 + }, + { + "epoch": 0.9203228570140235, + "flos": 24974089077120.0, + "grad_norm": 1.6066608690154598, + "learning_rate": 6.615935784703409e-08, + "loss": 0.6791, + "num_input_tokens_seen": 331710815, + "step": 10205 + }, + { + "epoch": 0.9204130405374938, + "flos": 26501969854560.0, + "grad_norm": 2.2104132472187694, + "learning_rate": 6.601042072237328e-08, + "loss": 0.6647, + "num_input_tokens_seen": 331740690, + "step": 10206 + }, + { + "epoch": 0.9205032240609641, + "flos": 24645642645120.0, + "grad_norm": 2.3106691484513173, + "learning_rate": 6.586164861636767e-08, + "loss": 0.7711, + "num_input_tokens_seen": 331770630, + "step": 10207 + }, + { + "epoch": 0.9205934075844343, + "flos": 20748374530080.0, + "grad_norm": 2.4266366724983204, + "learning_rate": 6.571304154171065e-08, + "loss": 0.6536, + "num_input_tokens_seen": 331798735, + "step": 10208 + }, + { + "epoch": 0.9206835911079045, + "flos": 26683758648960.0, + "grad_norm": 1.9875327389497806, + "learning_rate": 6.556459951108273e-08, + "loss": 0.7648, + "num_input_tokens_seen": 331828515, + "step": 10209 + }, + { + "epoch": 0.9207737746313749, + "flos": 19946496667680.0, + "grad_norm": 2.8845751989450963, + "learning_rate": 6.541632253714957e-08, + "loss": 0.7157, + "num_input_tokens_seen": 331854755, + "step": 10210 + }, + { + "epoch": 0.9208639581548451, + "flos": 22168929009600.0, + "grad_norm": 2.1680030613838905, + "learning_rate": 6.526821063256261e-08, + "loss": 0.6673, + "num_input_tokens_seen": 331883840, + "step": 10211 + }, + { + "epoch": 0.9209541416783154, + "flos": 19947128553600.0, + "grad_norm": 2.101390391963677, + "learning_rate": 6.512026380996016e-08, + "loss": 0.6473, + "num_input_tokens_seen": 331911010, + "step": 10212 + }, + { + "epoch": 0.9210443252017856, + "flos": 35357088442080.0, + "grad_norm": 1.8873574891511302, + "learning_rate": 6.49724820819657e-08, + "loss": 0.7084, + "num_input_tokens_seen": 331941440, + "step": 10213 + }, + { + "epoch": 0.9211345087252559, + "flos": 22278113795520.0, + "grad_norm": 1.9219841654717706, + "learning_rate": 6.48248654611887e-08, + "loss": 0.7719, + "num_input_tokens_seen": 331970770, + "step": 10214 + }, + { + "epoch": 0.9212246922487262, + "flos": 28616518352160.0, + "grad_norm": 1.5137247974619337, + "learning_rate": 6.467741396022419e-08, + "loss": 0.7859, + "num_input_tokens_seen": 332003690, + "step": 10215 + }, + { + "epoch": 0.9213148757721964, + "flos": 24828658041600.0, + "grad_norm": 1.8983261640891702, + "learning_rate": 6.453012759165455e-08, + "loss": 0.7458, + "num_input_tokens_seen": 332032830, + "step": 10216 + }, + { + "epoch": 0.9214050592956666, + "flos": 25522094513280.0, + "grad_norm": 1.6656231792504412, + "learning_rate": 6.438300636804639e-08, + "loss": 0.7057, + "num_input_tokens_seen": 332062120, + "step": 10217 + }, + { + "epoch": 0.921495242819137, + "flos": 24974906811840.0, + "grad_norm": 1.7374480918979407, + "learning_rate": 6.423605030195278e-08, + "loss": 0.7718, + "num_input_tokens_seen": 332092240, + "step": 10218 + }, + { + "epoch": 0.9215854263426072, + "flos": 17614061805120.0, + "grad_norm": 2.055947037382048, + "learning_rate": 6.408925940591304e-08, + "loss": 0.8105, + "num_input_tokens_seen": 332118770, + "step": 10219 + }, + { + "epoch": 0.9216756098660774, + "flos": 36845563540320.0, + "grad_norm": 2.040883270790835, + "learning_rate": 6.394263369245222e-08, + "loss": 0.6741, + "num_input_tokens_seen": 332152500, + "step": 10220 + }, + { + "epoch": 0.9217657933895478, + "flos": 74623419395520.0, + "grad_norm": 2.0063297719337063, + "learning_rate": 6.379617317408126e-08, + "loss": 0.7423, + "num_input_tokens_seen": 332186715, + "step": 10221 + }, + { + "epoch": 0.921855976913018, + "flos": 26103112429920.0, + "grad_norm": 2.3921046446361607, + "learning_rate": 6.364987786329723e-08, + "loss": 0.7349, + "num_input_tokens_seen": 332216115, + "step": 10222 + }, + { + "epoch": 0.9219461604364882, + "flos": 26869152910080.0, + "grad_norm": 2.4305284032364667, + "learning_rate": 6.350374777258193e-08, + "loss": 0.7324, + "num_input_tokens_seen": 332246920, + "step": 10223 + }, + { + "epoch": 0.9220363439599585, + "flos": 22569273224640.0, + "grad_norm": 1.7126495992097923, + "learning_rate": 6.335778291440519e-08, + "loss": 0.7181, + "num_input_tokens_seen": 332276130, + "step": 10224 + }, + { + "epoch": 0.9221265274834288, + "flos": 25372834992480.0, + "grad_norm": 1.8849135582933663, + "learning_rate": 6.321198330122057e-08, + "loss": 0.6433, + "num_input_tokens_seen": 332305935, + "step": 10225 + }, + { + "epoch": 0.9222167110068991, + "flos": 28721763143520.0, + "grad_norm": 1.7890289836559279, + "learning_rate": 6.306634894546902e-08, + "loss": 0.6757, + "num_input_tokens_seen": 332338425, + "step": 10226 + }, + { + "epoch": 0.9223068945303693, + "flos": 23880419899680.0, + "grad_norm": 1.589154518186359, + "learning_rate": 6.292087985957661e-08, + "loss": 0.6794, + "num_input_tokens_seen": 332369190, + "step": 10227 + }, + { + "epoch": 0.9223970780538395, + "flos": 20128694329440.0, + "grad_norm": 1.996121281323865, + "learning_rate": 6.277557605595585e-08, + "loss": 0.7769, + "num_input_tokens_seen": 332396245, + "step": 10228 + }, + { + "epoch": 0.9224872615773099, + "flos": 23735323392000.0, + "grad_norm": 1.5776355097585124, + "learning_rate": 6.263043754700481e-08, + "loss": 0.6605, + "num_input_tokens_seen": 332426015, + "step": 10229 + }, + { + "epoch": 0.9225774451007801, + "flos": 18088496819040.0, + "grad_norm": 2.229018934351722, + "learning_rate": 6.248546434510671e-08, + "loss": 0.7391, + "num_input_tokens_seen": 332454265, + "step": 10230 + }, + { + "epoch": 0.9226676286242503, + "flos": 35830928739840.0, + "grad_norm": 2.077579742260022, + "learning_rate": 6.234065646263298e-08, + "loss": 0.6698, + "num_input_tokens_seen": 332486115, + "step": 10231 + }, + { + "epoch": 0.9227578121477206, + "flos": 20273976685920.0, + "grad_norm": 2.0211348816616335, + "learning_rate": 6.219601391193796e-08, + "loss": 0.6501, + "num_input_tokens_seen": 332514685, + "step": 10232 + }, + { + "epoch": 0.9228479956711909, + "flos": 44026738428960.0, + "grad_norm": 1.6972104185408086, + "learning_rate": 6.205153670536423e-08, + "loss": 0.6652, + "num_input_tokens_seen": 332550860, + "step": 10233 + }, + { + "epoch": 0.9229381791946611, + "flos": 27267341279040.0, + "grad_norm": 1.7663595646421437, + "learning_rate": 6.190722485523902e-08, + "loss": 0.775, + "num_input_tokens_seen": 332583135, + "step": 10234 + }, + { + "epoch": 0.9230283627181314, + "flos": 23116572435360.0, + "grad_norm": 1.8788509241789477, + "learning_rate": 6.176307837387607e-08, + "loss": 0.716, + "num_input_tokens_seen": 332610275, + "step": 10235 + }, + { + "epoch": 0.9231185462416016, + "flos": 27557459954880.0, + "grad_norm": 1.509881847887469, + "learning_rate": 6.16190972735744e-08, + "loss": 0.6874, + "num_input_tokens_seen": 332639745, + "step": 10236 + }, + { + "epoch": 0.923208729765072, + "flos": 34409853883680.0, + "grad_norm": 2.4592737780514344, + "learning_rate": 6.147528156661974e-08, + "loss": 0.7636, + "num_input_tokens_seen": 332670200, + "step": 10237 + }, + { + "epoch": 0.9232989132885422, + "flos": 25009926459360.0, + "grad_norm": 2.003118944084547, + "learning_rate": 6.133163126528273e-08, + "loss": 0.6618, + "num_input_tokens_seen": 332700460, + "step": 10238 + }, + { + "epoch": 0.9233890968120124, + "flos": 20268103863840.0, + "grad_norm": 1.9789649325170167, + "learning_rate": 6.11881463818209e-08, + "loss": 0.7471, + "num_input_tokens_seen": 332727585, + "step": 10239 + }, + { + "epoch": 0.9234792803354827, + "flos": 21105930617760.0, + "grad_norm": 1.708210971761041, + "learning_rate": 6.104482692847668e-08, + "loss": 0.7249, + "num_input_tokens_seen": 332754645, + "step": 10240 + }, + { + "epoch": 0.923569463858953, + "flos": 22533101314560.0, + "grad_norm": 1.7783854020685181, + "learning_rate": 6.090167291747917e-08, + "loss": 0.7063, + "num_input_tokens_seen": 332784690, + "step": 10241 + }, + { + "epoch": 0.9236596473824232, + "flos": 24609470735040.0, + "grad_norm": 2.087214047154171, + "learning_rate": 6.075868436104303e-08, + "loss": 0.6392, + "num_input_tokens_seen": 332814090, + "step": 10242 + }, + { + "epoch": 0.9237498309058935, + "flos": 27305074319040.0, + "grad_norm": 2.1051442987995648, + "learning_rate": 6.061586127136875e-08, + "loss": 0.7235, + "num_input_tokens_seen": 332845605, + "step": 10243 + }, + { + "epoch": 0.9238400144293638, + "flos": 23696846956800.0, + "grad_norm": 1.7603443815600581, + "learning_rate": 6.047320366064324e-08, + "loss": 0.7048, + "num_input_tokens_seen": 332876895, + "step": 10244 + }, + { + "epoch": 0.923930197952834, + "flos": 20488554942240.0, + "grad_norm": 1.8473711176867016, + "learning_rate": 6.033071154103786e-08, + "loss": 0.7456, + "num_input_tokens_seen": 332905860, + "step": 10245 + }, + { + "epoch": 0.9240203814763043, + "flos": 21695720597760.0, + "grad_norm": 2.2135653027039246, + "learning_rate": 6.018838492471178e-08, + "loss": 0.7594, + "num_input_tokens_seen": 332933560, + "step": 10246 + }, + { + "epoch": 0.9241105649997745, + "flos": 70333532641920.0, + "grad_norm": 0.6206978069675374, + "learning_rate": 6.00462238238082e-08, + "loss": 0.5791, + "num_input_tokens_seen": 333030060, + "step": 10247 + }, + { + "epoch": 0.9242007485232449, + "flos": 25552802468640.0, + "grad_norm": 2.520560499406958, + "learning_rate": 5.990422825045827e-08, + "loss": 0.6992, + "num_input_tokens_seen": 333058255, + "step": 10248 + }, + { + "epoch": 0.9242909320467151, + "flos": 20381897700000.0, + "grad_norm": 3.0085689237566324, + "learning_rate": 5.976239821677675e-08, + "loss": 0.715, + "num_input_tokens_seen": 333086645, + "step": 10249 + }, + { + "epoch": 0.9243811155701853, + "flos": 26904990292320.0, + "grad_norm": 3.3294725152671605, + "learning_rate": 5.962073373486598e-08, + "loss": 0.6806, + "num_input_tokens_seen": 333117380, + "step": 10250 + }, + { + "epoch": 0.9244712990936556, + "flos": 34919866091520.0, + "grad_norm": 1.8455624481359287, + "learning_rate": 5.947923481681316e-08, + "loss": 0.7004, + "num_input_tokens_seen": 333149170, + "step": 10251 + }, + { + "epoch": 0.9245614826171259, + "flos": 26465314737600.0, + "grad_norm": 1.8355914499959456, + "learning_rate": 5.933790147469198e-08, + "loss": 0.7091, + "num_input_tokens_seen": 333177955, + "step": 10252 + }, + { + "epoch": 0.9246516661405961, + "flos": 22929914402400.0, + "grad_norm": 2.175019218096837, + "learning_rate": 5.9196733720561665e-08, + "loss": 0.6887, + "num_input_tokens_seen": 333204710, + "step": 10253 + }, + { + "epoch": 0.9247418496640664, + "flos": 19977799339200.0, + "grad_norm": 2.032014086869739, + "learning_rate": 5.905573156646793e-08, + "loss": 0.8208, + "num_input_tokens_seen": 333230990, + "step": 10254 + }, + { + "epoch": 0.9248320331875366, + "flos": 23370556370880.0, + "grad_norm": 2.428774548699538, + "learning_rate": 5.8914895024441134e-08, + "loss": 0.8138, + "num_input_tokens_seen": 333257450, + "step": 10255 + }, + { + "epoch": 0.9249222167110069, + "flos": 18810782758080.0, + "grad_norm": 2.3666532310140287, + "learning_rate": 5.877422410649857e-08, + "loss": 0.7725, + "num_input_tokens_seen": 333280985, + "step": 10256 + }, + { + "epoch": 0.9250124002344772, + "flos": 17395878082080.0, + "grad_norm": 2.4670868111974444, + "learning_rate": 5.863371882464285e-08, + "loss": 0.7535, + "num_input_tokens_seen": 333307550, + "step": 10257 + }, + { + "epoch": 0.9251025837579474, + "flos": 19472916558240.0, + "grad_norm": 1.949400318075514, + "learning_rate": 5.849337919086283e-08, + "loss": 0.7221, + "num_input_tokens_seen": 333333880, + "step": 10258 + }, + { + "epoch": 0.9251927672814176, + "flos": 26613124637760.0, + "grad_norm": 2.1006045515795444, + "learning_rate": 5.835320521713316e-08, + "loss": 0.7158, + "num_input_tokens_seen": 333362885, + "step": 10259 + }, + { + "epoch": 0.925282950804888, + "flos": 21475789896000.0, + "grad_norm": 1.5849688348203206, + "learning_rate": 5.8213196915414264e-08, + "loss": 0.7214, + "num_input_tokens_seen": 333391290, + "step": 10260 + }, + { + "epoch": 0.9253731343283582, + "flos": 23260999887360.0, + "grad_norm": 3.264255948540648, + "learning_rate": 5.807335429765237e-08, + "loss": 0.7941, + "num_input_tokens_seen": 333418195, + "step": 10261 + }, + { + "epoch": 0.9254633178518284, + "flos": 22164840336000.0, + "grad_norm": 1.921900865334236, + "learning_rate": 5.7933677375779034e-08, + "loss": 0.7864, + "num_input_tokens_seen": 333447745, + "step": 10262 + }, + { + "epoch": 0.9255535013752987, + "flos": 23331039182400.0, + "grad_norm": 2.5316801294400353, + "learning_rate": 5.77941661617134e-08, + "loss": 0.7877, + "num_input_tokens_seen": 333478495, + "step": 10263 + }, + { + "epoch": 0.925643684898769, + "flos": 26174861533920.0, + "grad_norm": 2.080269531613083, + "learning_rate": 5.765482066735816e-08, + "loss": 0.6706, + "num_input_tokens_seen": 333507235, + "step": 10264 + }, + { + "epoch": 0.9257338684222393, + "flos": 18051470004480.0, + "grad_norm": 2.1332032680598743, + "learning_rate": 5.7515640904604256e-08, + "loss": 0.7426, + "num_input_tokens_seen": 333534270, + "step": 10265 + }, + { + "epoch": 0.9258240519457095, + "flos": 23951537117760.0, + "grad_norm": 1.9237155939486916, + "learning_rate": 5.7376626885326187e-08, + "loss": 0.7218, + "num_input_tokens_seen": 333563960, + "step": 10266 + }, + { + "epoch": 0.9259142354691798, + "flos": 21767841399360.0, + "grad_norm": 1.8200344151607093, + "learning_rate": 5.723777862138601e-08, + "loss": 0.8124, + "num_input_tokens_seen": 333592700, + "step": 10267 + }, + { + "epoch": 0.9260044189926501, + "flos": 16558051328160.0, + "grad_norm": 1.820344916452073, + "learning_rate": 5.7099096124630705e-08, + "loss": 0.8181, + "num_input_tokens_seen": 333618385, + "step": 10268 + }, + { + "epoch": 0.9260946025161203, + "flos": 24936987923040.0, + "grad_norm": 1.3917133915456366, + "learning_rate": 5.696057940689347e-08, + "loss": 0.7474, + "num_input_tokens_seen": 333649165, + "step": 10269 + }, + { + "epoch": 0.9261847860395905, + "flos": 22532692447200.0, + "grad_norm": 2.222020170509952, + "learning_rate": 5.6822228479993736e-08, + "loss": 0.7571, + "num_input_tokens_seen": 333678055, + "step": 10270 + }, + { + "epoch": 0.9262749695630609, + "flos": 20672388073440.0, + "grad_norm": 2.0303249893872946, + "learning_rate": 5.668404335573584e-08, + "loss": 0.7806, + "num_input_tokens_seen": 333702030, + "step": 10271 + }, + { + "epoch": 0.9263651530865311, + "flos": 27013952059680.0, + "grad_norm": 1.7758549417762017, + "learning_rate": 5.654602404591058e-08, + "loss": 0.683, + "num_input_tokens_seen": 333732655, + "step": 10272 + }, + { + "epoch": 0.9264553366100013, + "flos": 29673904110240.0, + "grad_norm": 1.8588953120697091, + "learning_rate": 5.640817056229474e-08, + "loss": 0.7962, + "num_input_tokens_seen": 333762340, + "step": 10273 + }, + { + "epoch": 0.9265455201334716, + "flos": 22387187072160.0, + "grad_norm": 1.73100434361028, + "learning_rate": 5.6270482916650706e-08, + "loss": 0.7793, + "num_input_tokens_seen": 333791420, + "step": 10274 + }, + { + "epoch": 0.9266357036569419, + "flos": 27925535084640.0, + "grad_norm": 2.0426423083704766, + "learning_rate": 5.613296112072663e-08, + "loss": 0.8132, + "num_input_tokens_seen": 333822790, + "step": 10275 + }, + { + "epoch": 0.9267258871804122, + "flos": 20710641490080.0, + "grad_norm": 2.4511709023279944, + "learning_rate": 5.59956051862569e-08, + "loss": 0.7423, + "num_input_tokens_seen": 333848125, + "step": 10276 + }, + { + "epoch": 0.9268160707038824, + "flos": 22569682092000.0, + "grad_norm": 1.8062830885658034, + "learning_rate": 5.585841512496081e-08, + "loss": 0.8033, + "num_input_tokens_seen": 333878090, + "step": 10277 + }, + { + "epoch": 0.9269062542273526, + "flos": 26175530589600.0, + "grad_norm": 2.0707200135119983, + "learning_rate": 5.5721390948545e-08, + "loss": 0.8475, + "num_input_tokens_seen": 333904770, + "step": 10278 + }, + { + "epoch": 0.926996437750823, + "flos": 25302572678880.0, + "grad_norm": 2.5353832642845076, + "learning_rate": 5.558453266870056e-08, + "loss": 0.784, + "num_input_tokens_seen": 333935095, + "step": 10279 + }, + { + "epoch": 0.9270866212742932, + "flos": 28799868276480.0, + "grad_norm": 1.8494066008420207, + "learning_rate": 5.544784029710525e-08, + "loss": 0.6793, + "num_input_tokens_seen": 333964960, + "step": 10280 + }, + { + "epoch": 0.9271768047977634, + "flos": 16411988406720.0, + "grad_norm": 2.1620734262622343, + "learning_rate": 5.531131384542242e-08, + "loss": 0.8106, + "num_input_tokens_seen": 333991190, + "step": 10281 + }, + { + "epoch": 0.9272669883212337, + "flos": 25303130225280.0, + "grad_norm": 2.589273013003913, + "learning_rate": 5.51749533253012e-08, + "loss": 0.6889, + "num_input_tokens_seen": 334021830, + "step": 10282 + }, + { + "epoch": 0.927357171844704, + "flos": 35644419385920.0, + "grad_norm": 1.706427705334904, + "learning_rate": 5.503875874837649e-08, + "loss": 0.6822, + "num_input_tokens_seen": 334050575, + "step": 10283 + }, + { + "epoch": 0.9274473553681742, + "flos": 25010112308160.0, + "grad_norm": 2.316437745161625, + "learning_rate": 5.4902730126269225e-08, + "loss": 0.706, + "num_input_tokens_seen": 334079210, + "step": 10284 + }, + { + "epoch": 0.9275375388916445, + "flos": 26832014586240.0, + "grad_norm": 1.8928227537172482, + "learning_rate": 5.476686747058656e-08, + "loss": 0.7685, + "num_input_tokens_seen": 334108560, + "step": 10285 + }, + { + "epoch": 0.9276277224151147, + "flos": 22715745013440.0, + "grad_norm": 2.1897213148007943, + "learning_rate": 5.4631170792920124e-08, + "loss": 0.7139, + "num_input_tokens_seen": 334136195, + "step": 10286 + }, + { + "epoch": 0.927717905938585, + "flos": 20893508207520.0, + "grad_norm": 2.1389348395339716, + "learning_rate": 5.449564010484953e-08, + "loss": 0.7201, + "num_input_tokens_seen": 334162385, + "step": 10287 + }, + { + "epoch": 0.9278080894620553, + "flos": 24755533656480.0, + "grad_norm": 2.1474580508876686, + "learning_rate": 5.436027541793775e-08, + "loss": 0.7626, + "num_input_tokens_seen": 334190755, + "step": 10288 + }, + { + "epoch": 0.9278982729855255, + "flos": 25552616619840.0, + "grad_norm": 1.7337012142036543, + "learning_rate": 5.4225076743735554e-08, + "loss": 0.7135, + "num_input_tokens_seen": 334222160, + "step": 10289 + }, + { + "epoch": 0.9279884565089958, + "flos": 21768287436480.0, + "grad_norm": 1.925449242128143, + "learning_rate": 5.409004409377882e-08, + "loss": 0.6581, + "num_input_tokens_seen": 334250115, + "step": 10290 + }, + { + "epoch": 0.9280786400324661, + "flos": 48837782584800.0, + "grad_norm": 2.9129405789454896, + "learning_rate": 5.3955177479589e-08, + "loss": 0.6643, + "num_input_tokens_seen": 334285805, + "step": 10291 + }, + { + "epoch": 0.9281688235559363, + "flos": 21987363233760.0, + "grad_norm": 1.8431016460916982, + "learning_rate": 5.3820476912674e-08, + "loss": 0.7476, + "num_input_tokens_seen": 334316025, + "step": 10292 + }, + { + "epoch": 0.9282590070794066, + "flos": 28727301437760.0, + "grad_norm": 1.5065090890503234, + "learning_rate": 5.3685942404527063e-08, + "loss": 0.7674, + "num_input_tokens_seen": 334347520, + "step": 10293 + }, + { + "epoch": 0.9283491906028769, + "flos": 23189325122880.0, + "grad_norm": 1.9055544100029818, + "learning_rate": 5.355157396662702e-08, + "loss": 0.7235, + "num_input_tokens_seen": 334375255, + "step": 10294 + }, + { + "epoch": 0.9284393741263471, + "flos": 26281295757600.0, + "grad_norm": 1.7871640723357645, + "learning_rate": 5.34173716104398e-08, + "loss": 0.7229, + "num_input_tokens_seen": 334405020, + "step": 10295 + }, + { + "epoch": 0.9285295576498174, + "flos": 20377623177600.0, + "grad_norm": 1.82750430533219, + "learning_rate": 5.328333534741536e-08, + "loss": 0.7474, + "num_input_tokens_seen": 334429905, + "step": 10296 + }, + { + "epoch": 0.9286197411732876, + "flos": 23334495970080.0, + "grad_norm": 2.1084263681785327, + "learning_rate": 5.314946518899099e-08, + "loss": 0.713, + "num_input_tokens_seen": 334455135, + "step": 10297 + }, + { + "epoch": 0.928709924696758, + "flos": 25155729192480.0, + "grad_norm": 2.02913426471927, + "learning_rate": 5.301576114658912e-08, + "loss": 0.8125, + "num_input_tokens_seen": 334484225, + "step": 10298 + }, + { + "epoch": 0.9288001082202282, + "flos": 22095247078080.0, + "grad_norm": 1.924780595339298, + "learning_rate": 5.288222323161795e-08, + "loss": 0.7667, + "num_input_tokens_seen": 334511515, + "step": 10299 + }, + { + "epoch": 0.9288902917436984, + "flos": 12804504439680.0, + "grad_norm": 5.382435291289343, + "learning_rate": 5.274885145547214e-08, + "loss": 0.7842, + "num_input_tokens_seen": 334537320, + "step": 10300 + }, + { + "epoch": 0.9289804752671686, + "flos": 25263613036800.0, + "grad_norm": 1.5832524613590646, + "learning_rate": 5.261564582953082e-08, + "loss": 0.732, + "num_input_tokens_seen": 334570210, + "step": 10301 + }, + { + "epoch": 0.929070658790639, + "flos": 22569236054880.0, + "grad_norm": 2.3946425643860576, + "learning_rate": 5.248260636516066e-08, + "loss": 0.7187, + "num_input_tokens_seen": 334598095, + "step": 10302 + }, + { + "epoch": 0.9291608423141092, + "flos": 21002655823680.0, + "grad_norm": 1.995279164584709, + "learning_rate": 5.2349733073712824e-08, + "loss": 0.683, + "num_input_tokens_seen": 334626795, + "step": 10303 + }, + { + "epoch": 0.9292510258375795, + "flos": 26941013523360.0, + "grad_norm": 1.787351031198118, + "learning_rate": 5.221702596652533e-08, + "loss": 0.8286, + "num_input_tokens_seen": 334654825, + "step": 10304 + }, + { + "epoch": 0.9293412093610497, + "flos": 12950939058720.0, + "grad_norm": 2.77985904500923, + "learning_rate": 5.208448505492091e-08, + "loss": 0.6794, + "num_input_tokens_seen": 334679835, + "step": 10305 + }, + { + "epoch": 0.92943139288452, + "flos": 29819483824800.0, + "grad_norm": 2.029628752224576, + "learning_rate": 5.1952110350208965e-08, + "loss": 0.7227, + "num_input_tokens_seen": 334707495, + "step": 10306 + }, + { + "epoch": 0.9295215764079903, + "flos": 25338595909920.0, + "grad_norm": 1.750122679088801, + "learning_rate": 5.181990186368446e-08, + "loss": 0.6974, + "num_input_tokens_seen": 334736870, + "step": 10307 + }, + { + "epoch": 0.9296117599314605, + "flos": 21764310272160.0, + "grad_norm": 1.954112665315472, + "learning_rate": 5.1687859606627915e-08, + "loss": 0.8376, + "num_input_tokens_seen": 334765550, + "step": 10308 + }, + { + "epoch": 0.9297019434549307, + "flos": 36413730804960.0, + "grad_norm": 1.8463636719997487, + "learning_rate": 5.1555983590306327e-08, + "loss": 0.6507, + "num_input_tokens_seen": 334796980, + "step": 10309 + }, + { + "epoch": 0.9297921269784011, + "flos": 70912580561280.0, + "grad_norm": 0.6540854041130395, + "learning_rate": 5.1424273825971806e-08, + "loss": 0.5768, + "num_input_tokens_seen": 334885840, + "step": 10310 + }, + { + "epoch": 0.9298823105018713, + "flos": 34264906055040.0, + "grad_norm": 3.233841856723497, + "learning_rate": 5.1292730324862475e-08, + "loss": 0.7037, + "num_input_tokens_seen": 334916060, + "step": 10311 + }, + { + "epoch": 0.9299724940253415, + "flos": 20782056066240.0, + "grad_norm": 1.6071254229735297, + "learning_rate": 5.116135309820224e-08, + "loss": 0.7539, + "num_input_tokens_seen": 334944410, + "step": 10312 + }, + { + "epoch": 0.9300626775488118, + "flos": 21471664052640.0, + "grad_norm": 1.728250291362723, + "learning_rate": 5.103014215720147e-08, + "loss": 0.7751, + "num_input_tokens_seen": 334972900, + "step": 10313 + }, + { + "epoch": 0.9301528610722821, + "flos": 19216813946400.0, + "grad_norm": 2.496643101471063, + "learning_rate": 5.0899097513055214e-08, + "loss": 0.7955, + "num_input_tokens_seen": 334999965, + "step": 10314 + }, + { + "epoch": 0.9302430445957524, + "flos": 28143086921760.0, + "grad_norm": 2.126068856194577, + "learning_rate": 5.076821917694563e-08, + "loss": 0.7277, + "num_input_tokens_seen": 335028110, + "step": 10315 + }, + { + "epoch": 0.9303332281192226, + "flos": 19654593843360.0, + "grad_norm": 2.0674013795591026, + "learning_rate": 5.063750716003889e-08, + "loss": 0.7883, + "num_input_tokens_seen": 335054545, + "step": 10316 + }, + { + "epoch": 0.9304234116426929, + "flos": 20674246561440.0, + "grad_norm": 1.856472346286102, + "learning_rate": 5.050696147348921e-08, + "loss": 0.7262, + "num_input_tokens_seen": 335080910, + "step": 10317 + }, + { + "epoch": 0.9305135951661632, + "flos": 35758027373280.0, + "grad_norm": 2.110742535400927, + "learning_rate": 5.037658212843454e-08, + "loss": 0.7135, + "num_input_tokens_seen": 335112275, + "step": 10318 + }, + { + "epoch": 0.9306037786896334, + "flos": 34956261020160.0, + "grad_norm": 1.6826583494176315, + "learning_rate": 5.0246369136000444e-08, + "loss": 0.6305, + "num_input_tokens_seen": 335142560, + "step": 10319 + }, + { + "epoch": 0.9306939622131036, + "flos": 19101124452480.0, + "grad_norm": 2.0122666881209095, + "learning_rate": 5.011632250729691e-08, + "loss": 0.8108, + "num_input_tokens_seen": 335169365, + "step": 10320 + }, + { + "epoch": 0.930784145736574, + "flos": 23295387648960.0, + "grad_norm": 2.0011758923792757, + "learning_rate": 4.998644225342019e-08, + "loss": 0.8295, + "num_input_tokens_seen": 335196700, + "step": 10321 + }, + { + "epoch": 0.9308743292600442, + "flos": 21768844982880.0, + "grad_norm": 1.6913488787993072, + "learning_rate": 4.9856728385452296e-08, + "loss": 0.7427, + "num_input_tokens_seen": 335224685, + "step": 10322 + }, + { + "epoch": 0.9309645127835144, + "flos": 21035816983200.0, + "grad_norm": 1.8307633231726714, + "learning_rate": 4.9727180914461485e-08, + "loss": 0.8228, + "num_input_tokens_seen": 335253050, + "step": 10323 + }, + { + "epoch": 0.9310546963069847, + "flos": 26352264296640.0, + "grad_norm": 1.9277904003779318, + "learning_rate": 4.959779985150137e-08, + "loss": 0.749, + "num_input_tokens_seen": 335281375, + "step": 10324 + }, + { + "epoch": 0.931144879830455, + "flos": 16084025181600.0, + "grad_norm": 2.7371194200425335, + "learning_rate": 4.9468585207611105e-08, + "loss": 0.6862, + "num_input_tokens_seen": 335306935, + "step": 10325 + }, + { + "epoch": 0.9312350633539253, + "flos": 20488480602720.0, + "grad_norm": 2.6756322666097163, + "learning_rate": 4.9339536993816764e-08, + "loss": 0.7151, + "num_input_tokens_seen": 335332845, + "step": 10326 + }, + { + "epoch": 0.9313252468773955, + "flos": 26537807236800.0, + "grad_norm": 1.9056712663488353, + "learning_rate": 4.921065522112844e-08, + "loss": 0.6909, + "num_input_tokens_seen": 335361575, + "step": 10327 + }, + { + "epoch": 0.9314154304008657, + "flos": 30581101103520.0, + "grad_norm": 1.8194424100998756, + "learning_rate": 4.908193990054377e-08, + "loss": 0.7839, + "num_input_tokens_seen": 335390195, + "step": 10328 + }, + { + "epoch": 0.9315056139243361, + "flos": 29086827522720.0, + "grad_norm": 1.5723377282331445, + "learning_rate": 4.89533910430453e-08, + "loss": 0.7551, + "num_input_tokens_seen": 335418990, + "step": 10329 + }, + { + "epoch": 0.9315957974478063, + "flos": 23080065997440.0, + "grad_norm": 5.03420630373056, + "learning_rate": 4.8825008659601376e-08, + "loss": 0.7303, + "num_input_tokens_seen": 335448435, + "step": 10330 + }, + { + "epoch": 0.9316859809712765, + "flos": 19544851511040.0, + "grad_norm": 2.2621436370589465, + "learning_rate": 4.869679276116634e-08, + "loss": 0.7564, + "num_input_tokens_seen": 335472445, + "step": 10331 + }, + { + "epoch": 0.9317761644947468, + "flos": 24535156917600.0, + "grad_norm": 1.903633448138219, + "learning_rate": 4.856874335868055e-08, + "loss": 0.758, + "num_input_tokens_seen": 335503325, + "step": 10332 + }, + { + "epoch": 0.9318663480182171, + "flos": 24063249447360.0, + "grad_norm": 4.440472104516178, + "learning_rate": 4.844086046306928e-08, + "loss": 0.8432, + "num_input_tokens_seen": 335529155, + "step": 10333 + }, + { + "epoch": 0.9319565315416873, + "flos": 29783795121600.0, + "grad_norm": 1.8076292559995222, + "learning_rate": 4.8313144085244896e-08, + "loss": 0.7413, + "num_input_tokens_seen": 335559445, + "step": 10334 + }, + { + "epoch": 0.9320467150651576, + "flos": 62621561916000.0, + "grad_norm": 0.6401165828840487, + "learning_rate": 4.818559423610424e-08, + "loss": 0.5739, + "num_input_tokens_seen": 335649605, + "step": 10335 + }, + { + "epoch": 0.9321368985886278, + "flos": 22568641338720.0, + "grad_norm": 1.9186074617108546, + "learning_rate": 4.8058210926531284e-08, + "loss": 0.6816, + "num_input_tokens_seen": 335677965, + "step": 10336 + }, + { + "epoch": 0.9322270821120981, + "flos": 23807964570240.0, + "grad_norm": 1.754485512552704, + "learning_rate": 4.7930994167394435e-08, + "loss": 0.7724, + "num_input_tokens_seen": 335707630, + "step": 10337 + }, + { + "epoch": 0.9323172656355684, + "flos": 20530859862720.0, + "grad_norm": 1.5444872373755685, + "learning_rate": 4.7803943969548786e-08, + "loss": 0.7816, + "num_input_tokens_seen": 335738835, + "step": 10338 + }, + { + "epoch": 0.9324074491590386, + "flos": 22787382608160.0, + "grad_norm": 2.029389336162056, + "learning_rate": 4.7677060343834784e-08, + "loss": 0.6874, + "num_input_tokens_seen": 335768185, + "step": 10339 + }, + { + "epoch": 0.932497632682509, + "flos": 15355383213600.0, + "grad_norm": 1.9363672575344884, + "learning_rate": 4.75503433010791e-08, + "loss": 0.7339, + "num_input_tokens_seen": 335794550, + "step": 10340 + }, + { + "epoch": 0.9325878162059792, + "flos": 28099889927040.0, + "grad_norm": 2.241990137219748, + "learning_rate": 4.742379285209419e-08, + "loss": 0.7223, + "num_input_tokens_seen": 335821655, + "step": 10341 + }, + { + "epoch": 0.9326779997294494, + "flos": 37616361749760.0, + "grad_norm": 1.8539931684036888, + "learning_rate": 4.72974090076772e-08, + "loss": 0.8416, + "num_input_tokens_seen": 335853405, + "step": 10342 + }, + { + "epoch": 0.9327681832529197, + "flos": 26317207479360.0, + "grad_norm": 1.9496300517571532, + "learning_rate": 4.717119177861262e-08, + "loss": 0.6965, + "num_input_tokens_seen": 335883760, + "step": 10343 + }, + { + "epoch": 0.93285836677639, + "flos": 23550189319200.0, + "grad_norm": 1.825720811920049, + "learning_rate": 4.70451411756696e-08, + "loss": 0.7552, + "num_input_tokens_seen": 335911365, + "step": 10344 + }, + { + "epoch": 0.9329485502998602, + "flos": 21400807022880.0, + "grad_norm": 1.53961860950132, + "learning_rate": 4.691925720960355e-08, + "loss": 0.7795, + "num_input_tokens_seen": 335940035, + "step": 10345 + }, + { + "epoch": 0.9330387338233305, + "flos": 16953823662720.0, + "grad_norm": 1.9664275382294918, + "learning_rate": 4.6793539891155645e-08, + "loss": 0.7621, + "num_input_tokens_seen": 335966580, + "step": 10346 + }, + { + "epoch": 0.9331289173468007, + "flos": 36085321542720.0, + "grad_norm": 2.1872425472800963, + "learning_rate": 4.6667989231052864e-08, + "loss": 0.7194, + "num_input_tokens_seen": 335997530, + "step": 10347 + }, + { + "epoch": 0.933219100870271, + "flos": 22642100251680.0, + "grad_norm": 2.3297608024378658, + "learning_rate": 4.654260524000797e-08, + "loss": 0.7491, + "num_input_tokens_seen": 336025340, + "step": 10348 + }, + { + "epoch": 0.9333092843937413, + "flos": 23550895544640.0, + "grad_norm": 1.61231385568638, + "learning_rate": 4.6417387928719076e-08, + "loss": 0.7284, + "num_input_tokens_seen": 336054480, + "step": 10349 + }, + { + "epoch": 0.9333994679172115, + "flos": 25483952605920.0, + "grad_norm": 1.6846072516330441, + "learning_rate": 4.629233730787052e-08, + "loss": 0.6146, + "num_input_tokens_seen": 336084115, + "step": 10350 + }, + { + "epoch": 0.9334896514406817, + "flos": 19545260378400.0, + "grad_norm": 2.751900709639894, + "learning_rate": 4.616745338813266e-08, + "loss": 0.7036, + "num_input_tokens_seen": 336110710, + "step": 10351 + }, + { + "epoch": 0.9335798349641521, + "flos": 21984315313440.0, + "grad_norm": 1.621687520681431, + "learning_rate": 4.6042736180160744e-08, + "loss": 0.757, + "num_input_tokens_seen": 336140250, + "step": 10352 + }, + { + "epoch": 0.9336700184876223, + "flos": 19690988772000.0, + "grad_norm": 1.8600918325703957, + "learning_rate": 4.591818569459671e-08, + "loss": 0.8586, + "num_input_tokens_seen": 336168375, + "step": 10353 + }, + { + "epoch": 0.9337602020110926, + "flos": 21694493995680.0, + "grad_norm": 4.476188604109043, + "learning_rate": 4.5793801942067614e-08, + "loss": 0.7614, + "num_input_tokens_seen": 336197430, + "step": 10354 + }, + { + "epoch": 0.9338503855345628, + "flos": 29378098461120.0, + "grad_norm": 1.9107937675313742, + "learning_rate": 4.566958493318673e-08, + "loss": 0.61, + "num_input_tokens_seen": 336225940, + "step": 10355 + }, + { + "epoch": 0.9339405690580331, + "flos": 67994481562080.0, + "grad_norm": 0.6220249416986793, + "learning_rate": 4.554553467855316e-08, + "loss": 0.5986, + "num_input_tokens_seen": 336314730, + "step": 10356 + }, + { + "epoch": 0.9340307525815034, + "flos": 23772127188000.0, + "grad_norm": 2.5613534735000822, + "learning_rate": 4.5421651188751074e-08, + "loss": 0.7576, + "num_input_tokens_seen": 336342825, + "step": 10357 + }, + { + "epoch": 0.9341209361049736, + "flos": 22168705991040.0, + "grad_norm": 1.755988840797611, + "learning_rate": 4.529793447435137e-08, + "loss": 0.7534, + "num_input_tokens_seen": 336371675, + "step": 10358 + }, + { + "epoch": 0.9342111196284438, + "flos": 19472544860640.0, + "grad_norm": 1.7498634830087405, + "learning_rate": 4.5174384545909824e-08, + "loss": 0.741, + "num_input_tokens_seen": 336400155, + "step": 10359 + }, + { + "epoch": 0.9343013031519142, + "flos": 26175716438400.0, + "grad_norm": 2.309047906034182, + "learning_rate": 4.505100141396867e-08, + "loss": 0.7768, + "num_input_tokens_seen": 336426035, + "step": 10360 + }, + { + "epoch": 0.9343914866753844, + "flos": 13752519563040.0, + "grad_norm": 2.4028050917859627, + "learning_rate": 4.492778508905548e-08, + "loss": 0.7727, + "num_input_tokens_seen": 336451990, + "step": 10361 + }, + { + "epoch": 0.9344816701988546, + "flos": 33899692996800.0, + "grad_norm": 2.0259017961431316, + "learning_rate": 4.480473558168385e-08, + "loss": 0.6788, + "num_input_tokens_seen": 336481710, + "step": 10362 + }, + { + "epoch": 0.934571853722325, + "flos": 26208468730560.0, + "grad_norm": 3.3472971417566746, + "learning_rate": 4.4681852902352936e-08, + "loss": 0.6449, + "num_input_tokens_seen": 336511250, + "step": 10363 + }, + { + "epoch": 0.9346620372457952, + "flos": 24463853850720.0, + "grad_norm": 2.2848729667159926, + "learning_rate": 4.455913706154812e-08, + "loss": 0.654, + "num_input_tokens_seen": 336539370, + "step": 10364 + }, + { + "epoch": 0.9347522207692655, + "flos": 22350234597120.0, + "grad_norm": 2.4893913248203505, + "learning_rate": 4.443658806973949e-08, + "loss": 0.7659, + "num_input_tokens_seen": 336566585, + "step": 10365 + }, + { + "epoch": 0.9348424042927357, + "flos": 62092778979360.0, + "grad_norm": 0.661689467609781, + "learning_rate": 4.431420593738444e-08, + "loss": 0.6002, + "num_input_tokens_seen": 336657390, + "step": 10366 + }, + { + "epoch": 0.934932587816206, + "flos": 21948440761440.0, + "grad_norm": 1.8904486960184432, + "learning_rate": 4.419199067492485e-08, + "loss": 0.6524, + "num_input_tokens_seen": 336683915, + "step": 10367 + }, + { + "epoch": 0.9350227713396763, + "flos": 21840705596160.0, + "grad_norm": 1.7952382469924029, + "learning_rate": 4.4069942292788596e-08, + "loss": 0.7496, + "num_input_tokens_seen": 336712435, + "step": 10368 + }, + { + "epoch": 0.9351129548631465, + "flos": 29268839335680.0, + "grad_norm": 1.9005655049448902, + "learning_rate": 4.39480608013898e-08, + "loss": 0.7462, + "num_input_tokens_seen": 336744545, + "step": 10369 + }, + { + "epoch": 0.9352031383866167, + "flos": 30911740551360.0, + "grad_norm": 1.8387416770430314, + "learning_rate": 4.3826346211128126e-08, + "loss": 0.8312, + "num_input_tokens_seen": 336773910, + "step": 10370 + }, + { + "epoch": 0.9352933219100871, + "flos": 20343978811200.0, + "grad_norm": 1.5609706260366083, + "learning_rate": 4.370479853238884e-08, + "loss": 0.7444, + "num_input_tokens_seen": 336802645, + "step": 10371 + }, + { + "epoch": 0.9353835054335573, + "flos": 22022531560320.0, + "grad_norm": 2.5349796274784846, + "learning_rate": 4.3583417775542756e-08, + "loss": 0.7055, + "num_input_tokens_seen": 336831835, + "step": 10372 + }, + { + "epoch": 0.9354736889570275, + "flos": 36123723638400.0, + "grad_norm": 1.6054102383460218, + "learning_rate": 4.3462203950947575e-08, + "loss": 0.6944, + "num_input_tokens_seen": 336866585, + "step": 10373 + }, + { + "epoch": 0.9355638724804978, + "flos": 23407285827360.0, + "grad_norm": 1.6741332729661835, + "learning_rate": 4.3341157068944814e-08, + "loss": 0.7259, + "num_input_tokens_seen": 336894260, + "step": 10374 + }, + { + "epoch": 0.9356540560039681, + "flos": 27372511730880.0, + "grad_norm": 1.9881210412493007, + "learning_rate": 4.322027713986376e-08, + "loss": 0.7546, + "num_input_tokens_seen": 336922730, + "step": 10375 + }, + { + "epoch": 0.9357442395274383, + "flos": 24898846015680.0, + "grad_norm": 2.709791221459267, + "learning_rate": 4.309956417401816e-08, + "loss": 0.8103, + "num_input_tokens_seen": 336947925, + "step": 10376 + }, + { + "epoch": 0.9358344230509086, + "flos": 29161252849440.0, + "grad_norm": 1.8624983768328052, + "learning_rate": 4.297901818170801e-08, + "loss": 0.7105, + "num_input_tokens_seen": 336980095, + "step": 10377 + }, + { + "epoch": 0.9359246065743788, + "flos": 30403661171040.0, + "grad_norm": 1.8280359230184584, + "learning_rate": 4.285863917321886e-08, + "loss": 0.6749, + "num_input_tokens_seen": 337013240, + "step": 10378 + }, + { + "epoch": 0.9360147900978492, + "flos": 21434079691680.0, + "grad_norm": 1.753685816312898, + "learning_rate": 4.2738427158822253e-08, + "loss": 0.6765, + "num_input_tokens_seen": 337040650, + "step": 10379 + }, + { + "epoch": 0.9361049736213194, + "flos": 16411765388160.0, + "grad_norm": 2.215395688779169, + "learning_rate": 4.261838214877511e-08, + "loss": 0.7497, + "num_input_tokens_seen": 337067525, + "step": 10380 + }, + { + "epoch": 0.9361951571447896, + "flos": 28469303168160.0, + "grad_norm": 2.303964585542036, + "learning_rate": 4.249850415332079e-08, + "loss": 0.7468, + "num_input_tokens_seen": 337097500, + "step": 10381 + }, + { + "epoch": 0.9362853406682599, + "flos": 25917532320000.0, + "grad_norm": 1.890706525573913, + "learning_rate": 4.237879318268756e-08, + "loss": 0.8007, + "num_input_tokens_seen": 337127135, + "step": 10382 + }, + { + "epoch": 0.9363755241917302, + "flos": 25593620598720.0, + "grad_norm": 1.9407521302971824, + "learning_rate": 4.225924924708968e-08, + "loss": 0.6722, + "num_input_tokens_seen": 337157040, + "step": 10383 + }, + { + "epoch": 0.9364657077152004, + "flos": 29563901589600.0, + "grad_norm": 2.138436751309088, + "learning_rate": 4.2139872356727665e-08, + "loss": 0.7491, + "num_input_tokens_seen": 337186525, + "step": 10384 + }, + { + "epoch": 0.9365558912386707, + "flos": 27631141886400.0, + "grad_norm": 1.7071609273228914, + "learning_rate": 4.202066252178738e-08, + "loss": 0.7044, + "num_input_tokens_seen": 337217670, + "step": 10385 + }, + { + "epoch": 0.936646074762141, + "flos": 28617001559040.0, + "grad_norm": 1.775767941976778, + "learning_rate": 4.1901619752440445e-08, + "loss": 0.8322, + "num_input_tokens_seen": 337246800, + "step": 10386 + }, + { + "epoch": 0.9367362582856112, + "flos": 23474723239200.0, + "grad_norm": 1.9388550357662604, + "learning_rate": 4.178274405884363e-08, + "loss": 0.7401, + "num_input_tokens_seen": 337273775, + "step": 10387 + }, + { + "epoch": 0.9368264418090815, + "flos": 51277878273120.0, + "grad_norm": 1.8275927559500098, + "learning_rate": 4.166403545114105e-08, + "loss": 0.7498, + "num_input_tokens_seen": 337306810, + "step": 10388 + }, + { + "epoch": 0.9369166253325517, + "flos": 25119259924320.0, + "grad_norm": 2.383413570449479, + "learning_rate": 4.154549393946083e-08, + "loss": 0.7944, + "num_input_tokens_seen": 337334535, + "step": 10389 + }, + { + "epoch": 0.937006808856022, + "flos": 21768621964320.0, + "grad_norm": 2.1191124980132225, + "learning_rate": 4.14271195339182e-08, + "loss": 0.767, + "num_input_tokens_seen": 337364015, + "step": 10390 + }, + { + "epoch": 0.9370969923794923, + "flos": 36047588502720.0, + "grad_norm": 1.6089980785863391, + "learning_rate": 4.1308912244613084e-08, + "loss": 0.7074, + "num_input_tokens_seen": 337397320, + "step": 10391 + }, + { + "epoch": 0.9371871759029625, + "flos": 23514203257920.0, + "grad_norm": 1.5621838822068235, + "learning_rate": 4.1190872081631636e-08, + "loss": 0.698, + "num_input_tokens_seen": 337426515, + "step": 10392 + }, + { + "epoch": 0.9372773594264328, + "flos": 18561482212320.0, + "grad_norm": 3.8714010692136376, + "learning_rate": 4.107299905504558e-08, + "loss": 0.8076, + "num_input_tokens_seen": 337452230, + "step": 10393 + }, + { + "epoch": 0.9373675429499031, + "flos": 30074434174080.0, + "grad_norm": 1.7580053605137884, + "learning_rate": 4.095529317491286e-08, + "loss": 0.7335, + "num_input_tokens_seen": 337482840, + "step": 10394 + }, + { + "epoch": 0.9374577264733733, + "flos": 19106737086240.0, + "grad_norm": 2.1939444300900144, + "learning_rate": 4.0837754451276575e-08, + "loss": 0.7443, + "num_input_tokens_seen": 337510850, + "step": 10395 + }, + { + "epoch": 0.9375479099968436, + "flos": 25883404746720.0, + "grad_norm": 1.5614434293918298, + "learning_rate": 4.072038289416557e-08, + "loss": 0.7269, + "num_input_tokens_seen": 337543260, + "step": 10396 + }, + { + "epoch": 0.9376380935203138, + "flos": 22677417257280.0, + "grad_norm": 2.8604639199936517, + "learning_rate": 4.0603178513595185e-08, + "loss": 0.7452, + "num_input_tokens_seen": 337572855, + "step": 10397 + }, + { + "epoch": 0.9377282770437841, + "flos": 16776420900000.0, + "grad_norm": 3.105517577741137, + "learning_rate": 4.0486141319565624e-08, + "loss": 0.7489, + "num_input_tokens_seen": 337599315, + "step": 10398 + }, + { + "epoch": 0.9378184605672544, + "flos": 24864718442400.0, + "grad_norm": 2.0289239329616646, + "learning_rate": 4.0369271322062916e-08, + "loss": 0.6847, + "num_input_tokens_seen": 337625305, + "step": 10399 + }, + { + "epoch": 0.9379086440907246, + "flos": 17652017863680.0, + "grad_norm": 1.7983904874673933, + "learning_rate": 4.0252568531059295e-08, + "loss": 0.759, + "num_input_tokens_seen": 337653615, + "step": 10400 + }, + { + "epoch": 0.9379988276141948, + "flos": 20930200494240.0, + "grad_norm": 2.1524350238700407, + "learning_rate": 4.013603295651235e-08, + "loss": 0.7424, + "num_input_tokens_seen": 337682140, + "step": 10401 + }, + { + "epoch": 0.9380890111376652, + "flos": 20019063506400.0, + "grad_norm": 2.1462859876900353, + "learning_rate": 4.001966460836592e-08, + "loss": 0.7681, + "num_input_tokens_seen": 337708540, + "step": 10402 + }, + { + "epoch": 0.9381791946611354, + "flos": 24318348475680.0, + "grad_norm": 3.521069325018943, + "learning_rate": 3.990346349654894e-08, + "loss": 0.8143, + "num_input_tokens_seen": 337737770, + "step": 10403 + }, + { + "epoch": 0.9382693781846057, + "flos": 30872855248800.0, + "grad_norm": 1.726714390766543, + "learning_rate": 3.9787429630975924e-08, + "loss": 0.7173, + "num_input_tokens_seen": 337768330, + "step": 10404 + }, + { + "epoch": 0.9383595617080759, + "flos": 20019175015680.0, + "grad_norm": 2.507203310360698, + "learning_rate": 3.967156302154828e-08, + "loss": 0.7275, + "num_input_tokens_seen": 337797275, + "step": 10405 + }, + { + "epoch": 0.9384497452315462, + "flos": 26139841886400.0, + "grad_norm": 2.372404788712573, + "learning_rate": 3.955586367815189e-08, + "loss": 0.6966, + "num_input_tokens_seen": 337826645, + "step": 10406 + }, + { + "epoch": 0.9385399287550165, + "flos": 26832534962880.0, + "grad_norm": 1.9040325587798648, + "learning_rate": 3.944033161065907e-08, + "loss": 0.8105, + "num_input_tokens_seen": 337856510, + "step": 10407 + }, + { + "epoch": 0.9386301122784867, + "flos": 36775412736000.0, + "grad_norm": 1.8412692304281602, + "learning_rate": 3.93249668289275e-08, + "loss": 0.6263, + "num_input_tokens_seen": 337887590, + "step": 10408 + }, + { + "epoch": 0.9387202958019569, + "flos": 23589892356480.0, + "grad_norm": 2.7349195537556743, + "learning_rate": 3.920976934280063e-08, + "loss": 0.6668, + "num_input_tokens_seen": 337917795, + "step": 10409 + }, + { + "epoch": 0.9388104793254273, + "flos": 19654445164320.0, + "grad_norm": 1.7891260660531199, + "learning_rate": 3.909473916210815e-08, + "loss": 0.8271, + "num_input_tokens_seen": 337944670, + "step": 10410 + }, + { + "epoch": 0.9389006628488975, + "flos": 17322976715520.0, + "grad_norm": 2.4899251110044744, + "learning_rate": 3.897987629666488e-08, + "loss": 0.703, + "num_input_tokens_seen": 337971030, + "step": 10411 + }, + { + "epoch": 0.9389908463723677, + "flos": 23807741551680.0, + "grad_norm": 1.6579679540291254, + "learning_rate": 3.886518075627143e-08, + "loss": 0.8145, + "num_input_tokens_seen": 338002745, + "step": 10412 + }, + { + "epoch": 0.9390810298958381, + "flos": 21840222389280.0, + "grad_norm": 2.5632737084916966, + "learning_rate": 3.875065255071419e-08, + "loss": 0.7408, + "num_input_tokens_seen": 338027405, + "step": 10413 + }, + { + "epoch": 0.9391712134193083, + "flos": 21221248414080.0, + "grad_norm": 1.8106694020576959, + "learning_rate": 3.863629168976579e-08, + "loss": 0.7105, + "num_input_tokens_seen": 338054975, + "step": 10414 + }, + { + "epoch": 0.9392613969427785, + "flos": 19326890806560.0, + "grad_norm": 1.7563898729983243, + "learning_rate": 3.852209818318375e-08, + "loss": 0.798, + "num_input_tokens_seen": 338082415, + "step": 10415 + }, + { + "epoch": 0.9393515804662488, + "flos": 26358360137280.0, + "grad_norm": 2.0883316987262894, + "learning_rate": 3.840807204071161e-08, + "loss": 0.6484, + "num_input_tokens_seen": 338112570, + "step": 10416 + }, + { + "epoch": 0.9394417639897191, + "flos": 28361642342400.0, + "grad_norm": 1.6661832664731018, + "learning_rate": 3.829421327207894e-08, + "loss": 0.7272, + "num_input_tokens_seen": 338140370, + "step": 10417 + }, + { + "epoch": 0.9395319475131894, + "flos": 20927152573920.0, + "grad_norm": 1.9790360725550202, + "learning_rate": 3.8180521887000825e-08, + "loss": 0.6929, + "num_input_tokens_seen": 338170220, + "step": 10418 + }, + { + "epoch": 0.9396221310366596, + "flos": 68289320797440.0, + "grad_norm": 0.5876508458240725, + "learning_rate": 3.806699789517775e-08, + "loss": 0.6094, + "num_input_tokens_seen": 338264105, + "step": 10419 + }, + { + "epoch": 0.9397123145601298, + "flos": 20456062838400.0, + "grad_norm": 1.6900696439895706, + "learning_rate": 3.7953641306296635e-08, + "loss": 0.7262, + "num_input_tokens_seen": 338291340, + "step": 10420 + }, + { + "epoch": 0.9398024980836002, + "flos": 21621072252480.0, + "grad_norm": 2.8058061414382354, + "learning_rate": 3.784045213002951e-08, + "loss": 0.5846, + "num_input_tokens_seen": 338317915, + "step": 10421 + }, + { + "epoch": 0.9398926816070704, + "flos": 22275920779680.0, + "grad_norm": 1.983663416356893, + "learning_rate": 3.7727430376033986e-08, + "loss": 0.7564, + "num_input_tokens_seen": 338345635, + "step": 10422 + }, + { + "epoch": 0.9399828651305406, + "flos": 19472433351360.0, + "grad_norm": 2.262284756532898, + "learning_rate": 3.7614576053954126e-08, + "loss": 0.7448, + "num_input_tokens_seen": 338373295, + "step": 10423 + }, + { + "epoch": 0.9400730486540109, + "flos": 35063661657600.0, + "grad_norm": 1.9888195815462735, + "learning_rate": 3.75018891734189e-08, + "loss": 0.739, + "num_input_tokens_seen": 338403830, + "step": 10424 + }, + { + "epoch": 0.9401632321774812, + "flos": 31851615497280.0, + "grad_norm": 2.380040191343089, + "learning_rate": 3.738936974404372e-08, + "loss": 0.7888, + "num_input_tokens_seen": 338430765, + "step": 10425 + }, + { + "epoch": 0.9402534157009514, + "flos": 19399569154560.0, + "grad_norm": 1.3639781668025923, + "learning_rate": 3.7277017775429354e-08, + "loss": 0.8063, + "num_input_tokens_seen": 338459680, + "step": 10426 + }, + { + "epoch": 0.9403435992244217, + "flos": 17760830952000.0, + "grad_norm": 1.6237592361810078, + "learning_rate": 3.7164833277162136e-08, + "loss": 0.7655, + "num_input_tokens_seen": 338486540, + "step": 10427 + }, + { + "epoch": 0.9404337827478919, + "flos": 25229596972800.0, + "grad_norm": 1.4686812526765298, + "learning_rate": 3.705281625881418e-08, + "loss": 0.7237, + "num_input_tokens_seen": 338518055, + "step": 10428 + }, + { + "epoch": 0.9405239662713623, + "flos": 44062278453120.0, + "grad_norm": 2.1102135016009322, + "learning_rate": 3.694096672994362e-08, + "loss": 0.6511, + "num_input_tokens_seen": 338552875, + "step": 10429 + }, + { + "epoch": 0.9406141497948325, + "flos": 22491911486880.0, + "grad_norm": 1.8368201548671808, + "learning_rate": 3.682928470009394e-08, + "loss": 0.7974, + "num_input_tokens_seen": 338580060, + "step": 10430 + }, + { + "epoch": 0.9407043333183027, + "flos": 14845519684800.0, + "grad_norm": 2.4075281290884836, + "learning_rate": 3.6717770178794406e-08, + "loss": 0.6982, + "num_input_tokens_seen": 338602745, + "step": 10431 + }, + { + "epoch": 0.940794516841773, + "flos": 25190079784320.0, + "grad_norm": 4.65913496313815, + "learning_rate": 3.6606423175560287e-08, + "loss": 0.8463, + "num_input_tokens_seen": 338633445, + "step": 10432 + }, + { + "epoch": 0.9408847003652433, + "flos": 68815836378720.0, + "grad_norm": 0.7853976964978466, + "learning_rate": 3.649524369989221e-08, + "loss": 0.5447, + "num_input_tokens_seen": 338729240, + "step": 10433 + }, + { + "epoch": 0.9409748838887135, + "flos": 22930211760480.0, + "grad_norm": 1.8278874299970809, + "learning_rate": 3.638423176127636e-08, + "loss": 0.775, + "num_input_tokens_seen": 338758540, + "step": 10434 + }, + { + "epoch": 0.9410650674121838, + "flos": 24135518928000.0, + "grad_norm": 1.495047340281443, + "learning_rate": 3.6273387369185396e-08, + "loss": 0.7783, + "num_input_tokens_seen": 338787675, + "step": 10435 + }, + { + "epoch": 0.9411552509356541, + "flos": 31494505446720.0, + "grad_norm": 1.948723311519837, + "learning_rate": 3.616271053307685e-08, + "loss": 0.6885, + "num_input_tokens_seen": 338820425, + "step": 10436 + }, + { + "epoch": 0.9412454344591243, + "flos": 22714927278720.0, + "grad_norm": 2.128180311831389, + "learning_rate": 3.6052201262394275e-08, + "loss": 0.7635, + "num_input_tokens_seen": 338850500, + "step": 10437 + }, + { + "epoch": 0.9413356179825946, + "flos": 24536532198720.0, + "grad_norm": 1.8437005375838527, + "learning_rate": 3.5941859566566816e-08, + "loss": 0.798, + "num_input_tokens_seen": 338880245, + "step": 10438 + }, + { + "epoch": 0.9414258015060648, + "flos": 57023290516800.0, + "grad_norm": 0.6927483028458532, + "learning_rate": 3.583168545500981e-08, + "loss": 0.5666, + "num_input_tokens_seen": 338969715, + "step": 10439 + }, + { + "epoch": 0.9415159850295352, + "flos": 20638929555840.0, + "grad_norm": 2.044107677646885, + "learning_rate": 3.5721678937123746e-08, + "loss": 0.6883, + "num_input_tokens_seen": 338997570, + "step": 10440 + }, + { + "epoch": 0.9416061685530054, + "flos": 25594587012480.0, + "grad_norm": 1.9229151074481345, + "learning_rate": 3.561184002229467e-08, + "loss": 0.8261, + "num_input_tokens_seen": 339026170, + "step": 10441 + }, + { + "epoch": 0.9416963520764756, + "flos": 30074620022880.0, + "grad_norm": 2.0898106652608948, + "learning_rate": 3.550216871989531e-08, + "loss": 0.6898, + "num_input_tokens_seen": 339056605, + "step": 10442 + }, + { + "epoch": 0.9417865355999459, + "flos": 70415429090400.0, + "grad_norm": 0.6385620209362708, + "learning_rate": 3.539266503928262e-08, + "loss": 0.6454, + "num_input_tokens_seen": 339157000, + "step": 10443 + }, + { + "epoch": 0.9418767191234162, + "flos": 23516359104000.0, + "grad_norm": 1.6947568343165837, + "learning_rate": 3.528332898980091e-08, + "loss": 0.7537, + "num_input_tokens_seen": 339186675, + "step": 10444 + }, + { + "epoch": 0.9419669026468864, + "flos": 38490880790400.0, + "grad_norm": 2.467086521278267, + "learning_rate": 3.517416058077849e-08, + "loss": 0.7098, + "num_input_tokens_seen": 339218520, + "step": 10445 + }, + { + "epoch": 0.9420570861703567, + "flos": 18889779965280.0, + "grad_norm": 1.9688813907436502, + "learning_rate": 3.506515982153102e-08, + "loss": 0.6455, + "num_input_tokens_seen": 339245695, + "step": 10446 + }, + { + "epoch": 0.9421472696938269, + "flos": 30183544620480.0, + "grad_norm": 1.9374225354859769, + "learning_rate": 3.495632672135862e-08, + "loss": 0.7477, + "num_input_tokens_seen": 339276690, + "step": 10447 + }, + { + "epoch": 0.9422374532172972, + "flos": 26248246107360.0, + "grad_norm": 1.5591705708103991, + "learning_rate": 3.4847661289547417e-08, + "loss": 0.7716, + "num_input_tokens_seen": 339308845, + "step": 10448 + }, + { + "epoch": 0.9423276367407675, + "flos": 19836828674880.0, + "grad_norm": 2.4374497237277906, + "learning_rate": 3.473916353536932e-08, + "loss": 0.6993, + "num_input_tokens_seen": 339336550, + "step": 10449 + }, + { + "epoch": 0.9424178202642377, + "flos": 21510028978560.0, + "grad_norm": 1.8091935219946664, + "learning_rate": 3.463083346808249e-08, + "loss": 0.7089, + "num_input_tokens_seen": 339365435, + "step": 10450 + }, + { + "epoch": 0.9425080037877079, + "flos": 23115866209920.0, + "grad_norm": 1.7358599654907942, + "learning_rate": 3.452267109692975e-08, + "loss": 0.6387, + "num_input_tokens_seen": 339394115, + "step": 10451 + }, + { + "epoch": 0.9425981873111783, + "flos": 21834238057920.0, + "grad_norm": 3.3970255406912804, + "learning_rate": 3.441467643114016e-08, + "loss": 0.7434, + "num_input_tokens_seen": 339420425, + "step": 10452 + }, + { + "epoch": 0.9426883708346485, + "flos": 66687981107040.0, + "grad_norm": 0.5859988371215843, + "learning_rate": 3.430684947992857e-08, + "loss": 0.5565, + "num_input_tokens_seen": 339506505, + "step": 10453 + }, + { + "epoch": 0.9427785543581187, + "flos": 21876096941280.0, + "grad_norm": 2.007867122159488, + "learning_rate": 3.419919025249518e-08, + "loss": 0.7579, + "num_input_tokens_seen": 339535825, + "step": 10454 + }, + { + "epoch": 0.942868737881589, + "flos": 22461315040800.0, + "grad_norm": 1.7926207892502317, + "learning_rate": 3.40916987580262e-08, + "loss": 0.7637, + "num_input_tokens_seen": 339566200, + "step": 10455 + }, + { + "epoch": 0.9429589214050593, + "flos": 21836839941120.0, + "grad_norm": 1.8230533512279172, + "learning_rate": 3.398437500569362e-08, + "loss": 0.7207, + "num_input_tokens_seen": 339593890, + "step": 10456 + }, + { + "epoch": 0.9430491049285296, + "flos": 61770508461120.0, + "grad_norm": 2.0846945956565013, + "learning_rate": 3.3877219004654347e-08, + "loss": 0.6236, + "num_input_tokens_seen": 339626525, + "step": 10457 + }, + { + "epoch": 0.9431392884519998, + "flos": 22569273224640.0, + "grad_norm": 3.1114174733620175, + "learning_rate": 3.3770230764051946e-08, + "loss": 0.776, + "num_input_tokens_seen": 339654650, + "step": 10458 + }, + { + "epoch": 0.9432294719754701, + "flos": 25335101952480.0, + "grad_norm": 2.217589598762593, + "learning_rate": 3.366341029301534e-08, + "loss": 0.7604, + "num_input_tokens_seen": 339682525, + "step": 10459 + }, + { + "epoch": 0.9433196554989404, + "flos": 25483766757120.0, + "grad_norm": 2.53280949530638, + "learning_rate": 3.355675760065857e-08, + "loss": 0.7484, + "num_input_tokens_seen": 339709875, + "step": 10460 + }, + { + "epoch": 0.9434098390224106, + "flos": 62433194074080.0, + "grad_norm": 0.5778571358868833, + "learning_rate": 3.345027269608236e-08, + "loss": 0.525, + "num_input_tokens_seen": 339808940, + "step": 10461 + }, + { + "epoch": 0.9435000225458808, + "flos": 70542312415680.0, + "grad_norm": 0.6636037925210023, + "learning_rate": 3.334395558837211e-08, + "loss": 0.5915, + "num_input_tokens_seen": 339896430, + "step": 10462 + }, + { + "epoch": 0.9435902060693512, + "flos": 21508096151040.0, + "grad_norm": 1.6135380058008986, + "learning_rate": 3.3237806286599667e-08, + "loss": 0.6536, + "num_input_tokens_seen": 339924110, + "step": 10463 + }, + { + "epoch": 0.9436803895928214, + "flos": 20783914554240.0, + "grad_norm": 1.713721740289321, + "learning_rate": 3.313182479982224e-08, + "loss": 0.7292, + "num_input_tokens_seen": 339951910, + "step": 10464 + }, + { + "epoch": 0.9437705731162916, + "flos": 28434655218240.0, + "grad_norm": 1.8892884377282992, + "learning_rate": 3.302601113708259e-08, + "loss": 0.8089, + "num_input_tokens_seen": 339979910, + "step": 10465 + }, + { + "epoch": 0.9438607566397619, + "flos": 29559292539360.0, + "grad_norm": 1.3735215735280908, + "learning_rate": 3.292036530740972e-08, + "loss": 0.7877, + "num_input_tokens_seen": 340010485, + "step": 10466 + }, + { + "epoch": 0.9439509401632322, + "flos": 24936987923040.0, + "grad_norm": 2.1770995168373397, + "learning_rate": 3.2814887319817294e-08, + "loss": 0.7339, + "num_input_tokens_seen": 340039070, + "step": 10467 + }, + { + "epoch": 0.9440411236867025, + "flos": 21439915344000.0, + "grad_norm": 1.7874480713457295, + "learning_rate": 3.270957718330591e-08, + "loss": 0.7827, + "num_input_tokens_seen": 340067200, + "step": 10468 + }, + { + "epoch": 0.9441313072101727, + "flos": 21439357797600.0, + "grad_norm": 2.124403477135874, + "learning_rate": 3.260443490686082e-08, + "loss": 0.7854, + "num_input_tokens_seen": 340096735, + "step": 10469 + }, + { + "epoch": 0.9442214907336429, + "flos": 19545706415520.0, + "grad_norm": 1.9448963624070437, + "learning_rate": 3.249946049945351e-08, + "loss": 0.7572, + "num_input_tokens_seen": 340124100, + "step": 10470 + }, + { + "epoch": 0.9443116742571133, + "flos": 21803009725920.0, + "grad_norm": 2.0276043380814315, + "learning_rate": 3.239465397004082e-08, + "loss": 0.7567, + "num_input_tokens_seen": 340150960, + "step": 10471 + }, + { + "epoch": 0.9444018577805835, + "flos": 29637286163040.0, + "grad_norm": 3.1264176776979693, + "learning_rate": 3.229001532756559e-08, + "loss": 0.7022, + "num_input_tokens_seen": 340183370, + "step": 10472 + }, + { + "epoch": 0.9444920413040537, + "flos": 20637888802560.0, + "grad_norm": 2.2274499028284778, + "learning_rate": 3.218554458095602e-08, + "loss": 0.738, + "num_input_tokens_seen": 340211335, + "step": 10473 + }, + { + "epoch": 0.944582224827524, + "flos": 12076680206400.0, + "grad_norm": 2.5582695442555843, + "learning_rate": 3.20812417391263e-08, + "loss": 0.7492, + "num_input_tokens_seen": 340237825, + "step": 10474 + }, + { + "epoch": 0.9446724083509943, + "flos": 23077389774720.0, + "grad_norm": 1.4361622432129897, + "learning_rate": 3.1977106810975764e-08, + "loss": 0.7353, + "num_input_tokens_seen": 340269660, + "step": 10475 + }, + { + "epoch": 0.9447625918744645, + "flos": 54315866949120.0, + "grad_norm": 0.7835653328425601, + "learning_rate": 3.187313980539042e-08, + "loss": 0.6209, + "num_input_tokens_seen": 340338870, + "step": 10476 + }, + { + "epoch": 0.9448527753979348, + "flos": 22751210698080.0, + "grad_norm": 1.7127725871529864, + "learning_rate": 3.176934073124071e-08, + "loss": 0.823, + "num_input_tokens_seen": 340370305, + "step": 10477 + }, + { + "epoch": 0.944942958921405, + "flos": 20967227308800.0, + "grad_norm": 2.1936865381696133, + "learning_rate": 3.166570959738357e-08, + "loss": 0.7397, + "num_input_tokens_seen": 340398655, + "step": 10478 + }, + { + "epoch": 0.9450331424448754, + "flos": 41004546900960.0, + "grad_norm": 1.9064577958712736, + "learning_rate": 3.1562246412661476e-08, + "loss": 0.6795, + "num_input_tokens_seen": 340428240, + "step": 10479 + }, + { + "epoch": 0.9451233259683456, + "flos": 22606411548480.0, + "grad_norm": 1.7716004523927202, + "learning_rate": 3.145895118590225e-08, + "loss": 0.7744, + "num_input_tokens_seen": 340457710, + "step": 10480 + }, + { + "epoch": 0.9452135094918158, + "flos": 32110208483040.0, + "grad_norm": 1.7739026640129418, + "learning_rate": 3.135582392591996e-08, + "loss": 0.6525, + "num_input_tokens_seen": 340486985, + "step": 10481 + }, + { + "epoch": 0.9453036930152862, + "flos": 60248680620960.0, + "grad_norm": 0.5906035222583741, + "learning_rate": 3.125286464151333e-08, + "loss": 0.535, + "num_input_tokens_seen": 340584035, + "step": 10482 + }, + { + "epoch": 0.9453938765387564, + "flos": 24208531803840.0, + "grad_norm": 2.6839184516024877, + "learning_rate": 3.115007334146824e-08, + "loss": 0.7151, + "num_input_tokens_seen": 340611375, + "step": 10483 + }, + { + "epoch": 0.9454840600622266, + "flos": 59183229024960.0, + "grad_norm": 0.630492206879265, + "learning_rate": 3.104745003455478e-08, + "loss": 0.5564, + "num_input_tokens_seen": 340706240, + "step": 10484 + }, + { + "epoch": 0.9455742435856969, + "flos": 21293480724960.0, + "grad_norm": 1.8131837541698574, + "learning_rate": 3.094499472952972e-08, + "loss": 0.765, + "num_input_tokens_seen": 340736400, + "step": 10485 + }, + { + "epoch": 0.9456644271091672, + "flos": 24715942128480.0, + "grad_norm": 9.262785864162684, + "learning_rate": 3.084270743513495e-08, + "loss": 0.7099, + "num_input_tokens_seen": 340766285, + "step": 10486 + }, + { + "epoch": 0.9457546106326374, + "flos": 25374210273600.0, + "grad_norm": 2.498683964523211, + "learning_rate": 3.074058816009817e-08, + "loss": 0.6301, + "num_input_tokens_seen": 340794925, + "step": 10487 + }, + { + "epoch": 0.9458447941561077, + "flos": 25156509757440.0, + "grad_norm": 1.9081982521768004, + "learning_rate": 3.063863691313284e-08, + "loss": 0.691, + "num_input_tokens_seen": 340823905, + "step": 10488 + }, + { + "epoch": 0.9459349776795779, + "flos": 36045060959040.0, + "grad_norm": 1.874623769524823, + "learning_rate": 3.0536853702937794e-08, + "loss": 0.6811, + "num_input_tokens_seen": 340854765, + "step": 10489 + }, + { + "epoch": 0.9460251612030482, + "flos": 22568901527040.0, + "grad_norm": 2.3930690464144537, + "learning_rate": 3.043523853819807e-08, + "loss": 0.7443, + "num_input_tokens_seen": 340881790, + "step": 10490 + }, + { + "epoch": 0.9461153447265185, + "flos": 17095872250080.0, + "grad_norm": 2.3788056250526703, + "learning_rate": 3.0333791427583855e-08, + "loss": 0.7661, + "num_input_tokens_seen": 340907125, + "step": 10491 + }, + { + "epoch": 0.9462055282499887, + "flos": 19941144222240.0, + "grad_norm": 1.6834764196309644, + "learning_rate": 3.023251237975111e-08, + "loss": 0.8647, + "num_input_tokens_seen": 340934535, + "step": 10492 + }, + { + "epoch": 0.946295711773459, + "flos": 66951592010400.0, + "grad_norm": 0.753442532725477, + "learning_rate": 3.0131401403341584e-08, + "loss": 0.5931, + "num_input_tokens_seen": 341030860, + "step": 10493 + }, + { + "epoch": 0.9463858952969293, + "flos": 24064178691360.0, + "grad_norm": 1.587707884385965, + "learning_rate": 3.00304585069826e-08, + "loss": 0.7191, + "num_input_tokens_seen": 341061475, + "step": 10494 + }, + { + "epoch": 0.9464760788203995, + "flos": 20783951724000.0, + "grad_norm": 3.1751597266682436, + "learning_rate": 2.992968369928728e-08, + "loss": 0.7305, + "num_input_tokens_seen": 341090450, + "step": 10495 + }, + { + "epoch": 0.9465662623438698, + "flos": 18812195208960.0, + "grad_norm": 2.021983091598736, + "learning_rate": 2.982907698885429e-08, + "loss": 0.7575, + "num_input_tokens_seen": 341116265, + "step": 10496 + }, + { + "epoch": 0.94665644586734, + "flos": 23005380482400.0, + "grad_norm": 2.1706760308011375, + "learning_rate": 2.9728638384267645e-08, + "loss": 0.759, + "num_input_tokens_seen": 341140450, + "step": 10497 + }, + { + "epoch": 0.9467466293908103, + "flos": 18634160560320.0, + "grad_norm": 2.106474174545837, + "learning_rate": 2.962836789409784e-08, + "loss": 0.7492, + "num_input_tokens_seen": 341166710, + "step": 10498 + }, + { + "epoch": 0.9468368129142806, + "flos": 23480967758880.0, + "grad_norm": 1.670075919201644, + "learning_rate": 2.95282655268998e-08, + "loss": 0.835, + "num_input_tokens_seen": 341193200, + "step": 10499 + }, + { + "epoch": 0.9469269964377508, + "flos": 27342175473120.0, + "grad_norm": 1.8336032394209136, + "learning_rate": 2.942833129121558e-08, + "loss": 0.7436, + "num_input_tokens_seen": 341223690, + "step": 10500 + }, + { + "epoch": 0.947017179961221, + "flos": 25191640914240.0, + "grad_norm": 1.8642583880016463, + "learning_rate": 2.9328565195571475e-08, + "loss": 0.6687, + "num_input_tokens_seen": 341254330, + "step": 10501 + }, + { + "epoch": 0.9471073634846914, + "flos": 26794913432160.0, + "grad_norm": 1.7142957363516902, + "learning_rate": 2.9228967248480675e-08, + "loss": 0.8138, + "num_input_tokens_seen": 341282780, + "step": 10502 + }, + { + "epoch": 0.9471975470081616, + "flos": 15720336083520.0, + "grad_norm": 2.120176338203529, + "learning_rate": 2.912953745844082e-08, + "loss": 0.7867, + "num_input_tokens_seen": 341308775, + "step": 10503 + }, + { + "epoch": 0.9472877305316318, + "flos": 26389179601920.0, + "grad_norm": 1.7556865127559071, + "learning_rate": 2.9030275833936247e-08, + "loss": 0.732, + "num_input_tokens_seen": 341337860, + "step": 10504 + }, + { + "epoch": 0.9473779140551022, + "flos": 19653961957440.0, + "grad_norm": 2.600977376485286, + "learning_rate": 2.893118238343617e-08, + "loss": 0.6475, + "num_input_tokens_seen": 341362570, + "step": 10505 + }, + { + "epoch": 0.9474680975785724, + "flos": 23990459590080.0, + "grad_norm": 2.191954083567394, + "learning_rate": 2.8832257115396052e-08, + "loss": 0.6576, + "num_input_tokens_seen": 341391390, + "step": 10506 + }, + { + "epoch": 0.9475582811020427, + "flos": 21257383154400.0, + "grad_norm": 2.8578779126539424, + "learning_rate": 2.873350003825692e-08, + "loss": 0.725, + "num_input_tokens_seen": 341419875, + "step": 10507 + }, + { + "epoch": 0.9476484646255129, + "flos": 66514295320320.0, + "grad_norm": 0.5610645309739537, + "learning_rate": 2.8634911160444696e-08, + "loss": 0.4713, + "num_input_tokens_seen": 341511080, + "step": 10508 + }, + { + "epoch": 0.9477386481489832, + "flos": 25077847078080.0, + "grad_norm": 1.807612900855094, + "learning_rate": 2.853649049037199e-08, + "loss": 0.8017, + "num_input_tokens_seen": 341538485, + "step": 10509 + }, + { + "epoch": 0.9478288316724535, + "flos": 17906038968960.0, + "grad_norm": 2.20126339875302, + "learning_rate": 2.8438238036436525e-08, + "loss": 0.791, + "num_input_tokens_seen": 341564215, + "step": 10510 + }, + { + "epoch": 0.9479190151959237, + "flos": 30833152211520.0, + "grad_norm": 1.9499217506750093, + "learning_rate": 2.834015380702137e-08, + "loss": 0.7403, + "num_input_tokens_seen": 341592535, + "step": 10511 + }, + { + "epoch": 0.9480091987193939, + "flos": 61581465829920.0, + "grad_norm": 0.5801718279747698, + "learning_rate": 2.824223781049606e-08, + "loss": 0.5126, + "num_input_tokens_seen": 341688370, + "step": 10512 + }, + { + "epoch": 0.9480993822428643, + "flos": 17464393416960.0, + "grad_norm": 2.845056589968303, + "learning_rate": 2.8144490055215465e-08, + "loss": 0.8016, + "num_input_tokens_seen": 341714690, + "step": 10513 + }, + { + "epoch": 0.9481895657663345, + "flos": 23368735052640.0, + "grad_norm": 1.820823041187008, + "learning_rate": 2.8046910549519355e-08, + "loss": 0.7776, + "num_input_tokens_seen": 341743410, + "step": 10514 + }, + { + "epoch": 0.9482797492898047, + "flos": 26577473104320.0, + "grad_norm": 2.0651617503145263, + "learning_rate": 2.794949930173418e-08, + "loss": 0.7497, + "num_input_tokens_seen": 341775425, + "step": 10515 + }, + { + "epoch": 0.948369932813275, + "flos": 18594197334720.0, + "grad_norm": 2.0303389633635915, + "learning_rate": 2.7852256320171296e-08, + "loss": 0.7699, + "num_input_tokens_seen": 341801990, + "step": 10516 + }, + { + "epoch": 0.9484601163367453, + "flos": 29891976324000.0, + "grad_norm": 3.740682343033085, + "learning_rate": 2.775518161312851e-08, + "loss": 0.604, + "num_input_tokens_seen": 341831880, + "step": 10517 + }, + { + "epoch": 0.9485502998602156, + "flos": 26212111367040.0, + "grad_norm": 1.5794688630979243, + "learning_rate": 2.76582751888883e-08, + "loss": 0.7705, + "num_input_tokens_seen": 341863460, + "step": 10518 + }, + { + "epoch": 0.9486404833836858, + "flos": 17723766967680.0, + "grad_norm": 2.7252487908978216, + "learning_rate": 2.756153705571962e-08, + "loss": 0.7648, + "num_input_tokens_seen": 341888605, + "step": 10519 + }, + { + "epoch": 0.948730666907156, + "flos": 25559121327840.0, + "grad_norm": 1.4617251960290123, + "learning_rate": 2.74649672218763e-08, + "loss": 0.7264, + "num_input_tokens_seen": 341921065, + "step": 10520 + }, + { + "epoch": 0.9488208504306264, + "flos": 31422235966080.0, + "grad_norm": 1.7664608622582088, + "learning_rate": 2.7368565695598424e-08, + "loss": 0.7285, + "num_input_tokens_seen": 341951010, + "step": 10521 + }, + { + "epoch": 0.9489110339540966, + "flos": 27338793024960.0, + "grad_norm": 2.3548621666616945, + "learning_rate": 2.727233248511185e-08, + "loss": 0.671, + "num_input_tokens_seen": 341980090, + "step": 10522 + }, + { + "epoch": 0.9490012174775668, + "flos": 23844285159360.0, + "grad_norm": 2.337831475729406, + "learning_rate": 2.71762675986269e-08, + "loss": 0.6979, + "num_input_tokens_seen": 342006145, + "step": 10523 + }, + { + "epoch": 0.9490914010010371, + "flos": 26100845074560.0, + "grad_norm": 1.7854789592450706, + "learning_rate": 2.7080371044341242e-08, + "loss": 0.7809, + "num_input_tokens_seen": 342036480, + "step": 10524 + }, + { + "epoch": 0.9491815845245074, + "flos": 23005268973120.0, + "grad_norm": 1.928320767497161, + "learning_rate": 2.6984642830436556e-08, + "loss": 0.7737, + "num_input_tokens_seen": 342064765, + "step": 10525 + }, + { + "epoch": 0.9492717680479776, + "flos": 37761569766720.0, + "grad_norm": 1.7514370739800897, + "learning_rate": 2.688908296508141e-08, + "loss": 0.7279, + "num_input_tokens_seen": 342095740, + "step": 10526 + }, + { + "epoch": 0.9493619515714479, + "flos": 23443234718880.0, + "grad_norm": 2.005516788158466, + "learning_rate": 2.679369145642929e-08, + "loss": 0.7936, + "num_input_tokens_seen": 342123040, + "step": 10527 + }, + { + "epoch": 0.9494521350949181, + "flos": 23078913734880.0, + "grad_norm": 2.3178534530183716, + "learning_rate": 2.669846831261946e-08, + "loss": 0.7632, + "num_input_tokens_seen": 342149190, + "step": 10528 + }, + { + "epoch": 0.9495423186183884, + "flos": 17575176502560.0, + "grad_norm": 2.1750612009141648, + "learning_rate": 2.6603413541776976e-08, + "loss": 0.7341, + "num_input_tokens_seen": 342175900, + "step": 10529 + }, + { + "epoch": 0.9496325021418587, + "flos": 26537993085600.0, + "grad_norm": 1.792468298255298, + "learning_rate": 2.6508527152012683e-08, + "loss": 0.696, + "num_input_tokens_seen": 342206805, + "step": 10530 + }, + { + "epoch": 0.9497226856653289, + "flos": 31709381061120.0, + "grad_norm": 1.555541757492147, + "learning_rate": 2.641380915142233e-08, + "loss": 0.6453, + "num_input_tokens_seen": 342238325, + "step": 10531 + }, + { + "epoch": 0.9498128691887993, + "flos": 22643066665440.0, + "grad_norm": 1.8899290837501321, + "learning_rate": 2.6319259548088334e-08, + "loss": 0.7675, + "num_input_tokens_seen": 342267145, + "step": 10532 + }, + { + "epoch": 0.9499030527122695, + "flos": 20966669762400.0, + "grad_norm": 2.188760993446266, + "learning_rate": 2.6224878350077585e-08, + "loss": 0.7412, + "num_input_tokens_seen": 342297765, + "step": 10533 + }, + { + "epoch": 0.9499932362357397, + "flos": 31386361414080.0, + "grad_norm": 2.3445212557515136, + "learning_rate": 2.6130665565443633e-08, + "loss": 0.6599, + "num_input_tokens_seen": 342327135, + "step": 10534 + }, + { + "epoch": 0.95008341975921, + "flos": 19946236479360.0, + "grad_norm": 1.5792631067840528, + "learning_rate": 2.603662120222494e-08, + "loss": 0.7474, + "num_input_tokens_seen": 342355805, + "step": 10535 + }, + { + "epoch": 0.9501736032826803, + "flos": 31640679877440.0, + "grad_norm": 1.8728815190712838, + "learning_rate": 2.59427452684462e-08, + "loss": 0.5874, + "num_input_tokens_seen": 342383605, + "step": 10536 + }, + { + "epoch": 0.9502637868061505, + "flos": 27595378843680.0, + "grad_norm": 1.8408703703613254, + "learning_rate": 2.5849037772117443e-08, + "loss": 0.7189, + "num_input_tokens_seen": 342414050, + "step": 10537 + }, + { + "epoch": 0.9503539703296208, + "flos": 21873383548800.0, + "grad_norm": 2.0675087423266603, + "learning_rate": 2.575549872123384e-08, + "loss": 0.7654, + "num_input_tokens_seen": 342443300, + "step": 10538 + }, + { + "epoch": 0.950444153853091, + "flos": 22530536601120.0, + "grad_norm": 1.8099592751194549, + "learning_rate": 2.5662128123776994e-08, + "loss": 0.749, + "num_input_tokens_seen": 342470575, + "step": 10539 + }, + { + "epoch": 0.9505343373765613, + "flos": 22676934050400.0, + "grad_norm": 2.61313052804796, + "learning_rate": 2.5568925987713875e-08, + "loss": 0.7381, + "num_input_tokens_seen": 342502395, + "step": 10540 + }, + { + "epoch": 0.9506245209000316, + "flos": 18305862807360.0, + "grad_norm": 2.3313290866394607, + "learning_rate": 2.5475892320996785e-08, + "loss": 0.7142, + "num_input_tokens_seen": 342527500, + "step": 10541 + }, + { + "epoch": 0.9507147044235018, + "flos": 59869194374880.0, + "grad_norm": 0.6681935785281317, + "learning_rate": 2.5383027131564038e-08, + "loss": 0.56, + "num_input_tokens_seen": 342617755, + "step": 10542 + }, + { + "epoch": 0.950804887946972, + "flos": 21184927824960.0, + "grad_norm": 1.8298345544163972, + "learning_rate": 2.52903304273393e-08, + "loss": 0.7119, + "num_input_tokens_seen": 342646510, + "step": 10543 + }, + { + "epoch": 0.9508950714704424, + "flos": 20601828401760.0, + "grad_norm": 1.6101926929952948, + "learning_rate": 2.519780221623202e-08, + "loss": 0.7259, + "num_input_tokens_seen": 342676175, + "step": 10544 + }, + { + "epoch": 0.9509852549939126, + "flos": 31969126309440.0, + "grad_norm": 3.9901339928636306, + "learning_rate": 2.510544250613722e-08, + "loss": 0.6991, + "num_input_tokens_seen": 342706740, + "step": 10545 + }, + { + "epoch": 0.9510754385173829, + "flos": 21839887861440.0, + "grad_norm": 1.942843175986119, + "learning_rate": 2.501325130493548e-08, + "loss": 0.8406, + "num_input_tokens_seen": 342732370, + "step": 10546 + }, + { + "epoch": 0.9511656220408531, + "flos": 61348042505280.0, + "grad_norm": 0.7203260226025077, + "learning_rate": 2.4921228620493395e-08, + "loss": 0.5777, + "num_input_tokens_seen": 342821115, + "step": 10547 + }, + { + "epoch": 0.9512558055643234, + "flos": 61921955264160.0, + "grad_norm": 0.7316980472807134, + "learning_rate": 2.4829374460662244e-08, + "loss": 0.614, + "num_input_tokens_seen": 342914740, + "step": 10548 + }, + { + "epoch": 0.9513459890877937, + "flos": 24755756675040.0, + "grad_norm": 1.8933728638851823, + "learning_rate": 2.473768883327976e-08, + "loss": 0.7602, + "num_input_tokens_seen": 342944830, + "step": 10549 + }, + { + "epoch": 0.9514361726112639, + "flos": 28581201346560.0, + "grad_norm": 1.9752603697268623, + "learning_rate": 2.464617174616923e-08, + "loss": 0.7039, + "num_input_tokens_seen": 342975385, + "step": 10550 + }, + { + "epoch": 0.9515263561347341, + "flos": 23987560348800.0, + "grad_norm": 1.9786007590773362, + "learning_rate": 2.455482320713953e-08, + "loss": 0.7444, + "num_input_tokens_seen": 343004185, + "step": 10551 + }, + { + "epoch": 0.9516165396582045, + "flos": 28249223787360.0, + "grad_norm": 1.9620881371295174, + "learning_rate": 2.4463643223984643e-08, + "loss": 0.8011, + "num_input_tokens_seen": 343033370, + "step": 10552 + }, + { + "epoch": 0.9517067231816747, + "flos": 28106543314080.0, + "grad_norm": 2.3719518073591477, + "learning_rate": 2.4372631804484567e-08, + "loss": 0.8169, + "num_input_tokens_seen": 343056525, + "step": 10553 + }, + { + "epoch": 0.9517969067051449, + "flos": 25411125578880.0, + "grad_norm": 1.8450650196972465, + "learning_rate": 2.4281788956405313e-08, + "loss": 0.8259, + "num_input_tokens_seen": 343085760, + "step": 10554 + }, + { + "epoch": 0.9518870902286153, + "flos": 26280180664800.0, + "grad_norm": 2.6355545751703136, + "learning_rate": 2.4191114687497572e-08, + "loss": 0.6716, + "num_input_tokens_seen": 343113935, + "step": 10555 + }, + { + "epoch": 0.9519772737520855, + "flos": 24572815618080.0, + "grad_norm": 3.037822558697648, + "learning_rate": 2.4100609005498706e-08, + "loss": 0.7024, + "num_input_tokens_seen": 343143165, + "step": 10556 + }, + { + "epoch": 0.9520674572755558, + "flos": 25885263234720.0, + "grad_norm": 7.1747323122498035, + "learning_rate": 2.4010271918130764e-08, + "loss": 0.6739, + "num_input_tokens_seen": 343173940, + "step": 10557 + }, + { + "epoch": 0.952157640799026, + "flos": 18197458586400.0, + "grad_norm": 1.9807278014131613, + "learning_rate": 2.39201034331018e-08, + "loss": 0.705, + "num_input_tokens_seen": 343201920, + "step": 10558 + }, + { + "epoch": 0.9522478243224963, + "flos": 19399866512640.0, + "grad_norm": 7.302476844215443, + "learning_rate": 2.3830103558105663e-08, + "loss": 0.7348, + "num_input_tokens_seen": 343229935, + "step": 10559 + }, + { + "epoch": 0.9523380078459666, + "flos": 22314211366080.0, + "grad_norm": 2.165983153393043, + "learning_rate": 2.374027230082154e-08, + "loss": 0.7488, + "num_input_tokens_seen": 343255620, + "step": 10560 + }, + { + "epoch": 0.9524281913694368, + "flos": 67315466957280.0, + "grad_norm": 0.6881414585414668, + "learning_rate": 2.365060966891441e-08, + "loss": 0.5941, + "num_input_tokens_seen": 343339570, + "step": 10561 + }, + { + "epoch": 0.952518374892907, + "flos": 26066271464160.0, + "grad_norm": 1.7245598134644895, + "learning_rate": 2.3561115670034827e-08, + "loss": 0.7369, + "num_input_tokens_seen": 343372430, + "step": 10562 + }, + { + "epoch": 0.9526085584163774, + "flos": 21221657281440.0, + "grad_norm": 1.6649904484517828, + "learning_rate": 2.3471790311818675e-08, + "loss": 0.7634, + "num_input_tokens_seen": 343401300, + "step": 10563 + }, + { + "epoch": 0.9526987419398476, + "flos": 24493744071360.0, + "grad_norm": 1.6644077569686822, + "learning_rate": 2.338263360188808e-08, + "loss": 0.7325, + "num_input_tokens_seen": 343429675, + "step": 10564 + }, + { + "epoch": 0.9527889254633178, + "flos": 22381537268640.0, + "grad_norm": 2.2720303195085476, + "learning_rate": 2.329364554784985e-08, + "loss": 0.68, + "num_input_tokens_seen": 343457420, + "step": 10565 + }, + { + "epoch": 0.9528791089867881, + "flos": 19357710271200.0, + "grad_norm": 1.8894990452897507, + "learning_rate": 2.3204826157297465e-08, + "loss": 0.8026, + "num_input_tokens_seen": 343484770, + "step": 10566 + }, + { + "epoch": 0.9529692925102584, + "flos": 21804422176800.0, + "grad_norm": 1.7575564049559156, + "learning_rate": 2.3116175437809082e-08, + "loss": 0.6829, + "num_input_tokens_seen": 343513935, + "step": 10567 + }, + { + "epoch": 0.9530594760337286, + "flos": 69848467076640.0, + "grad_norm": 0.6462404649573399, + "learning_rate": 2.30276933969491e-08, + "loss": 0.5954, + "num_input_tokens_seen": 343603415, + "step": 10568 + }, + { + "epoch": 0.9531496595571989, + "flos": 25193164874400.0, + "grad_norm": 1.746349819446056, + "learning_rate": 2.2939380042267255e-08, + "loss": 0.7979, + "num_input_tokens_seen": 343633360, + "step": 10569 + }, + { + "epoch": 0.9532398430806691, + "flos": 19654928371200.0, + "grad_norm": 1.9695917176978146, + "learning_rate": 2.2851235381298627e-08, + "loss": 0.649, + "num_input_tokens_seen": 343659210, + "step": 10570 + }, + { + "epoch": 0.9533300266041395, + "flos": 23918004260640.0, + "grad_norm": 1.8736412211189404, + "learning_rate": 2.2763259421564986e-08, + "loss": 0.6697, + "num_input_tokens_seen": 343687610, + "step": 10571 + }, + { + "epoch": 0.9534202101276097, + "flos": 28725963326400.0, + "grad_norm": 1.955353038258484, + "learning_rate": 2.2675452170571873e-08, + "loss": 0.7154, + "num_input_tokens_seen": 343717685, + "step": 10572 + }, + { + "epoch": 0.9535103936510799, + "flos": 32114668854240.0, + "grad_norm": 1.9817279371580858, + "learning_rate": 2.2587813635812414e-08, + "loss": 0.7381, + "num_input_tokens_seen": 343747920, + "step": 10573 + }, + { + "epoch": 0.9536005771745502, + "flos": 21476198763360.0, + "grad_norm": 2.548658156300009, + "learning_rate": 2.2500343824763958e-08, + "loss": 0.8284, + "num_input_tokens_seen": 343777095, + "step": 10574 + }, + { + "epoch": 0.9536907606980205, + "flos": 27956912095680.0, + "grad_norm": 1.9723332206700683, + "learning_rate": 2.2413042744890088e-08, + "loss": 0.6926, + "num_input_tokens_seen": 343804935, + "step": 10575 + }, + { + "epoch": 0.9537809442214907, + "flos": 25556928312000.0, + "grad_norm": 2.2526388588293744, + "learning_rate": 2.2325910403639514e-08, + "loss": 0.7277, + "num_input_tokens_seen": 343832145, + "step": 10576 + }, + { + "epoch": 0.953871127744961, + "flos": 30618982822560.0, + "grad_norm": 1.758646517097105, + "learning_rate": 2.223894680844718e-08, + "loss": 0.6982, + "num_input_tokens_seen": 343863655, + "step": 10577 + }, + { + "epoch": 0.9539613112684313, + "flos": 33971850968160.0, + "grad_norm": 1.6057986920508094, + "learning_rate": 2.2152151966733146e-08, + "loss": 0.5776, + "num_input_tokens_seen": 343893730, + "step": 10578 + }, + { + "epoch": 0.9540514947919015, + "flos": 23115011305440.0, + "grad_norm": 1.5639250551057973, + "learning_rate": 2.2065525885903267e-08, + "loss": 0.7782, + "num_input_tokens_seen": 343924165, + "step": 10579 + }, + { + "epoch": 0.9541416783153718, + "flos": 21549174469440.0, + "grad_norm": 1.9893396105121817, + "learning_rate": 2.1979068573348747e-08, + "loss": 0.7876, + "num_input_tokens_seen": 343951315, + "step": 10580 + }, + { + "epoch": 0.954231861838842, + "flos": 25114353516000.0, + "grad_norm": 1.7586960669993839, + "learning_rate": 2.1892780036447013e-08, + "loss": 0.7045, + "num_input_tokens_seen": 343979740, + "step": 10581 + }, + { + "epoch": 0.9543220453623124, + "flos": 15902273556960.0, + "grad_norm": 2.2020740572596567, + "learning_rate": 2.1806660282560175e-08, + "loss": 0.8879, + "num_input_tokens_seen": 344003040, + "step": 10582 + }, + { + "epoch": 0.9544122288857826, + "flos": 26028129556800.0, + "grad_norm": 1.5558450501376548, + "learning_rate": 2.1720709319037024e-08, + "loss": 0.7648, + "num_input_tokens_seen": 344033455, + "step": 10583 + }, + { + "epoch": 0.9545024124092528, + "flos": 29309099919360.0, + "grad_norm": 1.887309169596186, + "learning_rate": 2.1634927153211023e-08, + "loss": 0.7076, + "num_input_tokens_seen": 344063595, + "step": 10584 + }, + { + "epoch": 0.954592595932723, + "flos": 21767432532000.0, + "grad_norm": 2.4713185143319314, + "learning_rate": 2.1549313792401437e-08, + "loss": 0.806, + "num_input_tokens_seen": 344086390, + "step": 10585 + }, + { + "epoch": 0.9546827794561934, + "flos": 24460025365440.0, + "grad_norm": 3.583919979837022, + "learning_rate": 2.1463869243913746e-08, + "loss": 0.835, + "num_input_tokens_seen": 344115035, + "step": 10586 + }, + { + "epoch": 0.9547729629796636, + "flos": 22933817227200.0, + "grad_norm": 2.0268096514297196, + "learning_rate": 2.1378593515037902e-08, + "loss": 0.7209, + "num_input_tokens_seen": 344143785, + "step": 10587 + }, + { + "epoch": 0.9548631465031339, + "flos": 23334793328160.0, + "grad_norm": 1.8516022239318994, + "learning_rate": 2.129348661305075e-08, + "loss": 0.7994, + "num_input_tokens_seen": 344171905, + "step": 10588 + }, + { + "epoch": 0.9549533300266041, + "flos": 25190488651680.0, + "grad_norm": 1.8893799899152406, + "learning_rate": 2.1208548545213813e-08, + "loss": 0.7756, + "num_input_tokens_seen": 344199185, + "step": 10589 + }, + { + "epoch": 0.9550435135500744, + "flos": 23990273741280.0, + "grad_norm": 2.8016581632298627, + "learning_rate": 2.1123779318774404e-08, + "loss": 0.6961, + "num_input_tokens_seen": 344227035, + "step": 10590 + }, + { + "epoch": 0.9551336970735447, + "flos": 23152335478080.0, + "grad_norm": 1.7034132827215829, + "learning_rate": 2.1039178940965408e-08, + "loss": 0.6879, + "num_input_tokens_seen": 344255680, + "step": 10591 + }, + { + "epoch": 0.9552238805970149, + "flos": 24898288469280.0, + "grad_norm": 1.5783795444178097, + "learning_rate": 2.0954747419005712e-08, + "loss": 0.8042, + "num_input_tokens_seen": 344283430, + "step": 10592 + }, + { + "epoch": 0.9553140641204851, + "flos": 19581729646560.0, + "grad_norm": 1.8511646986220758, + "learning_rate": 2.087048476009934e-08, + "loss": 0.7262, + "num_input_tokens_seen": 344311845, + "step": 10593 + }, + { + "epoch": 0.9554042476439555, + "flos": 23626101436320.0, + "grad_norm": 2.0298068275056433, + "learning_rate": 2.0786390971435862e-08, + "loss": 0.7175, + "num_input_tokens_seen": 344342095, + "step": 10594 + }, + { + "epoch": 0.9554944311674257, + "flos": 19509125638080.0, + "grad_norm": 2.4866385075131285, + "learning_rate": 2.070246606019088e-08, + "loss": 0.7125, + "num_input_tokens_seen": 344368735, + "step": 10595 + }, + { + "epoch": 0.955584614690896, + "flos": 20456100008160.0, + "grad_norm": 2.8143284258461008, + "learning_rate": 2.0618710033525112e-08, + "loss": 0.6554, + "num_input_tokens_seen": 344395065, + "step": 10596 + }, + { + "epoch": 0.9556747982143662, + "flos": 18379284550560.0, + "grad_norm": 1.9488339703705995, + "learning_rate": 2.053512289858528e-08, + "loss": 0.708, + "num_input_tokens_seen": 344422820, + "step": 10597 + }, + { + "epoch": 0.9557649817378365, + "flos": 21840445407840.0, + "grad_norm": 2.3067367494921305, + "learning_rate": 2.0451704662503456e-08, + "loss": 0.7831, + "num_input_tokens_seen": 344448995, + "step": 10598 + }, + { + "epoch": 0.9558551652613068, + "flos": 19873074924480.0, + "grad_norm": 1.9782201277050102, + "learning_rate": 2.0368455332397282e-08, + "loss": 0.8228, + "num_input_tokens_seen": 344473570, + "step": 10599 + }, + { + "epoch": 0.955945348784777, + "flos": 20711570734080.0, + "grad_norm": 2.1477811053393308, + "learning_rate": 2.0285374915369967e-08, + "loss": 0.7602, + "num_input_tokens_seen": 344501105, + "step": 10600 + }, + { + "epoch": 0.9560355323082473, + "flos": 25844333595360.0, + "grad_norm": 1.900045773306151, + "learning_rate": 2.020246341851073e-08, + "loss": 0.7843, + "num_input_tokens_seen": 344528260, + "step": 10601 + }, + { + "epoch": 0.9561257158317176, + "flos": 20347175410560.0, + "grad_norm": 1.6362735989634107, + "learning_rate": 2.0119720848893463e-08, + "loss": 0.786, + "num_input_tokens_seen": 344554040, + "step": 10602 + }, + { + "epoch": 0.9562158993551878, + "flos": 23951648627040.0, + "grad_norm": 1.8016073520764473, + "learning_rate": 2.0037147213578964e-08, + "loss": 0.6934, + "num_input_tokens_seen": 344581815, + "step": 10603 + }, + { + "epoch": 0.956306082878658, + "flos": 31349074411200.0, + "grad_norm": 1.6611881415085552, + "learning_rate": 1.9954742519612265e-08, + "loss": 0.769, + "num_input_tokens_seen": 344613645, + "step": 10604 + }, + { + "epoch": 0.9563962664021284, + "flos": 21622373194080.0, + "grad_norm": 1.8312934740133453, + "learning_rate": 1.9872506774024633e-08, + "loss": 0.7207, + "num_input_tokens_seen": 344640940, + "step": 10605 + }, + { + "epoch": 0.9564864499255986, + "flos": 15610593751200.0, + "grad_norm": 1.895050292341644, + "learning_rate": 1.979043998383334e-08, + "loss": 0.7514, + "num_input_tokens_seen": 344668100, + "step": 10606 + }, + { + "epoch": 0.9565766334490688, + "flos": 33355441706400.0, + "grad_norm": 1.3848906596434918, + "learning_rate": 1.970854215604034e-08, + "loss": 0.6851, + "num_input_tokens_seen": 344702400, + "step": 10607 + }, + { + "epoch": 0.9566668169725391, + "flos": 23873878021920.0, + "grad_norm": 7.571346184829823, + "learning_rate": 1.9626813297633826e-08, + "loss": 0.7122, + "num_input_tokens_seen": 344727170, + "step": 10608 + }, + { + "epoch": 0.9567570004960094, + "flos": 20419853758560.0, + "grad_norm": 2.483536680801243, + "learning_rate": 1.954525341558688e-08, + "loss": 0.6444, + "num_input_tokens_seen": 344754610, + "step": 10609 + }, + { + "epoch": 0.9568471840194797, + "flos": 21294112610880.0, + "grad_norm": 2.0802984129979003, + "learning_rate": 1.9463862516859498e-08, + "loss": 0.7072, + "num_input_tokens_seen": 344781550, + "step": 10610 + }, + { + "epoch": 0.9569373675429499, + "flos": 19797497335200.0, + "grad_norm": 1.8250062210813074, + "learning_rate": 1.938264060839545e-08, + "loss": 0.8276, + "num_input_tokens_seen": 344810055, + "step": 10611 + }, + { + "epoch": 0.9570275510664201, + "flos": 23254309330560.0, + "grad_norm": 2.0349369333603207, + "learning_rate": 1.9301587697126086e-08, + "loss": 0.7159, + "num_input_tokens_seen": 344833650, + "step": 10612 + }, + { + "epoch": 0.9571177345898905, + "flos": 24864309575040.0, + "grad_norm": 2.1619298458944813, + "learning_rate": 1.9220703789966318e-08, + "loss": 0.6954, + "num_input_tokens_seen": 344862080, + "step": 10613 + }, + { + "epoch": 0.9572079181133607, + "flos": 23407434506400.0, + "grad_norm": 1.8744123173656173, + "learning_rate": 1.913998889381818e-08, + "loss": 0.6987, + "num_input_tokens_seen": 344890910, + "step": 10614 + }, + { + "epoch": 0.9572981016368309, + "flos": 19545780755040.0, + "grad_norm": 1.8670739813736914, + "learning_rate": 1.9059443015568387e-08, + "loss": 0.756, + "num_input_tokens_seen": 344917090, + "step": 10615 + }, + { + "epoch": 0.9573882851603012, + "flos": 22970286495360.0, + "grad_norm": 2.147740809447036, + "learning_rate": 1.8979066162089884e-08, + "loss": 0.7549, + "num_input_tokens_seen": 344945390, + "step": 10616 + }, + { + "epoch": 0.9574784686837715, + "flos": 24646200191520.0, + "grad_norm": 2.0781402587330433, + "learning_rate": 1.889885834024052e-08, + "loss": 0.7476, + "num_input_tokens_seen": 344973345, + "step": 10617 + }, + { + "epoch": 0.9575686522072417, + "flos": 26284975563840.0, + "grad_norm": 1.8206626326144848, + "learning_rate": 1.8818819556864374e-08, + "loss": 0.693, + "num_input_tokens_seen": 345001090, + "step": 10618 + }, + { + "epoch": 0.957658835730712, + "flos": 13461248624640.0, + "grad_norm": 2.535018000190828, + "learning_rate": 1.873894981879065e-08, + "loss": 0.7226, + "num_input_tokens_seen": 345025695, + "step": 10619 + }, + { + "epoch": 0.9577490192541822, + "flos": 70560116730720.0, + "grad_norm": 0.6272558034245997, + "learning_rate": 1.8659249132834342e-08, + "loss": 0.5899, + "num_input_tokens_seen": 345123095, + "step": 10620 + }, + { + "epoch": 0.9578392027776526, + "flos": 21913978660320.0, + "grad_norm": 1.6811357508983646, + "learning_rate": 1.857971750579579e-08, + "loss": 0.7235, + "num_input_tokens_seen": 345152750, + "step": 10621 + }, + { + "epoch": 0.9579293863011228, + "flos": 33019226794560.0, + "grad_norm": 2.801073811891487, + "learning_rate": 1.8500354944461116e-08, + "loss": 0.6235, + "num_input_tokens_seen": 345182475, + "step": 10622 + }, + { + "epoch": 0.958019569824593, + "flos": 24171281970720.0, + "grad_norm": 1.7028356883849871, + "learning_rate": 1.8421161455602242e-08, + "loss": 0.7487, + "num_input_tokens_seen": 345214180, + "step": 10623 + }, + { + "epoch": 0.9581097533480634, + "flos": 18124371371040.0, + "grad_norm": 1.8997846933141997, + "learning_rate": 1.834213704597598e-08, + "loss": 0.7668, + "num_input_tokens_seen": 345241860, + "step": 10624 + }, + { + "epoch": 0.9581999368715336, + "flos": 21798512184960.0, + "grad_norm": 4.2381185614656935, + "learning_rate": 1.8263281722325385e-08, + "loss": 0.7532, + "num_input_tokens_seen": 345271470, + "step": 10625 + }, + { + "epoch": 0.9582901203950038, + "flos": 24354594725280.0, + "grad_norm": 2.036251542176845, + "learning_rate": 1.818459549137885e-08, + "loss": 0.7512, + "num_input_tokens_seen": 345298945, + "step": 10626 + }, + { + "epoch": 0.9583803039184741, + "flos": 16885903044000.0, + "grad_norm": 1.9530105557290736, + "learning_rate": 1.8106078359850117e-08, + "loss": 0.7314, + "num_input_tokens_seen": 345324600, + "step": 10627 + }, + { + "epoch": 0.9584704874419444, + "flos": 27597088652640.0, + "grad_norm": 2.244726677209766, + "learning_rate": 1.802773033443894e-08, + "loss": 0.7589, + "num_input_tokens_seen": 345353965, + "step": 10628 + }, + { + "epoch": 0.9585606709654146, + "flos": 27888768458400.0, + "grad_norm": 3.3903555345471905, + "learning_rate": 1.7949551421830413e-08, + "loss": 0.7085, + "num_input_tokens_seen": 345383060, + "step": 10629 + }, + { + "epoch": 0.9586508544888849, + "flos": 27050904534720.0, + "grad_norm": 1.7267349640338108, + "learning_rate": 1.7871541628694752e-08, + "loss": 0.7096, + "num_input_tokens_seen": 345411985, + "step": 10630 + }, + { + "epoch": 0.9587410380123551, + "flos": 25993890474240.0, + "grad_norm": 6.159849092691035, + "learning_rate": 1.779370096168864e-08, + "loss": 0.652, + "num_input_tokens_seen": 345441220, + "step": 10631 + }, + { + "epoch": 0.9588312215358255, + "flos": 24093994572480.0, + "grad_norm": 2.810297718363416, + "learning_rate": 1.771602942745387e-08, + "loss": 0.7052, + "num_input_tokens_seen": 345470160, + "step": 10632 + }, + { + "epoch": 0.9589214050592957, + "flos": 22237593023520.0, + "grad_norm": 3.2519661751333495, + "learning_rate": 1.763852703261759e-08, + "loss": 0.7281, + "num_input_tokens_seen": 345495170, + "step": 10633 + }, + { + "epoch": 0.9590115885827659, + "flos": 30730063266240.0, + "grad_norm": 2.774582403568269, + "learning_rate": 1.756119378379295e-08, + "loss": 0.8043, + "num_input_tokens_seen": 345527040, + "step": 10634 + }, + { + "epoch": 0.9591017721062362, + "flos": 22715521994880.0, + "grad_norm": 1.7992252404397764, + "learning_rate": 1.7484029687578005e-08, + "loss": 0.8344, + "num_input_tokens_seen": 345555870, + "step": 10635 + }, + { + "epoch": 0.9591919556297065, + "flos": 22168631651520.0, + "grad_norm": 1.861790154384815, + "learning_rate": 1.740703475055727e-08, + "loss": 0.7159, + "num_input_tokens_seen": 345585240, + "step": 10636 + }, + { + "epoch": 0.9592821391531767, + "flos": 21730480056960.0, + "grad_norm": 1.7119518739362114, + "learning_rate": 1.7330208979300153e-08, + "loss": 0.7799, + "num_input_tokens_seen": 345613430, + "step": 10637 + }, + { + "epoch": 0.959372322676647, + "flos": 25957384036320.0, + "grad_norm": 1.6038234475637274, + "learning_rate": 1.725355238036208e-08, + "loss": 0.784, + "num_input_tokens_seen": 345644155, + "step": 10638 + }, + { + "epoch": 0.9594625062001172, + "flos": 26321370492480.0, + "grad_norm": 2.0651022937921706, + "learning_rate": 1.7177064960283594e-08, + "loss": 0.7851, + "num_input_tokens_seen": 345670765, + "step": 10639 + }, + { + "epoch": 0.9595526897235875, + "flos": 18670220961120.0, + "grad_norm": 2.32958006991701, + "learning_rate": 1.7100746725591253e-08, + "loss": 0.7384, + "num_input_tokens_seen": 345697290, + "step": 10640 + }, + { + "epoch": 0.9596428732470578, + "flos": 37943395730880.0, + "grad_norm": 3.1896536040502212, + "learning_rate": 1.7024597682796517e-08, + "loss": 0.7352, + "num_input_tokens_seen": 345725515, + "step": 10641 + }, + { + "epoch": 0.959733056770528, + "flos": 24245261260320.0, + "grad_norm": 1.911771064419766, + "learning_rate": 1.6948617838397293e-08, + "loss": 0.8163, + "num_input_tokens_seen": 345755000, + "step": 10642 + }, + { + "epoch": 0.9598232402939982, + "flos": 21184630466880.0, + "grad_norm": 2.3678100573036387, + "learning_rate": 1.6872807198876404e-08, + "loss": 0.7711, + "num_input_tokens_seen": 345785665, + "step": 10643 + }, + { + "epoch": 0.9599134238174686, + "flos": 19763964478080.0, + "grad_norm": 2.0587487062914547, + "learning_rate": 1.679716577070245e-08, + "loss": 0.8407, + "num_input_tokens_seen": 345813980, + "step": 10644 + }, + { + "epoch": 0.9600036073409388, + "flos": 17432793387360.0, + "grad_norm": 2.047715458601549, + "learning_rate": 1.6721693560329596e-08, + "loss": 0.829, + "num_input_tokens_seen": 345839880, + "step": 10645 + }, + { + "epoch": 0.960093790864409, + "flos": 24244480695360.0, + "grad_norm": 17.146930991627645, + "learning_rate": 1.6646390574197366e-08, + "loss": 0.6528, + "num_input_tokens_seen": 345869135, + "step": 10646 + }, + { + "epoch": 0.9601839743878793, + "flos": 60923978249760.0, + "grad_norm": 0.6865080700127028, + "learning_rate": 1.6571256818731504e-08, + "loss": 0.5961, + "num_input_tokens_seen": 345940045, + "step": 10647 + }, + { + "epoch": 0.9602741579113496, + "flos": 26248320446880.0, + "grad_norm": 1.8636396211396906, + "learning_rate": 1.6496292300342218e-08, + "loss": 0.677, + "num_input_tokens_seen": 345968105, + "step": 10648 + }, + { + "epoch": 0.9603643414348199, + "flos": 28726706721600.0, + "grad_norm": 2.1894721813886906, + "learning_rate": 1.642149702542639e-08, + "loss": 0.6808, + "num_input_tokens_seen": 345996555, + "step": 10649 + }, + { + "epoch": 0.9604545249582901, + "flos": 22276366816800.0, + "grad_norm": 1.5737713729461498, + "learning_rate": 1.634687100036558e-08, + "loss": 0.7042, + "num_input_tokens_seen": 346025255, + "step": 10650 + }, + { + "epoch": 0.9605447084817604, + "flos": 25082418958560.0, + "grad_norm": 2.0780236929386313, + "learning_rate": 1.627241423152781e-08, + "loss": 0.6798, + "num_input_tokens_seen": 346052665, + "step": 10651 + }, + { + "epoch": 0.9606348920052307, + "flos": 20521195725120.0, + "grad_norm": 6.949795660953199, + "learning_rate": 1.619812672526555e-08, + "loss": 0.7945, + "num_input_tokens_seen": 346078560, + "step": 10652 + }, + { + "epoch": 0.9607250755287009, + "flos": 16594855124160.0, + "grad_norm": 2.189317220915879, + "learning_rate": 1.6124008487917727e-08, + "loss": 0.7271, + "num_input_tokens_seen": 346104095, + "step": 10653 + }, + { + "epoch": 0.9608152590521711, + "flos": 70655994742560.0, + "grad_norm": 0.6397527944452359, + "learning_rate": 1.6050059525808623e-08, + "loss": 0.5937, + "num_input_tokens_seen": 346189615, + "step": 10654 + }, + { + "epoch": 0.9609054425756415, + "flos": 65701266529920.0, + "grad_norm": 0.6269279070859841, + "learning_rate": 1.597627984524763e-08, + "loss": 0.589, + "num_input_tokens_seen": 346284535, + "step": 10655 + }, + { + "epoch": 0.9609956260991117, + "flos": 23735695089600.0, + "grad_norm": 2.5781225955836327, + "learning_rate": 1.590266945253038e-08, + "loss": 0.7787, + "num_input_tokens_seen": 346314125, + "step": 10656 + }, + { + "epoch": 0.9610858096225819, + "flos": 18741263839680.0, + "grad_norm": 1.5596151040941804, + "learning_rate": 1.582922835393763e-08, + "loss": 0.7566, + "num_input_tokens_seen": 346341655, + "step": 10657 + }, + { + "epoch": 0.9611759931460522, + "flos": 18015818471040.0, + "grad_norm": 1.940727928982547, + "learning_rate": 1.5755956555735473e-08, + "loss": 0.7203, + "num_input_tokens_seen": 346368530, + "step": 10658 + }, + { + "epoch": 0.9612661766695225, + "flos": 56500634856960.0, + "grad_norm": 0.7311955880029906, + "learning_rate": 1.5682854064176244e-08, + "loss": 0.6643, + "num_input_tokens_seen": 346451655, + "step": 10659 + }, + { + "epoch": 0.9613563601929928, + "flos": 21585755246880.0, + "grad_norm": 1.7170521065787558, + "learning_rate": 1.5609920885497395e-08, + "loss": 0.6761, + "num_input_tokens_seen": 346480530, + "step": 10660 + }, + { + "epoch": 0.961446543716463, + "flos": 25515366786720.0, + "grad_norm": 2.343695372898544, + "learning_rate": 1.5537157025921732e-08, + "loss": 0.7634, + "num_input_tokens_seen": 346510105, + "step": 10661 + }, + { + "epoch": 0.9615367272399332, + "flos": 23589148961280.0, + "grad_norm": 2.3188871893180676, + "learning_rate": 1.5464562491658285e-08, + "loss": 0.7268, + "num_input_tokens_seen": 346538150, + "step": 10662 + }, + { + "epoch": 0.9616269107634036, + "flos": 21804756704640.0, + "grad_norm": 1.5672735158300357, + "learning_rate": 1.5392137288900764e-08, + "loss": 0.7795, + "num_input_tokens_seen": 346566330, + "step": 10663 + }, + { + "epoch": 0.9617170942868738, + "flos": 70482160276800.0, + "grad_norm": 0.592364602291686, + "learning_rate": 1.531988142382934e-08, + "loss": 0.5793, + "num_input_tokens_seen": 346662455, + "step": 10664 + }, + { + "epoch": 0.961807277810344, + "flos": 46431294093120.0, + "grad_norm": 3.1707012842447226, + "learning_rate": 1.5247794902608634e-08, + "loss": 0.6743, + "num_input_tokens_seen": 346695700, + "step": 10665 + }, + { + "epoch": 0.9618974613338143, + "flos": 22744854669120.0, + "grad_norm": 2.56609343737273, + "learning_rate": 1.5175877731390398e-08, + "loss": 0.7013, + "num_input_tokens_seen": 346724570, + "step": 10666 + }, + { + "epoch": 0.9619876448572846, + "flos": 30872557890720.0, + "grad_norm": 2.0484509407260107, + "learning_rate": 1.510412991631016e-08, + "loss": 0.7701, + "num_input_tokens_seen": 346752665, + "step": 10667 + }, + { + "epoch": 0.9620778283807548, + "flos": 30329310183840.0, + "grad_norm": 1.5510493510380219, + "learning_rate": 1.503255146349014e-08, + "loss": 0.7685, + "num_input_tokens_seen": 346784400, + "step": 10668 + }, + { + "epoch": 0.9621680119042251, + "flos": 24937136602080.0, + "grad_norm": 2.680988051772394, + "learning_rate": 1.4961142379037893e-08, + "loss": 0.7054, + "num_input_tokens_seen": 346813835, + "step": 10669 + }, + { + "epoch": 0.9622581954276953, + "flos": 22460162778240.0, + "grad_norm": 1.7386765619890558, + "learning_rate": 1.4889902669046327e-08, + "loss": 0.782, + "num_input_tokens_seen": 346844180, + "step": 10670 + }, + { + "epoch": 0.9623483789511657, + "flos": 15829558039200.0, + "grad_norm": 2.4299376135144932, + "learning_rate": 1.4818832339594135e-08, + "loss": 0.7418, + "num_input_tokens_seen": 346868995, + "step": 10671 + }, + { + "epoch": 0.9624385624746359, + "flos": 21766280269440.0, + "grad_norm": 1.8294267502826282, + "learning_rate": 1.474793139674535e-08, + "loss": 0.6941, + "num_input_tokens_seen": 346898950, + "step": 10672 + }, + { + "epoch": 0.9625287459981061, + "flos": 29705950176960.0, + "grad_norm": 2.0601932457215746, + "learning_rate": 1.4677199846549581e-08, + "loss": 0.7397, + "num_input_tokens_seen": 346929955, + "step": 10673 + }, + { + "epoch": 0.9626189295215765, + "flos": 22387372920960.0, + "grad_norm": 1.9557974270533067, + "learning_rate": 1.4606637695042224e-08, + "loss": 0.7377, + "num_input_tokens_seen": 346957905, + "step": 10674 + }, + { + "epoch": 0.9627091130450467, + "flos": 23225050995840.0, + "grad_norm": 2.29427912375388, + "learning_rate": 1.4536244948243793e-08, + "loss": 0.8329, + "num_input_tokens_seen": 346982835, + "step": 10675 + }, + { + "epoch": 0.9627992965685169, + "flos": 25080263112480.0, + "grad_norm": 2.3427188893972684, + "learning_rate": 1.4466021612160595e-08, + "loss": 0.7118, + "num_input_tokens_seen": 347012550, + "step": 10676 + }, + { + "epoch": 0.9628894800919872, + "flos": 15465088376160.0, + "grad_norm": 1.8647685962327143, + "learning_rate": 1.4395967692784505e-08, + "loss": 0.7112, + "num_input_tokens_seen": 347040060, + "step": 10677 + }, + { + "epoch": 0.9629796636154575, + "flos": 33681434934240.0, + "grad_norm": 1.9049759030757945, + "learning_rate": 1.4326083196092963e-08, + "loss": 0.6728, + "num_input_tokens_seen": 347070090, + "step": 10678 + }, + { + "epoch": 0.9630698471389277, + "flos": 24281804868000.0, + "grad_norm": 1.6807091023662404, + "learning_rate": 1.42563681280492e-08, + "loss": 0.7501, + "num_input_tokens_seen": 347099000, + "step": 10679 + }, + { + "epoch": 0.963160030662398, + "flos": 21330619048800.0, + "grad_norm": 1.8770649665275476, + "learning_rate": 1.4186822494600902e-08, + "loss": 0.7156, + "num_input_tokens_seen": 347127210, + "step": 10680 + }, + { + "epoch": 0.9632502141858682, + "flos": 21948923968320.0, + "grad_norm": 2.107400379654625, + "learning_rate": 1.4117446301682877e-08, + "loss": 0.7287, + "num_input_tokens_seen": 347159370, + "step": 10681 + }, + { + "epoch": 0.9633403977093385, + "flos": 30146443466400.0, + "grad_norm": 2.0253534267618325, + "learning_rate": 1.4048239555214392e-08, + "loss": 0.6952, + "num_input_tokens_seen": 347191060, + "step": 10682 + }, + { + "epoch": 0.9634305812328088, + "flos": 24537349933440.0, + "grad_norm": 1.671099759341185, + "learning_rate": 1.3979202261100497e-08, + "loss": 0.7729, + "num_input_tokens_seen": 347220650, + "step": 10683 + }, + { + "epoch": 0.963520764756279, + "flos": 23261371584960.0, + "grad_norm": 4.254992525876358, + "learning_rate": 1.3910334425231817e-08, + "loss": 0.8034, + "num_input_tokens_seen": 347248360, + "step": 10684 + }, + { + "epoch": 0.9636109482797492, + "flos": 24715458921600.0, + "grad_norm": 1.8858734195824225, + "learning_rate": 1.384163605348454e-08, + "loss": 0.7381, + "num_input_tokens_seen": 347279190, + "step": 10685 + }, + { + "epoch": 0.9637011318032196, + "flos": 17537480632320.0, + "grad_norm": 2.216630917693484, + "learning_rate": 1.3773107151720642e-08, + "loss": 0.7966, + "num_input_tokens_seen": 347305820, + "step": 10686 + }, + { + "epoch": 0.9637913153266898, + "flos": 22529458678080.0, + "grad_norm": 1.882684093861164, + "learning_rate": 1.3704747725787003e-08, + "loss": 0.8037, + "num_input_tokens_seen": 347333080, + "step": 10687 + }, + { + "epoch": 0.9638814988501601, + "flos": 57607616542080.0, + "grad_norm": 0.7310131716526517, + "learning_rate": 1.3636557781516512e-08, + "loss": 0.5849, + "num_input_tokens_seen": 347427990, + "step": 10688 + }, + { + "epoch": 0.9639716823736303, + "flos": 34847039064480.0, + "grad_norm": 1.801622239881659, + "learning_rate": 1.3568537324727847e-08, + "loss": 0.7432, + "num_input_tokens_seen": 347456150, + "step": 10689 + }, + { + "epoch": 0.9640618658971006, + "flos": 32989448083200.0, + "grad_norm": 1.781438382543772, + "learning_rate": 1.3500686361224589e-08, + "loss": 0.7382, + "num_input_tokens_seen": 347485125, + "step": 10690 + }, + { + "epoch": 0.9641520494205709, + "flos": 16339644586560.0, + "grad_norm": 2.332994220705228, + "learning_rate": 1.3433004896796108e-08, + "loss": 0.7343, + "num_input_tokens_seen": 347510990, + "step": 10691 + }, + { + "epoch": 0.9642422329440411, + "flos": 24714157980000.0, + "grad_norm": 1.6323093204805459, + "learning_rate": 1.336549293721756e-08, + "loss": 0.6323, + "num_input_tokens_seen": 347538520, + "step": 10692 + }, + { + "epoch": 0.9643324164675113, + "flos": 63280684965600.0, + "grad_norm": 0.6604011126671191, + "learning_rate": 1.3298150488249227e-08, + "loss": 0.5715, + "num_input_tokens_seen": 347625785, + "step": 10693 + }, + { + "epoch": 0.9644225999909817, + "flos": 18743791383360.0, + "grad_norm": 2.5303082068059304, + "learning_rate": 1.3230977555637401e-08, + "loss": 0.8143, + "num_input_tokens_seen": 347651880, + "step": 10694 + }, + { + "epoch": 0.9645127835144519, + "flos": 27122690808480.0, + "grad_norm": 1.5245039117349426, + "learning_rate": 1.3163974145113499e-08, + "loss": 0.6959, + "num_input_tokens_seen": 347683365, + "step": 10695 + }, + { + "epoch": 0.9646029670379221, + "flos": 35538988745760.0, + "grad_norm": 2.379323036502196, + "learning_rate": 1.3097140262394723e-08, + "loss": 0.6411, + "num_input_tokens_seen": 347712900, + "step": 10696 + }, + { + "epoch": 0.9646931505613925, + "flos": 16375593478080.0, + "grad_norm": 2.326252741332901, + "learning_rate": 1.303047591318318e-08, + "loss": 0.774, + "num_input_tokens_seen": 347737335, + "step": 10697 + }, + { + "epoch": 0.9647833340848627, + "flos": 35502630986880.0, + "grad_norm": 1.77804611916046, + "learning_rate": 1.2963981103167875e-08, + "loss": 0.6722, + "num_input_tokens_seen": 347765760, + "step": 10698 + }, + { + "epoch": 0.964873517608333, + "flos": 19904451935520.0, + "grad_norm": 1.5172560262869939, + "learning_rate": 1.2897655838021825e-08, + "loss": 0.724, + "num_input_tokens_seen": 347793270, + "step": 10699 + }, + { + "epoch": 0.9649637011318032, + "flos": 42935113588320.0, + "grad_norm": 2.3393361555212735, + "learning_rate": 1.2831500123404726e-08, + "loss": 0.7482, + "num_input_tokens_seen": 347826640, + "step": 10700 + }, + { + "epoch": 0.9650538846552735, + "flos": 25702396517280.0, + "grad_norm": 2.4340264439333534, + "learning_rate": 1.2765513964961172e-08, + "loss": 0.8895, + "num_input_tokens_seen": 347852240, + "step": 10701 + }, + { + "epoch": 0.9651440681787438, + "flos": 22095841794240.0, + "grad_norm": 1.8959244786524758, + "learning_rate": 1.2699697368321549e-08, + "loss": 0.7365, + "num_input_tokens_seen": 347880945, + "step": 10702 + }, + { + "epoch": 0.965234251702214, + "flos": 25810280361600.0, + "grad_norm": 1.9190040792925773, + "learning_rate": 1.2634050339101366e-08, + "loss": 0.7266, + "num_input_tokens_seen": 347912675, + "step": 10703 + }, + { + "epoch": 0.9653244352256842, + "flos": 27814937847840.0, + "grad_norm": 1.5351133073619743, + "learning_rate": 1.2568572882902361e-08, + "loss": 0.718, + "num_input_tokens_seen": 347944535, + "step": 10704 + }, + { + "epoch": 0.9654146187491546, + "flos": 26135827552320.0, + "grad_norm": 1.6311732190065078, + "learning_rate": 1.2503265005311402e-08, + "loss": 0.7131, + "num_input_tokens_seen": 347973890, + "step": 10705 + }, + { + "epoch": 0.9655048022726248, + "flos": 21658284915840.0, + "grad_norm": 2.4973485237443036, + "learning_rate": 1.2438126711900698e-08, + "loss": 0.7073, + "num_input_tokens_seen": 348000205, + "step": 10706 + }, + { + "epoch": 0.965594985796095, + "flos": 44029117293600.0, + "grad_norm": 1.5894573970051953, + "learning_rate": 1.2373158008228247e-08, + "loss": 0.5899, + "num_input_tokens_seen": 348034140, + "step": 10707 + }, + { + "epoch": 0.9656851693195653, + "flos": 24135816286080.0, + "grad_norm": 1.7036579942944645, + "learning_rate": 1.2308358899837833e-08, + "loss": 0.7138, + "num_input_tokens_seen": 348064360, + "step": 10708 + }, + { + "epoch": 0.9657753528430356, + "flos": 20669191474080.0, + "grad_norm": 2.0160100997792916, + "learning_rate": 1.224372939225815e-08, + "loss": 0.7205, + "num_input_tokens_seen": 348092180, + "step": 10709 + }, + { + "epoch": 0.9658655363665059, + "flos": 24646163021760.0, + "grad_norm": 1.9583451967353498, + "learning_rate": 1.2179269491003674e-08, + "loss": 0.8225, + "num_input_tokens_seen": 348121495, + "step": 10710 + }, + { + "epoch": 0.9659557198899761, + "flos": 22424362565760.0, + "grad_norm": 1.6861793153735904, + "learning_rate": 1.2114979201574894e-08, + "loss": 0.6146, + "num_input_tokens_seen": 348149755, + "step": 10711 + }, + { + "epoch": 0.9660459034134463, + "flos": 54519777484320.0, + "grad_norm": 1.515980472368473, + "learning_rate": 1.2050858529456975e-08, + "loss": 0.7568, + "num_input_tokens_seen": 348185990, + "step": 10712 + }, + { + "epoch": 0.9661360869369167, + "flos": 23554017804480.0, + "grad_norm": 2.1616651566303346, + "learning_rate": 1.1986907480121545e-08, + "loss": 0.7419, + "num_input_tokens_seen": 348217410, + "step": 10713 + }, + { + "epoch": 0.9662262704603869, + "flos": 27123397033920.0, + "grad_norm": 1.8000383478954116, + "learning_rate": 1.192312605902468e-08, + "loss": 0.7615, + "num_input_tokens_seen": 348248360, + "step": 10714 + }, + { + "epoch": 0.9663164539838571, + "flos": 70783509953760.0, + "grad_norm": 0.796101288015749, + "learning_rate": 1.1859514271608917e-08, + "loss": 0.6018, + "num_input_tokens_seen": 348325990, + "step": 10715 + }, + { + "epoch": 0.9664066375073274, + "flos": 16877353999200.0, + "grad_norm": 2.82529446915412, + "learning_rate": 1.1796072123301914e-08, + "loss": 0.7761, + "num_input_tokens_seen": 348348765, + "step": 10716 + }, + { + "epoch": 0.9664968210307977, + "flos": 65884096077600.0, + "grad_norm": 0.6355299353613384, + "learning_rate": 1.1732799619516897e-08, + "loss": 0.5807, + "num_input_tokens_seen": 348435870, + "step": 10717 + }, + { + "epoch": 0.9665870045542679, + "flos": 12440926850880.0, + "grad_norm": 2.0968723075383235, + "learning_rate": 1.1669696765652659e-08, + "loss": 0.7315, + "num_input_tokens_seen": 348462675, + "step": 10718 + }, + { + "epoch": 0.9666771880777382, + "flos": 22784186008800.0, + "grad_norm": 1.8458674927845116, + "learning_rate": 1.1606763567093336e-08, + "loss": 0.7699, + "num_input_tokens_seen": 348490885, + "step": 10719 + }, + { + "epoch": 0.9667673716012085, + "flos": 59716775424480.0, + "grad_norm": 0.6897121731657702, + "learning_rate": 1.1544000029208857e-08, + "loss": 0.571, + "num_input_tokens_seen": 348575700, + "step": 10720 + }, + { + "epoch": 0.9668575551246787, + "flos": 25411162748640.0, + "grad_norm": 1.818100071824993, + "learning_rate": 1.148140615735449e-08, + "loss": 0.6055, + "num_input_tokens_seen": 348605280, + "step": 10721 + }, + { + "epoch": 0.966947738648149, + "flos": 22966011972960.0, + "grad_norm": 1.8406096401124221, + "learning_rate": 1.1418981956871076e-08, + "loss": 0.715, + "num_input_tokens_seen": 348633040, + "step": 10722 + }, + { + "epoch": 0.9670379221716192, + "flos": 17906076138720.0, + "grad_norm": 2.7628369259668735, + "learning_rate": 1.1356727433085245e-08, + "loss": 0.7249, + "num_input_tokens_seen": 348659655, + "step": 10723 + }, + { + "epoch": 0.9671281056950896, + "flos": 21804496516320.0, + "grad_norm": 1.5286485800861938, + "learning_rate": 1.1294642591308524e-08, + "loss": 0.7677, + "num_input_tokens_seen": 348688605, + "step": 10724 + }, + { + "epoch": 0.9672182892185598, + "flos": 24646869247200.0, + "grad_norm": 1.565207005664444, + "learning_rate": 1.1232727436838452e-08, + "loss": 0.7344, + "num_input_tokens_seen": 348719495, + "step": 10725 + }, + { + "epoch": 0.96730847274203, + "flos": 18956213793600.0, + "grad_norm": 2.0449919447238893, + "learning_rate": 1.1170981974958138e-08, + "loss": 0.7866, + "num_input_tokens_seen": 348746285, + "step": 10726 + }, + { + "epoch": 0.9673986562655003, + "flos": 19761362594880.0, + "grad_norm": 1.6440645765397677, + "learning_rate": 1.1109406210936035e-08, + "loss": 0.8034, + "num_input_tokens_seen": 348774325, + "step": 10727 + }, + { + "epoch": 0.9674888397889706, + "flos": 17906261987520.0, + "grad_norm": 3.8317759017561683, + "learning_rate": 1.1048000150025939e-08, + "loss": 0.862, + "num_input_tokens_seen": 348802345, + "step": 10728 + }, + { + "epoch": 0.9675790233124408, + "flos": 22127776351680.0, + "grad_norm": 3.9633982461586936, + "learning_rate": 1.0986763797467213e-08, + "loss": 0.6359, + "num_input_tokens_seen": 348829440, + "step": 10729 + }, + { + "epoch": 0.9676692068359111, + "flos": 58181008924320.0, + "grad_norm": 0.5840994680021816, + "learning_rate": 1.0925697158485459e-08, + "loss": 0.5425, + "num_input_tokens_seen": 348926050, + "step": 10730 + }, + { + "epoch": 0.9677593903593813, + "flos": 31492052242560.0, + "grad_norm": 1.75506533107174, + "learning_rate": 1.0864800238290727e-08, + "loss": 0.7438, + "num_input_tokens_seen": 348957950, + "step": 10731 + }, + { + "epoch": 0.9678495738828516, + "flos": 16630692506400.0, + "grad_norm": 2.1010265021343493, + "learning_rate": 1.0804073042079309e-08, + "loss": 0.7048, + "num_input_tokens_seen": 348984000, + "step": 10732 + }, + { + "epoch": 0.9679397574063219, + "flos": 65377800845760.0, + "grad_norm": 0.6380737490346655, + "learning_rate": 1.0743515575032392e-08, + "loss": 0.5886, + "num_input_tokens_seen": 349072685, + "step": 10733 + }, + { + "epoch": 0.9680299409297921, + "flos": 32625015589920.0, + "grad_norm": 2.091350261977636, + "learning_rate": 1.0683127842317619e-08, + "loss": 0.7507, + "num_input_tokens_seen": 349103090, + "step": 10734 + }, + { + "epoch": 0.9681201244532623, + "flos": 28324912885920.0, + "grad_norm": 2.175363825918774, + "learning_rate": 1.0622909849087314e-08, + "loss": 0.7721, + "num_input_tokens_seen": 349130805, + "step": 10735 + }, + { + "epoch": 0.9682103079767327, + "flos": 21360211911360.0, + "grad_norm": 1.7887882872082004, + "learning_rate": 1.0562861600479588e-08, + "loss": 0.8034, + "num_input_tokens_seen": 349158915, + "step": 10736 + }, + { + "epoch": 0.9683004915002029, + "flos": 22970174986080.0, + "grad_norm": 2.45796105965916, + "learning_rate": 1.0502983101618345e-08, + "loss": 0.8358, + "num_input_tokens_seen": 349187030, + "step": 10737 + }, + { + "epoch": 0.9683906750236732, + "flos": 34883471162880.0, + "grad_norm": 1.986212953062365, + "learning_rate": 1.0443274357612386e-08, + "loss": 0.6239, + "num_input_tokens_seen": 349218130, + "step": 10738 + }, + { + "epoch": 0.9684808585471434, + "flos": 25484287133760.0, + "grad_norm": 1.8250265354683406, + "learning_rate": 1.0383735373556524e-08, + "loss": 0.7574, + "num_input_tokens_seen": 349247270, + "step": 10739 + }, + { + "epoch": 0.9685710420706137, + "flos": 25592914373280.0, + "grad_norm": 1.5416339750518142, + "learning_rate": 1.0324366154531139e-08, + "loss": 0.7701, + "num_input_tokens_seen": 349278780, + "step": 10740 + }, + { + "epoch": 0.968661225594084, + "flos": 12878149201440.0, + "grad_norm": 2.5490596271483272, + "learning_rate": 1.0265166705601735e-08, + "loss": 0.7037, + "num_input_tokens_seen": 349304145, + "step": 10741 + }, + { + "epoch": 0.9687514091175542, + "flos": 71137268992320.0, + "grad_norm": 0.6809525391597228, + "learning_rate": 1.0206137031819606e-08, + "loss": 0.6101, + "num_input_tokens_seen": 349397265, + "step": 10742 + }, + { + "epoch": 0.9688415926410244, + "flos": 25229373954240.0, + "grad_norm": 1.7495878960529576, + "learning_rate": 1.0147277138221388e-08, + "loss": 0.7529, + "num_input_tokens_seen": 349427695, + "step": 10743 + }, + { + "epoch": 0.9689317761644948, + "flos": 21985913613120.0, + "grad_norm": 2.2694599186361812, + "learning_rate": 1.0088587029829287e-08, + "loss": 0.7757, + "num_input_tokens_seen": 349454770, + "step": 10744 + }, + { + "epoch": 0.969021959687965, + "flos": 23298472739040.0, + "grad_norm": 1.6556566003107245, + "learning_rate": 1.003006671165152e-08, + "loss": 0.8456, + "num_input_tokens_seen": 349484270, + "step": 10745 + }, + { + "epoch": 0.9691121432114352, + "flos": 18962495483040.0, + "grad_norm": 1.9515177102366532, + "learning_rate": 9.971716188680978e-09, + "loss": 0.7373, + "num_input_tokens_seen": 349512075, + "step": 10746 + }, + { + "epoch": 0.9692023267349056, + "flos": 22022977597440.0, + "grad_norm": 2.011438324280726, + "learning_rate": 9.91353546589635e-09, + "loss": 0.7051, + "num_input_tokens_seen": 349540760, + "step": 10747 + }, + { + "epoch": 0.9692925102583758, + "flos": 29454828312960.0, + "grad_norm": 3.2255375968505295, + "learning_rate": 9.855524548262106e-09, + "loss": 0.6791, + "num_input_tokens_seen": 349574380, + "step": 10748 + }, + { + "epoch": 0.969382693781846, + "flos": 18160914978720.0, + "grad_norm": 1.7016526603807285, + "learning_rate": 9.797683440728288e-09, + "loss": 0.6857, + "num_input_tokens_seen": 349602300, + "step": 10749 + }, + { + "epoch": 0.9694728773053163, + "flos": 21144109694880.0, + "grad_norm": 1.7828488452359348, + "learning_rate": 9.740012148229836e-09, + "loss": 0.6541, + "num_input_tokens_seen": 349631070, + "step": 10750 + }, + { + "epoch": 0.9695630608287866, + "flos": 27889734872160.0, + "grad_norm": 1.6013299540585435, + "learning_rate": 9.682510675687705e-09, + "loss": 0.6906, + "num_input_tokens_seen": 349660500, + "step": 10751 + }, + { + "epoch": 0.9696532443522569, + "flos": 23662050327840.0, + "grad_norm": 2.144112301120082, + "learning_rate": 9.625179028008191e-09, + "loss": 0.6648, + "num_input_tokens_seen": 349689665, + "step": 10752 + }, + { + "epoch": 0.9697434278757271, + "flos": 28866487953600.0, + "grad_norm": 2.03416280972955, + "learning_rate": 9.568017210083379e-09, + "loss": 0.5866, + "num_input_tokens_seen": 349720135, + "step": 10753 + }, + { + "epoch": 0.9698336113991973, + "flos": 29120546228640.0, + "grad_norm": 2.278879894042722, + "learning_rate": 9.511025226790259e-09, + "loss": 0.7176, + "num_input_tokens_seen": 349749920, + "step": 10754 + }, + { + "epoch": 0.9699237949226677, + "flos": 22095767454720.0, + "grad_norm": 1.6253817183313108, + "learning_rate": 9.454203082992052e-09, + "loss": 0.715, + "num_input_tokens_seen": 349776995, + "step": 10755 + }, + { + "epoch": 0.9700139784461379, + "flos": 21075668699520.0, + "grad_norm": 2.269610352705215, + "learning_rate": 9.3975507835371e-09, + "loss": 0.6477, + "num_input_tokens_seen": 349804420, + "step": 10756 + }, + { + "epoch": 0.9701041619696081, + "flos": 70114159486560.0, + "grad_norm": 0.6219678521925354, + "learning_rate": 9.341068333259094e-09, + "loss": 0.5646, + "num_input_tokens_seen": 349904840, + "step": 10757 + }, + { + "epoch": 0.9701943454930784, + "flos": 22605779662560.0, + "grad_norm": 1.6496807309687649, + "learning_rate": 9.28475573697729e-09, + "loss": 0.715, + "num_input_tokens_seen": 349932395, + "step": 10758 + }, + { + "epoch": 0.9702845290165487, + "flos": 19218672434400.0, + "grad_norm": 1.6900937668333929, + "learning_rate": 9.228612999497177e-09, + "loss": 0.7426, + "num_input_tokens_seen": 349960050, + "step": 10759 + }, + { + "epoch": 0.970374712540019, + "flos": 23948600706720.0, + "grad_norm": 2.4712219417293184, + "learning_rate": 9.172640125608478e-09, + "loss": 0.6832, + "num_input_tokens_seen": 349988450, + "step": 10760 + }, + { + "epoch": 0.9704648960634892, + "flos": 23080177506720.0, + "grad_norm": 1.3696764715030518, + "learning_rate": 9.116837120087817e-09, + "loss": 0.754, + "num_input_tokens_seen": 350020825, + "step": 10761 + }, + { + "epoch": 0.9705550795869594, + "flos": 18124631559360.0, + "grad_norm": 2.2824173122440703, + "learning_rate": 9.061203987695832e-09, + "loss": 0.807, + "num_input_tokens_seen": 350047495, + "step": 10762 + }, + { + "epoch": 0.9706452631104298, + "flos": 21840631256640.0, + "grad_norm": 2.4405393968875013, + "learning_rate": 9.005740733180055e-09, + "loss": 0.6879, + "num_input_tokens_seen": 350076535, + "step": 10763 + }, + { + "epoch": 0.9707354466339, + "flos": 20201335507680.0, + "grad_norm": 1.8610728870407778, + "learning_rate": 8.950447361272483e-09, + "loss": 0.8252, + "num_input_tokens_seen": 350103430, + "step": 10764 + }, + { + "epoch": 0.9708256301573702, + "flos": 19836196788960.0, + "grad_norm": 2.5709817588986983, + "learning_rate": 8.895323876691784e-09, + "loss": 0.7455, + "num_input_tokens_seen": 350129425, + "step": 10765 + }, + { + "epoch": 0.9709158136808405, + "flos": 14079590713920.0, + "grad_norm": 5.194397901756043, + "learning_rate": 8.840370284140419e-09, + "loss": 0.7918, + "num_input_tokens_seen": 350152995, + "step": 10766 + }, + { + "epoch": 0.9710059972043108, + "flos": 21549137299680.0, + "grad_norm": 2.303940577621563, + "learning_rate": 8.78558658830797e-09, + "loss": 0.6584, + "num_input_tokens_seen": 350179750, + "step": 10767 + }, + { + "epoch": 0.971096180727781, + "flos": 67828341236640.0, + "grad_norm": 0.6931295954394576, + "learning_rate": 8.730972793868696e-09, + "loss": 0.5766, + "num_input_tokens_seen": 350268855, + "step": 10768 + }, + { + "epoch": 0.9711863642512513, + "flos": 27336860197440.0, + "grad_norm": 2.0532977259426075, + "learning_rate": 8.67652890548265e-09, + "loss": 0.6996, + "num_input_tokens_seen": 350297830, + "step": 10769 + }, + { + "epoch": 0.9712765477747216, + "flos": 24900741673440.0, + "grad_norm": 1.5756966821146448, + "learning_rate": 8.622254927795004e-09, + "loss": 0.7234, + "num_input_tokens_seen": 350330795, + "step": 10770 + }, + { + "epoch": 0.9713667312981918, + "flos": 36627268308000.0, + "grad_norm": 1.847173906075245, + "learning_rate": 8.568150865436941e-09, + "loss": 0.6912, + "num_input_tokens_seen": 350360635, + "step": 10771 + }, + { + "epoch": 0.9714569148216621, + "flos": 42462611401920.0, + "grad_norm": 3.321386545035517, + "learning_rate": 8.514216723024991e-09, + "loss": 0.7528, + "num_input_tokens_seen": 350395705, + "step": 10772 + }, + { + "epoch": 0.9715470983451323, + "flos": 23661938818560.0, + "grad_norm": 2.795465209951873, + "learning_rate": 8.460452505161031e-09, + "loss": 0.6627, + "num_input_tokens_seen": 350422715, + "step": 10773 + }, + { + "epoch": 0.9716372818686027, + "flos": 21622150175520.0, + "grad_norm": 1.8852982317009779, + "learning_rate": 8.4068582164325e-09, + "loss": 0.7632, + "num_input_tokens_seen": 350451285, + "step": 10774 + }, + { + "epoch": 0.9717274653920729, + "flos": 19470797881920.0, + "grad_norm": 1.8819626279033033, + "learning_rate": 8.353433861412406e-09, + "loss": 0.722, + "num_input_tokens_seen": 350478780, + "step": 10775 + }, + { + "epoch": 0.9718176489155431, + "flos": 30507976718400.0, + "grad_norm": 1.6676879932402977, + "learning_rate": 8.300179444658883e-09, + "loss": 0.6947, + "num_input_tokens_seen": 350510810, + "step": 10776 + }, + { + "epoch": 0.9719078324390134, + "flos": 19144878993600.0, + "grad_norm": 2.284552809113454, + "learning_rate": 8.247094970716296e-09, + "loss": 0.8315, + "num_input_tokens_seen": 350537785, + "step": 10777 + }, + { + "epoch": 0.9719980159624837, + "flos": 23402230740000.0, + "grad_norm": 1.8445348847033094, + "learning_rate": 8.19418044411413e-09, + "loss": 0.7757, + "num_input_tokens_seen": 350568300, + "step": 10778 + }, + { + "epoch": 0.9720881994859539, + "flos": 23327359376160.0, + "grad_norm": 1.8517832564082626, + "learning_rate": 8.141435869367219e-09, + "loss": 0.7415, + "num_input_tokens_seen": 350596440, + "step": 10779 + }, + { + "epoch": 0.9721783830094242, + "flos": 22601096272800.0, + "grad_norm": 1.918063035189683, + "learning_rate": 8.088861250975742e-09, + "loss": 0.776, + "num_input_tokens_seen": 350623900, + "step": 10780 + }, + { + "epoch": 0.9722685665328944, + "flos": 22054057250400.0, + "grad_norm": 2.2084004973748352, + "learning_rate": 8.036456593426111e-09, + "loss": 0.7095, + "num_input_tokens_seen": 350650730, + "step": 10781 + }, + { + "epoch": 0.9723587500563647, + "flos": 15355457553120.0, + "grad_norm": 2.0735651915498527, + "learning_rate": 7.984221901189415e-09, + "loss": 0.7736, + "num_input_tokens_seen": 350675960, + "step": 10782 + }, + { + "epoch": 0.972448933579835, + "flos": 21908812063680.0, + "grad_norm": 1.9707996658557185, + "learning_rate": 7.932157178722976e-09, + "loss": 0.6617, + "num_input_tokens_seen": 350704975, + "step": 10783 + }, + { + "epoch": 0.9725391171033052, + "flos": 29272965179040.0, + "grad_norm": 1.8509827684067182, + "learning_rate": 7.880262430468799e-09, + "loss": 0.7062, + "num_input_tokens_seen": 350734150, + "step": 10784 + }, + { + "epoch": 0.9726293006267754, + "flos": 21221285583840.0, + "grad_norm": 2.1424715194678683, + "learning_rate": 7.828537660855339e-09, + "loss": 0.6712, + "num_input_tokens_seen": 350762025, + "step": 10785 + }, + { + "epoch": 0.9727194841502458, + "flos": 23149584915840.0, + "grad_norm": 3.062648224649579, + "learning_rate": 7.776982874295512e-09, + "loss": 0.7557, + "num_input_tokens_seen": 350791945, + "step": 10786 + }, + { + "epoch": 0.972809667673716, + "flos": 29266832168640.0, + "grad_norm": 2.189525102832538, + "learning_rate": 7.725598075188688e-09, + "loss": 0.647, + "num_input_tokens_seen": 350821270, + "step": 10787 + }, + { + "epoch": 0.9728998511971862, + "flos": 20673837694080.0, + "grad_norm": 1.9282225326031868, + "learning_rate": 7.674383267918916e-09, + "loss": 0.6905, + "num_input_tokens_seen": 350851145, + "step": 10788 + }, + { + "epoch": 0.9729900347206565, + "flos": 22132162383360.0, + "grad_norm": 1.8480455823082234, + "learning_rate": 7.623338456856476e-09, + "loss": 0.7255, + "num_input_tokens_seen": 350880800, + "step": 10789 + }, + { + "epoch": 0.9730802182441268, + "flos": 21840631256640.0, + "grad_norm": 2.322975659726105, + "learning_rate": 7.572463646356554e-09, + "loss": 0.6316, + "num_input_tokens_seen": 350907300, + "step": 10790 + }, + { + "epoch": 0.9731704017675971, + "flos": 20055384095520.0, + "grad_norm": 3.7629836066317504, + "learning_rate": 7.521758840760339e-09, + "loss": 0.7088, + "num_input_tokens_seen": 350936485, + "step": 10791 + }, + { + "epoch": 0.9732605852910673, + "flos": 18707470794240.0, + "grad_norm": 2.582483572302949, + "learning_rate": 7.471224044393931e-09, + "loss": 0.6958, + "num_input_tokens_seen": 350959700, + "step": 10792 + }, + { + "epoch": 0.9733507688145376, + "flos": 26973691476000.0, + "grad_norm": 1.7542657632389995, + "learning_rate": 7.420859261569434e-09, + "loss": 0.7971, + "num_input_tokens_seen": 350989520, + "step": 10793 + }, + { + "epoch": 0.9734409523380079, + "flos": 27734491020000.0, + "grad_norm": 2.0906434427539824, + "learning_rate": 7.370664496584078e-09, + "loss": 0.7975, + "num_input_tokens_seen": 351015665, + "step": 10794 + }, + { + "epoch": 0.9735311358614781, + "flos": 35320656343680.0, + "grad_norm": 2.5916466514902106, + "learning_rate": 7.3206397537211026e-09, + "loss": 0.6632, + "num_input_tokens_seen": 351049335, + "step": 10795 + }, + { + "epoch": 0.9736213193849483, + "flos": 35867174989440.0, + "grad_norm": 2.0827915886003066, + "learning_rate": 7.270785037248428e-09, + "loss": 0.7044, + "num_input_tokens_seen": 351079760, + "step": 10796 + }, + { + "epoch": 0.9737115029084187, + "flos": 23079917318400.0, + "grad_norm": 2.09416798077608, + "learning_rate": 7.221100351420428e-09, + "loss": 0.7001, + "num_input_tokens_seen": 351108635, + "step": 10797 + }, + { + "epoch": 0.9738016864318889, + "flos": 38048491843200.0, + "grad_norm": 1.9819215641187644, + "learning_rate": 7.171585700475935e-09, + "loss": 0.6654, + "num_input_tokens_seen": 351142030, + "step": 10798 + }, + { + "epoch": 0.9738918699553591, + "flos": 22054057250400.0, + "grad_norm": 2.06430668162598, + "learning_rate": 7.122241088640235e-09, + "loss": 0.7934, + "num_input_tokens_seen": 351170125, + "step": 10799 + }, + { + "epoch": 0.9739820534788294, + "flos": 27341878115040.0, + "grad_norm": 2.1369903224499236, + "learning_rate": 7.073066520123516e-09, + "loss": 0.8613, + "num_input_tokens_seen": 351198685, + "step": 10800 + }, + { + "epoch": 0.9740722370022997, + "flos": 20638372009440.0, + "grad_norm": 2.573119604676293, + "learning_rate": 7.0240619991217555e-09, + "loss": 0.6324, + "num_input_tokens_seen": 351224855, + "step": 10801 + }, + { + "epoch": 0.97416242052577, + "flos": 19983486312480.0, + "grad_norm": 5.1864556789551015, + "learning_rate": 6.975227529816052e-09, + "loss": 0.8135, + "num_input_tokens_seen": 351251455, + "step": 10802 + }, + { + "epoch": 0.9742526040492402, + "flos": 19509014128800.0, + "grad_norm": 2.00253730996339, + "learning_rate": 6.926563116373296e-09, + "loss": 0.7822, + "num_input_tokens_seen": 351280050, + "step": 10803 + }, + { + "epoch": 0.9743427875727104, + "flos": 23075791475040.0, + "grad_norm": 1.9741337143850537, + "learning_rate": 6.878068762945943e-09, + "loss": 0.6736, + "num_input_tokens_seen": 351308090, + "step": 10804 + }, + { + "epoch": 0.9744329710961808, + "flos": 29417987347200.0, + "grad_norm": 1.7913569101292388, + "learning_rate": 6.829744473671794e-09, + "loss": 0.7732, + "num_input_tokens_seen": 351338770, + "step": 10805 + }, + { + "epoch": 0.974523154619651, + "flos": 63117226596480.0, + "grad_norm": 0.5850808047868563, + "learning_rate": 6.781590252674219e-09, + "loss": 0.477, + "num_input_tokens_seen": 351432795, + "step": 10806 + }, + { + "epoch": 0.9746133381431212, + "flos": 33719688350880.0, + "grad_norm": 1.6081103198274282, + "learning_rate": 6.733606104061484e-09, + "loss": 0.7452, + "num_input_tokens_seen": 351470125, + "step": 10807 + }, + { + "epoch": 0.9747035216665915, + "flos": 25957421206080.0, + "grad_norm": 1.6705119914012971, + "learning_rate": 6.6857920319283165e-09, + "loss": 0.6807, + "num_input_tokens_seen": 351502885, + "step": 10808 + }, + { + "epoch": 0.9747937051900618, + "flos": 24572332411200.0, + "grad_norm": 1.9546103360857445, + "learning_rate": 6.638148040354563e-09, + "loss": 0.7114, + "num_input_tokens_seen": 351533750, + "step": 10809 + }, + { + "epoch": 0.974883888713532, + "flos": 22204729222080.0, + "grad_norm": 1.6776852761523782, + "learning_rate": 6.590674133405194e-09, + "loss": 0.6961, + "num_input_tokens_seen": 351560180, + "step": 10810 + }, + { + "epoch": 0.9749740722370023, + "flos": 30584855249280.0, + "grad_norm": 1.815289160023017, + "learning_rate": 6.5433703151311914e-09, + "loss": 0.7501, + "num_input_tokens_seen": 351591040, + "step": 10811 + }, + { + "epoch": 0.9750642557604725, + "flos": 35903644257600.0, + "grad_norm": 1.80239868344912, + "learning_rate": 6.49623658956866e-09, + "loss": 0.7383, + "num_input_tokens_seen": 351621290, + "step": 10812 + }, + { + "epoch": 0.9751544392839429, + "flos": 14590643675040.0, + "grad_norm": 2.7079925811812458, + "learning_rate": 6.44927296073905e-09, + "loss": 0.6166, + "num_input_tokens_seen": 351648110, + "step": 10813 + }, + { + "epoch": 0.9752446228074131, + "flos": 22459939759680.0, + "grad_norm": 1.7985131798802936, + "learning_rate": 6.402479432649821e-09, + "loss": 0.8237, + "num_input_tokens_seen": 351674760, + "step": 10814 + }, + { + "epoch": 0.9753348063308833, + "flos": 23113450175520.0, + "grad_norm": 1.861464152214065, + "learning_rate": 6.355856009293781e-09, + "loss": 0.7313, + "num_input_tokens_seen": 351701350, + "step": 10815 + }, + { + "epoch": 0.9754249898543537, + "flos": 25812250358880.0, + "grad_norm": 3.6917132860487905, + "learning_rate": 6.3094026946488575e-09, + "loss": 0.7105, + "num_input_tokens_seen": 351727825, + "step": 10816 + }, + { + "epoch": 0.9755151733778239, + "flos": 26322411245760.0, + "grad_norm": 2.2888636584059, + "learning_rate": 6.2631194926787704e-09, + "loss": 0.7373, + "num_input_tokens_seen": 351756410, + "step": 10817 + }, + { + "epoch": 0.9756053569012941, + "flos": 27123434203680.0, + "grad_norm": 1.7017843586981052, + "learning_rate": 6.217006407332581e-09, + "loss": 0.6353, + "num_input_tokens_seen": 351784950, + "step": 10818 + }, + { + "epoch": 0.9756955404247644, + "flos": 25737564843840.0, + "grad_norm": 2.3425765087942043, + "learning_rate": 6.1710634425453654e-09, + "loss": 0.6846, + "num_input_tokens_seen": 351814955, + "step": 10819 + }, + { + "epoch": 0.9757857239482347, + "flos": 25224393206400.0, + "grad_norm": 1.8068773599119026, + "learning_rate": 6.1252906022366544e-09, + "loss": 0.7577, + "num_input_tokens_seen": 351843490, + "step": 10820 + }, + { + "epoch": 0.9758759074717049, + "flos": 17760570763680.0, + "grad_norm": 6.204904142587633, + "learning_rate": 6.079687890312213e-09, + "loss": 0.7152, + "num_input_tokens_seen": 351870760, + "step": 10821 + }, + { + "epoch": 0.9759660909951752, + "flos": 26139879056160.0, + "grad_norm": 1.7427454854589073, + "learning_rate": 6.034255310663372e-09, + "loss": 0.7172, + "num_input_tokens_seen": 351902340, + "step": 10822 + }, + { + "epoch": 0.9760562745186454, + "flos": 29819037787680.0, + "grad_norm": 1.8316557607190131, + "learning_rate": 5.988992867166143e-09, + "loss": 0.8297, + "num_input_tokens_seen": 351932115, + "step": 10823 + }, + { + "epoch": 0.9761464580421157, + "flos": 24900964692000.0, + "grad_norm": 2.7493846378213855, + "learning_rate": 5.943900563682991e-09, + "loss": 0.7779, + "num_input_tokens_seen": 351959930, + "step": 10824 + }, + { + "epoch": 0.976236641565586, + "flos": 66262169872800.0, + "grad_norm": 0.7435119857129842, + "learning_rate": 5.898978404061506e-09, + "loss": 0.655, + "num_input_tokens_seen": 352050680, + "step": 10825 + }, + { + "epoch": 0.9763268250890562, + "flos": 21908143008000.0, + "grad_norm": 2.8708595647476263, + "learning_rate": 5.85422639213462e-09, + "loss": 0.7335, + "num_input_tokens_seen": 352079050, + "step": 10826 + }, + { + "epoch": 0.9764170086125264, + "flos": 30141722906880.0, + "grad_norm": 1.7367012301036218, + "learning_rate": 5.809644531720614e-09, + "loss": 0.7479, + "num_input_tokens_seen": 352110275, + "step": 10827 + }, + { + "epoch": 0.9765071921359968, + "flos": 23441450570400.0, + "grad_norm": 1.7541013069692082, + "learning_rate": 5.765232826623556e-09, + "loss": 0.7646, + "num_input_tokens_seen": 352141315, + "step": 10828 + }, + { + "epoch": 0.976597375659467, + "flos": 24169758010560.0, + "grad_norm": 1.8604485754314606, + "learning_rate": 5.720991280633081e-09, + "loss": 0.7174, + "num_input_tokens_seen": 352168770, + "step": 10829 + }, + { + "epoch": 0.9766875591829373, + "flos": 19654965540960.0, + "grad_norm": 1.725927382810254, + "learning_rate": 5.676919897523724e-09, + "loss": 0.7231, + "num_input_tokens_seen": 352196615, + "step": 10830 + }, + { + "epoch": 0.9767777427064075, + "flos": 45412310430720.0, + "grad_norm": 1.4764344497944346, + "learning_rate": 5.633018681056256e-09, + "loss": 0.7401, + "num_input_tokens_seen": 352229610, + "step": 10831 + }, + { + "epoch": 0.9768679262298778, + "flos": 26062294299840.0, + "grad_norm": 2.4846854345835787, + "learning_rate": 5.589287634976569e-09, + "loss": 0.7396, + "num_input_tokens_seen": 352256600, + "step": 10832 + }, + { + "epoch": 0.9769581097533481, + "flos": 56559528957600.0, + "grad_norm": 1.6634709555977187, + "learning_rate": 5.5457267630159014e-09, + "loss": 0.7086, + "num_input_tokens_seen": 352292935, + "step": 10833 + }, + { + "epoch": 0.9770482932768183, + "flos": 23803244010720.0, + "grad_norm": 1.8096826295123452, + "learning_rate": 5.5023360688910555e-09, + "loss": 0.7713, + "num_input_tokens_seen": 352318695, + "step": 10834 + }, + { + "epoch": 0.9771384768002885, + "flos": 22423396152000.0, + "grad_norm": 1.67150190236779, + "learning_rate": 5.459115556304183e-09, + "loss": 0.7842, + "num_input_tokens_seen": 352346175, + "step": 10835 + }, + { + "epoch": 0.9772286603237589, + "flos": 37106572560480.0, + "grad_norm": 2.31182335929194, + "learning_rate": 5.416065228943889e-09, + "loss": 0.6699, + "num_input_tokens_seen": 352375970, + "step": 10836 + }, + { + "epoch": 0.9773188438472291, + "flos": 22934003076000.0, + "grad_norm": 2.867844176386112, + "learning_rate": 5.373185090482568e-09, + "loss": 0.6884, + "num_input_tokens_seen": 352403920, + "step": 10837 + }, + { + "epoch": 0.9774090273706993, + "flos": 18562114098240.0, + "grad_norm": 1.891870009910674, + "learning_rate": 5.330475144579516e-09, + "loss": 0.7119, + "num_input_tokens_seen": 352431775, + "step": 10838 + }, + { + "epoch": 0.9774992108941697, + "flos": 25884110972160.0, + "grad_norm": 1.6764593997479618, + "learning_rate": 5.2879353948787065e-09, + "loss": 0.8197, + "num_input_tokens_seen": 352459060, + "step": 10839 + }, + { + "epoch": 0.9775893944176399, + "flos": 20164940579040.0, + "grad_norm": 2.18136553963869, + "learning_rate": 5.245565845010125e-09, + "loss": 0.7985, + "num_input_tokens_seen": 352485975, + "step": 10840 + }, + { + "epoch": 0.9776795779411102, + "flos": 29195343252960.0, + "grad_norm": 1.7055726268271545, + "learning_rate": 5.2033664985886575e-09, + "loss": 0.6676, + "num_input_tokens_seen": 352516820, + "step": 10841 + }, + { + "epoch": 0.9777697614645804, + "flos": 20449037753760.0, + "grad_norm": 1.9110373120819897, + "learning_rate": 5.161337359215201e-09, + "loss": 0.8813, + "num_input_tokens_seen": 352543995, + "step": 10842 + }, + { + "epoch": 0.9778599449880507, + "flos": 22096287831360.0, + "grad_norm": 1.9729366292081187, + "learning_rate": 5.119478430475999e-09, + "loss": 0.7476, + "num_input_tokens_seen": 352571775, + "step": 10843 + }, + { + "epoch": 0.977950128511521, + "flos": 23224790807520.0, + "grad_norm": 2.696126444553963, + "learning_rate": 5.077789715942416e-09, + "loss": 0.6892, + "num_input_tokens_seen": 352601345, + "step": 10844 + }, + { + "epoch": 0.9780403120349912, + "flos": 18962123785440.0, + "grad_norm": 2.102227882075649, + "learning_rate": 5.036271219171606e-09, + "loss": 0.7595, + "num_input_tokens_seen": 352627310, + "step": 10845 + }, + { + "epoch": 0.9781304955584614, + "flos": 19208004713280.0, + "grad_norm": 2.1528644243530004, + "learning_rate": 4.994922943706514e-09, + "loss": 0.7727, + "num_input_tokens_seen": 352649960, + "step": 10846 + }, + { + "epoch": 0.9782206790819318, + "flos": 19542138118560.0, + "grad_norm": 1.7114684792061354, + "learning_rate": 4.953744893074763e-09, + "loss": 0.72, + "num_input_tokens_seen": 352675815, + "step": 10847 + }, + { + "epoch": 0.978310862605402, + "flos": 18488803864320.0, + "grad_norm": 2.1039964839992953, + "learning_rate": 4.912737070789985e-09, + "loss": 0.6632, + "num_input_tokens_seen": 352702635, + "step": 10848 + }, + { + "epoch": 0.9784010461288722, + "flos": 22241904715680.0, + "grad_norm": 2.1386494674553886, + "learning_rate": 4.871899480351604e-09, + "loss": 0.807, + "num_input_tokens_seen": 352731750, + "step": 10849 + }, + { + "epoch": 0.9784912296523425, + "flos": 23735806598880.0, + "grad_norm": 1.7230462910770732, + "learning_rate": 4.831232125243501e-09, + "loss": 0.7499, + "num_input_tokens_seen": 352761905, + "step": 10850 + }, + { + "epoch": 0.9785814131758128, + "flos": 16485001282560.0, + "grad_norm": 2.142811792892867, + "learning_rate": 4.7907350089360086e-09, + "loss": 0.8381, + "num_input_tokens_seen": 352788860, + "step": 10851 + }, + { + "epoch": 0.978671596699283, + "flos": 17615028218880.0, + "grad_norm": 2.0799774861951303, + "learning_rate": 4.750408134884365e-09, + "loss": 0.6305, + "num_input_tokens_seen": 352816930, + "step": 10852 + }, + { + "epoch": 0.9787617802227533, + "flos": 29892125003040.0, + "grad_norm": 1.6705523811994938, + "learning_rate": 4.710251506529816e-09, + "loss": 0.6474, + "num_input_tokens_seen": 352848285, + "step": 10853 + }, + { + "epoch": 0.9788519637462235, + "flos": 31240001134560.0, + "grad_norm": 1.8068083484222854, + "learning_rate": 4.6702651272982894e-09, + "loss": 0.7502, + "num_input_tokens_seen": 352878940, + "step": 10854 + }, + { + "epoch": 0.9789421472696939, + "flos": 22670615191200.0, + "grad_norm": 2.3401803871464195, + "learning_rate": 4.630449000602166e-09, + "loss": 0.7087, + "num_input_tokens_seen": 352904920, + "step": 10855 + }, + { + "epoch": 0.9790323307931641, + "flos": 19355442915840.0, + "grad_norm": 3.2148780125661305, + "learning_rate": 4.590803129838283e-09, + "loss": 0.8005, + "num_input_tokens_seen": 352929320, + "step": 10856 + }, + { + "epoch": 0.9791225143166343, + "flos": 19945641763200.0, + "grad_norm": 3.4651989330087107, + "learning_rate": 4.551327518389714e-09, + "loss": 0.7435, + "num_input_tokens_seen": 352955020, + "step": 10857 + }, + { + "epoch": 0.9792126978401046, + "flos": 20269516314720.0, + "grad_norm": 2.607353203830027, + "learning_rate": 4.512022169624652e-09, + "loss": 0.7316, + "num_input_tokens_seen": 352980530, + "step": 10858 + }, + { + "epoch": 0.9793028813635749, + "flos": 20493349841280.0, + "grad_norm": 1.8633539917380166, + "learning_rate": 4.472887086896637e-09, + "loss": 0.7102, + "num_input_tokens_seen": 353009170, + "step": 10859 + }, + { + "epoch": 0.9793930648870451, + "flos": 27921929617920.0, + "grad_norm": 2.1397798080890738, + "learning_rate": 4.433922273545443e-09, + "loss": 0.6825, + "num_input_tokens_seen": 353038010, + "step": 10860 + }, + { + "epoch": 0.9794832484105154, + "flos": 69785973242880.0, + "grad_norm": 0.5986798316200826, + "learning_rate": 4.395127732895299e-09, + "loss": 0.5797, + "num_input_tokens_seen": 353129920, + "step": 10861 + }, + { + "epoch": 0.9795734319339856, + "flos": 25847195666880.0, + "grad_norm": 50.17446125769563, + "learning_rate": 4.356503468256445e-09, + "loss": 0.8166, + "num_input_tokens_seen": 353157735, + "step": 10862 + }, + { + "epoch": 0.979663615457456, + "flos": 19071866117760.0, + "grad_norm": 2.523688997662229, + "learning_rate": 4.318049482924913e-09, + "loss": 0.7508, + "num_input_tokens_seen": 353182395, + "step": 10863 + }, + { + "epoch": 0.9797537989809262, + "flos": 20165052088320.0, + "grad_norm": 5.81805023673128, + "learning_rate": 4.279765780181188e-09, + "loss": 0.705, + "num_input_tokens_seen": 353206270, + "step": 10864 + }, + { + "epoch": 0.9798439825043964, + "flos": 24900927522240.0, + "grad_norm": 2.2085113693542295, + "learning_rate": 4.241652363291992e-09, + "loss": 0.7516, + "num_input_tokens_seen": 353234695, + "step": 10865 + }, + { + "epoch": 0.9799341660278668, + "flos": 13789026000960.0, + "grad_norm": 3.942335883034403, + "learning_rate": 4.203709235509834e-09, + "loss": 0.7127, + "num_input_tokens_seen": 353259755, + "step": 10866 + }, + { + "epoch": 0.980024349551337, + "flos": 23443978114080.0, + "grad_norm": 1.7940192043845806, + "learning_rate": 4.165936400071679e-09, + "loss": 0.7598, + "num_input_tokens_seen": 353289070, + "step": 10867 + }, + { + "epoch": 0.9801145330748072, + "flos": 21476719140000.0, + "grad_norm": 1.7496535775762208, + "learning_rate": 4.12833386020095e-09, + "loss": 0.6863, + "num_input_tokens_seen": 353317870, + "step": 10868 + }, + { + "epoch": 0.9802047165982775, + "flos": 17940947107200.0, + "grad_norm": 2.21619579887918, + "learning_rate": 4.090901619105746e-09, + "loss": 0.7725, + "num_input_tokens_seen": 353344480, + "step": 10869 + }, + { + "epoch": 0.9802949001217478, + "flos": 13678205745600.0, + "grad_norm": 1.8617329752607101, + "learning_rate": 4.053639679980181e-09, + "loss": 0.6294, + "num_input_tokens_seen": 353370600, + "step": 10870 + }, + { + "epoch": 0.980385083645218, + "flos": 22861213218720.0, + "grad_norm": 1.9050079426637407, + "learning_rate": 4.01654804600371e-09, + "loss": 0.7446, + "num_input_tokens_seen": 353397435, + "step": 10871 + }, + { + "epoch": 0.9804752671686883, + "flos": 20456025668640.0, + "grad_norm": 1.7110213167745483, + "learning_rate": 3.9796267203409114e-09, + "loss": 0.7459, + "num_input_tokens_seen": 353424905, + "step": 10872 + }, + { + "epoch": 0.9805654506921585, + "flos": 25916565906240.0, + "grad_norm": 1.6531902336417301, + "learning_rate": 3.942875706142379e-09, + "loss": 0.7161, + "num_input_tokens_seen": 353452735, + "step": 10873 + }, + { + "epoch": 0.9806556342156288, + "flos": 20016201434880.0, + "grad_norm": 2.335553600188668, + "learning_rate": 3.906295006543825e-09, + "loss": 0.7982, + "num_input_tokens_seen": 353481650, + "step": 10874 + }, + { + "epoch": 0.9807458177390991, + "flos": 24752002529280.0, + "grad_norm": 1.9316362433002945, + "learning_rate": 3.8698846246665305e-09, + "loss": 0.7144, + "num_input_tokens_seen": 353510710, + "step": 10875 + }, + { + "epoch": 0.9808360012625693, + "flos": 19507229980320.0, + "grad_norm": 1.8326635509953404, + "learning_rate": 3.833644563617344e-09, + "loss": 0.7428, + "num_input_tokens_seen": 353537900, + "step": 10876 + }, + { + "epoch": 0.9809261847860395, + "flos": 23297655004320.0, + "grad_norm": 2.7348239203196587, + "learning_rate": 3.797574826488237e-09, + "loss": 0.7482, + "num_input_tokens_seen": 353567515, + "step": 10877 + }, + { + "epoch": 0.9810163683095099, + "flos": 29491446260160.0, + "grad_norm": 1.7304389210491689, + "learning_rate": 3.761675416356969e-09, + "loss": 0.6674, + "num_input_tokens_seen": 353596820, + "step": 10878 + }, + { + "epoch": 0.9811065518329801, + "flos": 34588334569440.0, + "grad_norm": 2.426141242253233, + "learning_rate": 3.725946336286867e-09, + "loss": 0.6456, + "num_input_tokens_seen": 353627125, + "step": 10879 + }, + { + "epoch": 0.9811967353564504, + "flos": 25014795697920.0, + "grad_norm": 1.7127073705376865, + "learning_rate": 3.6903875893261604e-09, + "loss": 0.7108, + "num_input_tokens_seen": 353660825, + "step": 10880 + }, + { + "epoch": 0.9812869188799206, + "flos": 24063509635680.0, + "grad_norm": 2.1480197569721686, + "learning_rate": 3.6549991785093105e-09, + "loss": 0.668, + "num_input_tokens_seen": 353690425, + "step": 10881 + }, + { + "epoch": 0.9813771024033909, + "flos": 27086295879840.0, + "grad_norm": 1.8875834692884137, + "learning_rate": 3.6197811068554575e-09, + "loss": 0.726, + "num_input_tokens_seen": 353719645, + "step": 10882 + }, + { + "epoch": 0.9814672859268612, + "flos": 23443048870080.0, + "grad_norm": 1.709332553817722, + "learning_rate": 3.584733377369975e-09, + "loss": 0.8214, + "num_input_tokens_seen": 353749520, + "step": 10883 + }, + { + "epoch": 0.9815574694503314, + "flos": 21804459346560.0, + "grad_norm": 2.0638096437205435, + "learning_rate": 3.549855993043138e-09, + "loss": 0.7163, + "num_input_tokens_seen": 353776300, + "step": 10884 + }, + { + "epoch": 0.9816476529738016, + "flos": 22380199157280.0, + "grad_norm": 1.645020473318037, + "learning_rate": 3.5151489568507887e-09, + "loss": 0.7238, + "num_input_tokens_seen": 353804210, + "step": 10885 + }, + { + "epoch": 0.981737836497272, + "flos": 26061811092960.0, + "grad_norm": 2.2806033955599023, + "learning_rate": 3.4806122717545572e-09, + "loss": 0.7503, + "num_input_tokens_seen": 353834270, + "step": 10886 + }, + { + "epoch": 0.9818280200207422, + "flos": 24610214130240.0, + "grad_norm": 3.0397438521135958, + "learning_rate": 3.446245940701198e-09, + "loss": 0.6967, + "num_input_tokens_seen": 353864130, + "step": 10887 + }, + { + "epoch": 0.9819182035442124, + "flos": 21804050479200.0, + "grad_norm": 1.8505661613595363, + "learning_rate": 3.41204996662281e-09, + "loss": 0.6776, + "num_input_tokens_seen": 353892350, + "step": 10888 + }, + { + "epoch": 0.9820083870676828, + "flos": 31131559743840.0, + "grad_norm": 2.049142644285565, + "learning_rate": 3.3780243524375028e-09, + "loss": 0.736, + "num_input_tokens_seen": 353924760, + "step": 10889 + }, + { + "epoch": 0.982098570591153, + "flos": 25442651268960.0, + "grad_norm": 2.8878525093542593, + "learning_rate": 3.3441691010485107e-09, + "loss": 0.7658, + "num_input_tokens_seen": 353953105, + "step": 10890 + }, + { + "epoch": 0.9821887541146233, + "flos": 22054205929440.0, + "grad_norm": 1.78400074897203, + "learning_rate": 3.3104842153444113e-09, + "loss": 0.8151, + "num_input_tokens_seen": 353977630, + "step": 10891 + }, + { + "epoch": 0.9822789376380935, + "flos": 26686100343840.0, + "grad_norm": 2.094870082678055, + "learning_rate": 3.27696969819935e-09, + "loss": 0.7312, + "num_input_tokens_seen": 354005290, + "step": 10892 + }, + { + "epoch": 0.9823691211615638, + "flos": 22241718866880.0, + "grad_norm": 1.6070083698133075, + "learning_rate": 3.2436255524732615e-09, + "loss": 0.6951, + "num_input_tokens_seen": 354033510, + "step": 10893 + }, + { + "epoch": 0.9824593046850341, + "flos": 22387298581440.0, + "grad_norm": 1.9367758105059896, + "learning_rate": 3.210451781010759e-09, + "loss": 0.721, + "num_input_tokens_seen": 354061695, + "step": 10894 + }, + { + "epoch": 0.9825494882085043, + "flos": 30071943800160.0, + "grad_norm": 1.7292571293098082, + "learning_rate": 3.1774483866426895e-09, + "loss": 0.6008, + "num_input_tokens_seen": 354091650, + "step": 10895 + }, + { + "epoch": 0.9826396717319745, + "flos": 41367641282880.0, + "grad_norm": 1.6235010863436492, + "learning_rate": 3.144615372185244e-09, + "loss": 0.7031, + "num_input_tokens_seen": 354123335, + "step": 10896 + }, + { + "epoch": 0.9827298552554449, + "flos": 25521871494720.0, + "grad_norm": 1.4255134226455746, + "learning_rate": 3.1119527404399604e-09, + "loss": 0.8285, + "num_input_tokens_seen": 354156290, + "step": 10897 + }, + { + "epoch": 0.9828200387789151, + "flos": 22933371190080.0, + "grad_norm": 2.2754797922107595, + "learning_rate": 3.0794604941932754e-09, + "loss": 0.7407, + "num_input_tokens_seen": 354182560, + "step": 10898 + }, + { + "epoch": 0.9829102223023853, + "flos": 34628260625280.0, + "grad_norm": 1.8806252338768055, + "learning_rate": 3.0471386362180827e-09, + "loss": 0.6371, + "num_input_tokens_seen": 354215595, + "step": 10899 + }, + { + "epoch": 0.9830004058258556, + "flos": 22638160257120.0, + "grad_norm": 1.9780251741456258, + "learning_rate": 3.0149871692719542e-09, + "loss": 0.786, + "num_input_tokens_seen": 354244945, + "step": 10900 + }, + { + "epoch": 0.9830905893493259, + "flos": 23589632168160.0, + "grad_norm": 1.6993142287516219, + "learning_rate": 2.9830060960984728e-09, + "loss": 0.7228, + "num_input_tokens_seen": 354276515, + "step": 10901 + }, + { + "epoch": 0.9831807728727961, + "flos": 62593981954080.0, + "grad_norm": 0.6665705497316443, + "learning_rate": 2.9511954194263442e-09, + "loss": 0.5834, + "num_input_tokens_seen": 354363285, + "step": 10902 + }, + { + "epoch": 0.9832709563962664, + "flos": 26504200040160.0, + "grad_norm": 2.0978154809283645, + "learning_rate": 2.9195551419698426e-09, + "loss": 0.7346, + "num_input_tokens_seen": 354395155, + "step": 10903 + }, + { + "epoch": 0.9833611399197366, + "flos": 16228118105760.0, + "grad_norm": 2.206787636063972, + "learning_rate": 2.888085266428808e-09, + "loss": 0.6997, + "num_input_tokens_seen": 354421480, + "step": 10904 + }, + { + "epoch": 0.983451323443207, + "flos": 23512865146560.0, + "grad_norm": 1.9951474807836924, + "learning_rate": 2.8567857954882037e-09, + "loss": 0.694, + "num_input_tokens_seen": 354448195, + "step": 10905 + }, + { + "epoch": 0.9835415069666772, + "flos": 25440681271680.0, + "grad_norm": 1.9924882848723084, + "learning_rate": 2.82565673181856e-09, + "loss": 0.7914, + "num_input_tokens_seen": 354475545, + "step": 10906 + }, + { + "epoch": 0.9836316904901474, + "flos": 25520682062400.0, + "grad_norm": 1.8711077026611498, + "learning_rate": 2.7946980780764184e-09, + "loss": 0.7558, + "num_input_tokens_seen": 354506375, + "step": 10907 + }, + { + "epoch": 0.9837218740136177, + "flos": 19145399370240.0, + "grad_norm": 2.7024695864226764, + "learning_rate": 2.763909836903e-09, + "loss": 0.7923, + "num_input_tokens_seen": 354532825, + "step": 10908 + }, + { + "epoch": 0.983812057537088, + "flos": 22605742492800.0, + "grad_norm": 1.609650597539055, + "learning_rate": 2.7332920109255364e-09, + "loss": 0.6527, + "num_input_tokens_seen": 354562070, + "step": 10909 + }, + { + "epoch": 0.9839022410605582, + "flos": 22894969094400.0, + "grad_norm": 6.547884114502419, + "learning_rate": 2.702844602756382e-09, + "loss": 0.7183, + "num_input_tokens_seen": 354592070, + "step": 10910 + }, + { + "epoch": 0.9839924245840285, + "flos": 20347138240800.0, + "grad_norm": 2.9453008433953682, + "learning_rate": 2.6725676149936814e-09, + "loss": 0.7148, + "num_input_tokens_seen": 354620310, + "step": 10911 + }, + { + "epoch": 0.9840826081074988, + "flos": 26133039820320.0, + "grad_norm": 1.9952650048107914, + "learning_rate": 2.642461050220479e-09, + "loss": 0.8033, + "num_input_tokens_seen": 354647640, + "step": 10912 + }, + { + "epoch": 0.984172791630969, + "flos": 33717235146720.0, + "grad_norm": 1.8451743362474562, + "learning_rate": 2.612524911005609e-09, + "loss": 0.7155, + "num_input_tokens_seen": 354679500, + "step": 10913 + }, + { + "epoch": 0.9842629751544393, + "flos": 29965175048640.0, + "grad_norm": 1.669598250234004, + "learning_rate": 2.582759199903917e-09, + "loss": 0.6145, + "num_input_tokens_seen": 354711450, + "step": 10914 + }, + { + "epoch": 0.9843531586779095, + "flos": 32588063114880.0, + "grad_norm": 2.127653820963446, + "learning_rate": 2.553163919454704e-09, + "loss": 0.6977, + "num_input_tokens_seen": 354740280, + "step": 10915 + }, + { + "epoch": 0.9844433422013799, + "flos": 37689486134880.0, + "grad_norm": 1.9816697671778845, + "learning_rate": 2.523739072183506e-09, + "loss": 0.7194, + "num_input_tokens_seen": 354771405, + "step": 10916 + }, + { + "epoch": 0.9845335257248501, + "flos": 30328641128160.0, + "grad_norm": 1.7391847469826094, + "learning_rate": 2.4944846606007597e-09, + "loss": 0.6978, + "num_input_tokens_seen": 354803075, + "step": 10917 + }, + { + "epoch": 0.9846237092483203, + "flos": 53824513960800.0, + "grad_norm": 0.6102139249106758, + "learning_rate": 2.46540068720269e-09, + "loss": 0.614, + "num_input_tokens_seen": 354894070, + "step": 10918 + }, + { + "epoch": 0.9847138927717906, + "flos": 22168408632960.0, + "grad_norm": 1.5738868836857176, + "learning_rate": 2.4364871544708674e-09, + "loss": 0.7727, + "num_input_tokens_seen": 354923035, + "step": 10919 + }, + { + "epoch": 0.9848040762952609, + "flos": 28617075898560.0, + "grad_norm": 1.639860865343254, + "learning_rate": 2.4077440648726523e-09, + "loss": 0.6865, + "num_input_tokens_seen": 354951480, + "step": 10920 + }, + { + "epoch": 0.9848942598187311, + "flos": 22860432653760.0, + "grad_norm": 1.8354012862855382, + "learning_rate": 2.379171420860082e-09, + "loss": 0.7988, + "num_input_tokens_seen": 354981540, + "step": 10921 + }, + { + "epoch": 0.9849844433422014, + "flos": 21476198763360.0, + "grad_norm": 1.9315891548296238, + "learning_rate": 2.3507692248714296e-09, + "loss": 0.6599, + "num_input_tokens_seen": 355010930, + "step": 10922 + }, + { + "epoch": 0.9850746268656716, + "flos": 29452077750720.0, + "grad_norm": 2.109325154033449, + "learning_rate": 2.322537479330089e-09, + "loss": 0.7489, + "num_input_tokens_seen": 355037495, + "step": 10923 + }, + { + "epoch": 0.9851648103891419, + "flos": 25046655915840.0, + "grad_norm": 18.874128511027504, + "learning_rate": 2.2944761866450223e-09, + "loss": 0.8142, + "num_input_tokens_seen": 355066305, + "step": 10924 + }, + { + "epoch": 0.9852549939126122, + "flos": 18124966087200.0, + "grad_norm": 2.0353708575067815, + "learning_rate": 2.266585349210315e-09, + "loss": 0.7173, + "num_input_tokens_seen": 355094370, + "step": 10925 + }, + { + "epoch": 0.9853451774360824, + "flos": 23334310121280.0, + "grad_norm": 2.342108494989296, + "learning_rate": 2.2388649694060623e-09, + "loss": 0.6961, + "num_input_tokens_seen": 355122020, + "step": 10926 + }, + { + "epoch": 0.9854353609595526, + "flos": 24316006780800.0, + "grad_norm": 1.8962487055674129, + "learning_rate": 2.211315049597262e-09, + "loss": 0.7974, + "num_input_tokens_seen": 355151475, + "step": 10927 + }, + { + "epoch": 0.985525544483023, + "flos": 28940764601280.0, + "grad_norm": 2.760500886220156, + "learning_rate": 2.1839355921349224e-09, + "loss": 0.7717, + "num_input_tokens_seen": 355179285, + "step": 10928 + }, + { + "epoch": 0.9856157280064932, + "flos": 23115568851840.0, + "grad_norm": 2.462859647406296, + "learning_rate": 2.156726599354952e-09, + "loss": 0.7335, + "num_input_tokens_seen": 355207185, + "step": 10929 + }, + { + "epoch": 0.9857059115299635, + "flos": 29017680301920.0, + "grad_norm": 1.9316865396780198, + "learning_rate": 2.129688073578828e-09, + "loss": 0.7608, + "num_input_tokens_seen": 355235490, + "step": 10930 + }, + { + "epoch": 0.9857960950534337, + "flos": 22056696303360.0, + "grad_norm": 2.681314224861127, + "learning_rate": 2.1028200171142597e-09, + "loss": 0.8671, + "num_input_tokens_seen": 355264245, + "step": 10931 + }, + { + "epoch": 0.985886278576904, + "flos": 22970026307040.0, + "grad_norm": 2.3733076792126977, + "learning_rate": 2.076122432253191e-09, + "loss": 0.7157, + "num_input_tokens_seen": 355292460, + "step": 10932 + }, + { + "epoch": 0.9859764621003743, + "flos": 29524495910400.0, + "grad_norm": 2.342407961312559, + "learning_rate": 2.0495953212738005e-09, + "loss": 0.7774, + "num_input_tokens_seen": 355320680, + "step": 10933 + }, + { + "epoch": 0.9860666456238445, + "flos": 33751548568800.0, + "grad_norm": 3.4963588848486586, + "learning_rate": 2.0232386864396102e-09, + "loss": 0.817, + "num_input_tokens_seen": 355353655, + "step": 10934 + }, + { + "epoch": 0.9861568291473148, + "flos": 26206981940160.0, + "grad_norm": 2.2700061399738103, + "learning_rate": 1.9970525299992656e-09, + "loss": 0.7613, + "num_input_tokens_seen": 355382745, + "step": 10935 + }, + { + "epoch": 0.9862470126707851, + "flos": 25441015799520.0, + "grad_norm": 2.0321874949317262, + "learning_rate": 1.9710368541874245e-09, + "loss": 0.697, + "num_input_tokens_seen": 355412640, + "step": 10936 + }, + { + "epoch": 0.9863371961942553, + "flos": 71215820162400.0, + "grad_norm": 0.5674758020779256, + "learning_rate": 1.945191661223644e-09, + "loss": 0.585, + "num_input_tokens_seen": 355515510, + "step": 10937 + }, + { + "epoch": 0.9864273797177255, + "flos": 27852150511200.0, + "grad_norm": 1.9430583471508875, + "learning_rate": 1.9195169533132714e-09, + "loss": 0.7834, + "num_input_tokens_seen": 355546160, + "step": 10938 + }, + { + "epoch": 0.9865175632411959, + "flos": 62937928176000.0, + "grad_norm": 0.5850714322844662, + "learning_rate": 1.894012732646999e-09, + "loss": 0.5923, + "num_input_tokens_seen": 355642920, + "step": 10939 + }, + { + "epoch": 0.9866077467646661, + "flos": 25666410456000.0, + "grad_norm": 3.2216064222780965, + "learning_rate": 1.8686790014010854e-09, + "loss": 0.684, + "num_input_tokens_seen": 355671980, + "step": 10940 + }, + { + "epoch": 0.9866979302881363, + "flos": 55971071355360.0, + "grad_norm": 0.6020519692810663, + "learning_rate": 1.8435157617369134e-09, + "loss": 0.5581, + "num_input_tokens_seen": 355759495, + "step": 10941 + }, + { + "epoch": 0.9867881138116066, + "flos": 22168668821280.0, + "grad_norm": 1.90320851059767, + "learning_rate": 1.818523015801876e-09, + "loss": 0.8791, + "num_input_tokens_seen": 355787025, + "step": 10942 + }, + { + "epoch": 0.9868782973350769, + "flos": 15938519806560.0, + "grad_norm": 3.0889649039534985, + "learning_rate": 1.7937007657282677e-09, + "loss": 0.698, + "num_input_tokens_seen": 355812020, + "step": 10943 + }, + { + "epoch": 0.9869684808585472, + "flos": 23587773680160.0, + "grad_norm": 1.776760600912777, + "learning_rate": 1.7690490136341718e-09, + "loss": 0.6781, + "num_input_tokens_seen": 355840520, + "step": 10944 + }, + { + "epoch": 0.9870586643820174, + "flos": 34883062295520.0, + "grad_norm": 2.233371962977105, + "learning_rate": 1.744567761622795e-09, + "loss": 0.7322, + "num_input_tokens_seen": 355874030, + "step": 10945 + }, + { + "epoch": 0.9871488479054876, + "flos": 29560704990240.0, + "grad_norm": 1.9924435320119427, + "learning_rate": 1.7202570117831327e-09, + "loss": 0.7308, + "num_input_tokens_seen": 355903775, + "step": 10946 + }, + { + "epoch": 0.987239031428958, + "flos": 21220170491040.0, + "grad_norm": 2.4567248906965786, + "learning_rate": 1.696116766189526e-09, + "loss": 0.7141, + "num_input_tokens_seen": 355930505, + "step": 10947 + }, + { + "epoch": 0.9873292149524282, + "flos": 26831048172480.0, + "grad_norm": 1.8784980568448033, + "learning_rate": 1.6721470269021042e-09, + "loss": 0.7749, + "num_input_tokens_seen": 355961490, + "step": 10948 + }, + { + "epoch": 0.9874193984758984, + "flos": 21184556127360.0, + "grad_norm": 2.2347118417222047, + "learning_rate": 1.6483477959654546e-09, + "loss": 0.7074, + "num_input_tokens_seen": 355989275, + "step": 10949 + }, + { + "epoch": 0.9875095819993687, + "flos": 40600745898240.0, + "grad_norm": 2.0461466417908354, + "learning_rate": 1.6247190754106187e-09, + "loss": 0.72, + "num_input_tokens_seen": 356023935, + "step": 10950 + }, + { + "epoch": 0.987599765522839, + "flos": 29960603168160.0, + "grad_norm": 1.566971522288522, + "learning_rate": 1.6012608672537619e-09, + "loss": 0.7002, + "num_input_tokens_seen": 356054085, + "step": 10951 + }, + { + "epoch": 0.9876899490463092, + "flos": 24573298824960.0, + "grad_norm": 2.9410902768015106, + "learning_rate": 1.5779731734963942e-09, + "loss": 0.7446, + "num_input_tokens_seen": 356080345, + "step": 10952 + }, + { + "epoch": 0.9877801325697795, + "flos": 26357839760640.0, + "grad_norm": 2.9559285414528027, + "learning_rate": 1.5548559961253705e-09, + "loss": 0.7968, + "num_input_tokens_seen": 356108350, + "step": 10953 + }, + { + "epoch": 0.9878703160932497, + "flos": 23077575623520.0, + "grad_norm": 1.9582780532609936, + "learning_rate": 1.5319093371135573e-09, + "loss": 0.6613, + "num_input_tokens_seen": 356137460, + "step": 10954 + }, + { + "epoch": 0.9879604996167201, + "flos": 28867826064960.0, + "grad_norm": 1.7597515346966421, + "learning_rate": 1.5091331984184997e-09, + "loss": 0.8264, + "num_input_tokens_seen": 356168770, + "step": 10955 + }, + { + "epoch": 0.9880506831401903, + "flos": 27153621782400.0, + "grad_norm": 1.7860231527513009, + "learning_rate": 1.486527581983754e-09, + "loss": 0.7356, + "num_input_tokens_seen": 356197475, + "step": 10956 + }, + { + "epoch": 0.9881408666636605, + "flos": 22022828918400.0, + "grad_norm": 1.9008350068380016, + "learning_rate": 1.4640924897382223e-09, + "loss": 0.7969, + "num_input_tokens_seen": 356225080, + "step": 10957 + }, + { + "epoch": 0.9882310501871309, + "flos": 23696289410400.0, + "grad_norm": 2.6305234109735522, + "learning_rate": 1.4418279235961506e-09, + "loss": 0.7773, + "num_input_tokens_seen": 356254910, + "step": 10958 + }, + { + "epoch": 0.9883212337106011, + "flos": 27086741916960.0, + "grad_norm": 1.9209762580750795, + "learning_rate": 1.4197338854573526e-09, + "loss": 0.6413, + "num_input_tokens_seen": 356284665, + "step": 10959 + }, + { + "epoch": 0.9884114172340713, + "flos": 17935669001280.0, + "grad_norm": 1.9416123708683735, + "learning_rate": 1.3978103772067651e-09, + "loss": 0.6931, + "num_input_tokens_seen": 356311090, + "step": 10960 + }, + { + "epoch": 0.9885016007575416, + "flos": 33536115408000.0, + "grad_norm": 1.7069818066173286, + "learning_rate": 1.3760574007153358e-09, + "loss": 0.6584, + "num_input_tokens_seen": 356341060, + "step": 10961 + }, + { + "epoch": 0.9885917842810119, + "flos": 20965926367200.0, + "grad_norm": 2.1666648449034245, + "learning_rate": 1.3544749578389137e-09, + "loss": 0.8187, + "num_input_tokens_seen": 356367475, + "step": 10962 + }, + { + "epoch": 0.9886819678044821, + "flos": 19797869032800.0, + "grad_norm": 1.9494618357904552, + "learning_rate": 1.3330630504189143e-09, + "loss": 0.7852, + "num_input_tokens_seen": 356393540, + "step": 10963 + }, + { + "epoch": 0.9887721513279524, + "flos": 35793939095040.0, + "grad_norm": 1.6671824690252188, + "learning_rate": 1.3118216802827652e-09, + "loss": 0.7049, + "num_input_tokens_seen": 356423635, + "step": 10964 + }, + { + "epoch": 0.9888623348514226, + "flos": 30876014678400.0, + "grad_norm": 2.2262177853017873, + "learning_rate": 1.2907508492425722e-09, + "loss": 0.7021, + "num_input_tokens_seen": 356452530, + "step": 10965 + }, + { + "epoch": 0.988952518374893, + "flos": 42461087441760.0, + "grad_norm": 1.9000970313743704, + "learning_rate": 1.2698505590962305e-09, + "loss": 0.6983, + "num_input_tokens_seen": 356483540, + "step": 10966 + }, + { + "epoch": 0.9890427018983632, + "flos": 25411497276480.0, + "grad_norm": 1.8019901611703073, + "learning_rate": 1.2491208116272022e-09, + "loss": 0.7289, + "num_input_tokens_seen": 356513790, + "step": 10967 + }, + { + "epoch": 0.9891328854218334, + "flos": 18342740942880.0, + "grad_norm": 1.9544767051122447, + "learning_rate": 1.2285616086040728e-09, + "loss": 0.7691, + "num_input_tokens_seen": 356539760, + "step": 10968 + }, + { + "epoch": 0.9892230689453037, + "flos": 24973791719040.0, + "grad_norm": 1.4698658215015354, + "learning_rate": 1.2081729517812167e-09, + "loss": 0.8025, + "num_input_tokens_seen": 356572735, + "step": 10969 + }, + { + "epoch": 0.989313252468774, + "flos": 28689494058240.0, + "grad_norm": 1.6827833876221212, + "learning_rate": 1.1879548428983533e-09, + "loss": 0.7309, + "num_input_tokens_seen": 356603645, + "step": 10970 + }, + { + "epoch": 0.9894034359922442, + "flos": 26030768609760.0, + "grad_norm": 2.266185717420588, + "learning_rate": 1.167907283680547e-09, + "loss": 0.6919, + "num_input_tokens_seen": 356633685, + "step": 10971 + }, + { + "epoch": 0.9894936195157145, + "flos": 26905027462080.0, + "grad_norm": 1.9992700675290171, + "learning_rate": 1.1480302758382077e-09, + "loss": 0.6781, + "num_input_tokens_seen": 356661165, + "step": 10972 + }, + { + "epoch": 0.9895838030391847, + "flos": 24135890625600.0, + "grad_norm": 3.55701854694326, + "learning_rate": 1.1283238210675338e-09, + "loss": 0.728, + "num_input_tokens_seen": 356692630, + "step": 10973 + }, + { + "epoch": 0.989673986562655, + "flos": 19981627824480.0, + "grad_norm": 2.511314632482177, + "learning_rate": 1.1087879210498474e-09, + "loss": 0.666, + "num_input_tokens_seen": 356720285, + "step": 10974 + }, + { + "epoch": 0.9897641700861253, + "flos": 14991991473600.0, + "grad_norm": 1.9564832578781612, + "learning_rate": 1.0894225774522592e-09, + "loss": 0.7834, + "num_input_tokens_seen": 356746730, + "step": 10975 + }, + { + "epoch": 0.9898543536095955, + "flos": 67494691038240.0, + "grad_norm": 0.6478090343023801, + "learning_rate": 1.070227791927003e-09, + "loss": 0.5941, + "num_input_tokens_seen": 356832875, + "step": 10976 + }, + { + "epoch": 0.9899445371330657, + "flos": 21476496121440.0, + "grad_norm": 2.9556507430686803, + "learning_rate": 1.0512035661118802e-09, + "loss": 0.8359, + "num_input_tokens_seen": 356859020, + "step": 10977 + }, + { + "epoch": 0.9900347206565361, + "flos": 22276887193440.0, + "grad_norm": 2.213432602532955, + "learning_rate": 1.0323499016300364e-09, + "loss": 0.7579, + "num_input_tokens_seen": 356886195, + "step": 10978 + }, + { + "epoch": 0.9901249041800063, + "flos": 22386852544320.0, + "grad_norm": 3.597059235548251, + "learning_rate": 1.013666800090407e-09, + "loss": 0.6909, + "num_input_tokens_seen": 356912935, + "step": 10979 + }, + { + "epoch": 0.9902150877034765, + "flos": 28034422512480.0, + "grad_norm": 1.7919541357506306, + "learning_rate": 9.951542630870502e-10, + "loss": 0.7198, + "num_input_tokens_seen": 356942975, + "step": 10980 + }, + { + "epoch": 0.9903052712269468, + "flos": 23841274408800.0, + "grad_norm": 2.0166969701628727, + "learning_rate": 9.768122921995915e-10, + "loss": 0.7754, + "num_input_tokens_seen": 356970560, + "step": 10981 + }, + { + "epoch": 0.9903954547504171, + "flos": 22641728554080.0, + "grad_norm": 2.5362763597407394, + "learning_rate": 9.58640888992779e-10, + "loss": 0.6756, + "num_input_tokens_seen": 356996120, + "step": 10982 + }, + { + "epoch": 0.9904856382738874, + "flos": 26977928828640.0, + "grad_norm": 2.0286638669610904, + "learning_rate": 9.40640055017594e-10, + "loss": 0.654, + "num_input_tokens_seen": 357026490, + "step": 10983 + }, + { + "epoch": 0.9905758217973576, + "flos": 24755050449600.0, + "grad_norm": 2.0367812919305908, + "learning_rate": 9.228097918094757e-10, + "loss": 0.7366, + "num_input_tokens_seen": 357056925, + "step": 10984 + }, + { + "epoch": 0.9906660053208279, + "flos": 28213312065600.0, + "grad_norm": 2.7593157510239523, + "learning_rate": 9.051501008900952e-10, + "loss": 0.7076, + "num_input_tokens_seen": 357089780, + "step": 10985 + }, + { + "epoch": 0.9907561888442982, + "flos": 25119371433600.0, + "grad_norm": 7.588243033965721, + "learning_rate": 8.876609837662475e-10, + "loss": 0.7545, + "num_input_tokens_seen": 357119240, + "step": 10986 + }, + { + "epoch": 0.9908463723677684, + "flos": 25700946896640.0, + "grad_norm": 1.6916599567950492, + "learning_rate": 8.70342441930294e-10, + "loss": 0.8233, + "num_input_tokens_seen": 357151345, + "step": 10987 + }, + { + "epoch": 0.9909365558912386, + "flos": 24822785219520.0, + "grad_norm": 2.6266002318731663, + "learning_rate": 8.531944768594979e-10, + "loss": 0.7087, + "num_input_tokens_seen": 357181070, + "step": 10988 + }, + { + "epoch": 0.991026739414709, + "flos": 24608429981760.0, + "grad_norm": 1.6372114728121605, + "learning_rate": 8.362170900175769e-10, + "loss": 0.818, + "num_input_tokens_seen": 357210720, + "step": 10989 + }, + { + "epoch": 0.9911169229381792, + "flos": 28981471222080.0, + "grad_norm": 1.8052610336450428, + "learning_rate": 8.194102828527061e-10, + "loss": 0.69, + "num_input_tokens_seen": 357239270, + "step": 10990 + }, + { + "epoch": 0.9912071064616494, + "flos": 17979089014560.0, + "grad_norm": 5.272574980099986, + "learning_rate": 8.027740567992936e-10, + "loss": 0.7812, + "num_input_tokens_seen": 357265965, + "step": 10991 + }, + { + "epoch": 0.9912972899851197, + "flos": 28325730620640.0, + "grad_norm": 1.7679030989031062, + "learning_rate": 7.863084132766484e-10, + "loss": 0.7324, + "num_input_tokens_seen": 357296820, + "step": 10992 + }, + { + "epoch": 0.99138747350859, + "flos": 25702359347520.0, + "grad_norm": 2.1717565490476365, + "learning_rate": 7.700133536896469e-10, + "loss": 0.7535, + "num_input_tokens_seen": 357325350, + "step": 10993 + }, + { + "epoch": 0.9914776570320603, + "flos": 24536495028960.0, + "grad_norm": 1.8357804900864325, + "learning_rate": 7.538888794287324e-10, + "loss": 0.7049, + "num_input_tokens_seen": 357355045, + "step": 10994 + }, + { + "epoch": 0.9915678405555305, + "flos": 25886341157760.0, + "grad_norm": 1.750901603051618, + "learning_rate": 7.379349918696931e-10, + "loss": 0.7548, + "num_input_tokens_seen": 357387635, + "step": 10995 + }, + { + "epoch": 0.9916580240790007, + "flos": 21439841004480.0, + "grad_norm": 2.390591390535548, + "learning_rate": 7.221516923738846e-10, + "loss": 0.7097, + "num_input_tokens_seen": 357414535, + "step": 10996 + }, + { + "epoch": 0.9917482076024711, + "flos": 20566065359040.0, + "grad_norm": 2.644564877723096, + "learning_rate": 7.065389822880075e-10, + "loss": 0.84, + "num_input_tokens_seen": 357441175, + "step": 10997 + }, + { + "epoch": 0.9918383911259413, + "flos": 22387484430240.0, + "grad_norm": 2.051613833038706, + "learning_rate": 6.910968629443292e-10, + "loss": 0.8132, + "num_input_tokens_seen": 357469735, + "step": 10998 + }, + { + "epoch": 0.9919285746494115, + "flos": 15282742035360.0, + "grad_norm": 2.0355048585785482, + "learning_rate": 6.758253356602406e-10, + "loss": 0.7343, + "num_input_tokens_seen": 357496855, + "step": 10999 + }, + { + "epoch": 0.9920187581728818, + "flos": 25520087346240.0, + "grad_norm": 1.7993602648996616, + "learning_rate": 6.607244017389213e-10, + "loss": 0.8333, + "num_input_tokens_seen": 357524210, + "step": 11000 + }, + { + "epoch": 0.9921089416963521, + "flos": 24390952484160.0, + "grad_norm": 2.299201254432989, + "learning_rate": 6.457940624686742e-10, + "loss": 0.7563, + "num_input_tokens_seen": 357553780, + "step": 11001 + }, + { + "epoch": 0.9921991252198223, + "flos": 25739274652800.0, + "grad_norm": 2.029058947305229, + "learning_rate": 6.310343191238132e-10, + "loss": 0.8232, + "num_input_tokens_seen": 357584435, + "step": 11002 + }, + { + "epoch": 0.9922893087432926, + "flos": 18853682394720.0, + "grad_norm": 1.889004856489087, + "learning_rate": 6.164451729635534e-10, + "loss": 0.6323, + "num_input_tokens_seen": 357612595, + "step": 11003 + }, + { + "epoch": 0.9923794922667628, + "flos": 17943474650880.0, + "grad_norm": 2.027042667912805, + "learning_rate": 6.020266252324546e-10, + "loss": 0.6073, + "num_input_tokens_seen": 357640345, + "step": 11004 + }, + { + "epoch": 0.9924696757902332, + "flos": 21986657008320.0, + "grad_norm": 2.026558608240348, + "learning_rate": 5.877786771610882e-10, + "loss": 0.7097, + "num_input_tokens_seen": 357669675, + "step": 11005 + }, + { + "epoch": 0.9925598593137034, + "flos": 67243234646400.0, + "grad_norm": 0.7083605078552148, + "learning_rate": 5.737013299651483e-10, + "loss": 0.6456, + "num_input_tokens_seen": 357765540, + "step": 11006 + }, + { + "epoch": 0.9926500428371736, + "flos": 20345279752800.0, + "grad_norm": 2.079487289547627, + "learning_rate": 5.597945848458963e-10, + "loss": 0.7404, + "num_input_tokens_seen": 357789685, + "step": 11007 + }, + { + "epoch": 0.992740226360644, + "flos": 19435964083200.0, + "grad_norm": 2.1542842584484365, + "learning_rate": 5.460584429894944e-10, + "loss": 0.7107, + "num_input_tokens_seen": 357817100, + "step": 11008 + }, + { + "epoch": 0.9928304098841142, + "flos": 20814771188640.0, + "grad_norm": 2.161111851386712, + "learning_rate": 5.32492905568338e-10, + "loss": 0.7383, + "num_input_tokens_seen": 357845610, + "step": 11009 + }, + { + "epoch": 0.9929205934075844, + "flos": 23480484552000.0, + "grad_norm": 1.7471437305528017, + "learning_rate": 5.190979737399459e-10, + "loss": 0.7682, + "num_input_tokens_seen": 357874510, + "step": 11010 + }, + { + "epoch": 0.9930107769310547, + "flos": 34625844590880.0, + "grad_norm": 1.9754522771494745, + "learning_rate": 5.058736486469594e-10, + "loss": 0.6275, + "num_input_tokens_seen": 357907545, + "step": 11011 + }, + { + "epoch": 0.993100960454525, + "flos": 20092262231040.0, + "grad_norm": 3.092048099839521, + "learning_rate": 4.928199314180314e-10, + "loss": 0.7696, + "num_input_tokens_seen": 357933965, + "step": 11012 + }, + { + "epoch": 0.9931911439779952, + "flos": 23734617166560.0, + "grad_norm": 1.9263622962652087, + "learning_rate": 4.799368231669376e-10, + "loss": 0.655, + "num_input_tokens_seen": 357964775, + "step": 11013 + }, + { + "epoch": 0.9932813275014655, + "flos": 23262040640640.0, + "grad_norm": 1.869529627836347, + "learning_rate": 4.672243249927988e-10, + "loss": 0.7233, + "num_input_tokens_seen": 357993000, + "step": 11014 + }, + { + "epoch": 0.9933715110249357, + "flos": 28945150632960.0, + "grad_norm": 1.757529002746642, + "learning_rate": 4.546824379803027e-10, + "loss": 0.7676, + "num_input_tokens_seen": 358023150, + "step": 11015 + }, + { + "epoch": 0.993461694548406, + "flos": 19727346530880.0, + "grad_norm": 2.2772491068991654, + "learning_rate": 4.4231116319970454e-10, + "loss": 0.7664, + "num_input_tokens_seen": 358049295, + "step": 11016 + }, + { + "epoch": 0.9935518780718763, + "flos": 24974237756160.0, + "grad_norm": 1.739443683919893, + "learning_rate": 4.3011050170660423e-10, + "loss": 0.7411, + "num_input_tokens_seen": 358078045, + "step": 11017 + }, + { + "epoch": 0.9936420615953465, + "flos": 25848979815360.0, + "grad_norm": 1.8014825543467587, + "learning_rate": 4.18080454542169e-10, + "loss": 0.7288, + "num_input_tokens_seen": 358107950, + "step": 11018 + }, + { + "epoch": 0.9937322451188167, + "flos": 63563592708000.0, + "grad_norm": 0.6567407771066653, + "learning_rate": 4.0622102273246694e-10, + "loss": 0.5751, + "num_input_tokens_seen": 358204455, + "step": 11019 + }, + { + "epoch": 0.9938224286422871, + "flos": 21838289561760.0, + "grad_norm": 1.6108977881789897, + "learning_rate": 3.945322072897994e-10, + "loss": 0.7967, + "num_input_tokens_seen": 358233065, + "step": 11020 + }, + { + "epoch": 0.9939126121657573, + "flos": 18853087678560.0, + "grad_norm": 2.808869116260185, + "learning_rate": 3.830140092111467e-10, + "loss": 0.7962, + "num_input_tokens_seen": 358260175, + "step": 11021 + }, + { + "epoch": 0.9940027956892276, + "flos": 19727718228480.0, + "grad_norm": 2.160777769991413, + "learning_rate": 3.7166642947972225e-10, + "loss": 0.7338, + "num_input_tokens_seen": 358286475, + "step": 11022 + }, + { + "epoch": 0.9940929792126978, + "flos": 24897322055520.0, + "grad_norm": 2.1832847308313563, + "learning_rate": 3.604894690634186e-10, + "loss": 0.7128, + "num_input_tokens_seen": 358314845, + "step": 11023 + }, + { + "epoch": 0.9941831627361681, + "flos": 24968253424800.0, + "grad_norm": 4.344943534387599, + "learning_rate": 3.494831289161393e-10, + "loss": 0.8224, + "num_input_tokens_seen": 358342060, + "step": 11024 + }, + { + "epoch": 0.9942733462596384, + "flos": 21258498247200.0, + "grad_norm": 1.9801051541543921, + "learning_rate": 3.3864740997668897e-10, + "loss": 0.7716, + "num_input_tokens_seen": 358370775, + "step": 11025 + }, + { + "epoch": 0.9943635297831086, + "flos": 23553534597600.0, + "grad_norm": 2.3185407765224793, + "learning_rate": 3.279823131701054e-10, + "loss": 0.7636, + "num_input_tokens_seen": 358403200, + "step": 11026 + }, + { + "epoch": 0.9944537133065788, + "flos": 27667536815040.0, + "grad_norm": 1.5497095358855555, + "learning_rate": 3.1748783940610536e-10, + "loss": 0.6932, + "num_input_tokens_seen": 358433815, + "step": 11027 + }, + { + "epoch": 0.9945438968300492, + "flos": 20674878447360.0, + "grad_norm": 1.7172188614540502, + "learning_rate": 3.071639895801947e-10, + "loss": 0.6981, + "num_input_tokens_seen": 358461645, + "step": 11028 + }, + { + "epoch": 0.9946340803535194, + "flos": 21148272708000.0, + "grad_norm": 2.2805330377856587, + "learning_rate": 2.9701076457322447e-10, + "loss": 0.8013, + "num_input_tokens_seen": 358487995, + "step": 11029 + }, + { + "epoch": 0.9947242638769896, + "flos": 18889333928160.0, + "grad_norm": 2.219947813984366, + "learning_rate": 2.870281652513906e-10, + "loss": 0.7794, + "num_input_tokens_seen": 358514315, + "step": 11030 + }, + { + "epoch": 0.99481444740046, + "flos": 24683003987520.0, + "grad_norm": 1.9019608317853847, + "learning_rate": 2.772161924669003e-10, + "loss": 0.7268, + "num_input_tokens_seen": 358543570, + "step": 11031 + }, + { + "epoch": 0.9949046309239302, + "flos": 27414816651360.0, + "grad_norm": 2.369506606399268, + "learning_rate": 2.6757484705641765e-10, + "loss": 0.7241, + "num_input_tokens_seen": 358574550, + "step": 11032 + }, + { + "epoch": 0.9949948144474005, + "flos": 22969877628000.0, + "grad_norm": 2.248758291300257, + "learning_rate": 2.58104129843062e-10, + "loss": 0.7363, + "num_input_tokens_seen": 358601585, + "step": 11033 + }, + { + "epoch": 0.9950849979708707, + "flos": 11923592200320.0, + "grad_norm": 2.3663656992569067, + "learning_rate": 2.4880404163463154e-10, + "loss": 0.6944, + "num_input_tokens_seen": 358626790, + "step": 11034 + }, + { + "epoch": 0.995175181494341, + "flos": 24169423482720.0, + "grad_norm": 2.693194632578376, + "learning_rate": 2.3967458322471377e-10, + "loss": 0.7266, + "num_input_tokens_seen": 358654255, + "step": 11035 + }, + { + "epoch": 0.9952653650178113, + "flos": 24132619686720.0, + "grad_norm": 1.723314985414938, + "learning_rate": 2.307157553922412e-10, + "loss": 0.75, + "num_input_tokens_seen": 358683510, + "step": 11036 + }, + { + "epoch": 0.9953555485412815, + "flos": 18671521902720.0, + "grad_norm": 1.71637902149243, + "learning_rate": 2.2192755890193538e-10, + "loss": 0.7834, + "num_input_tokens_seen": 358710705, + "step": 11037 + }, + { + "epoch": 0.9954457320647517, + "flos": 53517477310560.0, + "grad_norm": 0.6269417402237961, + "learning_rate": 2.133099945034189e-10, + "loss": 0.5444, + "num_input_tokens_seen": 358810460, + "step": 11038 + }, + { + "epoch": 0.9955359155882221, + "flos": 24675755884320.0, + "grad_norm": 2.1406120801104174, + "learning_rate": 2.048630629318815e-10, + "loss": 0.7603, + "num_input_tokens_seen": 358835520, + "step": 11039 + }, + { + "epoch": 0.9956260991116923, + "flos": 23880122541600.0, + "grad_norm": 1.672449815279459, + "learning_rate": 1.965867649080799e-10, + "loss": 0.7342, + "num_input_tokens_seen": 358866595, + "step": 11040 + }, + { + "epoch": 0.9957162826351625, + "flos": 65891901727200.0, + "grad_norm": 0.6070418318878011, + "learning_rate": 1.8848110113856008e-10, + "loss": 0.5626, + "num_input_tokens_seen": 358966380, + "step": 11041 + }, + { + "epoch": 0.9958064661586328, + "flos": 19144878993600.0, + "grad_norm": 2.3189193704857365, + "learning_rate": 1.8054607231454687e-10, + "loss": 0.7772, + "num_input_tokens_seen": 358995200, + "step": 11042 + }, + { + "epoch": 0.9958966496821031, + "flos": 20383719018240.0, + "grad_norm": 1.7637879934348437, + "learning_rate": 1.7278167911327635e-10, + "loss": 0.6914, + "num_input_tokens_seen": 359022925, + "step": 11043 + }, + { + "epoch": 0.9959868332055734, + "flos": 19763555610720.0, + "grad_norm": 1.9102982232315242, + "learning_rate": 1.6518792219710753e-10, + "loss": 0.7712, + "num_input_tokens_seen": 359050015, + "step": 11044 + }, + { + "epoch": 0.9960770167290436, + "flos": 21693416072640.0, + "grad_norm": 1.9606675041925032, + "learning_rate": 1.5776480221418865e-10, + "loss": 0.7423, + "num_input_tokens_seen": 359079395, + "step": 11045 + }, + { + "epoch": 0.9961672002525138, + "flos": 14735368485120.0, + "grad_norm": 3.291971474228443, + "learning_rate": 1.505123197977909e-10, + "loss": 0.7385, + "num_input_tokens_seen": 359102750, + "step": 11046 + }, + { + "epoch": 0.9962573837759842, + "flos": 41403813192960.0, + "grad_norm": 2.0158886404389094, + "learning_rate": 1.4343047556675258e-10, + "loss": 0.6725, + "num_input_tokens_seen": 359137715, + "step": 11047 + }, + { + "epoch": 0.9963475672994544, + "flos": 16083541974720.0, + "grad_norm": 2.332407960864596, + "learning_rate": 1.3651927012503506e-10, + "loss": 0.7417, + "num_input_tokens_seen": 359163620, + "step": 11048 + }, + { + "epoch": 0.9964377508229246, + "flos": 23183043433440.0, + "grad_norm": 2.029794494299094, + "learning_rate": 1.297787040630549e-10, + "loss": 0.7437, + "num_input_tokens_seen": 359191785, + "step": 11049 + }, + { + "epoch": 0.9965279343463949, + "flos": 19436187101760.0, + "grad_norm": 6.203679672540498, + "learning_rate": 1.2320877795524153e-10, + "loss": 0.7613, + "num_input_tokens_seen": 359216480, + "step": 11050 + }, + { + "epoch": 0.9966181178698652, + "flos": 52479802992480.0, + "grad_norm": 1.9803859234406103, + "learning_rate": 1.1680949236247962e-10, + "loss": 0.7198, + "num_input_tokens_seen": 359249600, + "step": 11051 + }, + { + "epoch": 0.9967083013933354, + "flos": 21731409300960.0, + "grad_norm": 1.8136060562695655, + "learning_rate": 1.1058084783099886e-10, + "loss": 0.7431, + "num_input_tokens_seen": 359278065, + "step": 11052 + }, + { + "epoch": 0.9967984849168057, + "flos": 25411980483360.0, + "grad_norm": 2.2352092759296944, + "learning_rate": 1.0452284489170793e-10, + "loss": 0.7498, + "num_input_tokens_seen": 359306610, + "step": 11053 + }, + { + "epoch": 0.996888668440276, + "flos": 21841300312320.0, + "grad_norm": 2.2542552975231382, + "learning_rate": 9.86354840621928e-11, + "loss": 0.7219, + "num_input_tokens_seen": 359337930, + "step": 11054 + }, + { + "epoch": 0.9969788519637462, + "flos": 29999265452160.0, + "grad_norm": 1.9884679440606743, + "learning_rate": 9.291876584427427e-11, + "loss": 0.6123, + "num_input_tokens_seen": 359368320, + "step": 11055 + }, + { + "epoch": 0.9970690354872165, + "flos": 21759701221920.0, + "grad_norm": 2.0935160947499427, + "learning_rate": 8.737269072578435e-11, + "loss": 0.7088, + "num_input_tokens_seen": 359392155, + "step": 11056 + }, + { + "epoch": 0.9971592190106867, + "flos": 23844285159360.0, + "grad_norm": 1.8144593007749448, + "learning_rate": 8.199725918012212e-11, + "loss": 0.7333, + "num_input_tokens_seen": 359420390, + "step": 11057 + }, + { + "epoch": 0.9972494025341571, + "flos": 19946868365280.0, + "grad_norm": 3.31867951571515, + "learning_rate": 7.679247166603175e-11, + "loss": 0.7463, + "num_input_tokens_seen": 359448290, + "step": 11058 + }, + { + "epoch": 0.9973395860576273, + "flos": 21840965784480.0, + "grad_norm": 2.0517823309978978, + "learning_rate": 7.17583286273804e-11, + "loss": 0.746, + "num_input_tokens_seen": 359476190, + "step": 11059 + }, + { + "epoch": 0.9974297695810975, + "flos": 25589383246080.0, + "grad_norm": 2.251914481421202, + "learning_rate": 6.689483049360233e-11, + "loss": 0.758, + "num_input_tokens_seen": 359504125, + "step": 11060 + }, + { + "epoch": 0.9975199531045678, + "flos": 16703928400800.0, + "grad_norm": 2.1777186098084065, + "learning_rate": 6.220197768014302e-11, + "loss": 0.7168, + "num_input_tokens_seen": 359531210, + "step": 11061 + }, + { + "epoch": 0.9976101366280381, + "flos": 24752076868800.0, + "grad_norm": 2.271595946414835, + "learning_rate": 5.7679770587126806e-11, + "loss": 0.7587, + "num_input_tokens_seen": 359559610, + "step": 11062 + }, + { + "epoch": 0.9977003201515083, + "flos": 68318653471680.0, + "grad_norm": 0.6123780567071877, + "learning_rate": 5.33282096002452e-11, + "loss": 0.6644, + "num_input_tokens_seen": 359648480, + "step": 11063 + }, + { + "epoch": 0.9977905036749786, + "flos": 22236403591200.0, + "grad_norm": 5.118487802946883, + "learning_rate": 4.914729509120086e-11, + "loss": 0.7083, + "num_input_tokens_seen": 359673525, + "step": 11064 + }, + { + "epoch": 0.9978806871984488, + "flos": 38378796763200.0, + "grad_norm": 1.6440046831096764, + "learning_rate": 4.513702741637537e-11, + "loss": 0.6615, + "num_input_tokens_seen": 359705890, + "step": 11065 + }, + { + "epoch": 0.9979708707219191, + "flos": 23844210819840.0, + "grad_norm": 2.6832053890488066, + "learning_rate": 4.129740691816153e-11, + "loss": 0.7991, + "num_input_tokens_seen": 359734380, + "step": 11066 + }, + { + "epoch": 0.9980610542453894, + "flos": 25521016590240.0, + "grad_norm": 1.7522433071457668, + "learning_rate": 3.762843392429715e-11, + "loss": 0.7572, + "num_input_tokens_seen": 359764085, + "step": 11067 + }, + { + "epoch": 0.9981512377688596, + "flos": 30875828829600.0, + "grad_norm": 2.6723332597693594, + "learning_rate": 3.413010874742106e-11, + "loss": 0.7175, + "num_input_tokens_seen": 359794820, + "step": 11068 + }, + { + "epoch": 0.9982414212923298, + "flos": 20165052088320.0, + "grad_norm": 2.406554462938073, + "learning_rate": 3.080243168618324e-11, + "loss": 0.7346, + "num_input_tokens_seen": 359823615, + "step": 11069 + }, + { + "epoch": 0.9983316048158002, + "flos": 24166635750720.0, + "grad_norm": 1.8152421399897132, + "learning_rate": 2.7645403024800783e-11, + "loss": 0.6857, + "num_input_tokens_seen": 359852190, + "step": 11070 + }, + { + "epoch": 0.9984217883392704, + "flos": 25771580907840.0, + "grad_norm": 1.4948916075079262, + "learning_rate": 2.4659023032391756e-11, + "loss": 0.8096, + "num_input_tokens_seen": 359881845, + "step": 11071 + }, + { + "epoch": 0.9985119718627407, + "flos": 31749715984320.0, + "grad_norm": 1.6260407122996736, + "learning_rate": 2.1843291963863364e-11, + "loss": 0.7854, + "num_input_tokens_seen": 359916640, + "step": 11072 + }, + { + "epoch": 0.9986021553862109, + "flos": 27997953244320.0, + "grad_norm": 2.355921055020005, + "learning_rate": 1.9198210059245822e-11, + "loss": 0.7577, + "num_input_tokens_seen": 359947255, + "step": 11073 + }, + { + "epoch": 0.9986923389096812, + "flos": 18889594116480.0, + "grad_norm": 2.141431559065539, + "learning_rate": 1.672377754458054e-11, + "loss": 0.7446, + "num_input_tokens_seen": 359973920, + "step": 11074 + }, + { + "epoch": 0.9987825224331515, + "flos": 27556270522560.0, + "grad_norm": 1.741781007703433, + "learning_rate": 1.4419994630809895e-11, + "loss": 0.7335, + "num_input_tokens_seen": 360002265, + "step": 11075 + }, + { + "epoch": 0.9988727059566217, + "flos": 28216397155680.0, + "grad_norm": 2.405348571768546, + "learning_rate": 1.2286861514443358e-11, + "loss": 0.6911, + "num_input_tokens_seen": 360033125, + "step": 11076 + }, + { + "epoch": 0.998962889480092, + "flos": 24717131560800.0, + "grad_norm": 1.5553175066652167, + "learning_rate": 1.0324378377779553e-11, + "loss": 0.777, + "num_input_tokens_seen": 360062410, + "step": 11077 + }, + { + "epoch": 0.9990530730035623, + "flos": 21589323543840.0, + "grad_norm": 1.836122544073976, + "learning_rate": 8.532545388018064e-12, + "loss": 0.7374, + "num_input_tokens_seen": 360095190, + "step": 11078 + }, + { + "epoch": 0.9991432565270325, + "flos": 28617410426400.0, + "grad_norm": 1.7601712825182325, + "learning_rate": 6.911362697925582e-12, + "loss": 0.575, + "num_input_tokens_seen": 360129270, + "step": 11079 + }, + { + "epoch": 0.9992334400505027, + "flos": 26030657100480.0, + "grad_norm": 1.8474848145517704, + "learning_rate": 5.46083044605794e-12, + "loss": 0.7094, + "num_input_tokens_seen": 360157605, + "step": 11080 + }, + { + "epoch": 0.9993236235739731, + "flos": 25479306385920.0, + "grad_norm": 1.8049620712606609, + "learning_rate": 4.1809487563160276e-12, + "loss": 0.7883, + "num_input_tokens_seen": 360186810, + "step": 11081 + }, + { + "epoch": 0.9994138070974433, + "flos": 24645531135840.0, + "grad_norm": 1.7686915141973738, + "learning_rate": 3.0717177375017e-12, + "loss": 0.7457, + "num_input_tokens_seen": 360217850, + "step": 11082 + }, + { + "epoch": 0.9995039906209136, + "flos": 24536643708000.0, + "grad_norm": 2.491157534156632, + "learning_rate": 2.1331374846500495e-12, + "loss": 0.7382, + "num_input_tokens_seen": 360244735, + "step": 11083 + }, + { + "epoch": 0.9995941741443838, + "flos": 21148086859200.0, + "grad_norm": 2.1108236354158825, + "learning_rate": 1.3652080774750885e-12, + "loss": 0.8361, + "num_input_tokens_seen": 360273275, + "step": 11084 + }, + { + "epoch": 0.9996843576678541, + "flos": 22496520537120.0, + "grad_norm": 1.4947101779041232, + "learning_rate": 7.679295817020204e-13, + "loss": 0.6566, + "num_input_tokens_seen": 360303005, + "step": 11085 + }, + { + "epoch": 0.9997745411913244, + "flos": 27520321631040.0, + "grad_norm": 1.884501795512068, + "learning_rate": 3.413020484011042e-13, + "loss": 0.7152, + "num_input_tokens_seen": 360331720, + "step": 11086 + }, + { + "epoch": 0.9998647247147946, + "flos": 17098845830880.0, + "grad_norm": 1.9920263487019265, + "learning_rate": 8.53255139876552e-14, + "loss": 0.7791, + "num_input_tokens_seen": 360355640, + "step": 11087 + }, + { + "epoch": 0.9999549082382648, + "flos": 22386034809600.0, + "grad_norm": 1.827759274147829, + "learning_rate": 0.0, + "loss": 0.6795, + "num_input_tokens_seen": 360384005, + "step": 11088 + }, + { + "epoch": 0.9999549082382648, + "num_input_tokens_seen": 360384005, + "step": 11088, + "total_flos": 1.3974821586704794e+18, + "train_loss": 0.7512187881860211, + "train_runtime": 130584.3731, + "train_samples_per_second": 5.095, + "train_steps_per_second": 0.085 + } + ], + "logging_steps": 1.0, + "max_steps": 11088, + "num_input_tokens_seen": 360384005, + "num_train_epochs": 1, + "save_steps": 832, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 1.3974821586704794e+18, + "train_batch_size": 5, + "trial_name": null, + "trial_params": null +}