{ "best_metric": 0.020646410062909126, "best_model_checkpoint": "/home/paperspace/Data/models/gommt/llm3br256/checkpoint-1200", "epoch": 7.741935483870968, "eval_steps": 25, "global_step": 1200, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0064516129032258064, "grad_norm": 0.5787440538406372, "learning_rate": 2.577319587628866e-07, "loss": 0.2947, "step": 1 }, { "epoch": 0.012903225806451613, "grad_norm": 0.5514878630638123, "learning_rate": 5.154639175257732e-07, "loss": 0.2774, "step": 2 }, { "epoch": 0.01935483870967742, "grad_norm": 0.5519412159919739, "learning_rate": 7.731958762886599e-07, "loss": 0.301, "step": 3 }, { "epoch": 0.025806451612903226, "grad_norm": 0.5313215851783752, "learning_rate": 1.0309278350515464e-06, "loss": 0.2792, "step": 4 }, { "epoch": 0.03225806451612903, "grad_norm": 0.5504052639007568, "learning_rate": 1.288659793814433e-06, "loss": 0.2918, "step": 5 }, { "epoch": 0.03870967741935484, "grad_norm": 0.5057755708694458, "learning_rate": 1.5463917525773197e-06, "loss": 0.2731, "step": 6 }, { "epoch": 0.04516129032258064, "grad_norm": 0.5207090377807617, "learning_rate": 1.804123711340206e-06, "loss": 0.3004, "step": 7 }, { "epoch": 0.05161290322580645, "grad_norm": 0.5149654746055603, "learning_rate": 2.061855670103093e-06, "loss": 0.2605, "step": 8 }, { "epoch": 0.05806451612903226, "grad_norm": 0.5021620988845825, "learning_rate": 2.3195876288659796e-06, "loss": 0.2932, "step": 9 }, { "epoch": 0.06451612903225806, "grad_norm": 0.5057784914970398, "learning_rate": 2.577319587628866e-06, "loss": 0.2761, "step": 10 }, { "epoch": 0.07096774193548387, "grad_norm": 0.49885252118110657, "learning_rate": 2.8350515463917527e-06, "loss": 0.2779, "step": 11 }, { "epoch": 0.07741935483870968, "grad_norm": 0.5161502957344055, "learning_rate": 3.0927835051546395e-06, "loss": 0.2666, "step": 12 }, { "epoch": 0.08387096774193549, "grad_norm": 0.4862574338912964, "learning_rate": 3.350515463917526e-06, "loss": 0.2891, "step": 13 }, { "epoch": 0.09032258064516129, "grad_norm": 0.42886313796043396, "learning_rate": 3.608247422680412e-06, "loss": 0.2605, "step": 14 }, { "epoch": 0.0967741935483871, "grad_norm": 0.41169312596321106, "learning_rate": 3.865979381443299e-06, "loss": 0.2515, "step": 15 }, { "epoch": 0.1032258064516129, "grad_norm": 0.39097175002098083, "learning_rate": 4.123711340206186e-06, "loss": 0.2676, "step": 16 }, { "epoch": 0.10967741935483871, "grad_norm": 0.3687884211540222, "learning_rate": 4.381443298969072e-06, "loss": 0.2564, "step": 17 }, { "epoch": 0.11612903225806452, "grad_norm": 0.36199015378952026, "learning_rate": 4.639175257731959e-06, "loss": 0.2631, "step": 18 }, { "epoch": 0.12258064516129032, "grad_norm": 0.33205243945121765, "learning_rate": 4.8969072164948455e-06, "loss": 0.2598, "step": 19 }, { "epoch": 0.12903225806451613, "grad_norm": 0.2873350977897644, "learning_rate": 5.154639175257732e-06, "loss": 0.2445, "step": 20 }, { "epoch": 0.13548387096774195, "grad_norm": 0.27828869223594666, "learning_rate": 5.412371134020619e-06, "loss": 0.2312, "step": 21 }, { "epoch": 0.14193548387096774, "grad_norm": 0.2796691358089447, "learning_rate": 5.670103092783505e-06, "loss": 0.2373, "step": 22 }, { "epoch": 0.14838709677419354, "grad_norm": 0.2518955171108246, "learning_rate": 5.927835051546392e-06, "loss": 0.2147, "step": 23 }, { "epoch": 0.15483870967741936, "grad_norm": 0.24648350477218628, "learning_rate": 6.185567010309279e-06, "loss": 0.209, "step": 24 }, { "epoch": 0.16129032258064516, "grad_norm": 0.24219118058681488, "learning_rate": 6.443298969072164e-06, "loss": 0.232, "step": 25 }, { "epoch": 0.16129032258064516, "eval_loss": 0.2192624807357788, "eval_runtime": 21.7174, "eval_samples_per_second": 4.605, "eval_steps_per_second": 0.092, "step": 25 }, { "epoch": 0.16774193548387098, "grad_norm": 0.23314689099788666, "learning_rate": 6.701030927835052e-06, "loss": 0.2115, "step": 26 }, { "epoch": 0.17419354838709677, "grad_norm": 0.2250155806541443, "learning_rate": 6.958762886597939e-06, "loss": 0.2171, "step": 27 }, { "epoch": 0.18064516129032257, "grad_norm": 0.21403469145298004, "learning_rate": 7.216494845360824e-06, "loss": 0.1991, "step": 28 }, { "epoch": 0.1870967741935484, "grad_norm": 0.2005668729543686, "learning_rate": 7.4742268041237115e-06, "loss": 0.2015, "step": 29 }, { "epoch": 0.1935483870967742, "grad_norm": 0.2055639624595642, "learning_rate": 7.731958762886599e-06, "loss": 0.1995, "step": 30 }, { "epoch": 0.2, "grad_norm": 0.20002567768096924, "learning_rate": 7.989690721649484e-06, "loss": 0.1987, "step": 31 }, { "epoch": 0.2064516129032258, "grad_norm": 0.19098369777202606, "learning_rate": 8.247422680412371e-06, "loss": 0.1842, "step": 32 }, { "epoch": 0.2129032258064516, "grad_norm": 0.17131243646144867, "learning_rate": 8.505154639175259e-06, "loss": 0.1934, "step": 33 }, { "epoch": 0.21935483870967742, "grad_norm": 0.19705425202846527, "learning_rate": 8.762886597938144e-06, "loss": 0.1983, "step": 34 }, { "epoch": 0.22580645161290322, "grad_norm": 0.1701938658952713, "learning_rate": 9.020618556701031e-06, "loss": 0.1854, "step": 35 }, { "epoch": 0.23225806451612904, "grad_norm": 0.16887512803077698, "learning_rate": 9.278350515463918e-06, "loss": 0.1883, "step": 36 }, { "epoch": 0.23870967741935484, "grad_norm": 0.16543127596378326, "learning_rate": 9.536082474226804e-06, "loss": 0.159, "step": 37 }, { "epoch": 0.24516129032258063, "grad_norm": 0.16382737457752228, "learning_rate": 9.793814432989691e-06, "loss": 0.1893, "step": 38 }, { "epoch": 0.25161290322580643, "grad_norm": 0.16113704442977905, "learning_rate": 1.0051546391752578e-05, "loss": 0.1746, "step": 39 }, { "epoch": 0.25806451612903225, "grad_norm": 0.15014876425266266, "learning_rate": 1.0309278350515464e-05, "loss": 0.1712, "step": 40 }, { "epoch": 0.2645161290322581, "grad_norm": 0.16049867868423462, "learning_rate": 1.0567010309278351e-05, "loss": 0.1589, "step": 41 }, { "epoch": 0.2709677419354839, "grad_norm": 0.13896311819553375, "learning_rate": 1.0824742268041238e-05, "loss": 0.1621, "step": 42 }, { "epoch": 0.27741935483870966, "grad_norm": 0.14771641790866852, "learning_rate": 1.1082474226804124e-05, "loss": 0.1783, "step": 43 }, { "epoch": 0.2838709677419355, "grad_norm": 0.14235621690750122, "learning_rate": 1.134020618556701e-05, "loss": 0.1637, "step": 44 }, { "epoch": 0.2903225806451613, "grad_norm": 0.13013780117034912, "learning_rate": 1.1597938144329898e-05, "loss": 0.1537, "step": 45 }, { "epoch": 0.2967741935483871, "grad_norm": 0.13013966381549835, "learning_rate": 1.1855670103092783e-05, "loss": 0.1574, "step": 46 }, { "epoch": 0.3032258064516129, "grad_norm": 0.1356808990240097, "learning_rate": 1.211340206185567e-05, "loss": 0.1585, "step": 47 }, { "epoch": 0.3096774193548387, "grad_norm": 0.1297629475593567, "learning_rate": 1.2371134020618558e-05, "loss": 0.1554, "step": 48 }, { "epoch": 0.3161290322580645, "grad_norm": 0.13573496043682098, "learning_rate": 1.2628865979381443e-05, "loss": 0.1584, "step": 49 }, { "epoch": 0.3225806451612903, "grad_norm": 0.14283044636249542, "learning_rate": 1.2886597938144329e-05, "loss": 0.1524, "step": 50 }, { "epoch": 0.3225806451612903, "eval_loss": 0.15066054463386536, "eval_runtime": 20.6503, "eval_samples_per_second": 4.843, "eval_steps_per_second": 0.097, "step": 50 }, { "epoch": 0.32903225806451614, "grad_norm": 0.1236417219042778, "learning_rate": 1.3144329896907218e-05, "loss": 0.1459, "step": 51 }, { "epoch": 0.33548387096774196, "grad_norm": 0.12115009129047394, "learning_rate": 1.3402061855670103e-05, "loss": 0.1408, "step": 52 }, { "epoch": 0.3419354838709677, "grad_norm": 0.1199135035276413, "learning_rate": 1.3659793814432989e-05, "loss": 0.1444, "step": 53 }, { "epoch": 0.34838709677419355, "grad_norm": 0.1241321787238121, "learning_rate": 1.3917525773195878e-05, "loss": 0.1416, "step": 54 }, { "epoch": 0.3548387096774194, "grad_norm": 0.11240442097187042, "learning_rate": 1.4175257731958763e-05, "loss": 0.1229, "step": 55 }, { "epoch": 0.36129032258064514, "grad_norm": 0.11829587817192078, "learning_rate": 1.4432989690721649e-05, "loss": 0.1387, "step": 56 }, { "epoch": 0.36774193548387096, "grad_norm": 0.1287028193473816, "learning_rate": 1.4690721649484537e-05, "loss": 0.1481, "step": 57 }, { "epoch": 0.3741935483870968, "grad_norm": 0.11497591435909271, "learning_rate": 1.4948453608247423e-05, "loss": 0.1259, "step": 58 }, { "epoch": 0.38064516129032255, "grad_norm": 0.11566972732543945, "learning_rate": 1.5206185567010308e-05, "loss": 0.1193, "step": 59 }, { "epoch": 0.3870967741935484, "grad_norm": 0.1135919988155365, "learning_rate": 1.5463917525773197e-05, "loss": 0.1215, "step": 60 }, { "epoch": 0.3935483870967742, "grad_norm": 0.1099628284573555, "learning_rate": 1.5721649484536083e-05, "loss": 0.1331, "step": 61 }, { "epoch": 0.4, "grad_norm": 0.10907468944787979, "learning_rate": 1.5979381443298968e-05, "loss": 0.117, "step": 62 }, { "epoch": 0.4064516129032258, "grad_norm": 0.11619539558887482, "learning_rate": 1.6237113402061857e-05, "loss": 0.1126, "step": 63 }, { "epoch": 0.4129032258064516, "grad_norm": 0.109222911298275, "learning_rate": 1.6494845360824743e-05, "loss": 0.1192, "step": 64 }, { "epoch": 0.41935483870967744, "grad_norm": 0.12224064767360687, "learning_rate": 1.6752577319587628e-05, "loss": 0.1245, "step": 65 }, { "epoch": 0.4258064516129032, "grad_norm": 0.12005683034658432, "learning_rate": 1.7010309278350517e-05, "loss": 0.1111, "step": 66 }, { "epoch": 0.432258064516129, "grad_norm": 0.11314112693071365, "learning_rate": 1.7268041237113403e-05, "loss": 0.1109, "step": 67 }, { "epoch": 0.43870967741935485, "grad_norm": 0.11626581847667694, "learning_rate": 1.7525773195876288e-05, "loss": 0.1224, "step": 68 }, { "epoch": 0.44516129032258067, "grad_norm": 0.11105289310216904, "learning_rate": 1.7783505154639177e-05, "loss": 0.1142, "step": 69 }, { "epoch": 0.45161290322580644, "grad_norm": 0.12696442008018494, "learning_rate": 1.8041237113402062e-05, "loss": 0.1077, "step": 70 }, { "epoch": 0.45806451612903226, "grad_norm": 0.11383359134197235, "learning_rate": 1.8298969072164948e-05, "loss": 0.1076, "step": 71 }, { "epoch": 0.4645161290322581, "grad_norm": 0.11222764849662781, "learning_rate": 1.8556701030927837e-05, "loss": 0.1212, "step": 72 }, { "epoch": 0.47096774193548385, "grad_norm": 0.12386021018028259, "learning_rate": 1.8814432989690722e-05, "loss": 0.1203, "step": 73 }, { "epoch": 0.4774193548387097, "grad_norm": 0.10532196611166, "learning_rate": 1.9072164948453608e-05, "loss": 0.1079, "step": 74 }, { "epoch": 0.4838709677419355, "grad_norm": 0.1084241271018982, "learning_rate": 1.9329896907216497e-05, "loss": 0.115, "step": 75 }, { "epoch": 0.4838709677419355, "eval_loss": 0.11645982414484024, "eval_runtime": 20.6021, "eval_samples_per_second": 4.854, "eval_steps_per_second": 0.097, "step": 75 }, { "epoch": 0.49032258064516127, "grad_norm": 0.12820391356945038, "learning_rate": 1.9587628865979382e-05, "loss": 0.1091, "step": 76 }, { "epoch": 0.4967741935483871, "grad_norm": 0.1124340295791626, "learning_rate": 1.9845360824742268e-05, "loss": 0.1129, "step": 77 }, { "epoch": 0.5032258064516129, "grad_norm": 0.10982845723628998, "learning_rate": 2.0103092783505157e-05, "loss": 0.126, "step": 78 }, { "epoch": 0.5096774193548387, "grad_norm": 0.1185258999466896, "learning_rate": 2.0360824742268042e-05, "loss": 0.1094, "step": 79 }, { "epoch": 0.5161290322580645, "grad_norm": 0.0972081646323204, "learning_rate": 2.0618556701030927e-05, "loss": 0.103, "step": 80 }, { "epoch": 0.5225806451612903, "grad_norm": 0.10331461578607559, "learning_rate": 2.0876288659793816e-05, "loss": 0.1107, "step": 81 }, { "epoch": 0.5290322580645161, "grad_norm": 0.11630548536777496, "learning_rate": 2.1134020618556702e-05, "loss": 0.1184, "step": 82 }, { "epoch": 0.535483870967742, "grad_norm": 0.10714004188776016, "learning_rate": 2.1391752577319587e-05, "loss": 0.1059, "step": 83 }, { "epoch": 0.5419354838709678, "grad_norm": 0.10682078450918198, "learning_rate": 2.1649484536082476e-05, "loss": 0.1047, "step": 84 }, { "epoch": 0.5483870967741935, "grad_norm": 0.10373492538928986, "learning_rate": 2.1907216494845362e-05, "loss": 0.0906, "step": 85 }, { "epoch": 0.5548387096774193, "grad_norm": 0.10939467698335648, "learning_rate": 2.2164948453608247e-05, "loss": 0.105, "step": 86 }, { "epoch": 0.5612903225806452, "grad_norm": 0.1066933274269104, "learning_rate": 2.2422680412371136e-05, "loss": 0.1046, "step": 87 }, { "epoch": 0.567741935483871, "grad_norm": 0.11235528439283371, "learning_rate": 2.268041237113402e-05, "loss": 0.1183, "step": 88 }, { "epoch": 0.5741935483870968, "grad_norm": 0.11120587587356567, "learning_rate": 2.2938144329896907e-05, "loss": 0.1012, "step": 89 }, { "epoch": 0.5806451612903226, "grad_norm": 0.11943212896585464, "learning_rate": 2.3195876288659796e-05, "loss": 0.1081, "step": 90 }, { "epoch": 0.5870967741935483, "grad_norm": 0.11067789793014526, "learning_rate": 2.345360824742268e-05, "loss": 0.106, "step": 91 }, { "epoch": 0.5935483870967742, "grad_norm": 0.10897871106863022, "learning_rate": 2.3711340206185567e-05, "loss": 0.1017, "step": 92 }, { "epoch": 0.6, "grad_norm": 0.10811934620141983, "learning_rate": 2.3969072164948456e-05, "loss": 0.1043, "step": 93 }, { "epoch": 0.6064516129032258, "grad_norm": 0.12143644690513611, "learning_rate": 2.422680412371134e-05, "loss": 0.0906, "step": 94 }, { "epoch": 0.6129032258064516, "grad_norm": 0.11472067981958389, "learning_rate": 2.4484536082474227e-05, "loss": 0.1064, "step": 95 }, { "epoch": 0.6193548387096774, "grad_norm": 0.11486558616161346, "learning_rate": 2.4742268041237116e-05, "loss": 0.1076, "step": 96 }, { "epoch": 0.6258064516129033, "grad_norm": 0.11682534962892532, "learning_rate": 2.5e-05, "loss": 0.1015, "step": 97 }, { "epoch": 0.632258064516129, "grad_norm": 0.10343407839536667, "learning_rate": 2.5257731958762887e-05, "loss": 0.0951, "step": 98 }, { "epoch": 0.6387096774193548, "grad_norm": 0.13064107298851013, "learning_rate": 2.5515463917525772e-05, "loss": 0.0952, "step": 99 }, { "epoch": 0.6451612903225806, "grad_norm": 0.11045321077108383, "learning_rate": 2.5773195876288658e-05, "loss": 0.0875, "step": 100 }, { "epoch": 0.6451612903225806, "eval_loss": 0.1003575325012207, "eval_runtime": 20.6227, "eval_samples_per_second": 4.849, "eval_steps_per_second": 0.097, "step": 100 }, { "epoch": 0.6516129032258065, "grad_norm": 0.10884646326303482, "learning_rate": 2.603092783505155e-05, "loss": 0.1082, "step": 101 }, { "epoch": 0.6580645161290323, "grad_norm": 0.12365655601024628, "learning_rate": 2.6288659793814435e-05, "loss": 0.1079, "step": 102 }, { "epoch": 0.6645161290322581, "grad_norm": 0.12333594262599945, "learning_rate": 2.654639175257732e-05, "loss": 0.0987, "step": 103 }, { "epoch": 0.6709677419354839, "grad_norm": 0.12649482488632202, "learning_rate": 2.6804123711340206e-05, "loss": 0.1008, "step": 104 }, { "epoch": 0.6774193548387096, "grad_norm": 0.12309876084327698, "learning_rate": 2.7061855670103092e-05, "loss": 0.0865, "step": 105 }, { "epoch": 0.6838709677419355, "grad_norm": 0.11590270698070526, "learning_rate": 2.7319587628865977e-05, "loss": 0.0943, "step": 106 }, { "epoch": 0.6903225806451613, "grad_norm": 0.13238589465618134, "learning_rate": 2.757731958762887e-05, "loss": 0.1092, "step": 107 }, { "epoch": 0.6967741935483871, "grad_norm": 0.10838130116462708, "learning_rate": 2.7835051546391755e-05, "loss": 0.0972, "step": 108 }, { "epoch": 0.7032258064516129, "grad_norm": 0.11450201272964478, "learning_rate": 2.809278350515464e-05, "loss": 0.0886, "step": 109 }, { "epoch": 0.7096774193548387, "grad_norm": 0.11476748436689377, "learning_rate": 2.8350515463917526e-05, "loss": 0.0934, "step": 110 }, { "epoch": 0.7161290322580646, "grad_norm": 0.1416608691215515, "learning_rate": 2.860824742268041e-05, "loss": 0.1044, "step": 111 }, { "epoch": 0.7225806451612903, "grad_norm": 0.11946123838424683, "learning_rate": 2.8865979381443297e-05, "loss": 0.1001, "step": 112 }, { "epoch": 0.7290322580645161, "grad_norm": 0.10268019139766693, "learning_rate": 2.912371134020619e-05, "loss": 0.0879, "step": 113 }, { "epoch": 0.7354838709677419, "grad_norm": 0.12969037890434265, "learning_rate": 2.9381443298969075e-05, "loss": 0.1046, "step": 114 }, { "epoch": 0.7419354838709677, "grad_norm": 0.112672358751297, "learning_rate": 2.963917525773196e-05, "loss": 0.0979, "step": 115 }, { "epoch": 0.7483870967741936, "grad_norm": 0.11950547248125076, "learning_rate": 2.9896907216494846e-05, "loss": 0.0983, "step": 116 }, { "epoch": 0.7548387096774194, "grad_norm": 0.11480306833982468, "learning_rate": 3.015463917525773e-05, "loss": 0.0772, "step": 117 }, { "epoch": 0.7612903225806451, "grad_norm": 0.11258739978075027, "learning_rate": 3.0412371134020617e-05, "loss": 0.0855, "step": 118 }, { "epoch": 0.7677419354838709, "grad_norm": 0.11581332236528397, "learning_rate": 3.0670103092783506e-05, "loss": 0.0864, "step": 119 }, { "epoch": 0.7741935483870968, "grad_norm": 0.13305814564228058, "learning_rate": 3.0927835051546395e-05, "loss": 0.0917, "step": 120 }, { "epoch": 0.7806451612903226, "grad_norm": 0.11415888369083405, "learning_rate": 3.118556701030928e-05, "loss": 0.0948, "step": 121 }, { "epoch": 0.7870967741935484, "grad_norm": 0.12214888632297516, "learning_rate": 3.1443298969072166e-05, "loss": 0.0931, "step": 122 }, { "epoch": 0.7935483870967742, "grad_norm": 0.13365623354911804, "learning_rate": 3.1701030927835054e-05, "loss": 0.0816, "step": 123 }, { "epoch": 0.8, "grad_norm": 0.12570084631443024, "learning_rate": 3.1958762886597937e-05, "loss": 0.0859, "step": 124 }, { "epoch": 0.8064516129032258, "grad_norm": 0.11142268776893616, "learning_rate": 3.2216494845360825e-05, "loss": 0.092, "step": 125 }, { "epoch": 0.8064516129032258, "eval_loss": 0.09087119251489639, "eval_runtime": 20.6603, "eval_samples_per_second": 4.84, "eval_steps_per_second": 0.097, "step": 125 }, { "epoch": 0.8129032258064516, "grad_norm": 0.13126152753829956, "learning_rate": 3.2474226804123714e-05, "loss": 0.085, "step": 126 }, { "epoch": 0.8193548387096774, "grad_norm": 0.12571650743484497, "learning_rate": 3.2731958762886596e-05, "loss": 0.092, "step": 127 }, { "epoch": 0.8258064516129032, "grad_norm": 0.11935378611087799, "learning_rate": 3.2989690721649485e-05, "loss": 0.0923, "step": 128 }, { "epoch": 0.832258064516129, "grad_norm": 0.1310584396123886, "learning_rate": 3.3247422680412374e-05, "loss": 0.0854, "step": 129 }, { "epoch": 0.8387096774193549, "grad_norm": 0.12715069949626923, "learning_rate": 3.3505154639175256e-05, "loss": 0.0917, "step": 130 }, { "epoch": 0.8451612903225807, "grad_norm": 0.13257332146167755, "learning_rate": 3.3762886597938145e-05, "loss": 0.083, "step": 131 }, { "epoch": 0.8516129032258064, "grad_norm": 0.11446800082921982, "learning_rate": 3.4020618556701034e-05, "loss": 0.0901, "step": 132 }, { "epoch": 0.8580645161290322, "grad_norm": 0.10843442380428314, "learning_rate": 3.4278350515463916e-05, "loss": 0.0892, "step": 133 }, { "epoch": 0.864516129032258, "grad_norm": 0.13865868747234344, "learning_rate": 3.4536082474226805e-05, "loss": 0.0843, "step": 134 }, { "epoch": 0.8709677419354839, "grad_norm": 0.11984317004680634, "learning_rate": 3.4793814432989694e-05, "loss": 0.0909, "step": 135 }, { "epoch": 0.8774193548387097, "grad_norm": 0.12147698551416397, "learning_rate": 3.5051546391752576e-05, "loss": 0.0818, "step": 136 }, { "epoch": 0.8838709677419355, "grad_norm": 0.14456403255462646, "learning_rate": 3.5309278350515465e-05, "loss": 0.0829, "step": 137 }, { "epoch": 0.8903225806451613, "grad_norm": 0.11969202756881714, "learning_rate": 3.5567010309278354e-05, "loss": 0.0846, "step": 138 }, { "epoch": 0.896774193548387, "grad_norm": 0.12115851789712906, "learning_rate": 3.5824742268041236e-05, "loss": 0.0736, "step": 139 }, { "epoch": 0.9032258064516129, "grad_norm": 0.1291874796152115, "learning_rate": 3.6082474226804125e-05, "loss": 0.0888, "step": 140 }, { "epoch": 0.9096774193548387, "grad_norm": 0.11851019412279129, "learning_rate": 3.6340206185567014e-05, "loss": 0.0847, "step": 141 }, { "epoch": 0.9161290322580645, "grad_norm": 0.12146991491317749, "learning_rate": 3.6597938144329896e-05, "loss": 0.0813, "step": 142 }, { "epoch": 0.9225806451612903, "grad_norm": 1.0139228105545044, "learning_rate": 3.6855670103092785e-05, "loss": 0.1059, "step": 143 }, { "epoch": 0.9290322580645162, "grad_norm": 0.21279872953891754, "learning_rate": 3.7113402061855674e-05, "loss": 0.0966, "step": 144 }, { "epoch": 0.9354838709677419, "grad_norm": 0.16659015417099, "learning_rate": 3.7371134020618556e-05, "loss": 0.0963, "step": 145 }, { "epoch": 0.9419354838709677, "grad_norm": 0.1789676696062088, "learning_rate": 3.7628865979381445e-05, "loss": 0.099, "step": 146 }, { "epoch": 0.9483870967741935, "grad_norm": 0.1633814126253128, "learning_rate": 3.7886597938144333e-05, "loss": 0.0958, "step": 147 }, { "epoch": 0.9548387096774194, "grad_norm": 0.1775261014699936, "learning_rate": 3.8144329896907216e-05, "loss": 0.1017, "step": 148 }, { "epoch": 0.9612903225806452, "grad_norm": 0.11921869218349457, "learning_rate": 3.8402061855670104e-05, "loss": 0.0945, "step": 149 }, { "epoch": 0.967741935483871, "grad_norm": 0.14354467391967773, "learning_rate": 3.865979381443299e-05, "loss": 0.1077, "step": 150 }, { "epoch": 0.967741935483871, "eval_loss": 0.08996601402759552, "eval_runtime": 20.6312, "eval_samples_per_second": 4.847, "eval_steps_per_second": 0.097, "step": 150 }, { "epoch": 0.9741935483870968, "grad_norm": 0.13735030591487885, "learning_rate": 3.8917525773195875e-05, "loss": 0.099, "step": 151 }, { "epoch": 0.9806451612903225, "grad_norm": 0.10479952394962311, "learning_rate": 3.9175257731958764e-05, "loss": 0.0775, "step": 152 }, { "epoch": 0.9870967741935484, "grad_norm": 0.1122341901063919, "learning_rate": 3.943298969072165e-05, "loss": 0.0723, "step": 153 }, { "epoch": 0.9935483870967742, "grad_norm": 0.1230883076786995, "learning_rate": 3.9690721649484535e-05, "loss": 0.0721, "step": 154 }, { "epoch": 1.0, "grad_norm": 0.14200343191623688, "learning_rate": 3.9948453608247424e-05, "loss": 0.0826, "step": 155 }, { "epoch": 1.0064516129032257, "grad_norm": 0.1111890971660614, "learning_rate": 4.020618556701031e-05, "loss": 0.0801, "step": 156 }, { "epoch": 1.0129032258064516, "grad_norm": 0.12783704698085785, "learning_rate": 4.0463917525773195e-05, "loss": 0.0947, "step": 157 }, { "epoch": 1.0193548387096774, "grad_norm": 0.12492270022630692, "learning_rate": 4.0721649484536084e-05, "loss": 0.0865, "step": 158 }, { "epoch": 1.0258064516129033, "grad_norm": 0.1070851981639862, "learning_rate": 4.097938144329897e-05, "loss": 0.0748, "step": 159 }, { "epoch": 1.032258064516129, "grad_norm": 0.1142066940665245, "learning_rate": 4.1237113402061855e-05, "loss": 0.0773, "step": 160 }, { "epoch": 1.038709677419355, "grad_norm": 0.13485798239707947, "learning_rate": 4.1494845360824744e-05, "loss": 0.0886, "step": 161 }, { "epoch": 1.0451612903225806, "grad_norm": 0.11550623178482056, "learning_rate": 4.175257731958763e-05, "loss": 0.0795, "step": 162 }, { "epoch": 1.0516129032258064, "grad_norm": 0.1292477399110794, "learning_rate": 4.2010309278350515e-05, "loss": 0.0721, "step": 163 }, { "epoch": 1.0580645161290323, "grad_norm": 0.1202891618013382, "learning_rate": 4.2268041237113404e-05, "loss": 0.0713, "step": 164 }, { "epoch": 1.064516129032258, "grad_norm": 0.1149148941040039, "learning_rate": 4.252577319587629e-05, "loss": 0.0776, "step": 165 }, { "epoch": 1.070967741935484, "grad_norm": 0.12351440638303757, "learning_rate": 4.2783505154639175e-05, "loss": 0.0802, "step": 166 }, { "epoch": 1.0774193548387097, "grad_norm": 0.1218574196100235, "learning_rate": 4.3041237113402064e-05, "loss": 0.0791, "step": 167 }, { "epoch": 1.0838709677419356, "grad_norm": 0.1170034110546112, "learning_rate": 4.329896907216495e-05, "loss": 0.0834, "step": 168 }, { "epoch": 1.0903225806451613, "grad_norm": 0.10807105898857117, "learning_rate": 4.3556701030927835e-05, "loss": 0.074, "step": 169 }, { "epoch": 1.096774193548387, "grad_norm": 0.14144091308116913, "learning_rate": 4.3814432989690723e-05, "loss": 0.0867, "step": 170 }, { "epoch": 1.103225806451613, "grad_norm": 0.11608368158340454, "learning_rate": 4.407216494845361e-05, "loss": 0.0702, "step": 171 }, { "epoch": 1.1096774193548387, "grad_norm": 0.11372846364974976, "learning_rate": 4.4329896907216494e-05, "loss": 0.0768, "step": 172 }, { "epoch": 1.1161290322580646, "grad_norm": 0.13371770083904266, "learning_rate": 4.458762886597938e-05, "loss": 0.0754, "step": 173 }, { "epoch": 1.1225806451612903, "grad_norm": 0.10914810746908188, "learning_rate": 4.484536082474227e-05, "loss": 0.0781, "step": 174 }, { "epoch": 1.129032258064516, "grad_norm": 0.1071656197309494, "learning_rate": 4.5103092783505154e-05, "loss": 0.0688, "step": 175 }, { "epoch": 1.129032258064516, "eval_loss": 0.07778502255678177, "eval_runtime": 20.6276, "eval_samples_per_second": 4.848, "eval_steps_per_second": 0.097, "step": 175 }, { "epoch": 1.135483870967742, "grad_norm": 0.11929682642221451, "learning_rate": 4.536082474226804e-05, "loss": 0.0865, "step": 176 }, { "epoch": 1.1419354838709677, "grad_norm": 0.11669743806123734, "learning_rate": 4.561855670103093e-05, "loss": 0.0656, "step": 177 }, { "epoch": 1.1483870967741936, "grad_norm": 0.11047486960887909, "learning_rate": 4.5876288659793814e-05, "loss": 0.0745, "step": 178 }, { "epoch": 1.1548387096774193, "grad_norm": 0.10313151776790619, "learning_rate": 4.61340206185567e-05, "loss": 0.0695, "step": 179 }, { "epoch": 1.1612903225806452, "grad_norm": 0.10674114525318146, "learning_rate": 4.639175257731959e-05, "loss": 0.0802, "step": 180 }, { "epoch": 1.167741935483871, "grad_norm": 0.12059634923934937, "learning_rate": 4.6649484536082474e-05, "loss": 0.0733, "step": 181 }, { "epoch": 1.1741935483870969, "grad_norm": 0.13753844797611237, "learning_rate": 4.690721649484536e-05, "loss": 0.0778, "step": 182 }, { "epoch": 1.1806451612903226, "grad_norm": 0.12316151708364487, "learning_rate": 4.716494845360825e-05, "loss": 0.0732, "step": 183 }, { "epoch": 1.1870967741935483, "grad_norm": 0.1259049028158188, "learning_rate": 4.7422680412371134e-05, "loss": 0.0768, "step": 184 }, { "epoch": 1.1935483870967742, "grad_norm": 0.12398969382047653, "learning_rate": 4.768041237113402e-05, "loss": 0.0784, "step": 185 }, { "epoch": 1.2, "grad_norm": 0.11158449947834015, "learning_rate": 4.793814432989691e-05, "loss": 0.063, "step": 186 }, { "epoch": 1.206451612903226, "grad_norm": 0.1099303588271141, "learning_rate": 4.8195876288659794e-05, "loss": 0.0715, "step": 187 }, { "epoch": 1.2129032258064516, "grad_norm": 0.12735599279403687, "learning_rate": 4.845360824742268e-05, "loss": 0.0727, "step": 188 }, { "epoch": 1.2193548387096773, "grad_norm": 0.12002066522836685, "learning_rate": 4.871134020618557e-05, "loss": 0.0662, "step": 189 }, { "epoch": 1.2258064516129032, "grad_norm": 0.11072317510843277, "learning_rate": 4.8969072164948454e-05, "loss": 0.0632, "step": 190 }, { "epoch": 1.232258064516129, "grad_norm": 0.11942410469055176, "learning_rate": 4.922680412371134e-05, "loss": 0.0787, "step": 191 }, { "epoch": 1.238709677419355, "grad_norm": 0.12091650813817978, "learning_rate": 4.948453608247423e-05, "loss": 0.068, "step": 192 }, { "epoch": 1.2451612903225806, "grad_norm": 0.12011866271495819, "learning_rate": 4.9742268041237114e-05, "loss": 0.0718, "step": 193 }, { "epoch": 1.2516129032258063, "grad_norm": 0.1052054762840271, "learning_rate": 5e-05, "loss": 0.0684, "step": 194 }, { "epoch": 1.2580645161290323, "grad_norm": 0.14064306020736694, "learning_rate": 5.025773195876289e-05, "loss": 0.1004, "step": 195 }, { "epoch": 1.2645161290322582, "grad_norm": 0.11315371841192245, "learning_rate": 5.051546391752577e-05, "loss": 0.0723, "step": 196 }, { "epoch": 1.270967741935484, "grad_norm": 0.10275565087795258, "learning_rate": 5.077319587628866e-05, "loss": 0.0624, "step": 197 }, { "epoch": 1.2774193548387096, "grad_norm": 0.10500184446573257, "learning_rate": 5.1030927835051544e-05, "loss": 0.0654, "step": 198 }, { "epoch": 1.2838709677419355, "grad_norm": 0.11868240684270859, "learning_rate": 5.128865979381443e-05, "loss": 0.0682, "step": 199 }, { "epoch": 1.2903225806451613, "grad_norm": 0.12519873678684235, "learning_rate": 5.1546391752577315e-05, "loss": 0.0682, "step": 200 }, { "epoch": 1.2903225806451613, "eval_loss": 0.07227456569671631, "eval_runtime": 21.4152, "eval_samples_per_second": 4.67, "eval_steps_per_second": 0.093, "step": 200 }, { "epoch": 1.2967741935483872, "grad_norm": 0.11801688373088837, "learning_rate": 5.180412371134021e-05, "loss": 0.0661, "step": 201 }, { "epoch": 1.303225806451613, "grad_norm": 0.12295070290565491, "learning_rate": 5.20618556701031e-05, "loss": 0.0572, "step": 202 }, { "epoch": 1.3096774193548386, "grad_norm": 0.11511899530887604, "learning_rate": 5.231958762886598e-05, "loss": 0.0687, "step": 203 }, { "epoch": 1.3161290322580645, "grad_norm": 0.10693535953760147, "learning_rate": 5.257731958762887e-05, "loss": 0.0606, "step": 204 }, { "epoch": 1.3225806451612903, "grad_norm": 0.11811978369951248, "learning_rate": 5.283505154639175e-05, "loss": 0.0692, "step": 205 }, { "epoch": 1.3290322580645162, "grad_norm": 0.11573677510023117, "learning_rate": 5.309278350515464e-05, "loss": 0.0778, "step": 206 }, { "epoch": 1.335483870967742, "grad_norm": 0.1057257205247879, "learning_rate": 5.335051546391753e-05, "loss": 0.0537, "step": 207 }, { "epoch": 1.3419354838709676, "grad_norm": 0.11804960668087006, "learning_rate": 5.360824742268041e-05, "loss": 0.0637, "step": 208 }, { "epoch": 1.3483870967741935, "grad_norm": 0.10735280811786652, "learning_rate": 5.38659793814433e-05, "loss": 0.0694, "step": 209 }, { "epoch": 1.3548387096774195, "grad_norm": 0.10669250041246414, "learning_rate": 5.4123711340206184e-05, "loss": 0.0581, "step": 210 }, { "epoch": 1.3612903225806452, "grad_norm": 0.12118824571371078, "learning_rate": 5.438144329896907e-05, "loss": 0.0711, "step": 211 }, { "epoch": 1.367741935483871, "grad_norm": 0.10448060184717178, "learning_rate": 5.4639175257731955e-05, "loss": 0.0641, "step": 212 }, { "epoch": 1.3741935483870968, "grad_norm": 0.11256902664899826, "learning_rate": 5.489690721649485e-05, "loss": 0.0617, "step": 213 }, { "epoch": 1.3806451612903226, "grad_norm": 0.12368633598089218, "learning_rate": 5.515463917525774e-05, "loss": 0.0698, "step": 214 }, { "epoch": 1.3870967741935485, "grad_norm": 0.1177687793970108, "learning_rate": 5.541237113402062e-05, "loss": 0.0748, "step": 215 }, { "epoch": 1.3935483870967742, "grad_norm": 0.11173977702856064, "learning_rate": 5.567010309278351e-05, "loss": 0.058, "step": 216 }, { "epoch": 1.4, "grad_norm": 0.10333617776632309, "learning_rate": 5.592783505154639e-05, "loss": 0.0544, "step": 217 }, { "epoch": 1.4064516129032258, "grad_norm": 0.12021996080875397, "learning_rate": 5.618556701030928e-05, "loss": 0.0654, "step": 218 }, { "epoch": 1.4129032258064516, "grad_norm": 0.10966646671295166, "learning_rate": 5.644329896907217e-05, "loss": 0.0654, "step": 219 }, { "epoch": 1.4193548387096775, "grad_norm": 0.1111980453133583, "learning_rate": 5.670103092783505e-05, "loss": 0.0683, "step": 220 }, { "epoch": 1.4258064516129032, "grad_norm": 0.10386307537555695, "learning_rate": 5.695876288659794e-05, "loss": 0.0514, "step": 221 }, { "epoch": 1.432258064516129, "grad_norm": 0.12473642826080322, "learning_rate": 5.721649484536082e-05, "loss": 0.0728, "step": 222 }, { "epoch": 1.4387096774193548, "grad_norm": 0.12365364283323288, "learning_rate": 5.747422680412371e-05, "loss": 0.0719, "step": 223 }, { "epoch": 1.4451612903225808, "grad_norm": 0.124308742582798, "learning_rate": 5.7731958762886594e-05, "loss": 0.0668, "step": 224 }, { "epoch": 1.4516129032258065, "grad_norm": 0.10575917363166809, "learning_rate": 5.798969072164949e-05, "loss": 0.0621, "step": 225 }, { "epoch": 1.4516129032258065, "eval_loss": 0.06675011664628983, "eval_runtime": 20.6214, "eval_samples_per_second": 4.849, "eval_steps_per_second": 0.097, "step": 225 }, { "epoch": 1.4580645161290322, "grad_norm": 0.11430371552705765, "learning_rate": 5.824742268041238e-05, "loss": 0.0544, "step": 226 }, { "epoch": 1.4645161290322581, "grad_norm": 0.11742933839559555, "learning_rate": 5.850515463917526e-05, "loss": 0.0628, "step": 227 }, { "epoch": 1.4709677419354839, "grad_norm": 0.11245157569646835, "learning_rate": 5.876288659793815e-05, "loss": 0.0601, "step": 228 }, { "epoch": 1.4774193548387098, "grad_norm": 0.11818389594554901, "learning_rate": 5.902061855670103e-05, "loss": 0.061, "step": 229 }, { "epoch": 1.4838709677419355, "grad_norm": 0.11650433391332626, "learning_rate": 5.927835051546392e-05, "loss": 0.0618, "step": 230 }, { "epoch": 1.4903225806451612, "grad_norm": 0.12074518948793411, "learning_rate": 5.953608247422681e-05, "loss": 0.0638, "step": 231 }, { "epoch": 1.4967741935483871, "grad_norm": 0.10723990947008133, "learning_rate": 5.979381443298969e-05, "loss": 0.0645, "step": 232 }, { "epoch": 1.5032258064516129, "grad_norm": 0.11135580390691757, "learning_rate": 6.005154639175258e-05, "loss": 0.0623, "step": 233 }, { "epoch": 1.5096774193548388, "grad_norm": 0.12259957939386368, "learning_rate": 6.030927835051546e-05, "loss": 0.0735, "step": 234 }, { "epoch": 1.5161290322580645, "grad_norm": 0.09547890722751617, "learning_rate": 6.056701030927835e-05, "loss": 0.0564, "step": 235 }, { "epoch": 1.5225806451612902, "grad_norm": 0.10515465587377548, "learning_rate": 6.0824742268041234e-05, "loss": 0.062, "step": 236 }, { "epoch": 1.5290322580645161, "grad_norm": 0.1125471293926239, "learning_rate": 6.108247422680413e-05, "loss": 0.0563, "step": 237 }, { "epoch": 1.535483870967742, "grad_norm": 0.11628098785877228, "learning_rate": 6.134020618556701e-05, "loss": 0.0649, "step": 238 }, { "epoch": 1.5419354838709678, "grad_norm": 0.1112600564956665, "learning_rate": 6.159793814432991e-05, "loss": 0.0539, "step": 239 }, { "epoch": 1.5483870967741935, "grad_norm": 0.13200533390045166, "learning_rate": 6.185567010309279e-05, "loss": 0.0713, "step": 240 }, { "epoch": 1.5548387096774192, "grad_norm": 0.11476446688175201, "learning_rate": 6.211340206185567e-05, "loss": 0.0626, "step": 241 }, { "epoch": 1.5612903225806452, "grad_norm": 0.13394683599472046, "learning_rate": 6.237113402061855e-05, "loss": 0.0672, "step": 242 }, { "epoch": 1.567741935483871, "grad_norm": 0.1322658658027649, "learning_rate": 6.262886597938145e-05, "loss": 0.0687, "step": 243 }, { "epoch": 1.5741935483870968, "grad_norm": 0.10362022370100021, "learning_rate": 6.288659793814433e-05, "loss": 0.057, "step": 244 }, { "epoch": 1.5806451612903225, "grad_norm": 0.1106092780828476, "learning_rate": 6.314432989690721e-05, "loss": 0.0541, "step": 245 }, { "epoch": 1.5870967741935482, "grad_norm": 0.12139949202537537, "learning_rate": 6.340206185567011e-05, "loss": 0.0641, "step": 246 }, { "epoch": 1.5935483870967742, "grad_norm": 0.09728733450174332, "learning_rate": 6.365979381443299e-05, "loss": 0.0491, "step": 247 }, { "epoch": 1.6, "grad_norm": 0.10788699984550476, "learning_rate": 6.391752577319587e-05, "loss": 0.0545, "step": 248 }, { "epoch": 1.6064516129032258, "grad_norm": 0.11605466157197952, "learning_rate": 6.417525773195877e-05, "loss": 0.0528, "step": 249 }, { "epoch": 1.6129032258064515, "grad_norm": 0.10914571583271027, "learning_rate": 6.443298969072165e-05, "loss": 0.0668, "step": 250 }, { "epoch": 1.6129032258064515, "eval_loss": 0.06458242982625961, "eval_runtime": 20.6082, "eval_samples_per_second": 4.852, "eval_steps_per_second": 0.097, "step": 250 }, { "epoch": 1.6193548387096774, "grad_norm": 0.10836615413427353, "learning_rate": 6.469072164948455e-05, "loss": 0.0647, "step": 251 }, { "epoch": 1.6258064516129034, "grad_norm": 0.09761244803667068, "learning_rate": 6.494845360824743e-05, "loss": 0.0545, "step": 252 }, { "epoch": 1.632258064516129, "grad_norm": 0.10434439033269882, "learning_rate": 6.520618556701031e-05, "loss": 0.058, "step": 253 }, { "epoch": 1.6387096774193548, "grad_norm": 0.10498231649398804, "learning_rate": 6.546391752577319e-05, "loss": 0.0613, "step": 254 }, { "epoch": 1.6451612903225805, "grad_norm": 0.09180215001106262, "learning_rate": 6.572164948453609e-05, "loss": 0.0499, "step": 255 }, { "epoch": 1.6516129032258065, "grad_norm": 0.10550273954868317, "learning_rate": 6.597938144329897e-05, "loss": 0.0546, "step": 256 }, { "epoch": 1.6580645161290324, "grad_norm": 0.11006518453359604, "learning_rate": 6.623711340206185e-05, "loss": 0.0573, "step": 257 }, { "epoch": 1.664516129032258, "grad_norm": 0.09517877548933029, "learning_rate": 6.649484536082475e-05, "loss": 0.0571, "step": 258 }, { "epoch": 1.6709677419354838, "grad_norm": 0.10255111753940582, "learning_rate": 6.675257731958763e-05, "loss": 0.067, "step": 259 }, { "epoch": 1.6774193548387095, "grad_norm": 0.10252523422241211, "learning_rate": 6.701030927835051e-05, "loss": 0.0547, "step": 260 }, { "epoch": 1.6838709677419355, "grad_norm": 0.10796947032213211, "learning_rate": 6.726804123711341e-05, "loss": 0.0555, "step": 261 }, { "epoch": 1.6903225806451614, "grad_norm": 0.09685982763767242, "learning_rate": 6.752577319587629e-05, "loss": 0.056, "step": 262 }, { "epoch": 1.696774193548387, "grad_norm": 0.10517261922359467, "learning_rate": 6.778350515463919e-05, "loss": 0.0721, "step": 263 }, { "epoch": 1.7032258064516128, "grad_norm": 0.10260171443223953, "learning_rate": 6.804123711340207e-05, "loss": 0.063, "step": 264 }, { "epoch": 1.7096774193548387, "grad_norm": 0.11811640858650208, "learning_rate": 6.829896907216495e-05, "loss": 0.0695, "step": 265 }, { "epoch": 1.7161290322580647, "grad_norm": 0.11343566328287125, "learning_rate": 6.855670103092783e-05, "loss": 0.0644, "step": 266 }, { "epoch": 1.7225806451612904, "grad_norm": 0.10600105673074722, "learning_rate": 6.881443298969073e-05, "loss": 0.058, "step": 267 }, { "epoch": 1.729032258064516, "grad_norm": 0.11247742176055908, "learning_rate": 6.907216494845361e-05, "loss": 0.0585, "step": 268 }, { "epoch": 1.7354838709677418, "grad_norm": 0.10289078205823898, "learning_rate": 6.932989690721649e-05, "loss": 0.06, "step": 269 }, { "epoch": 1.7419354838709677, "grad_norm": 0.11505839973688126, "learning_rate": 6.958762886597939e-05, "loss": 0.0622, "step": 270 }, { "epoch": 1.7483870967741937, "grad_norm": 0.10896378755569458, "learning_rate": 6.984536082474227e-05, "loss": 0.0564, "step": 271 }, { "epoch": 1.7548387096774194, "grad_norm": 0.10462034493684769, "learning_rate": 7.010309278350515e-05, "loss": 0.0478, "step": 272 }, { "epoch": 1.761290322580645, "grad_norm": 0.0960521250963211, "learning_rate": 7.036082474226805e-05, "loss": 0.0501, "step": 273 }, { "epoch": 1.7677419354838708, "grad_norm": 0.113226018846035, "learning_rate": 7.061855670103093e-05, "loss": 0.0487, "step": 274 }, { "epoch": 1.7741935483870968, "grad_norm": 0.12256605923175812, "learning_rate": 7.087628865979383e-05, "loss": 0.0672, "step": 275 }, { "epoch": 1.7741935483870968, "eval_loss": 0.0586564727127552, "eval_runtime": 20.6122, "eval_samples_per_second": 4.851, "eval_steps_per_second": 0.097, "step": 275 }, { "epoch": 1.7806451612903227, "grad_norm": 0.11341353505849838, "learning_rate": 7.113402061855671e-05, "loss": 0.0605, "step": 276 }, { "epoch": 1.7870967741935484, "grad_norm": 0.10689063370227814, "learning_rate": 7.139175257731959e-05, "loss": 0.0557, "step": 277 }, { "epoch": 1.793548387096774, "grad_norm": 0.10520404577255249, "learning_rate": 7.164948453608247e-05, "loss": 0.0587, "step": 278 }, { "epoch": 1.8, "grad_norm": 0.12727777659893036, "learning_rate": 7.190721649484537e-05, "loss": 0.0581, "step": 279 }, { "epoch": 1.8064516129032258, "grad_norm": 0.11627282202243805, "learning_rate": 7.216494845360825e-05, "loss": 0.0561, "step": 280 }, { "epoch": 1.8129032258064517, "grad_norm": 0.09967481344938278, "learning_rate": 7.242268041237113e-05, "loss": 0.0489, "step": 281 }, { "epoch": 1.8193548387096774, "grad_norm": 0.09930626302957535, "learning_rate": 7.268041237113403e-05, "loss": 0.054, "step": 282 }, { "epoch": 1.8258064516129031, "grad_norm": 0.09552939981222153, "learning_rate": 7.293814432989691e-05, "loss": 0.0555, "step": 283 }, { "epoch": 1.832258064516129, "grad_norm": 0.11425229907035828, "learning_rate": 7.319587628865979e-05, "loss": 0.0636, "step": 284 }, { "epoch": 1.838709677419355, "grad_norm": 0.1075424998998642, "learning_rate": 7.345360824742269e-05, "loss": 0.0533, "step": 285 }, { "epoch": 1.8451612903225807, "grad_norm": 0.09101930260658264, "learning_rate": 7.371134020618557e-05, "loss": 0.0393, "step": 286 }, { "epoch": 1.8516129032258064, "grad_norm": 0.11300979554653168, "learning_rate": 7.396907216494847e-05, "loss": 0.0512, "step": 287 }, { "epoch": 1.8580645161290321, "grad_norm": 0.12898088991641998, "learning_rate": 7.422680412371135e-05, "loss": 0.0657, "step": 288 }, { "epoch": 1.864516129032258, "grad_norm": 0.10775969177484512, "learning_rate": 7.448453608247423e-05, "loss": 0.0633, "step": 289 }, { "epoch": 1.870967741935484, "grad_norm": 0.12368246912956238, "learning_rate": 7.474226804123711e-05, "loss": 0.0639, "step": 290 }, { "epoch": 1.8774193548387097, "grad_norm": 0.11313799023628235, "learning_rate": 7.500000000000001e-05, "loss": 0.059, "step": 291 }, { "epoch": 1.8838709677419354, "grad_norm": 0.10701242834329605, "learning_rate": 7.525773195876289e-05, "loss": 0.0473, "step": 292 }, { "epoch": 1.8903225806451613, "grad_norm": 0.09166759997606277, "learning_rate": 7.551546391752577e-05, "loss": 0.0438, "step": 293 }, { "epoch": 1.896774193548387, "grad_norm": 0.1073215901851654, "learning_rate": 7.577319587628867e-05, "loss": 0.0618, "step": 294 }, { "epoch": 1.903225806451613, "grad_norm": 0.11972848325967789, "learning_rate": 7.603092783505155e-05, "loss": 0.06, "step": 295 }, { "epoch": 1.9096774193548387, "grad_norm": 0.10502435266971588, "learning_rate": 7.628865979381443e-05, "loss": 0.0517, "step": 296 }, { "epoch": 1.9161290322580644, "grad_norm": 0.10517352819442749, "learning_rate": 7.654639175257731e-05, "loss": 0.0499, "step": 297 }, { "epoch": 1.9225806451612903, "grad_norm": 0.08997687697410583, "learning_rate": 7.680412371134021e-05, "loss": 0.0426, "step": 298 }, { "epoch": 1.9290322580645163, "grad_norm": 0.10763172805309296, "learning_rate": 7.70618556701031e-05, "loss": 0.0624, "step": 299 }, { "epoch": 1.935483870967742, "grad_norm": 0.09808403253555298, "learning_rate": 7.731958762886599e-05, "loss": 0.0484, "step": 300 }, { "epoch": 1.935483870967742, "eval_loss": 0.05441552773118019, "eval_runtime": 20.6185, "eval_samples_per_second": 4.85, "eval_steps_per_second": 0.097, "step": 300 }, { "epoch": 1.9419354838709677, "grad_norm": 0.09616626799106598, "learning_rate": 7.757731958762887e-05, "loss": 0.0454, "step": 301 }, { "epoch": 1.9483870967741934, "grad_norm": 0.12124036252498627, "learning_rate": 7.783505154639175e-05, "loss": 0.0551, "step": 302 }, { "epoch": 1.9548387096774194, "grad_norm": 0.09645543247461319, "learning_rate": 7.809278350515465e-05, "loss": 0.0484, "step": 303 }, { "epoch": 1.9612903225806453, "grad_norm": 0.1068158969283104, "learning_rate": 7.835051546391753e-05, "loss": 0.0606, "step": 304 }, { "epoch": 1.967741935483871, "grad_norm": 0.09896648675203323, "learning_rate": 7.860824742268041e-05, "loss": 0.0437, "step": 305 }, { "epoch": 1.9741935483870967, "grad_norm": 0.11050526052713394, "learning_rate": 7.88659793814433e-05, "loss": 0.0593, "step": 306 }, { "epoch": 1.9806451612903224, "grad_norm": 0.10204251855611801, "learning_rate": 7.912371134020619e-05, "loss": 0.061, "step": 307 }, { "epoch": 1.9870967741935484, "grad_norm": 0.10084798187017441, "learning_rate": 7.938144329896907e-05, "loss": 0.0473, "step": 308 }, { "epoch": 1.9935483870967743, "grad_norm": 0.1033916026353836, "learning_rate": 7.963917525773195e-05, "loss": 0.0455, "step": 309 }, { "epoch": 2.0, "grad_norm": 0.16185474395751953, "learning_rate": 7.989690721649485e-05, "loss": 0.0757, "step": 310 }, { "epoch": 2.0064516129032257, "grad_norm": 0.09380825608968735, "learning_rate": 8.015463917525774e-05, "loss": 0.0454, "step": 311 }, { "epoch": 2.0129032258064514, "grad_norm": 0.08151062577962875, "learning_rate": 8.041237113402063e-05, "loss": 0.0312, "step": 312 }, { "epoch": 2.0193548387096776, "grad_norm": 0.08947654813528061, "learning_rate": 8.067010309278351e-05, "loss": 0.0428, "step": 313 }, { "epoch": 2.0258064516129033, "grad_norm": 0.10064241290092468, "learning_rate": 8.092783505154639e-05, "loss": 0.0375, "step": 314 }, { "epoch": 2.032258064516129, "grad_norm": 0.10445793718099594, "learning_rate": 8.118556701030929e-05, "loss": 0.0437, "step": 315 }, { "epoch": 2.0387096774193547, "grad_norm": 0.08970332890748978, "learning_rate": 8.144329896907217e-05, "loss": 0.0345, "step": 316 }, { "epoch": 2.0451612903225804, "grad_norm": 0.12089475244283676, "learning_rate": 8.170103092783505e-05, "loss": 0.0477, "step": 317 }, { "epoch": 2.0516129032258066, "grad_norm": 0.12545561790466309, "learning_rate": 8.195876288659795e-05, "loss": 0.0632, "step": 318 }, { "epoch": 2.0580645161290323, "grad_norm": 0.15998150408267975, "learning_rate": 8.221649484536083e-05, "loss": 0.0478, "step": 319 }, { "epoch": 2.064516129032258, "grad_norm": 0.09755123406648636, "learning_rate": 8.247422680412371e-05, "loss": 0.0395, "step": 320 }, { "epoch": 2.0709677419354837, "grad_norm": 0.10251940786838531, "learning_rate": 8.273195876288659e-05, "loss": 0.0474, "step": 321 }, { "epoch": 2.07741935483871, "grad_norm": 0.09739600867033005, "learning_rate": 8.298969072164949e-05, "loss": 0.0443, "step": 322 }, { "epoch": 2.0838709677419356, "grad_norm": 0.11511006206274033, "learning_rate": 8.324742268041238e-05, "loss": 0.0458, "step": 323 }, { "epoch": 2.0903225806451613, "grad_norm": 0.09681696444749832, "learning_rate": 8.350515463917527e-05, "loss": 0.0396, "step": 324 }, { "epoch": 2.096774193548387, "grad_norm": 0.1090191900730133, "learning_rate": 8.376288659793815e-05, "loss": 0.0468, "step": 325 }, { "epoch": 2.096774193548387, "eval_loss": 0.05155442655086517, "eval_runtime": 20.6145, "eval_samples_per_second": 4.851, "eval_steps_per_second": 0.097, "step": 325 }, { "epoch": 2.1032258064516127, "grad_norm": 0.10617301613092422, "learning_rate": 8.402061855670103e-05, "loss": 0.0437, "step": 326 }, { "epoch": 2.109677419354839, "grad_norm": 0.11813998222351074, "learning_rate": 8.427835051546393e-05, "loss": 0.0458, "step": 327 }, { "epoch": 2.1161290322580646, "grad_norm": 0.09822124242782593, "learning_rate": 8.453608247422681e-05, "loss": 0.0414, "step": 328 }, { "epoch": 2.1225806451612903, "grad_norm": 0.0994827002286911, "learning_rate": 8.479381443298969e-05, "loss": 0.0505, "step": 329 }, { "epoch": 2.129032258064516, "grad_norm": 0.10852402448654175, "learning_rate": 8.505154639175259e-05, "loss": 0.0401, "step": 330 }, { "epoch": 2.135483870967742, "grad_norm": 0.0973113551735878, "learning_rate": 8.530927835051547e-05, "loss": 0.0468, "step": 331 }, { "epoch": 2.141935483870968, "grad_norm": 0.10243872553110123, "learning_rate": 8.556701030927835e-05, "loss": 0.0457, "step": 332 }, { "epoch": 2.1483870967741936, "grad_norm": 0.11962252110242844, "learning_rate": 8.582474226804123e-05, "loss": 0.0548, "step": 333 }, { "epoch": 2.1548387096774193, "grad_norm": 0.10123623162508011, "learning_rate": 8.608247422680413e-05, "loss": 0.0369, "step": 334 }, { "epoch": 2.161290322580645, "grad_norm": 0.09829887747764587, "learning_rate": 8.634020618556702e-05, "loss": 0.0372, "step": 335 }, { "epoch": 2.167741935483871, "grad_norm": 0.09360591322183609, "learning_rate": 8.65979381443299e-05, "loss": 0.0479, "step": 336 }, { "epoch": 2.174193548387097, "grad_norm": 0.10151084512472153, "learning_rate": 8.685567010309279e-05, "loss": 0.0367, "step": 337 }, { "epoch": 2.1806451612903226, "grad_norm": 0.12664125859737396, "learning_rate": 8.711340206185567e-05, "loss": 0.043, "step": 338 }, { "epoch": 2.1870967741935483, "grad_norm": 0.11126257479190826, "learning_rate": 8.737113402061856e-05, "loss": 0.0393, "step": 339 }, { "epoch": 2.193548387096774, "grad_norm": 0.09164562076330185, "learning_rate": 8.762886597938145e-05, "loss": 0.0436, "step": 340 }, { "epoch": 2.2, "grad_norm": 0.10741306096315384, "learning_rate": 8.788659793814433e-05, "loss": 0.0463, "step": 341 }, { "epoch": 2.206451612903226, "grad_norm": 0.09051238000392914, "learning_rate": 8.814432989690722e-05, "loss": 0.043, "step": 342 }, { "epoch": 2.2129032258064516, "grad_norm": 0.07722020149230957, "learning_rate": 8.840206185567011e-05, "loss": 0.0354, "step": 343 }, { "epoch": 2.2193548387096773, "grad_norm": 0.08493324369192123, "learning_rate": 8.865979381443299e-05, "loss": 0.0525, "step": 344 }, { "epoch": 2.225806451612903, "grad_norm": 0.09371798485517502, "learning_rate": 8.891752577319587e-05, "loss": 0.0424, "step": 345 }, { "epoch": 2.232258064516129, "grad_norm": 0.10090696066617966, "learning_rate": 8.917525773195877e-05, "loss": 0.0469, "step": 346 }, { "epoch": 2.238709677419355, "grad_norm": 0.08301045000553131, "learning_rate": 8.943298969072166e-05, "loss": 0.0404, "step": 347 }, { "epoch": 2.2451612903225806, "grad_norm": 0.0968640148639679, "learning_rate": 8.969072164948454e-05, "loss": 0.0534, "step": 348 }, { "epoch": 2.2516129032258063, "grad_norm": 0.10685165971517563, "learning_rate": 8.994845360824743e-05, "loss": 0.0391, "step": 349 }, { "epoch": 2.258064516129032, "grad_norm": 0.09006287902593613, "learning_rate": 9.020618556701031e-05, "loss": 0.0438, "step": 350 }, { "epoch": 2.258064516129032, "eval_loss": 0.05031716451048851, "eval_runtime": 20.609, "eval_samples_per_second": 4.852, "eval_steps_per_second": 0.097, "step": 350 }, { "epoch": 2.264516129032258, "grad_norm": 0.10811453312635422, "learning_rate": 9.04639175257732e-05, "loss": 0.0488, "step": 351 }, { "epoch": 2.270967741935484, "grad_norm": 0.10823424160480499, "learning_rate": 9.072164948453609e-05, "loss": 0.0456, "step": 352 }, { "epoch": 2.2774193548387096, "grad_norm": 0.08600781857967377, "learning_rate": 9.097938144329897e-05, "loss": 0.0398, "step": 353 }, { "epoch": 2.2838709677419353, "grad_norm": 0.10172222554683685, "learning_rate": 9.123711340206186e-05, "loss": 0.0427, "step": 354 }, { "epoch": 2.2903225806451615, "grad_norm": 0.09986208379268646, "learning_rate": 9.149484536082475e-05, "loss": 0.0387, "step": 355 }, { "epoch": 2.296774193548387, "grad_norm": 0.10122333467006683, "learning_rate": 9.175257731958763e-05, "loss": 0.0442, "step": 356 }, { "epoch": 2.303225806451613, "grad_norm": 0.09510695189237595, "learning_rate": 9.201030927835051e-05, "loss": 0.0447, "step": 357 }, { "epoch": 2.3096774193548386, "grad_norm": 0.09774709492921829, "learning_rate": 9.22680412371134e-05, "loss": 0.0344, "step": 358 }, { "epoch": 2.3161290322580643, "grad_norm": 0.0921277180314064, "learning_rate": 9.25257731958763e-05, "loss": 0.0411, "step": 359 }, { "epoch": 2.3225806451612905, "grad_norm": 0.10613450407981873, "learning_rate": 9.278350515463918e-05, "loss": 0.0443, "step": 360 }, { "epoch": 2.329032258064516, "grad_norm": 0.10701876133680344, "learning_rate": 9.304123711340207e-05, "loss": 0.039, "step": 361 }, { "epoch": 2.335483870967742, "grad_norm": 0.09017759561538696, "learning_rate": 9.329896907216495e-05, "loss": 0.0369, "step": 362 }, { "epoch": 2.3419354838709676, "grad_norm": 0.10165076702833176, "learning_rate": 9.355670103092784e-05, "loss": 0.041, "step": 363 }, { "epoch": 2.3483870967741938, "grad_norm": 0.09364940971136093, "learning_rate": 9.381443298969073e-05, "loss": 0.0463, "step": 364 }, { "epoch": 2.3548387096774195, "grad_norm": 0.08863069862127304, "learning_rate": 9.407216494845361e-05, "loss": 0.0363, "step": 365 }, { "epoch": 2.361290322580645, "grad_norm": 0.10201136022806168, "learning_rate": 9.43298969072165e-05, "loss": 0.0384, "step": 366 }, { "epoch": 2.367741935483871, "grad_norm": 0.09771512448787689, "learning_rate": 9.458762886597939e-05, "loss": 0.0386, "step": 367 }, { "epoch": 2.3741935483870966, "grad_norm": 0.09068713337182999, "learning_rate": 9.484536082474227e-05, "loss": 0.0389, "step": 368 }, { "epoch": 2.3806451612903228, "grad_norm": 0.10729826241731644, "learning_rate": 9.510309278350515e-05, "loss": 0.0419, "step": 369 }, { "epoch": 2.3870967741935485, "grad_norm": 0.10047543793916702, "learning_rate": 9.536082474226805e-05, "loss": 0.046, "step": 370 }, { "epoch": 2.393548387096774, "grad_norm": 0.08138115704059601, "learning_rate": 9.561855670103094e-05, "loss": 0.03, "step": 371 }, { "epoch": 2.4, "grad_norm": 0.08465878665447235, "learning_rate": 9.587628865979382e-05, "loss": 0.0383, "step": 372 }, { "epoch": 2.4064516129032256, "grad_norm": 0.10496239364147186, "learning_rate": 9.61340206185567e-05, "loss": 0.0423, "step": 373 }, { "epoch": 2.412903225806452, "grad_norm": 0.09935601055622101, "learning_rate": 9.639175257731959e-05, "loss": 0.043, "step": 374 }, { "epoch": 2.4193548387096775, "grad_norm": 0.0851937010884285, "learning_rate": 9.664948453608248e-05, "loss": 0.0364, "step": 375 }, { "epoch": 2.4193548387096775, "eval_loss": 0.04934929683804512, "eval_runtime": 20.6137, "eval_samples_per_second": 4.851, "eval_steps_per_second": 0.097, "step": 375 }, { "epoch": 2.425806451612903, "grad_norm": 0.07887910306453705, "learning_rate": 9.690721649484537e-05, "loss": 0.0307, "step": 376 }, { "epoch": 2.432258064516129, "grad_norm": 0.10647827386856079, "learning_rate": 9.716494845360825e-05, "loss": 0.0455, "step": 377 }, { "epoch": 2.4387096774193546, "grad_norm": 0.09436062723398209, "learning_rate": 9.742268041237114e-05, "loss": 0.0476, "step": 378 }, { "epoch": 2.445161290322581, "grad_norm": 0.10967639833688736, "learning_rate": 9.768041237113403e-05, "loss": 0.0368, "step": 379 }, { "epoch": 2.4516129032258065, "grad_norm": 0.0899633839726448, "learning_rate": 9.793814432989691e-05, "loss": 0.0291, "step": 380 }, { "epoch": 2.458064516129032, "grad_norm": 0.10881899297237396, "learning_rate": 9.819587628865979e-05, "loss": 0.0452, "step": 381 }, { "epoch": 2.464516129032258, "grad_norm": 0.09735005348920822, "learning_rate": 9.845360824742269e-05, "loss": 0.0329, "step": 382 }, { "epoch": 2.4709677419354836, "grad_norm": 0.10843765735626221, "learning_rate": 9.871134020618558e-05, "loss": 0.0379, "step": 383 }, { "epoch": 2.47741935483871, "grad_norm": 0.09915640950202942, "learning_rate": 9.896907216494846e-05, "loss": 0.0416, "step": 384 }, { "epoch": 2.4838709677419355, "grad_norm": 0.08816944807767868, "learning_rate": 9.922680412371134e-05, "loss": 0.035, "step": 385 }, { "epoch": 2.490322580645161, "grad_norm": 0.08104445785284042, "learning_rate": 9.948453608247423e-05, "loss": 0.0356, "step": 386 }, { "epoch": 2.496774193548387, "grad_norm": 0.10272940993309021, "learning_rate": 9.974226804123712e-05, "loss": 0.0404, "step": 387 }, { "epoch": 2.5032258064516126, "grad_norm": 0.08973879367113113, "learning_rate": 0.0001, "loss": 0.0305, "step": 388 }, { "epoch": 2.509677419354839, "grad_norm": 0.1103263571858406, "learning_rate": 9.999997970748714e-05, "loss": 0.0459, "step": 389 }, { "epoch": 2.5161290322580645, "grad_norm": 0.09237506985664368, "learning_rate": 9.999991882996502e-05, "loss": 0.0363, "step": 390 }, { "epoch": 2.52258064516129, "grad_norm": 0.08599911630153656, "learning_rate": 9.999981736748307e-05, "loss": 0.0377, "step": 391 }, { "epoch": 2.5290322580645164, "grad_norm": 0.08956660330295563, "learning_rate": 9.999967532012362e-05, "loss": 0.0355, "step": 392 }, { "epoch": 2.535483870967742, "grad_norm": 0.08925971388816833, "learning_rate": 9.9999492688002e-05, "loss": 0.038, "step": 393 }, { "epoch": 2.541935483870968, "grad_norm": 0.08915024995803833, "learning_rate": 9.999926947126643e-05, "loss": 0.0376, "step": 394 }, { "epoch": 2.5483870967741935, "grad_norm": 0.09460075944662094, "learning_rate": 9.999900567009811e-05, "loss": 0.0367, "step": 395 }, { "epoch": 2.554838709677419, "grad_norm": 0.0868026539683342, "learning_rate": 9.999870128471116e-05, "loss": 0.0356, "step": 396 }, { "epoch": 2.5612903225806454, "grad_norm": 0.09221860021352768, "learning_rate": 9.999835631535266e-05, "loss": 0.0441, "step": 397 }, { "epoch": 2.567741935483871, "grad_norm": 0.1023736447095871, "learning_rate": 9.99979707623026e-05, "loss": 0.0451, "step": 398 }, { "epoch": 2.574193548387097, "grad_norm": 0.08648494631052017, "learning_rate": 9.999754462587395e-05, "loss": 0.0375, "step": 399 }, { "epoch": 2.5806451612903225, "grad_norm": 0.08478187024593353, "learning_rate": 9.99970779064126e-05, "loss": 0.0365, "step": 400 }, { "epoch": 2.5806451612903225, "eval_loss": 0.046045903116464615, "eval_runtime": 20.63, "eval_samples_per_second": 4.847, "eval_steps_per_second": 0.097, "step": 400 }, { "epoch": 2.587096774193548, "grad_norm": 0.09959060698747635, "learning_rate": 9.99965706042974e-05, "loss": 0.0467, "step": 401 }, { "epoch": 2.5935483870967744, "grad_norm": 0.09755048900842667, "learning_rate": 9.99960227199401e-05, "loss": 0.0533, "step": 402 }, { "epoch": 2.6, "grad_norm": 0.09167367219924927, "learning_rate": 9.999543425378546e-05, "loss": 0.0325, "step": 403 }, { "epoch": 2.606451612903226, "grad_norm": 0.09100550413131714, "learning_rate": 9.999480520631108e-05, "loss": 0.0325, "step": 404 }, { "epoch": 2.6129032258064515, "grad_norm": 0.09340760111808777, "learning_rate": 9.999413557802761e-05, "loss": 0.0342, "step": 405 }, { "epoch": 2.6193548387096772, "grad_norm": 0.08973444998264313, "learning_rate": 9.999342536947856e-05, "loss": 0.0276, "step": 406 }, { "epoch": 2.6258064516129034, "grad_norm": 0.09528621286153793, "learning_rate": 9.999267458124042e-05, "loss": 0.0375, "step": 407 }, { "epoch": 2.632258064516129, "grad_norm": 0.0875079482793808, "learning_rate": 9.999188321392261e-05, "loss": 0.0307, "step": 408 }, { "epoch": 2.638709677419355, "grad_norm": 0.08494490385055542, "learning_rate": 9.999105126816746e-05, "loss": 0.0331, "step": 409 }, { "epoch": 2.6451612903225805, "grad_norm": 0.08419065922498703, "learning_rate": 9.999017874465026e-05, "loss": 0.0299, "step": 410 }, { "epoch": 2.6516129032258062, "grad_norm": 0.1131298691034317, "learning_rate": 9.998926564407929e-05, "loss": 0.045, "step": 411 }, { "epoch": 2.6580645161290324, "grad_norm": 0.09079862385988235, "learning_rate": 9.998831196719565e-05, "loss": 0.0386, "step": 412 }, { "epoch": 2.664516129032258, "grad_norm": 0.082625612616539, "learning_rate": 9.998731771477347e-05, "loss": 0.0334, "step": 413 }, { "epoch": 2.670967741935484, "grad_norm": 0.09609875082969666, "learning_rate": 9.998628288761977e-05, "loss": 0.038, "step": 414 }, { "epoch": 2.6774193548387095, "grad_norm": 0.09008405357599258, "learning_rate": 9.998520748657454e-05, "loss": 0.0412, "step": 415 }, { "epoch": 2.6838709677419352, "grad_norm": 0.09257365018129349, "learning_rate": 9.998409151251069e-05, "loss": 0.0339, "step": 416 }, { "epoch": 2.6903225806451614, "grad_norm": 0.0978696420788765, "learning_rate": 9.998293496633402e-05, "loss": 0.0322, "step": 417 }, { "epoch": 2.696774193548387, "grad_norm": 0.10335182398557663, "learning_rate": 9.998173784898333e-05, "loss": 0.0361, "step": 418 }, { "epoch": 2.703225806451613, "grad_norm": 0.10199162364006042, "learning_rate": 9.99805001614303e-05, "loss": 0.0455, "step": 419 }, { "epoch": 2.709677419354839, "grad_norm": 0.09214220941066742, "learning_rate": 9.997922190467958e-05, "loss": 0.0348, "step": 420 }, { "epoch": 2.7161290322580647, "grad_norm": 0.08050487190485, "learning_rate": 9.997790307976873e-05, "loss": 0.0309, "step": 421 }, { "epoch": 2.7225806451612904, "grad_norm": 0.08799546957015991, "learning_rate": 9.997654368776823e-05, "loss": 0.0424, "step": 422 }, { "epoch": 2.729032258064516, "grad_norm": 0.09670068323612213, "learning_rate": 9.997514372978151e-05, "loss": 0.049, "step": 423 }, { "epoch": 2.735483870967742, "grad_norm": 0.08845686167478561, "learning_rate": 9.997370320694492e-05, "loss": 0.0328, "step": 424 }, { "epoch": 2.741935483870968, "grad_norm": 0.09527582675218582, "learning_rate": 9.997222212042771e-05, "loss": 0.0469, "step": 425 }, { "epoch": 2.741935483870968, "eval_loss": 0.04324241355061531, "eval_runtime": 20.6183, "eval_samples_per_second": 4.85, "eval_steps_per_second": 0.097, "step": 425 }, { "epoch": 2.7483870967741937, "grad_norm": 0.0964616984128952, "learning_rate": 9.997070047143209e-05, "loss": 0.0349, "step": 426 }, { "epoch": 2.7548387096774194, "grad_norm": 0.08241977542638779, "learning_rate": 9.996913826119319e-05, "loss": 0.0342, "step": 427 }, { "epoch": 2.761290322580645, "grad_norm": 0.08565547317266464, "learning_rate": 9.996753549097906e-05, "loss": 0.0272, "step": 428 }, { "epoch": 2.767741935483871, "grad_norm": 0.0907561182975769, "learning_rate": 9.996589216209065e-05, "loss": 0.0283, "step": 429 }, { "epoch": 2.774193548387097, "grad_norm": 0.09964805096387863, "learning_rate": 9.996420827586187e-05, "loss": 0.037, "step": 430 }, { "epoch": 2.7806451612903227, "grad_norm": 0.08956048637628555, "learning_rate": 9.996248383365951e-05, "loss": 0.0355, "step": 431 }, { "epoch": 2.7870967741935484, "grad_norm": 0.09447566419839859, "learning_rate": 9.996071883688334e-05, "loss": 0.0304, "step": 432 }, { "epoch": 2.793548387096774, "grad_norm": 0.07058817893266678, "learning_rate": 9.995891328696599e-05, "loss": 0.0315, "step": 433 }, { "epoch": 2.8, "grad_norm": 0.08722875267267227, "learning_rate": 9.995706718537299e-05, "loss": 0.0364, "step": 434 }, { "epoch": 2.806451612903226, "grad_norm": 0.08992493152618408, "learning_rate": 9.995518053360287e-05, "loss": 0.0448, "step": 435 }, { "epoch": 2.8129032258064517, "grad_norm": 0.0942826047539711, "learning_rate": 9.995325333318698e-05, "loss": 0.0385, "step": 436 }, { "epoch": 2.8193548387096774, "grad_norm": 0.08824922144412994, "learning_rate": 9.995128558568969e-05, "loss": 0.0322, "step": 437 }, { "epoch": 2.825806451612903, "grad_norm": 0.08112802356481552, "learning_rate": 9.994927729270818e-05, "loss": 0.0364, "step": 438 }, { "epoch": 2.832258064516129, "grad_norm": 0.09755432605743408, "learning_rate": 9.994722845587258e-05, "loss": 0.0386, "step": 439 }, { "epoch": 2.838709677419355, "grad_norm": 0.08299075067043304, "learning_rate": 9.994513907684597e-05, "loss": 0.0339, "step": 440 }, { "epoch": 2.8451612903225807, "grad_norm": 0.08259541541337967, "learning_rate": 9.994300915732426e-05, "loss": 0.0363, "step": 441 }, { "epoch": 2.8516129032258064, "grad_norm": 0.09055620431900024, "learning_rate": 9.994083869903631e-05, "loss": 0.0351, "step": 442 }, { "epoch": 2.858064516129032, "grad_norm": 0.08169237524271011, "learning_rate": 9.99386277037439e-05, "loss": 0.0347, "step": 443 }, { "epoch": 2.864516129032258, "grad_norm": 0.07610059529542923, "learning_rate": 9.993637617324169e-05, "loss": 0.0299, "step": 444 }, { "epoch": 2.870967741935484, "grad_norm": 0.076594777405262, "learning_rate": 9.993408410935724e-05, "loss": 0.0249, "step": 445 }, { "epoch": 2.8774193548387097, "grad_norm": 0.09088917821645737, "learning_rate": 9.993175151395103e-05, "loss": 0.0324, "step": 446 }, { "epoch": 2.8838709677419354, "grad_norm": 0.09005070477724075, "learning_rate": 9.992937838891643e-05, "loss": 0.0386, "step": 447 }, { "epoch": 2.8903225806451616, "grad_norm": 0.10479908436536789, "learning_rate": 9.99269647361797e-05, "loss": 0.0466, "step": 448 }, { "epoch": 2.896774193548387, "grad_norm": 0.08293581753969193, "learning_rate": 9.99245105577e-05, "loss": 0.0293, "step": 449 }, { "epoch": 2.903225806451613, "grad_norm": 0.07779109477996826, "learning_rate": 9.992201585546939e-05, "loss": 0.027, "step": 450 }, { "epoch": 2.903225806451613, "eval_loss": 0.037865012884140015, "eval_runtime": 20.634, "eval_samples_per_second": 4.846, "eval_steps_per_second": 0.097, "step": 450 }, { "epoch": 2.9096774193548387, "grad_norm": 0.09607955068349838, "learning_rate": 9.991948063151282e-05, "loss": 0.0331, "step": 451 }, { "epoch": 2.9161290322580644, "grad_norm": 0.07993067800998688, "learning_rate": 9.991690488788815e-05, "loss": 0.0293, "step": 452 }, { "epoch": 2.9225806451612906, "grad_norm": 0.08637774735689163, "learning_rate": 9.99142886266861e-05, "loss": 0.0273, "step": 453 }, { "epoch": 2.9290322580645163, "grad_norm": 0.08489525318145752, "learning_rate": 9.991163185003029e-05, "loss": 0.0295, "step": 454 }, { "epoch": 2.935483870967742, "grad_norm": 0.08981441706418991, "learning_rate": 9.990893456007721e-05, "loss": 0.0308, "step": 455 }, { "epoch": 2.9419354838709677, "grad_norm": 0.07571941614151001, "learning_rate": 9.99061967590163e-05, "loss": 0.0254, "step": 456 }, { "epoch": 2.9483870967741934, "grad_norm": 0.07855682075023651, "learning_rate": 9.990341844906977e-05, "loss": 0.0253, "step": 457 }, { "epoch": 2.9548387096774196, "grad_norm": 0.08273386210203171, "learning_rate": 9.99005996324928e-05, "loss": 0.0278, "step": 458 }, { "epoch": 2.9612903225806453, "grad_norm": 0.08668919652700424, "learning_rate": 9.989774031157346e-05, "loss": 0.0355, "step": 459 }, { "epoch": 2.967741935483871, "grad_norm": 0.07949639856815338, "learning_rate": 9.989484048863263e-05, "loss": 0.0311, "step": 460 }, { "epoch": 2.9741935483870967, "grad_norm": 0.07660870999097824, "learning_rate": 9.98919001660241e-05, "loss": 0.0281, "step": 461 }, { "epoch": 2.9806451612903224, "grad_norm": 0.0825650691986084, "learning_rate": 9.988891934613453e-05, "loss": 0.0334, "step": 462 }, { "epoch": 2.9870967741935486, "grad_norm": 0.09388405084609985, "learning_rate": 9.988589803138348e-05, "loss": 0.0389, "step": 463 }, { "epoch": 2.9935483870967743, "grad_norm": 0.09062936902046204, "learning_rate": 9.988283622422332e-05, "loss": 0.0383, "step": 464 }, { "epoch": 3.0, "grad_norm": 0.09347395598888397, "learning_rate": 9.987973392713932e-05, "loss": 0.0312, "step": 465 }, { "epoch": 3.0064516129032257, "grad_norm": 0.07513423264026642, "learning_rate": 9.987659114264962e-05, "loss": 0.0262, "step": 466 }, { "epoch": 3.0129032258064514, "grad_norm": 0.07603180408477783, "learning_rate": 9.987340787330525e-05, "loss": 0.0276, "step": 467 }, { "epoch": 3.0193548387096776, "grad_norm": 0.06836437433958054, "learning_rate": 9.987018412169004e-05, "loss": 0.021, "step": 468 }, { "epoch": 3.0258064516129033, "grad_norm": 0.07807572931051254, "learning_rate": 9.986691989042072e-05, "loss": 0.0207, "step": 469 }, { "epoch": 3.032258064516129, "grad_norm": 0.0833655521273613, "learning_rate": 9.986361518214686e-05, "loss": 0.0196, "step": 470 }, { "epoch": 3.0387096774193547, "grad_norm": 0.08109772205352783, "learning_rate": 9.98602699995509e-05, "loss": 0.0236, "step": 471 }, { "epoch": 3.0451612903225804, "grad_norm": 0.08900588750839233, "learning_rate": 9.985688434534812e-05, "loss": 0.0255, "step": 472 }, { "epoch": 3.0516129032258066, "grad_norm": 0.08419376611709595, "learning_rate": 9.985345822228669e-05, "loss": 0.02, "step": 473 }, { "epoch": 3.0580645161290323, "grad_norm": 0.0820038914680481, "learning_rate": 9.984999163314754e-05, "loss": 0.022, "step": 474 }, { "epoch": 3.064516129032258, "grad_norm": 0.07802214473485947, "learning_rate": 9.984648458074456e-05, "loss": 0.026, "step": 475 }, { "epoch": 3.064516129032258, "eval_loss": 0.03555970638990402, "eval_runtime": 20.6207, "eval_samples_per_second": 4.849, "eval_steps_per_second": 0.097, "step": 475 }, { "epoch": 3.0709677419354837, "grad_norm": 0.08685624599456787, "learning_rate": 9.984293706792438e-05, "loss": 0.0325, "step": 476 }, { "epoch": 3.07741935483871, "grad_norm": 0.06740245223045349, "learning_rate": 9.983934909756655e-05, "loss": 0.0228, "step": 477 }, { "epoch": 3.0838709677419356, "grad_norm": 0.0687062218785286, "learning_rate": 9.98357206725834e-05, "loss": 0.0205, "step": 478 }, { "epoch": 3.0903225806451613, "grad_norm": 0.07554581761360168, "learning_rate": 9.983205179592015e-05, "loss": 0.0231, "step": 479 }, { "epoch": 3.096774193548387, "grad_norm": 0.08018746972084045, "learning_rate": 9.982834247055481e-05, "loss": 0.0222, "step": 480 }, { "epoch": 3.1032258064516127, "grad_norm": 0.08768939971923828, "learning_rate": 9.982459269949826e-05, "loss": 0.025, "step": 481 }, { "epoch": 3.109677419354839, "grad_norm": 0.07732031494379044, "learning_rate": 9.982080248579417e-05, "loss": 0.021, "step": 482 }, { "epoch": 3.1161290322580646, "grad_norm": 0.08360020071268082, "learning_rate": 9.981697183251906e-05, "loss": 0.0256, "step": 483 }, { "epoch": 3.1225806451612903, "grad_norm": 0.0702652707695961, "learning_rate": 9.98131007427823e-05, "loss": 0.0204, "step": 484 }, { "epoch": 3.129032258064516, "grad_norm": 0.07399879395961761, "learning_rate": 9.980918921972602e-05, "loss": 0.019, "step": 485 }, { "epoch": 3.135483870967742, "grad_norm": 0.08124198019504547, "learning_rate": 9.980523726652523e-05, "loss": 0.0229, "step": 486 }, { "epoch": 3.141935483870968, "grad_norm": 0.07998952269554138, "learning_rate": 9.980124488638774e-05, "loss": 0.0246, "step": 487 }, { "epoch": 3.1483870967741936, "grad_norm": 0.09725457429885864, "learning_rate": 9.979721208255412e-05, "loss": 0.0253, "step": 488 }, { "epoch": 3.1548387096774193, "grad_norm": 0.08378903567790985, "learning_rate": 9.979313885829785e-05, "loss": 0.0265, "step": 489 }, { "epoch": 3.161290322580645, "grad_norm": 0.07236087322235107, "learning_rate": 9.978902521692515e-05, "loss": 0.0212, "step": 490 }, { "epoch": 3.167741935483871, "grad_norm": 0.06853360682725906, "learning_rate": 9.978487116177506e-05, "loss": 0.0193, "step": 491 }, { "epoch": 3.174193548387097, "grad_norm": 0.09208260476589203, "learning_rate": 9.978067669621943e-05, "loss": 0.0224, "step": 492 }, { "epoch": 3.1806451612903226, "grad_norm": 0.07864844053983688, "learning_rate": 9.977644182366292e-05, "loss": 0.0235, "step": 493 }, { "epoch": 3.1870967741935483, "grad_norm": 0.07481813430786133, "learning_rate": 9.977216654754296e-05, "loss": 0.0201, "step": 494 }, { "epoch": 3.193548387096774, "grad_norm": 0.08573871105909348, "learning_rate": 9.976785087132981e-05, "loss": 0.0219, "step": 495 }, { "epoch": 3.2, "grad_norm": 0.09021998196840286, "learning_rate": 9.97634947985265e-05, "loss": 0.0228, "step": 496 }, { "epoch": 3.206451612903226, "grad_norm": 0.07071111351251602, "learning_rate": 9.975909833266885e-05, "loss": 0.0207, "step": 497 }, { "epoch": 3.2129032258064516, "grad_norm": 0.08035361021757126, "learning_rate": 9.97546614773255e-05, "loss": 0.0199, "step": 498 }, { "epoch": 3.2193548387096773, "grad_norm": 0.07872354239225388, "learning_rate": 9.975018423609781e-05, "loss": 0.0238, "step": 499 }, { "epoch": 3.225806451612903, "grad_norm": 0.08297586441040039, "learning_rate": 9.974566661261999e-05, "loss": 0.0223, "step": 500 }, { "epoch": 3.225806451612903, "eval_loss": 0.0357317216694355, "eval_runtime": 20.6238, "eval_samples_per_second": 4.849, "eval_steps_per_second": 0.097, "step": 500 }, { "epoch": 3.232258064516129, "grad_norm": 0.09199152141809464, "learning_rate": 9.974110861055899e-05, "loss": 0.0235, "step": 501 }, { "epoch": 3.238709677419355, "grad_norm": 0.0826394334435463, "learning_rate": 9.973651023361452e-05, "loss": 0.0201, "step": 502 }, { "epoch": 3.2451612903225806, "grad_norm": 0.11516128480434418, "learning_rate": 9.973187148551911e-05, "loss": 0.0253, "step": 503 }, { "epoch": 3.2516129032258063, "grad_norm": 0.09063907712697983, "learning_rate": 9.972719237003802e-05, "loss": 0.0223, "step": 504 }, { "epoch": 3.258064516129032, "grad_norm": 0.08523458242416382, "learning_rate": 9.972247289096932e-05, "loss": 0.0211, "step": 505 }, { "epoch": 3.264516129032258, "grad_norm": 0.10173799842596054, "learning_rate": 9.971771305214377e-05, "loss": 0.0327, "step": 506 }, { "epoch": 3.270967741935484, "grad_norm": 0.0834076777100563, "learning_rate": 9.971291285742497e-05, "loss": 0.0257, "step": 507 }, { "epoch": 3.2774193548387096, "grad_norm": 0.07420746237039566, "learning_rate": 9.970807231070921e-05, "loss": 0.0209, "step": 508 }, { "epoch": 3.2838709677419353, "grad_norm": 0.08463042974472046, "learning_rate": 9.97031914159256e-05, "loss": 0.0212, "step": 509 }, { "epoch": 3.2903225806451615, "grad_norm": 0.08214029669761658, "learning_rate": 9.969827017703593e-05, "loss": 0.0264, "step": 510 }, { "epoch": 3.296774193548387, "grad_norm": 0.08387239277362823, "learning_rate": 9.969330859803478e-05, "loss": 0.0199, "step": 511 }, { "epoch": 3.303225806451613, "grad_norm": 0.08673185110092163, "learning_rate": 9.968830668294947e-05, "loss": 0.0214, "step": 512 }, { "epoch": 3.3096774193548386, "grad_norm": 0.08083758503198624, "learning_rate": 9.968326443584007e-05, "loss": 0.0213, "step": 513 }, { "epoch": 3.3161290322580643, "grad_norm": 0.08173336833715439, "learning_rate": 9.967818186079936e-05, "loss": 0.0217, "step": 514 }, { "epoch": 3.3225806451612905, "grad_norm": 0.07553200423717499, "learning_rate": 9.967305896195288e-05, "loss": 0.019, "step": 515 }, { "epoch": 3.329032258064516, "grad_norm": 0.08780141919851303, "learning_rate": 9.966789574345888e-05, "loss": 0.025, "step": 516 }, { "epoch": 3.335483870967742, "grad_norm": 0.08397892117500305, "learning_rate": 9.966269220950835e-05, "loss": 0.0259, "step": 517 }, { "epoch": 3.3419354838709676, "grad_norm": 0.10555091500282288, "learning_rate": 9.9657448364325e-05, "loss": 0.0269, "step": 518 }, { "epoch": 3.3483870967741938, "grad_norm": 0.07310711592435837, "learning_rate": 9.965216421216526e-05, "loss": 0.0214, "step": 519 }, { "epoch": 3.3548387096774195, "grad_norm": 0.07622754573822021, "learning_rate": 9.964683975731828e-05, "loss": 0.0202, "step": 520 }, { "epoch": 3.361290322580645, "grad_norm": 0.0859362930059433, "learning_rate": 9.964147500410592e-05, "loss": 0.0235, "step": 521 }, { "epoch": 3.367741935483871, "grad_norm": 0.08965717256069183, "learning_rate": 9.963606995688277e-05, "loss": 0.0189, "step": 522 }, { "epoch": 3.3741935483870966, "grad_norm": 0.0818329006433487, "learning_rate": 9.963062462003609e-05, "loss": 0.0199, "step": 523 }, { "epoch": 3.3806451612903228, "grad_norm": 0.07498865574598312, "learning_rate": 9.962513899798587e-05, "loss": 0.0213, "step": 524 }, { "epoch": 3.3870967741935485, "grad_norm": 0.08389492332935333, "learning_rate": 9.961961309518479e-05, "loss": 0.0228, "step": 525 }, { "epoch": 3.3870967741935485, "eval_loss": 0.035191066563129425, "eval_runtime": 20.6312, "eval_samples_per_second": 4.847, "eval_steps_per_second": 0.097, "step": 525 }, { "epoch": 3.393548387096774, "grad_norm": 0.0822795033454895, "learning_rate": 9.961404691611825e-05, "loss": 0.0215, "step": 526 }, { "epoch": 3.4, "grad_norm": 0.08398600667715073, "learning_rate": 9.960844046530428e-05, "loss": 0.0244, "step": 527 }, { "epoch": 3.4064516129032256, "grad_norm": 0.08129465579986572, "learning_rate": 9.960279374729367e-05, "loss": 0.0239, "step": 528 }, { "epoch": 3.412903225806452, "grad_norm": 0.07414031773805618, "learning_rate": 9.959710676666985e-05, "loss": 0.0244, "step": 529 }, { "epoch": 3.4193548387096775, "grad_norm": 0.08356715738773346, "learning_rate": 9.959137952804894e-05, "loss": 0.0188, "step": 530 }, { "epoch": 3.425806451612903, "grad_norm": 0.0784551128745079, "learning_rate": 9.958561203607975e-05, "loss": 0.023, "step": 531 }, { "epoch": 3.432258064516129, "grad_norm": 0.08018410205841064, "learning_rate": 9.957980429544377e-05, "loss": 0.0228, "step": 532 }, { "epoch": 3.4387096774193546, "grad_norm": 0.07876982539892197, "learning_rate": 9.957395631085513e-05, "loss": 0.0232, "step": 533 }, { "epoch": 3.445161290322581, "grad_norm": 0.08994334936141968, "learning_rate": 9.956806808706064e-05, "loss": 0.0248, "step": 534 }, { "epoch": 3.4516129032258065, "grad_norm": 0.097404845058918, "learning_rate": 9.956213962883977e-05, "loss": 0.025, "step": 535 }, { "epoch": 3.458064516129032, "grad_norm": 0.09073007851839066, "learning_rate": 9.955617094100469e-05, "loss": 0.0264, "step": 536 }, { "epoch": 3.464516129032258, "grad_norm": 0.08371078222990036, "learning_rate": 9.955016202840014e-05, "loss": 0.0239, "step": 537 }, { "epoch": 3.4709677419354836, "grad_norm": 0.08191969245672226, "learning_rate": 9.954411289590358e-05, "loss": 0.023, "step": 538 }, { "epoch": 3.47741935483871, "grad_norm": 0.07514638453722, "learning_rate": 9.95380235484251e-05, "loss": 0.0209, "step": 539 }, { "epoch": 3.4838709677419355, "grad_norm": 0.07831292599439621, "learning_rate": 9.95318939909074e-05, "loss": 0.0223, "step": 540 }, { "epoch": 3.490322580645161, "grad_norm": 0.0708436518907547, "learning_rate": 9.952572422832586e-05, "loss": 0.0245, "step": 541 }, { "epoch": 3.496774193548387, "grad_norm": 0.08529362827539444, "learning_rate": 9.951951426568851e-05, "loss": 0.0192, "step": 542 }, { "epoch": 3.5032258064516126, "grad_norm": 0.06932336837053299, "learning_rate": 9.951326410803593e-05, "loss": 0.0211, "step": 543 }, { "epoch": 3.509677419354839, "grad_norm": 0.0799434706568718, "learning_rate": 9.95069737604414e-05, "loss": 0.0216, "step": 544 }, { "epoch": 3.5161290322580645, "grad_norm": 0.08105307072401047, "learning_rate": 9.95006432280108e-05, "loss": 0.0217, "step": 545 }, { "epoch": 3.52258064516129, "grad_norm": 0.0905054360628128, "learning_rate": 9.949427251588263e-05, "loss": 0.0229, "step": 546 }, { "epoch": 3.5290322580645164, "grad_norm": 0.0961836725473404, "learning_rate": 9.948786162922799e-05, "loss": 0.0198, "step": 547 }, { "epoch": 3.535483870967742, "grad_norm": 0.08218283206224442, "learning_rate": 9.948141057325057e-05, "loss": 0.0205, "step": 548 }, { "epoch": 3.541935483870968, "grad_norm": 0.07585642486810684, "learning_rate": 9.947491935318676e-05, "loss": 0.0209, "step": 549 }, { "epoch": 3.5483870967741935, "grad_norm": 0.08271976560354233, "learning_rate": 9.946838797430546e-05, "loss": 0.0199, "step": 550 }, { "epoch": 3.5483870967741935, "eval_loss": 0.03360068425536156, "eval_runtime": 20.7001, "eval_samples_per_second": 4.831, "eval_steps_per_second": 0.097, "step": 550 }, { "epoch": 3.554838709677419, "grad_norm": 0.07116498053073883, "learning_rate": 9.946181644190817e-05, "loss": 0.0212, "step": 551 }, { "epoch": 3.5612903225806454, "grad_norm": 0.09358586370944977, "learning_rate": 9.945520476132903e-05, "loss": 0.0226, "step": 552 }, { "epoch": 3.567741935483871, "grad_norm": 0.07434282451868057, "learning_rate": 9.944855293793475e-05, "loss": 0.0173, "step": 553 }, { "epoch": 3.574193548387097, "grad_norm": 0.08777125179767609, "learning_rate": 9.94418609771246e-05, "loss": 0.0288, "step": 554 }, { "epoch": 3.5806451612903225, "grad_norm": 0.0815097913146019, "learning_rate": 9.943512888433047e-05, "loss": 0.0215, "step": 555 }, { "epoch": 3.587096774193548, "grad_norm": 0.0928550437092781, "learning_rate": 9.942835666501676e-05, "loss": 0.0279, "step": 556 }, { "epoch": 3.5935483870967744, "grad_norm": 0.07668986171483994, "learning_rate": 9.942154432468054e-05, "loss": 0.0209, "step": 557 }, { "epoch": 3.6, "grad_norm": 0.06721030920743942, "learning_rate": 9.941469186885135e-05, "loss": 0.0168, "step": 558 }, { "epoch": 3.606451612903226, "grad_norm": 0.07109752297401428, "learning_rate": 9.940779930309134e-05, "loss": 0.0181, "step": 559 }, { "epoch": 3.6129032258064515, "grad_norm": 0.08308808505535126, "learning_rate": 9.940086663299522e-05, "loss": 0.018, "step": 560 }, { "epoch": 3.6193548387096772, "grad_norm": 0.09721226990222931, "learning_rate": 9.939389386419024e-05, "loss": 0.0267, "step": 561 }, { "epoch": 3.6258064516129034, "grad_norm": 0.0851580798625946, "learning_rate": 9.93868810023362e-05, "loss": 0.021, "step": 562 }, { "epoch": 3.632258064516129, "grad_norm": 0.0647260770201683, "learning_rate": 9.937982805312543e-05, "loss": 0.0135, "step": 563 }, { "epoch": 3.638709677419355, "grad_norm": 0.06940098106861115, "learning_rate": 9.937273502228281e-05, "loss": 0.0199, "step": 564 }, { "epoch": 3.6451612903225805, "grad_norm": 0.07867590337991714, "learning_rate": 9.936560191556579e-05, "loss": 0.0191, "step": 565 }, { "epoch": 3.6516129032258062, "grad_norm": 0.07656869292259216, "learning_rate": 9.935842873876429e-05, "loss": 0.0247, "step": 566 }, { "epoch": 3.6580645161290324, "grad_norm": 0.09072941541671753, "learning_rate": 9.935121549770078e-05, "loss": 0.0239, "step": 567 }, { "epoch": 3.664516129032258, "grad_norm": 0.06793881952762604, "learning_rate": 9.934396219823026e-05, "loss": 0.0179, "step": 568 }, { "epoch": 3.670967741935484, "grad_norm": 0.07314693927764893, "learning_rate": 9.933666884624025e-05, "loss": 0.0215, "step": 569 }, { "epoch": 3.6774193548387095, "grad_norm": 0.06927098333835602, "learning_rate": 9.932933544765073e-05, "loss": 0.0218, "step": 570 }, { "epoch": 3.6838709677419352, "grad_norm": 0.07695493847131729, "learning_rate": 9.932196200841426e-05, "loss": 0.0237, "step": 571 }, { "epoch": 3.6903225806451614, "grad_norm": 0.07632816582918167, "learning_rate": 9.931454853451586e-05, "loss": 0.0164, "step": 572 }, { "epoch": 3.696774193548387, "grad_norm": 0.07444462180137634, "learning_rate": 9.930709503197302e-05, "loss": 0.0192, "step": 573 }, { "epoch": 3.703225806451613, "grad_norm": 0.08099304139614105, "learning_rate": 9.929960150683577e-05, "loss": 0.0204, "step": 574 }, { "epoch": 3.709677419354839, "grad_norm": 0.07179006189107895, "learning_rate": 9.929206796518662e-05, "loss": 0.0227, "step": 575 }, { "epoch": 3.709677419354839, "eval_loss": 0.03084612637758255, "eval_runtime": 20.6116, "eval_samples_per_second": 4.852, "eval_steps_per_second": 0.097, "step": 575 }, { "epoch": 3.7161290322580647, "grad_norm": 0.07330773770809174, "learning_rate": 9.928449441314055e-05, "loss": 0.0195, "step": 576 }, { "epoch": 3.7225806451612904, "grad_norm": 0.08855849504470825, "learning_rate": 9.927688085684499e-05, "loss": 0.028, "step": 577 }, { "epoch": 3.729032258064516, "grad_norm": 0.06930747628211975, "learning_rate": 9.92692273024799e-05, "loss": 0.0196, "step": 578 }, { "epoch": 3.735483870967742, "grad_norm": 0.08203380554914474, "learning_rate": 9.926153375625765e-05, "loss": 0.0192, "step": 579 }, { "epoch": 3.741935483870968, "grad_norm": 0.07359786331653595, "learning_rate": 9.925380022442311e-05, "loss": 0.0251, "step": 580 }, { "epoch": 3.7483870967741937, "grad_norm": 0.08028274029493332, "learning_rate": 9.924602671325358e-05, "loss": 0.024, "step": 581 }, { "epoch": 3.7548387096774194, "grad_norm": 0.07106206566095352, "learning_rate": 9.923821322905883e-05, "loss": 0.0189, "step": 582 }, { "epoch": 3.761290322580645, "grad_norm": 0.088934987783432, "learning_rate": 9.923035977818107e-05, "loss": 0.0202, "step": 583 }, { "epoch": 3.767741935483871, "grad_norm": 0.09591435641050339, "learning_rate": 9.922246636699494e-05, "loss": 0.0219, "step": 584 }, { "epoch": 3.774193548387097, "grad_norm": 0.08910396695137024, "learning_rate": 9.921453300190755e-05, "loss": 0.0206, "step": 585 }, { "epoch": 3.7806451612903227, "grad_norm": 0.08721159398555756, "learning_rate": 9.920655968935838e-05, "loss": 0.0229, "step": 586 }, { "epoch": 3.7870967741935484, "grad_norm": 0.0917913094162941, "learning_rate": 9.919854643581941e-05, "loss": 0.0236, "step": 587 }, { "epoch": 3.793548387096774, "grad_norm": 0.0874079242348671, "learning_rate": 9.919049324779497e-05, "loss": 0.018, "step": 588 }, { "epoch": 3.8, "grad_norm": 0.07590654492378235, "learning_rate": 9.918240013182185e-05, "loss": 0.0175, "step": 589 }, { "epoch": 3.806451612903226, "grad_norm": 0.07079242169857025, "learning_rate": 9.917426709446925e-05, "loss": 0.019, "step": 590 }, { "epoch": 3.8129032258064517, "grad_norm": 0.09207978844642639, "learning_rate": 9.916609414233873e-05, "loss": 0.0203, "step": 591 }, { "epoch": 3.8193548387096774, "grad_norm": 0.0762355849146843, "learning_rate": 9.915788128206432e-05, "loss": 0.0161, "step": 592 }, { "epoch": 3.825806451612903, "grad_norm": 0.09506982564926147, "learning_rate": 9.914962852031237e-05, "loss": 0.0213, "step": 593 }, { "epoch": 3.832258064516129, "grad_norm": 0.09398619085550308, "learning_rate": 9.914133586378166e-05, "loss": 0.0243, "step": 594 }, { "epoch": 3.838709677419355, "grad_norm": 0.09069452434778214, "learning_rate": 9.913300331920334e-05, "loss": 0.0286, "step": 595 }, { "epoch": 3.8451612903225807, "grad_norm": 0.0756390318274498, "learning_rate": 9.912463089334097e-05, "loss": 0.0209, "step": 596 }, { "epoch": 3.8516129032258064, "grad_norm": 0.08355611562728882, "learning_rate": 9.91162185929904e-05, "loss": 0.0224, "step": 597 }, { "epoch": 3.858064516129032, "grad_norm": 0.0704779401421547, "learning_rate": 9.910776642497994e-05, "loss": 0.0169, "step": 598 }, { "epoch": 3.864516129032258, "grad_norm": 0.08580672740936279, "learning_rate": 9.90992743961702e-05, "loss": 0.0255, "step": 599 }, { "epoch": 3.870967741935484, "grad_norm": 0.06796646118164062, "learning_rate": 9.909074251345417e-05, "loss": 0.0207, "step": 600 }, { "epoch": 3.870967741935484, "eval_loss": 0.02920311503112316, "eval_runtime": 20.6688, "eval_samples_per_second": 4.838, "eval_steps_per_second": 0.097, "step": 600 }, { "epoch": 3.8774193548387097, "grad_norm": 0.07997417449951172, "learning_rate": 9.908217078375718e-05, "loss": 0.0232, "step": 601 }, { "epoch": 3.8838709677419354, "grad_norm": 0.06856135278940201, "learning_rate": 9.907355921403691e-05, "loss": 0.0183, "step": 602 }, { "epoch": 3.8903225806451616, "grad_norm": 0.06765090674161911, "learning_rate": 9.906490781128339e-05, "loss": 0.0205, "step": 603 }, { "epoch": 3.896774193548387, "grad_norm": 0.08263225853443146, "learning_rate": 9.905621658251896e-05, "loss": 0.0204, "step": 604 }, { "epoch": 3.903225806451613, "grad_norm": 0.06636311858892441, "learning_rate": 9.904748553479827e-05, "loss": 0.0182, "step": 605 }, { "epoch": 3.9096774193548387, "grad_norm": 0.07616009563207626, "learning_rate": 9.903871467520835e-05, "loss": 0.0175, "step": 606 }, { "epoch": 3.9161290322580644, "grad_norm": 0.08746340870857239, "learning_rate": 9.902990401086849e-05, "loss": 0.0242, "step": 607 }, { "epoch": 3.9225806451612906, "grad_norm": 0.07934951782226562, "learning_rate": 9.90210535489303e-05, "loss": 0.0199, "step": 608 }, { "epoch": 3.9290322580645163, "grad_norm": 0.07786279171705246, "learning_rate": 9.901216329657774e-05, "loss": 0.0199, "step": 609 }, { "epoch": 3.935483870967742, "grad_norm": 0.08316848427057266, "learning_rate": 9.900323326102702e-05, "loss": 0.0208, "step": 610 }, { "epoch": 3.9419354838709677, "grad_norm": 0.07723607867956161, "learning_rate": 9.899426344952663e-05, "loss": 0.0201, "step": 611 }, { "epoch": 3.9483870967741934, "grad_norm": 0.07270605862140656, "learning_rate": 9.89852538693574e-05, "loss": 0.0187, "step": 612 }, { "epoch": 3.9548387096774196, "grad_norm": 0.0690370425581932, "learning_rate": 9.897620452783241e-05, "loss": 0.0162, "step": 613 }, { "epoch": 3.9612903225806453, "grad_norm": 0.07613126933574677, "learning_rate": 9.896711543229698e-05, "loss": 0.0179, "step": 614 }, { "epoch": 3.967741935483871, "grad_norm": 0.0803714394569397, "learning_rate": 9.895798659012879e-05, "loss": 0.0209, "step": 615 }, { "epoch": 3.9741935483870967, "grad_norm": 0.0816982090473175, "learning_rate": 9.894881800873768e-05, "loss": 0.0215, "step": 616 }, { "epoch": 3.9806451612903224, "grad_norm": 0.10295988619327545, "learning_rate": 9.89396096955658e-05, "loss": 0.0242, "step": 617 }, { "epoch": 3.9870967741935486, "grad_norm": 0.0716472640633583, "learning_rate": 9.893036165808755e-05, "loss": 0.0222, "step": 618 }, { "epoch": 3.9935483870967743, "grad_norm": 0.07746679335832596, "learning_rate": 9.892107390380958e-05, "loss": 0.0206, "step": 619 }, { "epoch": 4.0, "grad_norm": 0.07583904266357422, "learning_rate": 9.891174644027073e-05, "loss": 0.0172, "step": 620 }, { "epoch": 4.006451612903226, "grad_norm": 0.07562244683504105, "learning_rate": 9.890237927504214e-05, "loss": 0.0155, "step": 621 }, { "epoch": 4.012903225806451, "grad_norm": 0.05778181925415993, "learning_rate": 9.889297241572714e-05, "loss": 0.0141, "step": 622 }, { "epoch": 4.019354838709678, "grad_norm": 0.06385020166635513, "learning_rate": 9.888352586996126e-05, "loss": 0.0114, "step": 623 }, { "epoch": 4.025806451612903, "grad_norm": 0.071347676217556, "learning_rate": 9.887403964541228e-05, "loss": 0.0133, "step": 624 }, { "epoch": 4.032258064516129, "grad_norm": 0.07100800424814224, "learning_rate": 9.886451374978018e-05, "loss": 0.0125, "step": 625 }, { "epoch": 4.032258064516129, "eval_loss": 0.030386490747332573, "eval_runtime": 20.6264, "eval_samples_per_second": 4.848, "eval_steps_per_second": 0.097, "step": 625 }, { "epoch": 4.038709677419355, "grad_norm": 0.07996921241283417, "learning_rate": 9.885494819079713e-05, "loss": 0.0154, "step": 626 }, { "epoch": 4.04516129032258, "grad_norm": 0.07129693031311035, "learning_rate": 9.884534297622749e-05, "loss": 0.0132, "step": 627 }, { "epoch": 4.051612903225807, "grad_norm": 0.06520695984363556, "learning_rate": 9.883569811386782e-05, "loss": 0.012, "step": 628 }, { "epoch": 4.058064516129032, "grad_norm": 0.06700082868337631, "learning_rate": 9.882601361154687e-05, "loss": 0.0112, "step": 629 }, { "epoch": 4.064516129032258, "grad_norm": 0.07065930962562561, "learning_rate": 9.881628947712556e-05, "loss": 0.0125, "step": 630 }, { "epoch": 4.070967741935484, "grad_norm": 0.07708729058504105, "learning_rate": 9.880652571849695e-05, "loss": 0.0152, "step": 631 }, { "epoch": 4.077419354838709, "grad_norm": 0.07915516197681427, "learning_rate": 9.879672234358631e-05, "loss": 0.0132, "step": 632 }, { "epoch": 4.083870967741936, "grad_norm": 0.08086412400007248, "learning_rate": 9.878687936035103e-05, "loss": 0.0128, "step": 633 }, { "epoch": 4.090322580645161, "grad_norm": 0.07018841058015823, "learning_rate": 9.877699677678067e-05, "loss": 0.0126, "step": 634 }, { "epoch": 4.096774193548387, "grad_norm": 0.08451451361179352, "learning_rate": 9.876707460089692e-05, "loss": 0.0149, "step": 635 }, { "epoch": 4.103225806451613, "grad_norm": 0.08706196397542953, "learning_rate": 9.875711284075364e-05, "loss": 0.0155, "step": 636 }, { "epoch": 4.109677419354838, "grad_norm": 0.0794413611292839, "learning_rate": 9.874711150443678e-05, "loss": 0.0144, "step": 637 }, { "epoch": 4.116129032258065, "grad_norm": 0.0688992291688919, "learning_rate": 9.873707060006442e-05, "loss": 0.0127, "step": 638 }, { "epoch": 4.122580645161291, "grad_norm": 0.0626332089304924, "learning_rate": 9.872699013578679e-05, "loss": 0.012, "step": 639 }, { "epoch": 4.129032258064516, "grad_norm": 0.07945838570594788, "learning_rate": 9.871687011978618e-05, "loss": 0.0168, "step": 640 }, { "epoch": 4.135483870967742, "grad_norm": 0.07975637167692184, "learning_rate": 9.870671056027705e-05, "loss": 0.0129, "step": 641 }, { "epoch": 4.141935483870967, "grad_norm": 0.0599512979388237, "learning_rate": 9.86965114655059e-05, "loss": 0.0121, "step": 642 }, { "epoch": 4.148387096774194, "grad_norm": 0.08625789731740952, "learning_rate": 9.868627284375132e-05, "loss": 0.0158, "step": 643 }, { "epoch": 4.15483870967742, "grad_norm": 0.06459569931030273, "learning_rate": 9.867599470332402e-05, "loss": 0.0136, "step": 644 }, { "epoch": 4.161290322580645, "grad_norm": 0.06582548469305038, "learning_rate": 9.866567705256678e-05, "loss": 0.0131, "step": 645 }, { "epoch": 4.167741935483871, "grad_norm": 0.07506062835454941, "learning_rate": 9.865531989985443e-05, "loss": 0.0118, "step": 646 }, { "epoch": 4.174193548387096, "grad_norm": 0.06777074187994003, "learning_rate": 9.86449232535939e-05, "loss": 0.0124, "step": 647 }, { "epoch": 4.180645161290323, "grad_norm": 0.06837598979473114, "learning_rate": 9.863448712222411e-05, "loss": 0.0122, "step": 648 }, { "epoch": 4.187096774193549, "grad_norm": 0.07759612798690796, "learning_rate": 9.862401151421612e-05, "loss": 0.0152, "step": 649 }, { "epoch": 4.193548387096774, "grad_norm": 0.08142206817865372, "learning_rate": 9.861349643807295e-05, "loss": 0.0146, "step": 650 }, { "epoch": 4.193548387096774, "eval_loss": 0.027878066524863243, "eval_runtime": 20.628, "eval_samples_per_second": 4.848, "eval_steps_per_second": 0.097, "step": 650 }, { "epoch": 4.2, "grad_norm": 0.10191861540079117, "learning_rate": 9.860294190232972e-05, "loss": 0.0168, "step": 651 }, { "epoch": 4.2064516129032254, "grad_norm": 0.07277819514274597, "learning_rate": 9.859234791555355e-05, "loss": 0.0106, "step": 652 }, { "epoch": 4.212903225806452, "grad_norm": 0.06661704927682877, "learning_rate": 9.858171448634357e-05, "loss": 0.0116, "step": 653 }, { "epoch": 4.219354838709678, "grad_norm": 0.07337664812803268, "learning_rate": 9.857104162333092e-05, "loss": 0.0158, "step": 654 }, { "epoch": 4.225806451612903, "grad_norm": 0.060086801648139954, "learning_rate": 9.856032933517883e-05, "loss": 0.0113, "step": 655 }, { "epoch": 4.232258064516129, "grad_norm": 0.0661039799451828, "learning_rate": 9.854957763058241e-05, "loss": 0.016, "step": 656 }, { "epoch": 4.2387096774193544, "grad_norm": 0.06683111935853958, "learning_rate": 9.853878651826886e-05, "loss": 0.015, "step": 657 }, { "epoch": 4.245161290322581, "grad_norm": 0.0577155277132988, "learning_rate": 9.852795600699731e-05, "loss": 0.0102, "step": 658 }, { "epoch": 4.251612903225807, "grad_norm": 0.06298208236694336, "learning_rate": 9.85170861055589e-05, "loss": 0.0116, "step": 659 }, { "epoch": 4.258064516129032, "grad_norm": 0.06872560828924179, "learning_rate": 9.850617682277674e-05, "loss": 0.0144, "step": 660 }, { "epoch": 4.264516129032258, "grad_norm": 0.07436688244342804, "learning_rate": 9.84952281675059e-05, "loss": 0.0138, "step": 661 }, { "epoch": 4.270967741935484, "grad_norm": 0.07236282527446747, "learning_rate": 9.848424014863337e-05, "loss": 0.0137, "step": 662 }, { "epoch": 4.27741935483871, "grad_norm": 0.06464178115129471, "learning_rate": 9.84732127750782e-05, "loss": 0.0129, "step": 663 }, { "epoch": 4.283870967741936, "grad_norm": 0.07325076311826706, "learning_rate": 9.846214605579127e-05, "loss": 0.0144, "step": 664 }, { "epoch": 4.290322580645161, "grad_norm": 0.06621360778808594, "learning_rate": 9.845103999975547e-05, "loss": 0.0136, "step": 665 }, { "epoch": 4.296774193548387, "grad_norm": 0.06366943567991257, "learning_rate": 9.843989461598555e-05, "loss": 0.0124, "step": 666 }, { "epoch": 4.3032258064516125, "grad_norm": 0.05970180034637451, "learning_rate": 9.842870991352825e-05, "loss": 0.0113, "step": 667 }, { "epoch": 4.309677419354839, "grad_norm": 0.07171665132045746, "learning_rate": 9.84174859014622e-05, "loss": 0.0142, "step": 668 }, { "epoch": 4.316129032258065, "grad_norm": 0.07346490025520325, "learning_rate": 9.840622258889794e-05, "loss": 0.0152, "step": 669 }, { "epoch": 4.32258064516129, "grad_norm": 0.06221551075577736, "learning_rate": 9.839491998497787e-05, "loss": 0.0116, "step": 670 }, { "epoch": 4.329032258064516, "grad_norm": 0.07008364796638489, "learning_rate": 9.838357809887637e-05, "loss": 0.0135, "step": 671 }, { "epoch": 4.335483870967742, "grad_norm": 0.07009419053792953, "learning_rate": 9.837219693979963e-05, "loss": 0.0135, "step": 672 }, { "epoch": 4.341935483870968, "grad_norm": 0.06749093532562256, "learning_rate": 9.836077651698574e-05, "loss": 0.0098, "step": 673 }, { "epoch": 4.348387096774194, "grad_norm": 0.057939253747463226, "learning_rate": 9.834931683970468e-05, "loss": 0.0105, "step": 674 }, { "epoch": 4.354838709677419, "grad_norm": 0.08264298737049103, "learning_rate": 9.833781791725825e-05, "loss": 0.0126, "step": 675 }, { "epoch": 4.354838709677419, "eval_loss": 0.028300685808062553, "eval_runtime": 20.6434, "eval_samples_per_second": 4.844, "eval_steps_per_second": 0.097, "step": 675 }, { "epoch": 4.361290322580645, "grad_norm": 0.0651557594537735, "learning_rate": 9.832627975898015e-05, "loss": 0.0143, "step": 676 }, { "epoch": 4.367741935483871, "grad_norm": 0.07030880451202393, "learning_rate": 9.831470237423592e-05, "loss": 0.0143, "step": 677 }, { "epoch": 4.374193548387097, "grad_norm": 0.06443504244089127, "learning_rate": 9.83030857724229e-05, "loss": 0.0128, "step": 678 }, { "epoch": 4.380645161290323, "grad_norm": 0.08351744711399078, "learning_rate": 9.82914299629703e-05, "loss": 0.0133, "step": 679 }, { "epoch": 4.387096774193548, "grad_norm": 0.09203293919563293, "learning_rate": 9.827973495533915e-05, "loss": 0.0196, "step": 680 }, { "epoch": 4.393548387096774, "grad_norm": 0.06061858683824539, "learning_rate": 9.826800075902232e-05, "loss": 0.0117, "step": 681 }, { "epoch": 4.4, "grad_norm": 0.0718703344464302, "learning_rate": 9.825622738354441e-05, "loss": 0.016, "step": 682 }, { "epoch": 4.406451612903226, "grad_norm": 0.07132145017385483, "learning_rate": 9.824441483846191e-05, "loss": 0.0149, "step": 683 }, { "epoch": 4.412903225806452, "grad_norm": 0.05910409241914749, "learning_rate": 9.823256313336308e-05, "loss": 0.0125, "step": 684 }, { "epoch": 4.419354838709677, "grad_norm": 0.06985270231962204, "learning_rate": 9.822067227786793e-05, "loss": 0.0155, "step": 685 }, { "epoch": 4.425806451612903, "grad_norm": 0.0629580095410347, "learning_rate": 9.820874228162827e-05, "loss": 0.0119, "step": 686 }, { "epoch": 4.432258064516129, "grad_norm": 0.07799185812473297, "learning_rate": 9.81967731543277e-05, "loss": 0.0143, "step": 687 }, { "epoch": 4.438709677419355, "grad_norm": 0.08001742511987686, "learning_rate": 9.818476490568157e-05, "loss": 0.0119, "step": 688 }, { "epoch": 4.445161290322581, "grad_norm": 0.058317631483078, "learning_rate": 9.817271754543698e-05, "loss": 0.0114, "step": 689 }, { "epoch": 4.451612903225806, "grad_norm": 0.08109927922487259, "learning_rate": 9.816063108337274e-05, "loss": 0.013, "step": 690 }, { "epoch": 4.458064516129032, "grad_norm": 0.07449546456336975, "learning_rate": 9.81485055292995e-05, "loss": 0.0129, "step": 691 }, { "epoch": 4.464516129032258, "grad_norm": 0.07314932346343994, "learning_rate": 9.813634089305955e-05, "loss": 0.0128, "step": 692 }, { "epoch": 4.470967741935484, "grad_norm": 0.07322119176387787, "learning_rate": 9.81241371845269e-05, "loss": 0.0141, "step": 693 }, { "epoch": 4.47741935483871, "grad_norm": 0.07292065024375916, "learning_rate": 9.811189441360737e-05, "loss": 0.0131, "step": 694 }, { "epoch": 4.483870967741936, "grad_norm": 0.07420699298381805, "learning_rate": 9.809961259023838e-05, "loss": 0.0132, "step": 695 }, { "epoch": 4.490322580645161, "grad_norm": 0.07527197152376175, "learning_rate": 9.808729172438908e-05, "loss": 0.0114, "step": 696 }, { "epoch": 4.496774193548387, "grad_norm": 0.07561551779508591, "learning_rate": 9.807493182606037e-05, "loss": 0.014, "step": 697 }, { "epoch": 4.503225806451613, "grad_norm": 0.0866156667470932, "learning_rate": 9.806253290528476e-05, "loss": 0.0153, "step": 698 }, { "epoch": 4.509677419354839, "grad_norm": 0.07429295778274536, "learning_rate": 9.805009497212645e-05, "loss": 0.0154, "step": 699 }, { "epoch": 4.516129032258064, "grad_norm": 0.06641758978366852, "learning_rate": 9.803761803668135e-05, "loss": 0.0141, "step": 700 }, { "epoch": 4.516129032258064, "eval_loss": 0.026979686692357063, "eval_runtime": 20.638, "eval_samples_per_second": 4.845, "eval_steps_per_second": 0.097, "step": 700 }, { "epoch": 4.52258064516129, "grad_norm": 0.0701785758137703, "learning_rate": 9.802510210907694e-05, "loss": 0.0155, "step": 701 }, { "epoch": 4.529032258064516, "grad_norm": 0.06845016777515411, "learning_rate": 9.801254719947246e-05, "loss": 0.0134, "step": 702 }, { "epoch": 4.535483870967742, "grad_norm": 0.05970216915011406, "learning_rate": 9.799995331805871e-05, "loss": 0.0129, "step": 703 }, { "epoch": 4.541935483870968, "grad_norm": 0.0771360918879509, "learning_rate": 9.798732047505815e-05, "loss": 0.0234, "step": 704 }, { "epoch": 4.548387096774194, "grad_norm": 0.06995438039302826, "learning_rate": 9.797464868072488e-05, "loss": 0.0149, "step": 705 }, { "epoch": 4.554838709677419, "grad_norm": 0.055232495069503784, "learning_rate": 9.796193794534458e-05, "loss": 0.0117, "step": 706 }, { "epoch": 4.561290322580645, "grad_norm": 0.06754062324762344, "learning_rate": 9.794918827923458e-05, "loss": 0.0113, "step": 707 }, { "epoch": 4.567741935483871, "grad_norm": 0.06377311795949936, "learning_rate": 9.793639969274378e-05, "loss": 0.012, "step": 708 }, { "epoch": 4.574193548387097, "grad_norm": 0.07187410444021225, "learning_rate": 9.792357219625268e-05, "loss": 0.0134, "step": 709 }, { "epoch": 4.580645161290323, "grad_norm": 0.06659362465143204, "learning_rate": 9.791070580017337e-05, "loss": 0.0125, "step": 710 }, { "epoch": 4.587096774193548, "grad_norm": 0.07713113725185394, "learning_rate": 9.789780051494952e-05, "loss": 0.0127, "step": 711 }, { "epoch": 4.593548387096774, "grad_norm": 0.07113997638225555, "learning_rate": 9.788485635105635e-05, "loss": 0.0143, "step": 712 }, { "epoch": 4.6, "grad_norm": 0.06326919794082642, "learning_rate": 9.787187331900062e-05, "loss": 0.0101, "step": 713 }, { "epoch": 4.606451612903226, "grad_norm": 0.054118432104587555, "learning_rate": 9.785885142932072e-05, "loss": 0.0109, "step": 714 }, { "epoch": 4.612903225806452, "grad_norm": 0.0691436380147934, "learning_rate": 9.784579069258646e-05, "loss": 0.0132, "step": 715 }, { "epoch": 4.619354838709677, "grad_norm": 0.06021921709179878, "learning_rate": 9.78326911193993e-05, "loss": 0.0114, "step": 716 }, { "epoch": 4.625806451612903, "grad_norm": 0.06648483127355576, "learning_rate": 9.781955272039213e-05, "loss": 0.0126, "step": 717 }, { "epoch": 4.632258064516129, "grad_norm": 0.0594315268099308, "learning_rate": 9.780637550622941e-05, "loss": 0.0123, "step": 718 }, { "epoch": 4.638709677419355, "grad_norm": 0.05935937538743019, "learning_rate": 9.77931594876071e-05, "loss": 0.0099, "step": 719 }, { "epoch": 4.645161290322581, "grad_norm": 0.07264725118875504, "learning_rate": 9.777990467525265e-05, "loss": 0.0133, "step": 720 }, { "epoch": 4.651612903225806, "grad_norm": 0.0625627338886261, "learning_rate": 9.7766611079925e-05, "loss": 0.0137, "step": 721 }, { "epoch": 4.658064516129032, "grad_norm": 0.07260245829820633, "learning_rate": 9.775327871241453e-05, "loss": 0.0119, "step": 722 }, { "epoch": 4.664516129032258, "grad_norm": 0.06769385188817978, "learning_rate": 9.773990758354317e-05, "loss": 0.011, "step": 723 }, { "epoch": 4.670967741935484, "grad_norm": 0.07323268800973892, "learning_rate": 9.772649770416428e-05, "loss": 0.0104, "step": 724 }, { "epoch": 4.67741935483871, "grad_norm": 0.07127388566732407, "learning_rate": 9.771304908516263e-05, "loss": 0.0133, "step": 725 }, { "epoch": 4.67741935483871, "eval_loss": 0.02536088228225708, "eval_runtime": 20.6309, "eval_samples_per_second": 4.847, "eval_steps_per_second": 0.097, "step": 725 }, { "epoch": 4.683870967741935, "grad_norm": 0.07082170993089676, "learning_rate": 9.769956173745449e-05, "loss": 0.0098, "step": 726 }, { "epoch": 4.690322580645161, "grad_norm": 0.06554052233695984, "learning_rate": 9.768603567198755e-05, "loss": 0.0122, "step": 727 }, { "epoch": 4.6967741935483875, "grad_norm": 0.06847799569368362, "learning_rate": 9.767247089974091e-05, "loss": 0.0111, "step": 728 }, { "epoch": 4.703225806451613, "grad_norm": 0.06368619948625565, "learning_rate": 9.765886743172511e-05, "loss": 0.0112, "step": 729 }, { "epoch": 4.709677419354839, "grad_norm": 0.07584304362535477, "learning_rate": 9.76452252789821e-05, "loss": 0.0132, "step": 730 }, { "epoch": 4.716129032258064, "grad_norm": 0.0686490386724472, "learning_rate": 9.763154445258522e-05, "loss": 0.0145, "step": 731 }, { "epoch": 4.72258064516129, "grad_norm": 0.07918071746826172, "learning_rate": 9.76178249636392e-05, "loss": 0.0116, "step": 732 }, { "epoch": 4.729032258064516, "grad_norm": 0.07499564439058304, "learning_rate": 9.760406682328015e-05, "loss": 0.0112, "step": 733 }, { "epoch": 4.735483870967742, "grad_norm": 0.07125238329172134, "learning_rate": 9.759027004267558e-05, "loss": 0.0129, "step": 734 }, { "epoch": 4.741935483870968, "grad_norm": 0.0714704766869545, "learning_rate": 9.757643463302433e-05, "loss": 0.013, "step": 735 }, { "epoch": 4.748387096774193, "grad_norm": 0.057601042091846466, "learning_rate": 9.756256060555659e-05, "loss": 0.0105, "step": 736 }, { "epoch": 4.754838709677419, "grad_norm": 0.07206641882658005, "learning_rate": 9.754864797153394e-05, "loss": 0.0112, "step": 737 }, { "epoch": 4.7612903225806456, "grad_norm": 0.07657714933156967, "learning_rate": 9.753469674224928e-05, "loss": 0.0173, "step": 738 }, { "epoch": 4.767741935483871, "grad_norm": 0.07161399722099304, "learning_rate": 9.752070692902679e-05, "loss": 0.0128, "step": 739 }, { "epoch": 4.774193548387097, "grad_norm": 0.059535328298807144, "learning_rate": 9.750667854322207e-05, "loss": 0.012, "step": 740 }, { "epoch": 4.780645161290322, "grad_norm": 0.05769206956028938, "learning_rate": 9.749261159622192e-05, "loss": 0.0108, "step": 741 }, { "epoch": 4.787096774193548, "grad_norm": 0.0739009827375412, "learning_rate": 9.747850609944449e-05, "loss": 0.0124, "step": 742 }, { "epoch": 4.7935483870967746, "grad_norm": 0.059828855097293854, "learning_rate": 9.746436206433923e-05, "loss": 0.01, "step": 743 }, { "epoch": 4.8, "grad_norm": 0.05995893478393555, "learning_rate": 9.745017950238687e-05, "loss": 0.0091, "step": 744 }, { "epoch": 4.806451612903226, "grad_norm": 0.08231929689645767, "learning_rate": 9.74359584250994e-05, "loss": 0.0148, "step": 745 }, { "epoch": 4.812903225806451, "grad_norm": 0.06564637273550034, "learning_rate": 9.742169884402006e-05, "loss": 0.011, "step": 746 }, { "epoch": 4.819354838709677, "grad_norm": 0.06673679500818253, "learning_rate": 9.740740077072337e-05, "loss": 0.0119, "step": 747 }, { "epoch": 4.825806451612904, "grad_norm": 0.07423175126314163, "learning_rate": 9.739306421681506e-05, "loss": 0.0114, "step": 748 }, { "epoch": 4.832258064516129, "grad_norm": 0.06966380029916763, "learning_rate": 9.737868919393218e-05, "loss": 0.0124, "step": 749 }, { "epoch": 4.838709677419355, "grad_norm": 0.07534418255090714, "learning_rate": 9.736427571374288e-05, "loss": 0.0098, "step": 750 }, { "epoch": 4.838709677419355, "eval_loss": 0.025038301944732666, "eval_runtime": 20.6375, "eval_samples_per_second": 4.846, "eval_steps_per_second": 0.097, "step": 750 }, { "epoch": 4.845161290322581, "grad_norm": 0.06416190415620804, "learning_rate": 9.734982378794661e-05, "loss": 0.0092, "step": 751 }, { "epoch": 4.851612903225806, "grad_norm": 0.0716695487499237, "learning_rate": 9.733533342827403e-05, "loss": 0.0118, "step": 752 }, { "epoch": 4.858064516129033, "grad_norm": 0.058158546686172485, "learning_rate": 9.732080464648694e-05, "loss": 0.0107, "step": 753 }, { "epoch": 4.864516129032258, "grad_norm": 0.07941026240587234, "learning_rate": 9.730623745437836e-05, "loss": 0.0106, "step": 754 }, { "epoch": 4.870967741935484, "grad_norm": 0.06183311715722084, "learning_rate": 9.72916318637725e-05, "loss": 0.0101, "step": 755 }, { "epoch": 4.877419354838709, "grad_norm": 0.07516051828861237, "learning_rate": 9.727698788652474e-05, "loss": 0.013, "step": 756 }, { "epoch": 4.883870967741935, "grad_norm": 0.06480251252651215, "learning_rate": 9.726230553452158e-05, "loss": 0.0109, "step": 757 }, { "epoch": 4.890322580645162, "grad_norm": 0.08192919194698334, "learning_rate": 9.72475848196807e-05, "loss": 0.0128, "step": 758 }, { "epoch": 4.896774193548387, "grad_norm": 0.07532630115747452, "learning_rate": 9.72328257539509e-05, "loss": 0.0143, "step": 759 }, { "epoch": 4.903225806451613, "grad_norm": 0.06316731870174408, "learning_rate": 9.721802834931215e-05, "loss": 0.0116, "step": 760 }, { "epoch": 4.909677419354839, "grad_norm": 0.0601482093334198, "learning_rate": 9.720319261777549e-05, "loss": 0.0113, "step": 761 }, { "epoch": 4.916129032258064, "grad_norm": 0.05570144206285477, "learning_rate": 9.718831857138309e-05, "loss": 0.0124, "step": 762 }, { "epoch": 4.922580645161291, "grad_norm": 0.06473618745803833, "learning_rate": 9.717340622220823e-05, "loss": 0.0127, "step": 763 }, { "epoch": 4.929032258064516, "grad_norm": 0.05129361152648926, "learning_rate": 9.715845558235525e-05, "loss": 0.0097, "step": 764 }, { "epoch": 4.935483870967742, "grad_norm": 0.05756600573658943, "learning_rate": 9.714346666395963e-05, "loss": 0.0114, "step": 765 }, { "epoch": 4.941935483870967, "grad_norm": 0.057772960513830185, "learning_rate": 9.712843947918786e-05, "loss": 0.0097, "step": 766 }, { "epoch": 4.948387096774193, "grad_norm": 0.06774067878723145, "learning_rate": 9.711337404023752e-05, "loss": 0.011, "step": 767 }, { "epoch": 4.95483870967742, "grad_norm": 0.06922607868909836, "learning_rate": 9.709827035933722e-05, "loss": 0.0112, "step": 768 }, { "epoch": 4.961290322580645, "grad_norm": 0.06101693958044052, "learning_rate": 9.708312844874664e-05, "loss": 0.0108, "step": 769 }, { "epoch": 4.967741935483871, "grad_norm": 0.07150861620903015, "learning_rate": 9.706794832075647e-05, "loss": 0.0144, "step": 770 }, { "epoch": 4.974193548387097, "grad_norm": 0.07714319974184036, "learning_rate": 9.705272998768844e-05, "loss": 0.0138, "step": 771 }, { "epoch": 4.980645161290322, "grad_norm": 0.06570860743522644, "learning_rate": 9.703747346189527e-05, "loss": 0.0113, "step": 772 }, { "epoch": 4.987096774193549, "grad_norm": 0.06901867687702179, "learning_rate": 9.702217875576068e-05, "loss": 0.0119, "step": 773 }, { "epoch": 4.993548387096774, "grad_norm": 0.06524281948804855, "learning_rate": 9.700684588169939e-05, "loss": 0.0112, "step": 774 }, { "epoch": 5.0, "grad_norm": 0.07196003198623657, "learning_rate": 9.699147485215713e-05, "loss": 0.0093, "step": 775 }, { "epoch": 5.0, "eval_loss": 0.023430930450558662, "eval_runtime": 20.6296, "eval_samples_per_second": 4.847, "eval_steps_per_second": 0.097, "step": 775 }, { "epoch": 5.006451612903226, "grad_norm": 0.05366189032793045, "learning_rate": 9.697606567961056e-05, "loss": 0.009, "step": 776 }, { "epoch": 5.012903225806451, "grad_norm": 0.06412062793970108, "learning_rate": 9.696061837656727e-05, "loss": 0.0091, "step": 777 }, { "epoch": 5.019354838709678, "grad_norm": 0.06010841950774193, "learning_rate": 9.694513295556591e-05, "loss": 0.0103, "step": 778 }, { "epoch": 5.025806451612903, "grad_norm": 0.057299621403217316, "learning_rate": 9.692960942917597e-05, "loss": 0.0083, "step": 779 }, { "epoch": 5.032258064516129, "grad_norm": 0.06616218388080597, "learning_rate": 9.69140478099979e-05, "loss": 0.0096, "step": 780 }, { "epoch": 5.038709677419355, "grad_norm": 0.06639255583286285, "learning_rate": 9.689844811066311e-05, "loss": 0.0081, "step": 781 }, { "epoch": 5.04516129032258, "grad_norm": 0.05673486739397049, "learning_rate": 9.688281034383384e-05, "loss": 0.0069, "step": 782 }, { "epoch": 5.051612903225807, "grad_norm": 0.0635608583688736, "learning_rate": 9.68671345222033e-05, "loss": 0.0078, "step": 783 }, { "epoch": 5.058064516129032, "grad_norm": 0.07250364869832993, "learning_rate": 9.685142065849556e-05, "loss": 0.0084, "step": 784 }, { "epoch": 5.064516129032258, "grad_norm": 0.0699780210852623, "learning_rate": 9.683566876546555e-05, "loss": 0.0065, "step": 785 }, { "epoch": 5.070967741935484, "grad_norm": 0.07433956116437912, "learning_rate": 9.68198788558991e-05, "loss": 0.0084, "step": 786 }, { "epoch": 5.077419354838709, "grad_norm": 0.06806059181690216, "learning_rate": 9.680405094261291e-05, "loss": 0.0074, "step": 787 }, { "epoch": 5.083870967741936, "grad_norm": 0.058462999761104584, "learning_rate": 9.678818503845448e-05, "loss": 0.0082, "step": 788 }, { "epoch": 5.090322580645161, "grad_norm": 0.05572250112891197, "learning_rate": 9.677228115630218e-05, "loss": 0.0065, "step": 789 }, { "epoch": 5.096774193548387, "grad_norm": 0.07094363868236542, "learning_rate": 9.67563393090652e-05, "loss": 0.0072, "step": 790 }, { "epoch": 5.103225806451613, "grad_norm": 0.05128363147377968, "learning_rate": 9.674035950968354e-05, "loss": 0.0068, "step": 791 }, { "epoch": 5.109677419354838, "grad_norm": 0.06875258684158325, "learning_rate": 9.672434177112803e-05, "loss": 0.0078, "step": 792 }, { "epoch": 5.116129032258065, "grad_norm": 0.05699928104877472, "learning_rate": 9.670828610640026e-05, "loss": 0.0085, "step": 793 }, { "epoch": 5.122580645161291, "grad_norm": 0.06592380255460739, "learning_rate": 9.669219252853261e-05, "loss": 0.0078, "step": 794 }, { "epoch": 5.129032258064516, "grad_norm": 0.07786811143159866, "learning_rate": 9.667606105058828e-05, "loss": 0.0092, "step": 795 }, { "epoch": 5.135483870967742, "grad_norm": 0.05944572389125824, "learning_rate": 9.665989168566116e-05, "loss": 0.0065, "step": 796 }, { "epoch": 5.141935483870967, "grad_norm": 0.06329312175512314, "learning_rate": 9.664368444687597e-05, "loss": 0.0087, "step": 797 }, { "epoch": 5.148387096774194, "grad_norm": 0.073185496032238, "learning_rate": 9.662743934738811e-05, "loss": 0.0081, "step": 798 }, { "epoch": 5.15483870967742, "grad_norm": 0.08725328743457794, "learning_rate": 9.661115640038373e-05, "loss": 0.008, "step": 799 }, { "epoch": 5.161290322580645, "grad_norm": 0.059760626405477524, "learning_rate": 9.659483561907974e-05, "loss": 0.0073, "step": 800 }, { "epoch": 5.161290322580645, "eval_loss": 0.02465805411338806, "eval_runtime": 20.6536, "eval_samples_per_second": 4.842, "eval_steps_per_second": 0.097, "step": 800 }, { "epoch": 5.167741935483871, "grad_norm": 0.06096651777625084, "learning_rate": 9.657847701672368e-05, "loss": 0.0071, "step": 801 }, { "epoch": 5.174193548387096, "grad_norm": 0.05319224298000336, "learning_rate": 9.656208060659387e-05, "loss": 0.0065, "step": 802 }, { "epoch": 5.180645161290323, "grad_norm": 0.052184510976076126, "learning_rate": 9.654564640199927e-05, "loss": 0.0069, "step": 803 }, { "epoch": 5.187096774193549, "grad_norm": 0.06279890239238739, "learning_rate": 9.652917441627954e-05, "loss": 0.0104, "step": 804 }, { "epoch": 5.193548387096774, "grad_norm": 0.07697561383247375, "learning_rate": 9.651266466280499e-05, "loss": 0.0091, "step": 805 }, { "epoch": 5.2, "grad_norm": 0.07125485688447952, "learning_rate": 9.649611715497662e-05, "loss": 0.008, "step": 806 }, { "epoch": 5.2064516129032254, "grad_norm": 0.051803454756736755, "learning_rate": 9.647953190622602e-05, "loss": 0.0062, "step": 807 }, { "epoch": 5.212903225806452, "grad_norm": 0.05186665058135986, "learning_rate": 9.646290893001547e-05, "loss": 0.0066, "step": 808 }, { "epoch": 5.219354838709678, "grad_norm": 0.056108903139829636, "learning_rate": 9.644624823983782e-05, "loss": 0.0075, "step": 809 }, { "epoch": 5.225806451612903, "grad_norm": 0.07074058055877686, "learning_rate": 9.642954984921657e-05, "loss": 0.0103, "step": 810 }, { "epoch": 5.232258064516129, "grad_norm": 0.05524521321058273, "learning_rate": 9.641281377170581e-05, "loss": 0.007, "step": 811 }, { "epoch": 5.2387096774193544, "grad_norm": 0.05544526129961014, "learning_rate": 9.639604002089025e-05, "loss": 0.0084, "step": 812 }, { "epoch": 5.245161290322581, "grad_norm": 0.06004936620593071, "learning_rate": 9.637922861038513e-05, "loss": 0.0069, "step": 813 }, { "epoch": 5.251612903225807, "grad_norm": 0.06896119564771652, "learning_rate": 9.636237955383629e-05, "loss": 0.0137, "step": 814 }, { "epoch": 5.258064516129032, "grad_norm": 0.07235223799943924, "learning_rate": 9.63454928649201e-05, "loss": 0.0106, "step": 815 }, { "epoch": 5.264516129032258, "grad_norm": 0.06411406397819519, "learning_rate": 9.632856855734352e-05, "loss": 0.0083, "step": 816 }, { "epoch": 5.270967741935484, "grad_norm": 0.057488951832056046, "learning_rate": 9.631160664484398e-05, "loss": 0.0079, "step": 817 }, { "epoch": 5.27741935483871, "grad_norm": 0.04805266857147217, "learning_rate": 9.629460714118952e-05, "loss": 0.0061, "step": 818 }, { "epoch": 5.283870967741936, "grad_norm": 0.06258268654346466, "learning_rate": 9.627757006017861e-05, "loss": 0.0092, "step": 819 }, { "epoch": 5.290322580645161, "grad_norm": 0.051763035356998444, "learning_rate": 9.626049541564028e-05, "loss": 0.0067, "step": 820 }, { "epoch": 5.296774193548387, "grad_norm": 0.06160223111510277, "learning_rate": 9.624338322143403e-05, "loss": 0.0091, "step": 821 }, { "epoch": 5.3032258064516125, "grad_norm": 0.06290455907583237, "learning_rate": 9.622623349144982e-05, "loss": 0.008, "step": 822 }, { "epoch": 5.309677419354839, "grad_norm": 0.059179775416851044, "learning_rate": 9.620904623960808e-05, "loss": 0.0081, "step": 823 }, { "epoch": 5.316129032258065, "grad_norm": 0.05398336052894592, "learning_rate": 9.619182147985975e-05, "loss": 0.0072, "step": 824 }, { "epoch": 5.32258064516129, "grad_norm": 0.06514342129230499, "learning_rate": 9.617455922618614e-05, "loss": 0.0087, "step": 825 }, { "epoch": 5.32258064516129, "eval_loss": 0.02535240724682808, "eval_runtime": 20.6322, "eval_samples_per_second": 4.847, "eval_steps_per_second": 0.097, "step": 825 }, { "epoch": 5.329032258064516, "grad_norm": 0.0695829913020134, "learning_rate": 9.615725949259906e-05, "loss": 0.0089, "step": 826 }, { "epoch": 5.335483870967742, "grad_norm": 0.06091795116662979, "learning_rate": 9.61399222931407e-05, "loss": 0.0076, "step": 827 }, { "epoch": 5.341935483870968, "grad_norm": 0.055582962930202484, "learning_rate": 9.612254764188369e-05, "loss": 0.0079, "step": 828 }, { "epoch": 5.348387096774194, "grad_norm": 0.07405589520931244, "learning_rate": 9.610513555293101e-05, "loss": 0.0105, "step": 829 }, { "epoch": 5.354838709677419, "grad_norm": 0.050850674510002136, "learning_rate": 9.608768604041609e-05, "loss": 0.0066, "step": 830 }, { "epoch": 5.361290322580645, "grad_norm": 0.05596979707479477, "learning_rate": 9.607019911850269e-05, "loss": 0.0083, "step": 831 }, { "epoch": 5.367741935483871, "grad_norm": 0.06340134888887405, "learning_rate": 9.605267480138496e-05, "loss": 0.0109, "step": 832 }, { "epoch": 5.374193548387097, "grad_norm": 0.053815796971321106, "learning_rate": 9.60351131032874e-05, "loss": 0.0074, "step": 833 }, { "epoch": 5.380645161290323, "grad_norm": 0.07047566026449203, "learning_rate": 9.601751403846486e-05, "loss": 0.007, "step": 834 }, { "epoch": 5.387096774193548, "grad_norm": 0.06638946384191513, "learning_rate": 9.599987762120248e-05, "loss": 0.0095, "step": 835 }, { "epoch": 5.393548387096774, "grad_norm": 0.07727184891700745, "learning_rate": 9.598220386581577e-05, "loss": 0.0098, "step": 836 }, { "epoch": 5.4, "grad_norm": 0.07882851362228394, "learning_rate": 9.596449278665053e-05, "loss": 0.0081, "step": 837 }, { "epoch": 5.406451612903226, "grad_norm": 0.055430445820093155, "learning_rate": 9.594674439808284e-05, "loss": 0.0074, "step": 838 }, { "epoch": 5.412903225806452, "grad_norm": 0.07451054453849792, "learning_rate": 9.592895871451908e-05, "loss": 0.0071, "step": 839 }, { "epoch": 5.419354838709677, "grad_norm": 0.054919954389333725, "learning_rate": 9.591113575039589e-05, "loss": 0.008, "step": 840 }, { "epoch": 5.425806451612903, "grad_norm": 0.0726809874176979, "learning_rate": 9.589327552018021e-05, "loss": 0.0079, "step": 841 }, { "epoch": 5.432258064516129, "grad_norm": 0.06910938024520874, "learning_rate": 9.587537803836918e-05, "loss": 0.007, "step": 842 }, { "epoch": 5.438709677419355, "grad_norm": 0.052713364362716675, "learning_rate": 9.585744331949018e-05, "loss": 0.0067, "step": 843 }, { "epoch": 5.445161290322581, "grad_norm": 0.06306347250938416, "learning_rate": 9.583947137810085e-05, "loss": 0.0085, "step": 844 }, { "epoch": 5.451612903225806, "grad_norm": 0.06878506392240524, "learning_rate": 9.5821462228789e-05, "loss": 0.0091, "step": 845 }, { "epoch": 5.458064516129032, "grad_norm": 0.07872054725885391, "learning_rate": 9.58034158861727e-05, "loss": 0.0108, "step": 846 }, { "epoch": 5.464516129032258, "grad_norm": 0.06477072834968567, "learning_rate": 9.578533236490015e-05, "loss": 0.0091, "step": 847 }, { "epoch": 5.470967741935484, "grad_norm": 0.07100635021924973, "learning_rate": 9.576721167964975e-05, "loss": 0.0094, "step": 848 }, { "epoch": 5.47741935483871, "grad_norm": 0.06277081370353699, "learning_rate": 9.57490538451301e-05, "loss": 0.0087, "step": 849 }, { "epoch": 5.483870967741936, "grad_norm": 0.06297729909420013, "learning_rate": 9.573085887607991e-05, "loss": 0.0102, "step": 850 }, { "epoch": 5.483870967741936, "eval_loss": 0.02419852837920189, "eval_runtime": 20.6309, "eval_samples_per_second": 4.847, "eval_steps_per_second": 0.097, "step": 850 }, { "epoch": 5.490322580645161, "grad_norm": 0.05889495834708214, "learning_rate": 9.571262678726802e-05, "loss": 0.0075, "step": 851 }, { "epoch": 5.496774193548387, "grad_norm": 0.07738583534955978, "learning_rate": 9.569435759349347e-05, "loss": 0.0098, "step": 852 }, { "epoch": 5.503225806451613, "grad_norm": 0.06201765313744545, "learning_rate": 9.567605130958532e-05, "loss": 0.008, "step": 853 }, { "epoch": 5.509677419354839, "grad_norm": 0.0606846958398819, "learning_rate": 9.565770795040286e-05, "loss": 0.0086, "step": 854 }, { "epoch": 5.516129032258064, "grad_norm": 0.06933721899986267, "learning_rate": 9.563932753083534e-05, "loss": 0.009, "step": 855 }, { "epoch": 5.52258064516129, "grad_norm": 0.05704152211546898, "learning_rate": 9.562091006580219e-05, "loss": 0.0081, "step": 856 }, { "epoch": 5.529032258064516, "grad_norm": 0.046698153018951416, "learning_rate": 9.560245557025285e-05, "loss": 0.0069, "step": 857 }, { "epoch": 5.535483870967742, "grad_norm": 0.05271074175834656, "learning_rate": 9.558396405916686e-05, "loss": 0.0069, "step": 858 }, { "epoch": 5.541935483870968, "grad_norm": 0.053661540150642395, "learning_rate": 9.55654355475538e-05, "loss": 0.0092, "step": 859 }, { "epoch": 5.548387096774194, "grad_norm": 0.0608212947845459, "learning_rate": 9.554687005045327e-05, "loss": 0.0082, "step": 860 }, { "epoch": 5.554838709677419, "grad_norm": 0.06301247328519821, "learning_rate": 9.552826758293487e-05, "loss": 0.0084, "step": 861 }, { "epoch": 5.561290322580645, "grad_norm": 0.06680440157651901, "learning_rate": 9.550962816009825e-05, "loss": 0.0094, "step": 862 }, { "epoch": 5.567741935483871, "grad_norm": 0.06409584730863571, "learning_rate": 9.549095179707304e-05, "loss": 0.01, "step": 863 }, { "epoch": 5.574193548387097, "grad_norm": 0.07452081888914108, "learning_rate": 9.547223850901885e-05, "loss": 0.0091, "step": 864 }, { "epoch": 5.580645161290323, "grad_norm": 0.07207295298576355, "learning_rate": 9.545348831112525e-05, "loss": 0.0081, "step": 865 }, { "epoch": 5.587096774193548, "grad_norm": 0.05220545455813408, "learning_rate": 9.543470121861183e-05, "loss": 0.0078, "step": 866 }, { "epoch": 5.593548387096774, "grad_norm": 0.05606219172477722, "learning_rate": 9.541587724672805e-05, "loss": 0.0072, "step": 867 }, { "epoch": 5.6, "grad_norm": 0.060226067900657654, "learning_rate": 9.539701641075333e-05, "loss": 0.0078, "step": 868 }, { "epoch": 5.606451612903226, "grad_norm": 0.05184217914938927, "learning_rate": 9.537811872599702e-05, "loss": 0.0082, "step": 869 }, { "epoch": 5.612903225806452, "grad_norm": 0.052946384996175766, "learning_rate": 9.535918420779841e-05, "loss": 0.0063, "step": 870 }, { "epoch": 5.619354838709677, "grad_norm": 0.057897213846445084, "learning_rate": 9.534021287152663e-05, "loss": 0.0084, "step": 871 }, { "epoch": 5.625806451612903, "grad_norm": 0.06058947369456291, "learning_rate": 9.532120473258075e-05, "loss": 0.0089, "step": 872 }, { "epoch": 5.632258064516129, "grad_norm": 0.06276161968708038, "learning_rate": 9.530215980638966e-05, "loss": 0.009, "step": 873 }, { "epoch": 5.638709677419355, "grad_norm": 0.05766284838318825, "learning_rate": 9.528307810841214e-05, "loss": 0.0094, "step": 874 }, { "epoch": 5.645161290322581, "grad_norm": 0.0866272896528244, "learning_rate": 9.526395965413682e-05, "loss": 0.0077, "step": 875 }, { "epoch": 5.645161290322581, "eval_loss": 0.023009639233350754, "eval_runtime": 20.6229, "eval_samples_per_second": 4.849, "eval_steps_per_second": 0.097, "step": 875 }, { "epoch": 5.651612903225806, "grad_norm": 0.056562308222055435, "learning_rate": 9.524480445908216e-05, "loss": 0.0073, "step": 876 }, { "epoch": 5.658064516129032, "grad_norm": 0.07527109235525131, "learning_rate": 9.522561253879644e-05, "loss": 0.0086, "step": 877 }, { "epoch": 5.664516129032258, "grad_norm": 0.06116427481174469, "learning_rate": 9.520638390885777e-05, "loss": 0.0086, "step": 878 }, { "epoch": 5.670967741935484, "grad_norm": 0.0696583017706871, "learning_rate": 9.5187118584874e-05, "loss": 0.0091, "step": 879 }, { "epoch": 5.67741935483871, "grad_norm": 0.05933855101466179, "learning_rate": 9.516781658248284e-05, "loss": 0.0076, "step": 880 }, { "epoch": 5.683870967741935, "grad_norm": 0.05984939634799957, "learning_rate": 9.514847791735171e-05, "loss": 0.0078, "step": 881 }, { "epoch": 5.690322580645161, "grad_norm": 0.05205286294221878, "learning_rate": 9.512910260517782e-05, "loss": 0.0087, "step": 882 }, { "epoch": 5.6967741935483875, "grad_norm": 0.05857549607753754, "learning_rate": 9.510969066168813e-05, "loss": 0.0098, "step": 883 }, { "epoch": 5.703225806451613, "grad_norm": 0.05208706110715866, "learning_rate": 9.509024210263933e-05, "loss": 0.0081, "step": 884 }, { "epoch": 5.709677419354839, "grad_norm": 0.0637330710887909, "learning_rate": 9.507075694381782e-05, "loss": 0.0097, "step": 885 }, { "epoch": 5.716129032258064, "grad_norm": 0.055275216698646545, "learning_rate": 9.505123520103968e-05, "loss": 0.0066, "step": 886 }, { "epoch": 5.72258064516129, "grad_norm": 0.05725083872675896, "learning_rate": 9.503167689015077e-05, "loss": 0.0074, "step": 887 }, { "epoch": 5.729032258064516, "grad_norm": 0.061076343059539795, "learning_rate": 9.501208202702656e-05, "loss": 0.0091, "step": 888 }, { "epoch": 5.735483870967742, "grad_norm": 0.060000624507665634, "learning_rate": 9.49924506275722e-05, "loss": 0.01, "step": 889 }, { "epoch": 5.741935483870968, "grad_norm": 0.12105455994606018, "learning_rate": 9.497278270772251e-05, "loss": 0.0075, "step": 890 }, { "epoch": 5.748387096774193, "grad_norm": 0.06688644737005234, "learning_rate": 9.495307828344198e-05, "loss": 0.0097, "step": 891 }, { "epoch": 5.754838709677419, "grad_norm": 0.07520955801010132, "learning_rate": 9.493333737072465e-05, "loss": 0.012, "step": 892 }, { "epoch": 5.7612903225806456, "grad_norm": 0.06276785582304001, "learning_rate": 9.491355998559426e-05, "loss": 0.0072, "step": 893 }, { "epoch": 5.767741935483871, "grad_norm": 0.0585506446659565, "learning_rate": 9.489374614410413e-05, "loss": 0.0073, "step": 894 }, { "epoch": 5.774193548387097, "grad_norm": 0.054901156574487686, "learning_rate": 9.487389586233717e-05, "loss": 0.0076, "step": 895 }, { "epoch": 5.780645161290322, "grad_norm": 0.05712338536977768, "learning_rate": 9.485400915640583e-05, "loss": 0.0084, "step": 896 }, { "epoch": 5.787096774193548, "grad_norm": 0.07083632797002792, "learning_rate": 9.483408604245218e-05, "loss": 0.0085, "step": 897 }, { "epoch": 5.7935483870967746, "grad_norm": 0.07239817827939987, "learning_rate": 9.481412653664784e-05, "loss": 0.0105, "step": 898 }, { "epoch": 5.8, "grad_norm": 0.06843467801809311, "learning_rate": 9.47941306551939e-05, "loss": 0.0065, "step": 899 }, { "epoch": 5.806451612903226, "grad_norm": 0.06935566663742065, "learning_rate": 9.477409841432108e-05, "loss": 0.0085, "step": 900 }, { "epoch": 5.806451612903226, "eval_loss": 0.022989710792899132, "eval_runtime": 20.6155, "eval_samples_per_second": 4.851, "eval_steps_per_second": 0.097, "step": 900 }, { "epoch": 5.812903225806451, "grad_norm": 0.0664864331483841, "learning_rate": 9.475402983028953e-05, "loss": 0.0079, "step": 901 }, { "epoch": 5.819354838709677, "grad_norm": 0.08055754005908966, "learning_rate": 9.473392491938895e-05, "loss": 0.0075, "step": 902 }, { "epoch": 5.825806451612904, "grad_norm": 0.08719275146722794, "learning_rate": 9.47137836979385e-05, "loss": 0.0101, "step": 903 }, { "epoch": 5.832258064516129, "grad_norm": 0.06438102573156357, "learning_rate": 9.46936061822868e-05, "loss": 0.0082, "step": 904 }, { "epoch": 5.838709677419355, "grad_norm": 0.0579158253967762, "learning_rate": 9.467339238881199e-05, "loss": 0.0086, "step": 905 }, { "epoch": 5.845161290322581, "grad_norm": 0.06194271147251129, "learning_rate": 9.465314233392158e-05, "loss": 0.0104, "step": 906 }, { "epoch": 5.851612903225806, "grad_norm": 0.05593113228678703, "learning_rate": 9.463285603405255e-05, "loss": 0.0091, "step": 907 }, { "epoch": 5.858064516129033, "grad_norm": 0.06721027195453644, "learning_rate": 9.461253350567132e-05, "loss": 0.0085, "step": 908 }, { "epoch": 5.864516129032258, "grad_norm": 0.05705774575471878, "learning_rate": 9.459217476527368e-05, "loss": 0.0075, "step": 909 }, { "epoch": 5.870967741935484, "grad_norm": 0.05105244368314743, "learning_rate": 9.457177982938485e-05, "loss": 0.0085, "step": 910 }, { "epoch": 5.877419354838709, "grad_norm": 0.06674633175134659, "learning_rate": 9.455134871455941e-05, "loss": 0.0077, "step": 911 }, { "epoch": 5.883870967741935, "grad_norm": 0.058289535343647, "learning_rate": 9.453088143738129e-05, "loss": 0.0086, "step": 912 }, { "epoch": 5.890322580645162, "grad_norm": 0.07466110587120056, "learning_rate": 9.451037801446378e-05, "loss": 0.0105, "step": 913 }, { "epoch": 5.896774193548387, "grad_norm": 0.0619179792702198, "learning_rate": 9.448983846244955e-05, "loss": 0.0085, "step": 914 }, { "epoch": 5.903225806451613, "grad_norm": 0.06765176355838776, "learning_rate": 9.446926279801053e-05, "loss": 0.0102, "step": 915 }, { "epoch": 5.909677419354839, "grad_norm": 0.060166824609041214, "learning_rate": 9.444865103784804e-05, "loss": 0.0075, "step": 916 }, { "epoch": 5.916129032258064, "grad_norm": 0.06444356590509415, "learning_rate": 9.442800319869262e-05, "loss": 0.0071, "step": 917 }, { "epoch": 5.922580645161291, "grad_norm": 0.07145165652036667, "learning_rate": 9.440731929730415e-05, "loss": 0.0096, "step": 918 }, { "epoch": 5.929032258064516, "grad_norm": 0.05392426252365112, "learning_rate": 9.438659935047175e-05, "loss": 0.0064, "step": 919 }, { "epoch": 5.935483870967742, "grad_norm": 0.06370414048433304, "learning_rate": 9.436584337501382e-05, "loss": 0.0083, "step": 920 }, { "epoch": 5.941935483870967, "grad_norm": 0.0562657006084919, "learning_rate": 9.4345051387778e-05, "loss": 0.0086, "step": 921 }, { "epoch": 5.948387096774193, "grad_norm": 0.06523040682077408, "learning_rate": 9.432422340564115e-05, "loss": 0.0115, "step": 922 }, { "epoch": 5.95483870967742, "grad_norm": 0.08145349472761154, "learning_rate": 9.430335944550936e-05, "loss": 0.0127, "step": 923 }, { "epoch": 5.961290322580645, "grad_norm": 0.05838509649038315, "learning_rate": 9.42824595243179e-05, "loss": 0.0068, "step": 924 }, { "epoch": 5.967741935483871, "grad_norm": 0.05700171738862991, "learning_rate": 9.426152365903127e-05, "loss": 0.0069, "step": 925 }, { "epoch": 5.967741935483871, "eval_loss": 0.021346276625990868, "eval_runtime": 20.6536, "eval_samples_per_second": 4.842, "eval_steps_per_second": 0.097, "step": 925 }, { "epoch": 5.974193548387097, "grad_norm": 0.05546183884143829, "learning_rate": 9.424055186664311e-05, "loss": 0.0086, "step": 926 }, { "epoch": 5.980645161290322, "grad_norm": 0.051414016634225845, "learning_rate": 9.421954416417624e-05, "loss": 0.0089, "step": 927 }, { "epoch": 5.987096774193549, "grad_norm": 0.06089739874005318, "learning_rate": 9.419850056868262e-05, "loss": 0.0085, "step": 928 }, { "epoch": 5.993548387096774, "grad_norm": 0.056536007672548294, "learning_rate": 9.417742109724334e-05, "loss": 0.0086, "step": 929 }, { "epoch": 6.0, "grad_norm": 0.09017127752304077, "learning_rate": 9.415630576696863e-05, "loss": 0.0097, "step": 930 }, { "epoch": 6.006451612903226, "grad_norm": 0.04831506684422493, "learning_rate": 9.413515459499782e-05, "loss": 0.0065, "step": 931 }, { "epoch": 6.012903225806451, "grad_norm": 0.04744061455130577, "learning_rate": 9.411396759849931e-05, "loss": 0.0062, "step": 932 }, { "epoch": 6.019354838709678, "grad_norm": 0.048048291355371475, "learning_rate": 9.409274479467058e-05, "loss": 0.006, "step": 933 }, { "epoch": 6.025806451612903, "grad_norm": 0.04612509161233902, "learning_rate": 9.407148620073824e-05, "loss": 0.0054, "step": 934 }, { "epoch": 6.032258064516129, "grad_norm": 0.04895896464586258, "learning_rate": 9.405019183395787e-05, "loss": 0.0061, "step": 935 }, { "epoch": 6.038709677419355, "grad_norm": 0.04923215135931969, "learning_rate": 9.402886171161411e-05, "loss": 0.0056, "step": 936 }, { "epoch": 6.04516129032258, "grad_norm": 0.05468977242708206, "learning_rate": 9.400749585102067e-05, "loss": 0.0056, "step": 937 }, { "epoch": 6.051612903225807, "grad_norm": 0.04829660430550575, "learning_rate": 9.398609426952019e-05, "loss": 0.0049, "step": 938 }, { "epoch": 6.058064516129032, "grad_norm": 0.0703238770365715, "learning_rate": 9.396465698448435e-05, "loss": 0.0064, "step": 939 }, { "epoch": 6.064516129032258, "grad_norm": 0.06474682688713074, "learning_rate": 9.394318401331383e-05, "loss": 0.0071, "step": 940 }, { "epoch": 6.070967741935484, "grad_norm": 0.0701901838183403, "learning_rate": 9.392167537343823e-05, "loss": 0.0072, "step": 941 }, { "epoch": 6.077419354838709, "grad_norm": 0.05554347485303879, "learning_rate": 9.390013108231614e-05, "loss": 0.0069, "step": 942 }, { "epoch": 6.083870967741936, "grad_norm": 0.06273981183767319, "learning_rate": 9.387855115743505e-05, "loss": 0.0065, "step": 943 }, { "epoch": 6.090322580645161, "grad_norm": 0.05572091042995453, "learning_rate": 9.385693561631141e-05, "loss": 0.0059, "step": 944 }, { "epoch": 6.096774193548387, "grad_norm": 0.03943298012018204, "learning_rate": 9.383528447649057e-05, "loss": 0.0054, "step": 945 }, { "epoch": 6.103225806451613, "grad_norm": 0.04861641675233841, "learning_rate": 9.381359775554675e-05, "loss": 0.0055, "step": 946 }, { "epoch": 6.109677419354838, "grad_norm": 0.0495055690407753, "learning_rate": 9.379187547108312e-05, "loss": 0.0055, "step": 947 }, { "epoch": 6.116129032258065, "grad_norm": 0.048694632947444916, "learning_rate": 9.377011764073162e-05, "loss": 0.0058, "step": 948 }, { "epoch": 6.122580645161291, "grad_norm": 0.06473204493522644, "learning_rate": 9.37483242821531e-05, "loss": 0.0055, "step": 949 }, { "epoch": 6.129032258064516, "grad_norm": 0.05870223045349121, "learning_rate": 9.372649541303726e-05, "loss": 0.0056, "step": 950 }, { "epoch": 6.129032258064516, "eval_loss": 0.022576194256544113, "eval_runtime": 20.6217, "eval_samples_per_second": 4.849, "eval_steps_per_second": 0.097, "step": 950 }, { "epoch": 6.135483870967742, "grad_norm": 0.049414485692977905, "learning_rate": 9.37046310511026e-05, "loss": 0.0057, "step": 951 }, { "epoch": 6.141935483870967, "grad_norm": 0.05746641382575035, "learning_rate": 9.368273121409644e-05, "loss": 0.0068, "step": 952 }, { "epoch": 6.148387096774194, "grad_norm": 0.05265411362051964, "learning_rate": 9.366079591979486e-05, "loss": 0.0057, "step": 953 }, { "epoch": 6.15483870967742, "grad_norm": 0.07672982662916183, "learning_rate": 9.363882518600275e-05, "loss": 0.0068, "step": 954 }, { "epoch": 6.161290322580645, "grad_norm": 0.060838665813207626, "learning_rate": 9.361681903055382e-05, "loss": 0.0054, "step": 955 }, { "epoch": 6.167741935483871, "grad_norm": 0.052103929221630096, "learning_rate": 9.359477747131042e-05, "loss": 0.005, "step": 956 }, { "epoch": 6.174193548387096, "grad_norm": 0.05221958085894585, "learning_rate": 9.357270052616372e-05, "loss": 0.0065, "step": 957 }, { "epoch": 6.180645161290323, "grad_norm": 0.050105560570955276, "learning_rate": 9.355058821303359e-05, "loss": 0.0054, "step": 958 }, { "epoch": 6.187096774193549, "grad_norm": 0.0577901154756546, "learning_rate": 9.352844054986857e-05, "loss": 0.0067, "step": 959 }, { "epoch": 6.193548387096774, "grad_norm": 0.05385618656873703, "learning_rate": 9.3506257554646e-05, "loss": 0.0069, "step": 960 }, { "epoch": 6.2, "grad_norm": 0.06024151295423508, "learning_rate": 9.348403924537177e-05, "loss": 0.007, "step": 961 }, { "epoch": 6.2064516129032254, "grad_norm": 0.05268295854330063, "learning_rate": 9.346178564008052e-05, "loss": 0.0067, "step": 962 }, { "epoch": 6.212903225806452, "grad_norm": 0.04031063988804817, "learning_rate": 9.343949675683549e-05, "loss": 0.0049, "step": 963 }, { "epoch": 6.219354838709678, "grad_norm": 0.06445210427045822, "learning_rate": 9.34171726137286e-05, "loss": 0.0069, "step": 964 }, { "epoch": 6.225806451612903, "grad_norm": 0.0537705235183239, "learning_rate": 9.339481322888035e-05, "loss": 0.0068, "step": 965 }, { "epoch": 6.232258064516129, "grad_norm": 0.05740392953157425, "learning_rate": 9.337241862043988e-05, "loss": 0.0056, "step": 966 }, { "epoch": 6.2387096774193544, "grad_norm": 0.06430184841156006, "learning_rate": 9.33499888065849e-05, "loss": 0.0063, "step": 967 }, { "epoch": 6.245161290322581, "grad_norm": 0.07137993723154068, "learning_rate": 9.33275238055217e-05, "loss": 0.0071, "step": 968 }, { "epoch": 6.251612903225807, "grad_norm": 0.05400635674595833, "learning_rate": 9.330502363548515e-05, "loss": 0.007, "step": 969 }, { "epoch": 6.258064516129032, "grad_norm": 0.05895206704735756, "learning_rate": 9.32824883147386e-05, "loss": 0.0063, "step": 970 }, { "epoch": 6.264516129032258, "grad_norm": 0.05703943967819214, "learning_rate": 9.325991786157404e-05, "loss": 0.0063, "step": 971 }, { "epoch": 6.270967741935484, "grad_norm": 0.05772950127720833, "learning_rate": 9.32373122943119e-05, "loss": 0.0068, "step": 972 }, { "epoch": 6.27741935483871, "grad_norm": 0.06098498776555061, "learning_rate": 9.32146716313011e-05, "loss": 0.0063, "step": 973 }, { "epoch": 6.283870967741936, "grad_norm": 0.05442085489630699, "learning_rate": 9.319199589091911e-05, "loss": 0.0059, "step": 974 }, { "epoch": 6.290322580645161, "grad_norm": 0.06356194615364075, "learning_rate": 9.316928509157185e-05, "loss": 0.0063, "step": 975 }, { "epoch": 6.290322580645161, "eval_loss": 0.022397752851247787, "eval_runtime": 20.6405, "eval_samples_per_second": 4.845, "eval_steps_per_second": 0.097, "step": 975 }, { "epoch": 6.296774193548387, "grad_norm": 0.05127127468585968, "learning_rate": 9.314653925169365e-05, "loss": 0.0068, "step": 976 }, { "epoch": 6.3032258064516125, "grad_norm": 0.05523466691374779, "learning_rate": 9.312375838974735e-05, "loss": 0.0054, "step": 977 }, { "epoch": 6.309677419354839, "grad_norm": 0.05204270780086517, "learning_rate": 9.310094252422417e-05, "loss": 0.0054, "step": 978 }, { "epoch": 6.316129032258065, "grad_norm": 0.07963524013757706, "learning_rate": 9.307809167364377e-05, "loss": 0.0066, "step": 979 }, { "epoch": 6.32258064516129, "grad_norm": 0.05632827803492546, "learning_rate": 9.30552058565542e-05, "loss": 0.0049, "step": 980 }, { "epoch": 6.329032258064516, "grad_norm": 0.07022556662559509, "learning_rate": 9.303228509153188e-05, "loss": 0.0071, "step": 981 }, { "epoch": 6.335483870967742, "grad_norm": 0.05617734044790268, "learning_rate": 9.300932939718159e-05, "loss": 0.007, "step": 982 }, { "epoch": 6.341935483870968, "grad_norm": 0.0606408566236496, "learning_rate": 9.298633879213652e-05, "loss": 0.0055, "step": 983 }, { "epoch": 6.348387096774194, "grad_norm": 0.05535956844687462, "learning_rate": 9.296331329505813e-05, "loss": 0.0068, "step": 984 }, { "epoch": 6.354838709677419, "grad_norm": 0.05563861131668091, "learning_rate": 9.294025292463623e-05, "loss": 0.0049, "step": 985 }, { "epoch": 6.361290322580645, "grad_norm": 0.05694695562124252, "learning_rate": 9.291715769958895e-05, "loss": 0.0056, "step": 986 }, { "epoch": 6.367741935483871, "grad_norm": 0.06230534613132477, "learning_rate": 9.289402763866266e-05, "loss": 0.0058, "step": 987 }, { "epoch": 6.374193548387097, "grad_norm": 0.05194805935025215, "learning_rate": 9.287086276063208e-05, "loss": 0.0058, "step": 988 }, { "epoch": 6.380645161290323, "grad_norm": 0.05432300642132759, "learning_rate": 9.284766308430014e-05, "loss": 0.0054, "step": 989 }, { "epoch": 6.387096774193548, "grad_norm": 0.04841138422489166, "learning_rate": 9.282442862849803e-05, "loss": 0.0056, "step": 990 }, { "epoch": 6.393548387096774, "grad_norm": 0.05156128108501434, "learning_rate": 9.280115941208516e-05, "loss": 0.0065, "step": 991 }, { "epoch": 6.4, "grad_norm": 0.05393640324473381, "learning_rate": 9.277785545394917e-05, "loss": 0.0063, "step": 992 }, { "epoch": 6.406451612903226, "grad_norm": 0.04842933639883995, "learning_rate": 9.275451677300591e-05, "loss": 0.0059, "step": 993 }, { "epoch": 6.412903225806452, "grad_norm": 0.05428517609834671, "learning_rate": 9.273114338819938e-05, "loss": 0.0065, "step": 994 }, { "epoch": 6.419354838709677, "grad_norm": 0.05844871327280998, "learning_rate": 9.270773531850179e-05, "loss": 0.006, "step": 995 }, { "epoch": 6.425806451612903, "grad_norm": 0.06882674247026443, "learning_rate": 9.268429258291345e-05, "loss": 0.006, "step": 996 }, { "epoch": 6.432258064516129, "grad_norm": 0.05051298066973686, "learning_rate": 9.266081520046286e-05, "loss": 0.0063, "step": 997 }, { "epoch": 6.438709677419355, "grad_norm": 0.05807333439588547, "learning_rate": 9.263730319020663e-05, "loss": 0.0058, "step": 998 }, { "epoch": 6.445161290322581, "grad_norm": 0.07875418663024902, "learning_rate": 9.261375657122946e-05, "loss": 0.0076, "step": 999 }, { "epoch": 6.451612903225806, "grad_norm": 0.05101834610104561, "learning_rate": 9.259017536264414e-05, "loss": 0.0055, "step": 1000 }, { "epoch": 6.451612903225806, "eval_loss": 0.022700941190123558, "eval_runtime": 20.6129, "eval_samples_per_second": 4.851, "eval_steps_per_second": 0.097, "step": 1000 }, { "epoch": 6.458064516129032, "grad_norm": 0.06487668305635452, "learning_rate": 9.256655958359159e-05, "loss": 0.0065, "step": 1001 }, { "epoch": 6.464516129032258, "grad_norm": 0.05859556794166565, "learning_rate": 9.25429092532407e-05, "loss": 0.0063, "step": 1002 }, { "epoch": 6.470967741935484, "grad_norm": 0.060135431587696075, "learning_rate": 9.25192243907885e-05, "loss": 0.0071, "step": 1003 }, { "epoch": 6.47741935483871, "grad_norm": 0.049589548259973526, "learning_rate": 9.249550501545997e-05, "loss": 0.0056, "step": 1004 }, { "epoch": 6.483870967741936, "grad_norm": 0.0466407835483551, "learning_rate": 9.247175114650817e-05, "loss": 0.0064, "step": 1005 }, { "epoch": 6.490322580645161, "grad_norm": 0.056752149015665054, "learning_rate": 9.24479628032141e-05, "loss": 0.0054, "step": 1006 }, { "epoch": 6.496774193548387, "grad_norm": 0.06155959144234657, "learning_rate": 9.242414000488677e-05, "loss": 0.0069, "step": 1007 }, { "epoch": 6.503225806451613, "grad_norm": 0.06714151054620743, "learning_rate": 9.240028277086319e-05, "loss": 0.0071, "step": 1008 }, { "epoch": 6.509677419354839, "grad_norm": 0.05602521076798439, "learning_rate": 9.237639112050827e-05, "loss": 0.0059, "step": 1009 }, { "epoch": 6.516129032258064, "grad_norm": 0.05501384660601616, "learning_rate": 9.235246507321487e-05, "loss": 0.0057, "step": 1010 }, { "epoch": 6.52258064516129, "grad_norm": 0.04764142259955406, "learning_rate": 9.232850464840377e-05, "loss": 0.0057, "step": 1011 }, { "epoch": 6.529032258064516, "grad_norm": 0.04876996576786041, "learning_rate": 9.230450986552368e-05, "loss": 0.0064, "step": 1012 }, { "epoch": 6.535483870967742, "grad_norm": 0.05286463722586632, "learning_rate": 9.228048074405115e-05, "loss": 0.0055, "step": 1013 }, { "epoch": 6.541935483870968, "grad_norm": 0.055027544498443604, "learning_rate": 9.225641730349066e-05, "loss": 0.006, "step": 1014 }, { "epoch": 6.548387096774194, "grad_norm": 0.06131773442029953, "learning_rate": 9.22323195633745e-05, "loss": 0.0058, "step": 1015 }, { "epoch": 6.554838709677419, "grad_norm": 0.049361925572156906, "learning_rate": 9.220818754326282e-05, "loss": 0.0052, "step": 1016 }, { "epoch": 6.561290322580645, "grad_norm": 0.047937195748090744, "learning_rate": 9.218402126274358e-05, "loss": 0.0055, "step": 1017 }, { "epoch": 6.567741935483871, "grad_norm": 0.05250125750899315, "learning_rate": 9.215982074143259e-05, "loss": 0.0061, "step": 1018 }, { "epoch": 6.574193548387097, "grad_norm": 0.056667257100343704, "learning_rate": 9.213558599897341e-05, "loss": 0.0053, "step": 1019 }, { "epoch": 6.580645161290323, "grad_norm": 0.06139816343784332, "learning_rate": 9.211131705503739e-05, "loss": 0.0062, "step": 1020 }, { "epoch": 6.587096774193548, "grad_norm": 0.05592657998204231, "learning_rate": 9.208701392932367e-05, "loss": 0.0065, "step": 1021 }, { "epoch": 6.593548387096774, "grad_norm": 0.05883883312344551, "learning_rate": 9.206267664155907e-05, "loss": 0.0061, "step": 1022 }, { "epoch": 6.6, "grad_norm": 0.057344142347574234, "learning_rate": 9.20383052114982e-05, "loss": 0.0065, "step": 1023 }, { "epoch": 6.606451612903226, "grad_norm": 0.044084545224905014, "learning_rate": 9.201389965892336e-05, "loss": 0.0056, "step": 1024 }, { "epoch": 6.612903225806452, "grad_norm": 0.05402151495218277, "learning_rate": 9.198946000364454e-05, "loss": 0.0067, "step": 1025 }, { "epoch": 6.612903225806452, "eval_loss": 0.022883424535393715, "eval_runtime": 20.6396, "eval_samples_per_second": 4.845, "eval_steps_per_second": 0.097, "step": 1025 }, { "epoch": 6.619354838709677, "grad_norm": 0.06240704283118248, "learning_rate": 9.196498626549945e-05, "loss": 0.0076, "step": 1026 }, { "epoch": 6.625806451612903, "grad_norm": 0.06148531660437584, "learning_rate": 9.194047846435341e-05, "loss": 0.0049, "step": 1027 }, { "epoch": 6.632258064516129, "grad_norm": 0.05714572221040726, "learning_rate": 9.191593662009941e-05, "loss": 0.0053, "step": 1028 }, { "epoch": 6.638709677419355, "grad_norm": 0.05543628707528114, "learning_rate": 9.18913607526581e-05, "loss": 0.0052, "step": 1029 }, { "epoch": 6.645161290322581, "grad_norm": 0.0643729567527771, "learning_rate": 9.186675088197772e-05, "loss": 0.0069, "step": 1030 }, { "epoch": 6.651612903225806, "grad_norm": 0.06447968631982803, "learning_rate": 9.184210702803409e-05, "loss": 0.0069, "step": 1031 }, { "epoch": 6.658064516129032, "grad_norm": 0.04606137052178383, "learning_rate": 9.181742921083065e-05, "loss": 0.0053, "step": 1032 }, { "epoch": 6.664516129032258, "grad_norm": 0.07696260511875153, "learning_rate": 9.179271745039843e-05, "loss": 0.0079, "step": 1033 }, { "epoch": 6.670967741935484, "grad_norm": 0.06481795012950897, "learning_rate": 9.176797176679593e-05, "loss": 0.0065, "step": 1034 }, { "epoch": 6.67741935483871, "grad_norm": 0.05713411420583725, "learning_rate": 9.174319218010928e-05, "loss": 0.0054, "step": 1035 }, { "epoch": 6.683870967741935, "grad_norm": 0.055928200483322144, "learning_rate": 9.171837871045204e-05, "loss": 0.0064, "step": 1036 }, { "epoch": 6.690322580645161, "grad_norm": 0.054814040660858154, "learning_rate": 9.169353137796534e-05, "loss": 0.0069, "step": 1037 }, { "epoch": 6.6967741935483875, "grad_norm": 0.062209051102399826, "learning_rate": 9.166865020281777e-05, "loss": 0.0063, "step": 1038 }, { "epoch": 6.703225806451613, "grad_norm": 0.057959817349910736, "learning_rate": 9.16437352052054e-05, "loss": 0.0071, "step": 1039 }, { "epoch": 6.709677419354839, "grad_norm": 0.06060996651649475, "learning_rate": 9.16187864053517e-05, "loss": 0.0064, "step": 1040 }, { "epoch": 6.716129032258064, "grad_norm": 0.05503141134977341, "learning_rate": 9.15938038235077e-05, "loss": 0.0052, "step": 1041 }, { "epoch": 6.72258064516129, "grad_norm": 0.0538954995572567, "learning_rate": 9.156878747995172e-05, "loss": 0.0064, "step": 1042 }, { "epoch": 6.729032258064516, "grad_norm": 0.05267870053648949, "learning_rate": 9.154373739498955e-05, "loss": 0.0053, "step": 1043 }, { "epoch": 6.735483870967742, "grad_norm": 0.07564899325370789, "learning_rate": 9.151865358895439e-05, "loss": 0.007, "step": 1044 }, { "epoch": 6.741935483870968, "grad_norm": 0.05886682868003845, "learning_rate": 9.149353608220672e-05, "loss": 0.0055, "step": 1045 }, { "epoch": 6.748387096774193, "grad_norm": 0.033152226358652115, "learning_rate": 9.146838489513447e-05, "loss": 0.005, "step": 1046 }, { "epoch": 6.754838709677419, "grad_norm": 0.05886426195502281, "learning_rate": 9.144320004815286e-05, "loss": 0.0059, "step": 1047 }, { "epoch": 6.7612903225806456, "grad_norm": 0.07072407007217407, "learning_rate": 9.141798156170446e-05, "loss": 0.0097, "step": 1048 }, { "epoch": 6.767741935483871, "grad_norm": 0.05597013607621193, "learning_rate": 9.13927294562591e-05, "loss": 0.006, "step": 1049 }, { "epoch": 6.774193548387097, "grad_norm": 0.046950407326221466, "learning_rate": 9.136744375231397e-05, "loss": 0.0052, "step": 1050 }, { "epoch": 6.774193548387097, "eval_loss": 0.0224434956908226, "eval_runtime": 20.6125, "eval_samples_per_second": 4.851, "eval_steps_per_second": 0.097, "step": 1050 }, { "epoch": 6.780645161290322, "grad_norm": 0.05880718678236008, "learning_rate": 9.134212447039343e-05, "loss": 0.007, "step": 1051 }, { "epoch": 6.787096774193548, "grad_norm": 0.04981274530291557, "learning_rate": 9.13167716310492e-05, "loss": 0.005, "step": 1052 }, { "epoch": 6.7935483870967746, "grad_norm": 0.05875200033187866, "learning_rate": 9.129138525486019e-05, "loss": 0.0056, "step": 1053 }, { "epoch": 6.8, "grad_norm": 0.046490080654621124, "learning_rate": 9.12659653624325e-05, "loss": 0.0051, "step": 1054 }, { "epoch": 6.806451612903226, "grad_norm": 0.05880575627088547, "learning_rate": 9.12405119743995e-05, "loss": 0.006, "step": 1055 }, { "epoch": 6.812903225806451, "grad_norm": 0.046049751341342926, "learning_rate": 9.121502511142172e-05, "loss": 0.0058, "step": 1056 }, { "epoch": 6.819354838709677, "grad_norm": 0.04743240028619766, "learning_rate": 9.118950479418684e-05, "loss": 0.0063, "step": 1057 }, { "epoch": 6.825806451612904, "grad_norm": 0.056094031780958176, "learning_rate": 9.116395104340974e-05, "loss": 0.0067, "step": 1058 }, { "epoch": 6.832258064516129, "grad_norm": 0.052488215267658234, "learning_rate": 9.113836387983239e-05, "loss": 0.0044, "step": 1059 }, { "epoch": 6.838709677419355, "grad_norm": 0.05668174475431442, "learning_rate": 9.11127433242239e-05, "loss": 0.0069, "step": 1060 }, { "epoch": 6.845161290322581, "grad_norm": 0.06447296589612961, "learning_rate": 9.108708939738051e-05, "loss": 0.0056, "step": 1061 }, { "epoch": 6.851612903225806, "grad_norm": 0.05771500989794731, "learning_rate": 9.106140212012553e-05, "loss": 0.006, "step": 1062 }, { "epoch": 6.858064516129033, "grad_norm": 0.04775027185678482, "learning_rate": 9.10356815133093e-05, "loss": 0.005, "step": 1063 }, { "epoch": 6.864516129032258, "grad_norm": 0.060601286590099335, "learning_rate": 9.100992759780928e-05, "loss": 0.0082, "step": 1064 }, { "epoch": 6.870967741935484, "grad_norm": 0.05034959316253662, "learning_rate": 9.098414039452993e-05, "loss": 0.0067, "step": 1065 }, { "epoch": 6.877419354838709, "grad_norm": 0.061979107558727264, "learning_rate": 9.095831992440273e-05, "loss": 0.0075, "step": 1066 }, { "epoch": 6.883870967741935, "grad_norm": 0.04176148772239685, "learning_rate": 9.093246620838618e-05, "loss": 0.0057, "step": 1067 }, { "epoch": 6.890322580645162, "grad_norm": 0.05120096728205681, "learning_rate": 9.090657926746575e-05, "loss": 0.0063, "step": 1068 }, { "epoch": 6.896774193548387, "grad_norm": 0.043810728937387466, "learning_rate": 9.088065912265386e-05, "loss": 0.006, "step": 1069 }, { "epoch": 6.903225806451613, "grad_norm": 0.046423133462667465, "learning_rate": 9.085470579498995e-05, "loss": 0.0056, "step": 1070 }, { "epoch": 6.909677419354839, "grad_norm": 0.04729986563324928, "learning_rate": 9.082871930554031e-05, "loss": 0.0069, "step": 1071 }, { "epoch": 6.916129032258064, "grad_norm": 0.046129655092954636, "learning_rate": 9.080269967539823e-05, "loss": 0.0054, "step": 1072 }, { "epoch": 6.922580645161291, "grad_norm": 0.05045049265027046, "learning_rate": 9.07766469256838e-05, "loss": 0.006, "step": 1073 }, { "epoch": 6.929032258064516, "grad_norm": 0.06024559587240219, "learning_rate": 9.07505610775441e-05, "loss": 0.0081, "step": 1074 }, { "epoch": 6.935483870967742, "grad_norm": 0.061700381338596344, "learning_rate": 9.072444215215299e-05, "loss": 0.008, "step": 1075 }, { "epoch": 6.935483870967742, "eval_loss": 0.021896785125136375, "eval_runtime": 20.6536, "eval_samples_per_second": 4.842, "eval_steps_per_second": 0.097, "step": 1075 }, { "epoch": 6.941935483870967, "grad_norm": 0.04216156154870987, "learning_rate": 9.069829017071123e-05, "loss": 0.0054, "step": 1076 }, { "epoch": 6.948387096774193, "grad_norm": 0.05374651029706001, "learning_rate": 9.06721051544464e-05, "loss": 0.0071, "step": 1077 }, { "epoch": 6.95483870967742, "grad_norm": 0.04941989853978157, "learning_rate": 9.06458871246129e-05, "loss": 0.0058, "step": 1078 }, { "epoch": 6.961290322580645, "grad_norm": 0.06097535789012909, "learning_rate": 9.06196361024919e-05, "loss": 0.0055, "step": 1079 }, { "epoch": 6.967741935483871, "grad_norm": 0.04807058721780777, "learning_rate": 9.059335210939137e-05, "loss": 0.0044, "step": 1080 }, { "epoch": 6.974193548387097, "grad_norm": 0.04932430759072304, "learning_rate": 9.056703516664606e-05, "loss": 0.006, "step": 1081 }, { "epoch": 6.980645161290322, "grad_norm": 0.043658170849084854, "learning_rate": 9.054068529561744e-05, "loss": 0.0051, "step": 1082 }, { "epoch": 6.987096774193549, "grad_norm": 0.06009669229388237, "learning_rate": 9.051430251769368e-05, "loss": 0.0049, "step": 1083 }, { "epoch": 6.993548387096774, "grad_norm": 0.04545913636684418, "learning_rate": 9.048788685428974e-05, "loss": 0.0045, "step": 1084 }, { "epoch": 7.0, "grad_norm": 0.07388494163751602, "learning_rate": 9.046143832684719e-05, "loss": 0.0065, "step": 1085 }, { "epoch": 7.006451612903226, "grad_norm": 0.052344489842653275, "learning_rate": 9.043495695683435e-05, "loss": 0.0053, "step": 1086 }, { "epoch": 7.012903225806451, "grad_norm": 0.03753964230418205, "learning_rate": 9.040844276574614e-05, "loss": 0.0041, "step": 1087 }, { "epoch": 7.019354838709678, "grad_norm": 0.050138022750616074, "learning_rate": 9.038189577510415e-05, "loss": 0.0048, "step": 1088 }, { "epoch": 7.025806451612903, "grad_norm": 0.037361446768045425, "learning_rate": 9.035531600645657e-05, "loss": 0.0037, "step": 1089 }, { "epoch": 7.032258064516129, "grad_norm": 0.0569661408662796, "learning_rate": 9.032870348137823e-05, "loss": 0.0048, "step": 1090 }, { "epoch": 7.038709677419355, "grad_norm": 0.055356767028570175, "learning_rate": 9.030205822147052e-05, "loss": 0.0033, "step": 1091 }, { "epoch": 7.04516129032258, "grad_norm": 0.037388719618320465, "learning_rate": 9.027538024836142e-05, "loss": 0.0048, "step": 1092 }, { "epoch": 7.051612903225807, "grad_norm": 0.04436417669057846, "learning_rate": 9.024866958370544e-05, "loss": 0.0055, "step": 1093 }, { "epoch": 7.058064516129032, "grad_norm": 0.04340614750981331, "learning_rate": 9.022192624918366e-05, "loss": 0.0043, "step": 1094 }, { "epoch": 7.064516129032258, "grad_norm": 0.0408780612051487, "learning_rate": 9.019515026650365e-05, "loss": 0.0039, "step": 1095 }, { "epoch": 7.070967741935484, "grad_norm": 0.04257588088512421, "learning_rate": 9.016834165739948e-05, "loss": 0.0049, "step": 1096 }, { "epoch": 7.077419354838709, "grad_norm": 0.053617022931575775, "learning_rate": 9.014150044363171e-05, "loss": 0.0044, "step": 1097 }, { "epoch": 7.083870967741936, "grad_norm": 0.049007218331098557, "learning_rate": 9.01146266469874e-05, "loss": 0.0041, "step": 1098 }, { "epoch": 7.090322580645161, "grad_norm": 0.04696111008524895, "learning_rate": 9.008772028927998e-05, "loss": 0.0034, "step": 1099 }, { "epoch": 7.096774193548387, "grad_norm": 0.0659966990351677, "learning_rate": 9.006078139234939e-05, "loss": 0.0053, "step": 1100 }, { "epoch": 7.096774193548387, "eval_loss": 0.022749263793230057, "eval_runtime": 20.6454, "eval_samples_per_second": 4.844, "eval_steps_per_second": 0.097, "step": 1100 }, { "epoch": 7.103225806451613, "grad_norm": 0.05406612157821655, "learning_rate": 9.003380997806193e-05, "loss": 0.0045, "step": 1101 }, { "epoch": 7.109677419354838, "grad_norm": 0.05338094010949135, "learning_rate": 9.000680606831032e-05, "loss": 0.0059, "step": 1102 }, { "epoch": 7.116129032258065, "grad_norm": 0.055592648684978485, "learning_rate": 8.997976968501361e-05, "loss": 0.0056, "step": 1103 }, { "epoch": 7.122580645161291, "grad_norm": 0.053097788244485855, "learning_rate": 8.99527008501173e-05, "loss": 0.0047, "step": 1104 }, { "epoch": 7.129032258064516, "grad_norm": 0.048222865909338, "learning_rate": 8.992559958559315e-05, "loss": 0.0047, "step": 1105 }, { "epoch": 7.135483870967742, "grad_norm": 0.05399785935878754, "learning_rate": 8.989846591343927e-05, "loss": 0.0047, "step": 1106 }, { "epoch": 7.141935483870967, "grad_norm": 0.04064812511205673, "learning_rate": 8.987129985568007e-05, "loss": 0.0046, "step": 1107 }, { "epoch": 7.148387096774194, "grad_norm": 0.048960182815790176, "learning_rate": 8.984410143436628e-05, "loss": 0.0048, "step": 1108 }, { "epoch": 7.15483870967742, "grad_norm": 0.04570115730166435, "learning_rate": 8.981687067157484e-05, "loss": 0.0043, "step": 1109 }, { "epoch": 7.161290322580645, "grad_norm": 0.057650115340948105, "learning_rate": 8.9789607589409e-05, "loss": 0.0052, "step": 1110 }, { "epoch": 7.167741935483871, "grad_norm": 0.04060333967208862, "learning_rate": 8.976231220999821e-05, "loss": 0.0048, "step": 1111 }, { "epoch": 7.174193548387096, "grad_norm": 0.04383286461234093, "learning_rate": 8.973498455549814e-05, "loss": 0.0054, "step": 1112 }, { "epoch": 7.180645161290323, "grad_norm": 0.049976591020822525, "learning_rate": 8.970762464809064e-05, "loss": 0.0051, "step": 1113 }, { "epoch": 7.187096774193549, "grad_norm": 0.05424259603023529, "learning_rate": 8.968023250998381e-05, "loss": 0.0046, "step": 1114 }, { "epoch": 7.193548387096774, "grad_norm": 0.053604692220687866, "learning_rate": 8.965280816341183e-05, "loss": 0.004, "step": 1115 }, { "epoch": 7.2, "grad_norm": 0.052061185240745544, "learning_rate": 8.962535163063506e-05, "loss": 0.0049, "step": 1116 }, { "epoch": 7.2064516129032254, "grad_norm": 0.03969616815447807, "learning_rate": 8.959786293393997e-05, "loss": 0.004, "step": 1117 }, { "epoch": 7.212903225806452, "grad_norm": 0.051511142402887344, "learning_rate": 8.957034209563917e-05, "loss": 0.0042, "step": 1118 }, { "epoch": 7.219354838709678, "grad_norm": 0.04875602200627327, "learning_rate": 8.954278913807133e-05, "loss": 0.0048, "step": 1119 }, { "epoch": 7.225806451612903, "grad_norm": 0.044596847146749496, "learning_rate": 8.95152040836012e-05, "loss": 0.004, "step": 1120 }, { "epoch": 7.232258064516129, "grad_norm": 0.04940240457653999, "learning_rate": 8.948758695461958e-05, "loss": 0.0049, "step": 1121 }, { "epoch": 7.2387096774193544, "grad_norm": 0.04639221727848053, "learning_rate": 8.945993777354333e-05, "loss": 0.004, "step": 1122 }, { "epoch": 7.245161290322581, "grad_norm": 0.041882410645484924, "learning_rate": 8.943225656281527e-05, "loss": 0.0038, "step": 1123 }, { "epoch": 7.251612903225807, "grad_norm": 0.05632800981402397, "learning_rate": 8.940454334490427e-05, "loss": 0.0045, "step": 1124 }, { "epoch": 7.258064516129032, "grad_norm": 0.05960334837436676, "learning_rate": 8.937679814230517e-05, "loss": 0.0049, "step": 1125 }, { "epoch": 7.258064516129032, "eval_loss": 0.0219831932336092, "eval_runtime": 20.6196, "eval_samples_per_second": 4.85, "eval_steps_per_second": 0.097, "step": 1125 }, { "epoch": 7.264516129032258, "grad_norm": 0.04046253114938736, "learning_rate": 8.934902097753874e-05, "loss": 0.0036, "step": 1126 }, { "epoch": 7.270967741935484, "grad_norm": 0.04227912425994873, "learning_rate": 8.932121187315176e-05, "loss": 0.0043, "step": 1127 }, { "epoch": 7.27741935483871, "grad_norm": 0.05613262206315994, "learning_rate": 8.929337085171686e-05, "loss": 0.0047, "step": 1128 }, { "epoch": 7.283870967741936, "grad_norm": 0.04614945501089096, "learning_rate": 8.926549793583263e-05, "loss": 0.0044, "step": 1129 }, { "epoch": 7.290322580645161, "grad_norm": 0.04451045021414757, "learning_rate": 8.923759314812351e-05, "loss": 0.0041, "step": 1130 }, { "epoch": 7.296774193548387, "grad_norm": 0.07014819979667664, "learning_rate": 8.920965651123984e-05, "loss": 0.0061, "step": 1131 }, { "epoch": 7.3032258064516125, "grad_norm": 0.044604312628507614, "learning_rate": 8.918168804785782e-05, "loss": 0.0048, "step": 1132 }, { "epoch": 7.309677419354839, "grad_norm": 0.04796329513192177, "learning_rate": 8.915368778067945e-05, "loss": 0.0052, "step": 1133 }, { "epoch": 7.316129032258065, "grad_norm": 0.054540954530239105, "learning_rate": 8.912565573243256e-05, "loss": 0.0044, "step": 1134 }, { "epoch": 7.32258064516129, "grad_norm": 0.06614997982978821, "learning_rate": 8.909759192587078e-05, "loss": 0.0049, "step": 1135 }, { "epoch": 7.329032258064516, "grad_norm": 0.04955561086535454, "learning_rate": 8.90694963837735e-05, "loss": 0.004, "step": 1136 }, { "epoch": 7.335483870967742, "grad_norm": 0.039983998984098434, "learning_rate": 8.904136912894593e-05, "loss": 0.0042, "step": 1137 }, { "epoch": 7.341935483870968, "grad_norm": 0.043588194996118546, "learning_rate": 8.901321018421893e-05, "loss": 0.0043, "step": 1138 }, { "epoch": 7.348387096774194, "grad_norm": 0.049632057547569275, "learning_rate": 8.898501957244917e-05, "loss": 0.0042, "step": 1139 }, { "epoch": 7.354838709677419, "grad_norm": 0.03852184861898422, "learning_rate": 8.895679731651895e-05, "loss": 0.0043, "step": 1140 }, { "epoch": 7.361290322580645, "grad_norm": 0.05421644076704979, "learning_rate": 8.892854343933631e-05, "loss": 0.0043, "step": 1141 }, { "epoch": 7.367741935483871, "grad_norm": 0.0643831118941307, "learning_rate": 8.890025796383492e-05, "loss": 0.0056, "step": 1142 }, { "epoch": 7.374193548387097, "grad_norm": 0.04557755962014198, "learning_rate": 8.887194091297415e-05, "loss": 0.0058, "step": 1143 }, { "epoch": 7.380645161290323, "grad_norm": 0.09635678678750992, "learning_rate": 8.884359230973891e-05, "loss": 0.005, "step": 1144 }, { "epoch": 7.387096774193548, "grad_norm": 0.05057625100016594, "learning_rate": 8.881521217713983e-05, "loss": 0.0054, "step": 1145 }, { "epoch": 7.393548387096774, "grad_norm": 0.04902059584856033, "learning_rate": 8.878680053821304e-05, "loss": 0.0043, "step": 1146 }, { "epoch": 7.4, "grad_norm": 0.049218758940696716, "learning_rate": 8.87583574160203e-05, "loss": 0.0047, "step": 1147 }, { "epoch": 7.406451612903226, "grad_norm": 0.05688267946243286, "learning_rate": 8.872988283364892e-05, "loss": 0.0055, "step": 1148 }, { "epoch": 7.412903225806452, "grad_norm": 0.06391454488039017, "learning_rate": 8.870137681421169e-05, "loss": 0.0051, "step": 1149 }, { "epoch": 7.419354838709677, "grad_norm": 0.054065387696027756, "learning_rate": 8.867283938084701e-05, "loss": 0.0059, "step": 1150 }, { "epoch": 7.419354838709677, "eval_loss": 0.021773455664515495, "eval_runtime": 20.6268, "eval_samples_per_second": 4.848, "eval_steps_per_second": 0.097, "step": 1150 }, { "epoch": 7.425806451612903, "grad_norm": 0.045604806393384933, "learning_rate": 8.864427055671869e-05, "loss": 0.0044, "step": 1151 }, { "epoch": 7.432258064516129, "grad_norm": 0.04603537544608116, "learning_rate": 8.861567036501608e-05, "loss": 0.0052, "step": 1152 }, { "epoch": 7.438709677419355, "grad_norm": 0.0699162408709526, "learning_rate": 8.858703882895397e-05, "loss": 0.006, "step": 1153 }, { "epoch": 7.445161290322581, "grad_norm": 0.04083402082324028, "learning_rate": 8.855837597177258e-05, "loss": 0.0044, "step": 1154 }, { "epoch": 7.451612903225806, "grad_norm": 0.04157475382089615, "learning_rate": 8.852968181673757e-05, "loss": 0.0042, "step": 1155 }, { "epoch": 7.458064516129032, "grad_norm": 0.03213357925415039, "learning_rate": 8.850095638714003e-05, "loss": 0.0039, "step": 1156 }, { "epoch": 7.464516129032258, "grad_norm": 0.05405053496360779, "learning_rate": 8.847219970629636e-05, "loss": 0.0058, "step": 1157 }, { "epoch": 7.470967741935484, "grad_norm": 0.0503312386572361, "learning_rate": 8.844341179754839e-05, "loss": 0.0045, "step": 1158 }, { "epoch": 7.47741935483871, "grad_norm": 0.05273361876606941, "learning_rate": 8.841459268426329e-05, "loss": 0.0046, "step": 1159 }, { "epoch": 7.483870967741936, "grad_norm": 0.05701204389333725, "learning_rate": 8.838574238983354e-05, "loss": 0.005, "step": 1160 }, { "epoch": 7.490322580645161, "grad_norm": 0.07972869277000427, "learning_rate": 8.835686093767696e-05, "loss": 0.0044, "step": 1161 }, { "epoch": 7.496774193548387, "grad_norm": 0.04704023525118828, "learning_rate": 8.832794835123661e-05, "loss": 0.0044, "step": 1162 }, { "epoch": 7.503225806451613, "grad_norm": 0.0568983368575573, "learning_rate": 8.829900465398087e-05, "loss": 0.0052, "step": 1163 }, { "epoch": 7.509677419354839, "grad_norm": 0.046636976301670074, "learning_rate": 8.827002986940333e-05, "loss": 0.0046, "step": 1164 }, { "epoch": 7.516129032258064, "grad_norm": 0.06755772978067398, "learning_rate": 8.824102402102285e-05, "loss": 0.0052, "step": 1165 }, { "epoch": 7.52258064516129, "grad_norm": 0.04916645213961601, "learning_rate": 8.821198713238352e-05, "loss": 0.0053, "step": 1166 }, { "epoch": 7.529032258064516, "grad_norm": 0.049249373376369476, "learning_rate": 8.818291922705456e-05, "loss": 0.0049, "step": 1167 }, { "epoch": 7.535483870967742, "grad_norm": 0.046701643615961075, "learning_rate": 8.815382032863042e-05, "loss": 0.0042, "step": 1168 }, { "epoch": 7.541935483870968, "grad_norm": 0.04675468057394028, "learning_rate": 8.812469046073069e-05, "loss": 0.0053, "step": 1169 }, { "epoch": 7.548387096774194, "grad_norm": 0.040924135595560074, "learning_rate": 8.80955296470001e-05, "loss": 0.0046, "step": 1170 }, { "epoch": 7.554838709677419, "grad_norm": 0.06187480688095093, "learning_rate": 8.806633791110847e-05, "loss": 0.0047, "step": 1171 }, { "epoch": 7.561290322580645, "grad_norm": 0.04417719691991806, "learning_rate": 8.80371152767508e-05, "loss": 0.0048, "step": 1172 }, { "epoch": 7.567741935483871, "grad_norm": 0.0677109807729721, "learning_rate": 8.800786176764708e-05, "loss": 0.0048, "step": 1173 }, { "epoch": 7.574193548387097, "grad_norm": 0.045606911182403564, "learning_rate": 8.797857740754241e-05, "loss": 0.0053, "step": 1174 }, { "epoch": 7.580645161290323, "grad_norm": 0.04456802085042, "learning_rate": 8.79492622202069e-05, "loss": 0.0045, "step": 1175 }, { "epoch": 7.580645161290323, "eval_loss": 0.02147463709115982, "eval_runtime": 20.6465, "eval_samples_per_second": 4.843, "eval_steps_per_second": 0.097, "step": 1175 }, { "epoch": 7.587096774193548, "grad_norm": 0.05277775600552559, "learning_rate": 8.791991622943574e-05, "loss": 0.0055, "step": 1176 }, { "epoch": 7.593548387096774, "grad_norm": 0.05247287079691887, "learning_rate": 8.789053945904906e-05, "loss": 0.0048, "step": 1177 }, { "epoch": 7.6, "grad_norm": 0.057342078536748886, "learning_rate": 8.786113193289201e-05, "loss": 0.0056, "step": 1178 }, { "epoch": 7.606451612903226, "grad_norm": 0.054789651185274124, "learning_rate": 8.783169367483467e-05, "loss": 0.0044, "step": 1179 }, { "epoch": 7.612903225806452, "grad_norm": 0.05523642152547836, "learning_rate": 8.780222470877214e-05, "loss": 0.0047, "step": 1180 }, { "epoch": 7.619354838709677, "grad_norm": 0.0644751563668251, "learning_rate": 8.777272505862435e-05, "loss": 0.0051, "step": 1181 }, { "epoch": 7.625806451612903, "grad_norm": 0.044002220034599304, "learning_rate": 8.77431947483362e-05, "loss": 0.0043, "step": 1182 }, { "epoch": 7.632258064516129, "grad_norm": 0.06370144337415695, "learning_rate": 8.771363380187743e-05, "loss": 0.0049, "step": 1183 }, { "epoch": 7.638709677419355, "grad_norm": 0.05078444629907608, "learning_rate": 8.768404224324271e-05, "loss": 0.005, "step": 1184 }, { "epoch": 7.645161290322581, "grad_norm": 0.051172371953725815, "learning_rate": 8.76544200964515e-05, "loss": 0.0049, "step": 1185 }, { "epoch": 7.651612903225806, "grad_norm": 0.04795413091778755, "learning_rate": 8.762476738554815e-05, "loss": 0.0045, "step": 1186 }, { "epoch": 7.658064516129032, "grad_norm": 0.06591526418924332, "learning_rate": 8.759508413460174e-05, "loss": 0.0051, "step": 1187 }, { "epoch": 7.664516129032258, "grad_norm": 0.06435294449329376, "learning_rate": 8.75653703677062e-05, "loss": 0.0057, "step": 1188 }, { "epoch": 7.670967741935484, "grad_norm": 0.051989562809467316, "learning_rate": 8.753562610898021e-05, "loss": 0.0049, "step": 1189 }, { "epoch": 7.67741935483871, "grad_norm": 0.039358485490083694, "learning_rate": 8.750585138256717e-05, "loss": 0.0043, "step": 1190 }, { "epoch": 7.683870967741935, "grad_norm": 0.03751049563288689, "learning_rate": 8.74760462126353e-05, "loss": 0.004, "step": 1191 }, { "epoch": 7.690322580645161, "grad_norm": 0.058525778353214264, "learning_rate": 8.744621062337741e-05, "loss": 0.0053, "step": 1192 }, { "epoch": 7.6967741935483875, "grad_norm": 0.05404181033372879, "learning_rate": 8.74163446390111e-05, "loss": 0.0051, "step": 1193 }, { "epoch": 7.703225806451613, "grad_norm": 0.043138835579156876, "learning_rate": 8.73864482837786e-05, "loss": 0.004, "step": 1194 }, { "epoch": 7.709677419354839, "grad_norm": 0.046529654413461685, "learning_rate": 8.735652158194678e-05, "loss": 0.0051, "step": 1195 }, { "epoch": 7.716129032258064, "grad_norm": 0.05675065889954567, "learning_rate": 8.732656455780717e-05, "loss": 0.0041, "step": 1196 }, { "epoch": 7.72258064516129, "grad_norm": 0.04349643364548683, "learning_rate": 8.72965772356759e-05, "loss": 0.0041, "step": 1197 }, { "epoch": 7.729032258064516, "grad_norm": 0.04814918711781502, "learning_rate": 8.726655963989371e-05, "loss": 0.0046, "step": 1198 }, { "epoch": 7.735483870967742, "grad_norm": 0.05350608006119728, "learning_rate": 8.723651179482588e-05, "loss": 0.0041, "step": 1199 }, { "epoch": 7.741935483870968, "grad_norm": 0.0744265466928482, "learning_rate": 8.720643372486228e-05, "loss": 0.0058, "step": 1200 }, { "epoch": 7.741935483870968, "eval_loss": 0.020646410062909126, "eval_runtime": 20.6232, "eval_samples_per_second": 4.849, "eval_steps_per_second": 0.097, "step": 1200 } ], "logging_steps": 1, "max_steps": 3875, "num_input_tokens_seen": 0, "num_train_epochs": 25, "save_steps": 100, "stateful_callbacks": { "EarlyStoppingCallback": { "args": { "early_stopping_patience": 5, "early_stopping_threshold": 0.0 }, "attributes": { "early_stopping_patience_counter": 0 } }, "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 6.043424643958702e+18, "train_batch_size": 64, "trial_name": null, "trial_params": null }