{ "best_metric": null, "best_model_checkpoint": null, "epoch": 49.6, "eval_steps": 500, "global_step": 3100, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.32, "grad_norm": 2.214611053466797, "learning_rate": 3.064516129032258e-06, "loss": 2.8022, "step": 20 }, { "epoch": 0.64, "grad_norm": 0.9293187260627747, "learning_rate": 6.290322580645161e-06, "loss": 2.6463, "step": 40 }, { "epoch": 0.96, "grad_norm": 2.630194902420044, "learning_rate": 9.516129032258064e-06, "loss": 2.4026, "step": 60 }, { "epoch": 1.28, "grad_norm": 1.7517138719558716, "learning_rate": 1.2741935483870968e-05, "loss": 1.9639, "step": 80 }, { "epoch": 1.6, "grad_norm": 1.6100302934646606, "learning_rate": 1.596774193548387e-05, "loss": 1.7115, "step": 100 }, { "epoch": 1.92, "grad_norm": 3.468630790710449, "learning_rate": 1.9193548387096774e-05, "loss": 1.7534, "step": 120 }, { "epoch": 2.24, "grad_norm": 2.6692144870758057, "learning_rate": 2.2419354838709678e-05, "loss": 1.3422, "step": 140 }, { "epoch": 2.56, "grad_norm": 1.5186293125152588, "learning_rate": 2.5645161290322582e-05, "loss": 1.1582, "step": 160 }, { "epoch": 2.88, "grad_norm": 4.727887153625488, "learning_rate": 2.8870967741935483e-05, "loss": 1.1084, "step": 180 }, { "epoch": 3.2, "grad_norm": 2.800290107727051, "learning_rate": 3.2096774193548393e-05, "loss": 0.844, "step": 200 }, { "epoch": 3.52, "grad_norm": 1.1538563966751099, "learning_rate": 3.532258064516129e-05, "loss": 0.6051, "step": 220 }, { "epoch": 3.84, "grad_norm": 3.282315492630005, "learning_rate": 3.8548387096774195e-05, "loss": 0.7587, "step": 240 }, { "epoch": 4.16, "grad_norm": 0.9393389821052551, "learning_rate": 4.17741935483871e-05, "loss": 0.537, "step": 260 }, { "epoch": 4.48, "grad_norm": 1.5951660871505737, "learning_rate": 4.5e-05, "loss": 0.4063, "step": 280 }, { "epoch": 4.8, "grad_norm": 2.4143407344818115, "learning_rate": 4.822580645161291e-05, "loss": 0.4817, "step": 300 }, { "epoch": 5.12, "grad_norm": 1.1503572463989258, "learning_rate": 4.9998716243505096e-05, "loss": 0.3556, "step": 320 }, { "epoch": 5.44, "grad_norm": 1.5514352321624756, "learning_rate": 4.9986672191133314e-05, "loss": 0.2663, "step": 340 }, { "epoch": 5.76, "grad_norm": 2.0045723915100098, "learning_rate": 4.9961956248762694e-05, "loss": 0.3273, "step": 360 }, { "epoch": 6.08, "grad_norm": 0.716200053691864, "learning_rate": 4.992458095098368e-05, "loss": 0.2174, "step": 380 }, { "epoch": 6.4, "grad_norm": 3.2870705127716064, "learning_rate": 4.9874565252527765e-05, "loss": 0.1886, "step": 400 }, { "epoch": 6.72, "grad_norm": 2.3181071281433105, "learning_rate": 4.981193451865465e-05, "loss": 0.2278, "step": 420 }, { "epoch": 7.04, "grad_norm": 1.6296441555023193, "learning_rate": 4.9736720512288334e-05, "loss": 0.168, "step": 440 }, { "epoch": 7.36, "grad_norm": 1.757379174232483, "learning_rate": 4.964896137790873e-05, "loss": 0.1227, "step": 460 }, { "epoch": 7.68, "grad_norm": 1.2622524499893188, "learning_rate": 4.954870162220679e-05, "loss": 0.1261, "step": 480 }, { "epoch": 8.0, "grad_norm": 1.1865277290344238, "learning_rate": 4.943599209151314e-05, "loss": 0.1167, "step": 500 }, { "epoch": 8.32, "grad_norm": 1.244333028793335, "learning_rate": 4.931088994601157e-05, "loss": 0.1049, "step": 520 }, { "epoch": 8.64, "grad_norm": 2.6851558685302734, "learning_rate": 4.917345863075048e-05, "loss": 0.1016, "step": 540 }, { "epoch": 8.96, "grad_norm": 1.0457267761230469, "learning_rate": 4.902376784346697e-05, "loss": 0.1157, "step": 560 }, { "epoch": 9.28, "grad_norm": 0.2963043749332428, "learning_rate": 4.886189349923992e-05, "loss": 0.073, "step": 580 }, { "epoch": 9.6, "grad_norm": 0.45336633920669556, "learning_rate": 4.868791769198995e-05, "loss": 0.0908, "step": 600 }, { "epoch": 9.92, "grad_norm": 1.0918829441070557, "learning_rate": 4.8501928652845854e-05, "loss": 0.0557, "step": 620 }, { "epoch": 10.24, "grad_norm": 2.7920358180999756, "learning_rate": 4.83040207053985e-05, "loss": 0.0779, "step": 640 }, { "epoch": 10.56, "grad_norm": 0.8492644429206848, "learning_rate": 4.809429421786502e-05, "loss": 0.048, "step": 660 }, { "epoch": 10.88, "grad_norm": 0.6128495335578918, "learning_rate": 4.787285555218748e-05, "loss": 0.0747, "step": 680 }, { "epoch": 11.2, "grad_norm": 0.5186921954154968, "learning_rate": 4.763981701009184e-05, "loss": 0.0629, "step": 700 }, { "epoch": 11.52, "grad_norm": 1.6753857135772705, "learning_rate": 4.739529677613456e-05, "loss": 0.051, "step": 720 }, { "epoch": 11.84, "grad_norm": 0.18876530230045319, "learning_rate": 4.713941885776586e-05, "loss": 0.0699, "step": 740 }, { "epoch": 12.16, "grad_norm": 0.4428744912147522, "learning_rate": 4.687231302243975e-05, "loss": 0.0526, "step": 760 }, { "epoch": 12.48, "grad_norm": 0.4765178859233856, "learning_rate": 4.659411473180304e-05, "loss": 0.0412, "step": 780 }, { "epoch": 12.8, "grad_norm": 1.4559166431427002, "learning_rate": 4.6304965072996495e-05, "loss": 0.0495, "step": 800 }, { "epoch": 13.12, "grad_norm": 1.1842377185821533, "learning_rate": 4.6005010687103076e-05, "loss": 0.063, "step": 820 }, { "epoch": 13.44, "grad_norm": 0.5502442717552185, "learning_rate": 4.569440369477951e-05, "loss": 0.0425, "step": 840 }, { "epoch": 13.76, "grad_norm": 1.819698452949524, "learning_rate": 4.5373301619108854e-05, "loss": 0.0451, "step": 860 }, { "epoch": 14.08, "grad_norm": 0.045023053884506226, "learning_rate": 4.5041867305713384e-05, "loss": 0.0445, "step": 880 }, { "epoch": 14.4, "grad_norm": 0.10354474931955338, "learning_rate": 4.4700268840168045e-05, "loss": 0.0214, "step": 900 }, { "epoch": 14.72, "grad_norm": 0.23762211203575134, "learning_rate": 4.4348679462756556e-05, "loss": 0.0552, "step": 920 }, { "epoch": 15.04, "grad_norm": 0.845052182674408, "learning_rate": 4.398727748061324e-05, "loss": 0.0524, "step": 940 }, { "epoch": 15.36, "grad_norm": 0.29455024003982544, "learning_rate": 4.361624617729536e-05, "loss": 0.0318, "step": 960 }, { "epoch": 15.68, "grad_norm": 0.05131356045603752, "learning_rate": 4.323577371983155e-05, "loss": 0.0347, "step": 980 }, { "epoch": 16.0, "grad_norm": 0.2307986468076706, "learning_rate": 4.28460530632937e-05, "loss": 0.0541, "step": 1000 }, { "epoch": 16.32, "grad_norm": 0.9982470870018005, "learning_rate": 4.2447281852940525e-05, "loss": 0.0327, "step": 1020 }, { "epoch": 16.64, "grad_norm": 2.9247119426727295, "learning_rate": 4.203966232398261e-05, "loss": 0.0251, "step": 1040 }, { "epoch": 16.96, "grad_norm": 0.5199835896492004, "learning_rate": 4.162340119901961e-05, "loss": 0.0451, "step": 1060 }, { "epoch": 17.28, "grad_norm": 0.6802399754524231, "learning_rate": 4.1198709583201754e-05, "loss": 0.0272, "step": 1080 }, { "epoch": 17.6, "grad_norm": 0.9806874394416809, "learning_rate": 4.0765802857168687e-05, "loss": 0.0517, "step": 1100 }, { "epoch": 17.92, "grad_norm": 0.12717130780220032, "learning_rate": 4.0324900567820046e-05, "loss": 0.0286, "step": 1120 }, { "epoch": 18.24, "grad_norm": 0.4048568308353424, "learning_rate": 3.987622631697316e-05, "loss": 0.0258, "step": 1140 }, { "epoch": 18.56, "grad_norm": 1.2760275602340698, "learning_rate": 3.942000764796427e-05, "loss": 0.0289, "step": 1160 }, { "epoch": 18.88, "grad_norm": 0.30493712425231934, "learning_rate": 3.895647593025088e-05, "loss": 0.0457, "step": 1180 }, { "epoch": 19.2, "grad_norm": 0.6177706122398376, "learning_rate": 3.8485866242073584e-05, "loss": 0.0316, "step": 1200 }, { "epoch": 19.52, "grad_norm": 0.10721703618764877, "learning_rate": 3.80084172512372e-05, "loss": 0.0326, "step": 1220 }, { "epoch": 19.84, "grad_norm": 0.06823062896728516, "learning_rate": 3.7524371094071266e-05, "loss": 0.0238, "step": 1240 }, { "epoch": 20.16, "grad_norm": 0.9792996644973755, "learning_rate": 3.703397325263162e-05, "loss": 0.0286, "step": 1260 }, { "epoch": 20.48, "grad_norm": 0.20131415128707886, "learning_rate": 3.653747243020515e-05, "loss": 0.0294, "step": 1280 }, { "epoch": 20.8, "grad_norm": 0.530823826789856, "learning_rate": 3.603512042518093e-05, "loss": 0.0364, "step": 1300 }, { "epoch": 21.12, "grad_norm": 0.006112121045589447, "learning_rate": 3.552717200335171e-05, "loss": 0.0265, "step": 1320 }, { "epoch": 21.44, "grad_norm": 0.005943021737039089, "learning_rate": 3.501388476871039e-05, "loss": 0.0319, "step": 1340 }, { "epoch": 21.76, "grad_norm": 0.07666248083114624, "learning_rate": 3.449551903280729e-05, "loss": 0.0137, "step": 1360 }, { "epoch": 22.08, "grad_norm": 0.6763716340065002, "learning_rate": 3.397233768273415e-05, "loss": 0.0416, "step": 1380 }, { "epoch": 22.4, "grad_norm": 0.07458912581205368, "learning_rate": 3.344460604780202e-05, "loss": 0.0179, "step": 1400 }, { "epoch": 22.72, "grad_norm": 0.25514400005340576, "learning_rate": 3.291259176498052e-05, "loss": 0.0276, "step": 1420 }, { "epoch": 23.04, "grad_norm": 0.006252670660614967, "learning_rate": 3.237656464316693e-05, "loss": 0.0352, "step": 1440 }, { "epoch": 23.36, "grad_norm": 0.36337536573410034, "learning_rate": 3.183679652635357e-05, "loss": 0.0212, "step": 1460 }, { "epoch": 23.68, "grad_norm": 0.28176209330558777, "learning_rate": 3.129356115576332e-05, "loss": 0.0338, "step": 1480 }, { "epoch": 24.0, "grad_norm": 0.012418941594660282, "learning_rate": 3.074713403102284e-05, "loss": 0.0295, "step": 1500 }, { "epoch": 24.32, "grad_norm": 0.16376622021198273, "learning_rate": 3.0197792270443982e-05, "loss": 0.0185, "step": 1520 }, { "epoch": 24.64, "grad_norm": 0.45608577132225037, "learning_rate": 2.9645814470484452e-05, "loss": 0.0328, "step": 1540 }, { "epoch": 24.96, "grad_norm": 0.1006656065583229, "learning_rate": 2.9091480564458666e-05, "loss": 0.025, "step": 1560 }, { "epoch": 25.28, "grad_norm": 0.3901682496070862, "learning_rate": 2.8535071680570734e-05, "loss": 0.0294, "step": 1580 }, { "epoch": 25.6, "grad_norm": 0.12243347615003586, "learning_rate": 2.7976869999341426e-05, "loss": 0.0282, "step": 1600 }, { "epoch": 25.92, "grad_norm": 0.0033943182788789272, "learning_rate": 2.741715861050143e-05, "loss": 0.0294, "step": 1620 }, { "epoch": 26.24, "grad_norm": 0.0014275741996243596, "learning_rate": 2.685622136942359e-05, "loss": 0.0354, "step": 1640 }, { "epoch": 26.56, "grad_norm": 0.018641650676727295, "learning_rate": 2.629434275316673e-05, "loss": 0.0162, "step": 1660 }, { "epoch": 26.88, "grad_norm": 0.13316482305526733, "learning_rate": 2.573180771620432e-05, "loss": 0.0205, "step": 1680 }, { "epoch": 27.2, "grad_norm": 0.003175324061885476, "learning_rate": 2.516890154591095e-05, "loss": 0.0129, "step": 1700 }, { "epoch": 27.52, "grad_norm": 0.28820428252220154, "learning_rate": 2.4605909717879964e-05, "loss": 0.0333, "step": 1720 }, { "epoch": 27.84, "grad_norm": 0.08302447199821472, "learning_rate": 2.4043117751145694e-05, "loss": 0.0261, "step": 1740 }, { "epoch": 28.16, "grad_norm": 0.37718892097473145, "learning_rate": 2.34808110633836e-05, "loss": 0.0418, "step": 1760 }, { "epoch": 28.48, "grad_norm": 0.16185913980007172, "learning_rate": 2.291927482616191e-05, "loss": 0.0111, "step": 1780 }, { "epoch": 28.8, "grad_norm": 0.33071696758270264, "learning_rate": 2.235879382031794e-05, "loss": 0.0263, "step": 1800 }, { "epoch": 29.12, "grad_norm": 0.002949915360659361, "learning_rate": 2.179965229153265e-05, "loss": 0.0299, "step": 1820 }, { "epoch": 29.44, "grad_norm": 0.0035819699987769127, "learning_rate": 2.1242133806176667e-05, "loss": 0.0267, "step": 1840 }, { "epoch": 29.76, "grad_norm": 0.0039703804068267345, "learning_rate": 2.0686521107500638e-05, "loss": 0.0204, "step": 1860 }, { "epoch": 30.08, "grad_norm": 0.004021900240331888, "learning_rate": 2.0133095972243233e-05, "loss": 0.0308, "step": 1880 }, { "epoch": 30.4, "grad_norm": 0.2821226418018341, "learning_rate": 1.9582139067729117e-05, "loss": 0.024, "step": 1900 }, { "epoch": 30.72, "grad_norm": 0.5529562830924988, "learning_rate": 1.90339298095297e-05, "loss": 0.0113, "step": 1920 }, { "epoch": 31.04, "grad_norm": 0.1492016613483429, "learning_rate": 1.8488746219758674e-05, "loss": 0.0301, "step": 1940 }, { "epoch": 31.36, "grad_norm": 0.19194553792476654, "learning_rate": 1.7946864786074165e-05, "loss": 0.0293, "step": 1960 }, { "epoch": 31.68, "grad_norm": 0.1448647677898407, "learning_rate": 1.740856032145917e-05, "loss": 0.0242, "step": 1980 }, { "epoch": 32.0, "grad_norm": 0.06869282573461533, "learning_rate": 1.6874105824851267e-05, "loss": 0.022, "step": 2000 }, { "epoch": 32.32, "grad_norm": 0.002635813085362315, "learning_rate": 1.634377234269226e-05, "loss": 0.0264, "step": 2020 }, { "epoch": 32.64, "grad_norm": 0.0027471587527543306, "learning_rate": 1.5817828831468144e-05, "loss": 0.0155, "step": 2040 }, { "epoch": 32.96, "grad_norm": 0.28300318121910095, "learning_rate": 1.5296542021308825e-05, "loss": 0.0208, "step": 2060 }, { "epoch": 33.28, "grad_norm": 0.0016769981011748314, "learning_rate": 1.478017628071706e-05, "loss": 0.0264, "step": 2080 }, { "epoch": 33.6, "grad_norm": 0.0014146752655506134, "learning_rate": 1.4268993482495055e-05, "loss": 0.0174, "step": 2100 }, { "epoch": 33.92, "grad_norm": 0.05964767187833786, "learning_rate": 1.3763252870936649e-05, "loss": 0.0214, "step": 2120 }, { "epoch": 34.24, "grad_norm": 0.0016291196225211024, "learning_rate": 1.3263210930352737e-05, "loss": 0.0306, "step": 2140 }, { "epoch": 34.56, "grad_norm": 0.034806057810783386, "learning_rate": 1.2769121254996159e-05, "loss": 0.0146, "step": 2160 }, { "epoch": 34.88, "grad_norm": 0.315729558467865, "learning_rate": 1.228123442045249e-05, "loss": 0.0254, "step": 2180 }, { "epoch": 35.2, "grad_norm": 0.5809018015861511, "learning_rate": 1.1799797856561606e-05, "loss": 0.0176, "step": 2200 }, { "epoch": 35.52, "grad_norm": 0.18777510523796082, "learning_rate": 1.1325055721934637e-05, "loss": 0.0205, "step": 2220 }, { "epoch": 35.84, "grad_norm": 0.11027589440345764, "learning_rate": 1.0857248780129928e-05, "loss": 0.0153, "step": 2240 }, { "epoch": 36.16, "grad_norm": 0.18985402584075928, "learning_rate": 1.0396614277550752e-05, "loss": 0.0251, "step": 2260 }, { "epoch": 36.48, "grad_norm": 0.011078303679823875, "learning_rate": 9.943385823126775e-06, "loss": 0.0224, "step": 2280 }, { "epoch": 36.8, "grad_norm": 0.15364831686019897, "learning_rate": 9.497793269840211e-06, "loss": 0.0219, "step": 2300 }, { "epoch": 37.12, "grad_norm": 0.16373781859874725, "learning_rate": 9.06006259815683e-06, "loss": 0.021, "step": 2320 }, { "epoch": 37.44, "grad_norm": 0.41937291622161865, "learning_rate": 8.630415801420835e-06, "loss": 0.0236, "step": 2340 }, { "epoch": 37.76, "grad_norm": 0.4237123727798462, "learning_rate": 8.209070773271894e-06, "loss": 0.0216, "step": 2360 }, { "epoch": 38.08, "grad_norm": 0.08532612025737762, "learning_rate": 7.79624119714121e-06, "loss": 0.0268, "step": 2380 }, { "epoch": 38.4, "grad_norm": 0.2872686982154846, "learning_rate": 7.392136437882855e-06, "loss": 0.0306, "step": 2400 }, { "epoch": 38.72, "grad_norm": 0.2834513485431671, "learning_rate": 6.996961435595223e-06, "loss": 0.0224, "step": 2420 }, { "epoch": 39.04, "grad_norm": 0.0006759735988453031, "learning_rate": 6.610916601686481e-06, "loss": 0.0138, "step": 2440 }, { "epoch": 39.36, "grad_norm": 0.24796371161937714, "learning_rate": 6.234197717236742e-06, "loss": 0.0234, "step": 2460 }, { "epoch": 39.68, "grad_norm": 0.2657662332057953, "learning_rate": 5.866995833708464e-06, "loss": 0.0164, "step": 2480 }, { "epoch": 40.0, "grad_norm": 0.0017559522530063987, "learning_rate": 5.509497176055492e-06, "loss": 0.0169, "step": 2500 }, { "epoch": 40.32, "grad_norm": 0.18896016478538513, "learning_rate": 5.161883048279817e-06, "loss": 0.0116, "step": 2520 }, { "epoch": 40.64, "grad_norm": 0.2336779683828354, "learning_rate": 4.824329741483949e-06, "loss": 0.0219, "step": 2540 }, { "epoch": 40.96, "grad_norm": 0.16521525382995605, "learning_rate": 4.497008444465681e-06, "loss": 0.0366, "step": 2560 }, { "epoch": 41.28, "grad_norm": 0.0013411182444542646, "learning_rate": 4.180085156900274e-06, "loss": 0.0228, "step": 2580 }, { "epoch": 41.6, "grad_norm": 0.0015802403213456273, "learning_rate": 3.873720605154468e-06, "loss": 0.0135, "step": 2600 }, { "epoch": 41.92, "grad_norm": 0.002210975391790271, "learning_rate": 3.578070160774724e-06, "loss": 0.0267, "step": 2620 }, { "epoch": 42.24, "grad_norm": 0.23047080636024475, "learning_rate": 3.293283761691182e-06, "loss": 0.0218, "step": 2640 }, { "epoch": 42.56, "grad_norm": 0.2625073492527008, "learning_rate": 3.0195058361772277e-06, "loss": 0.0315, "step": 2660 }, { "epoch": 42.88, "grad_norm": 0.22485554218292236, "learning_rate": 2.756875229603295e-06, "loss": 0.022, "step": 2680 }, { "epoch": 43.2, "grad_norm": 0.001550053246319294, "learning_rate": 2.5055251340219855e-06, "loss": 0.016, "step": 2700 }, { "epoch": 43.52, "grad_norm": 0.22042883932590485, "learning_rate": 2.2655830206202655e-06, "loss": 0.0267, "step": 2720 }, { "epoch": 43.84, "grad_norm": 0.08822102099657059, "learning_rate": 2.037170575072944e-06, "loss": 0.0167, "step": 2740 }, { "epoch": 44.16, "grad_norm": 0.0016505387611687183, "learning_rate": 1.8204036358303173e-06, "loss": 0.0492, "step": 2760 }, { "epoch": 44.48, "grad_norm": 0.3908143639564514, "learning_rate": 1.615392135371116e-06, "loss": 0.0254, "step": 2780 }, { "epoch": 44.8, "grad_norm": 0.3156016767024994, "learning_rate": 1.4222400444507318e-06, "loss": 0.0136, "step": 2800 }, { "epoch": 45.12, "grad_norm": 0.001897096517495811, "learning_rate": 1.2410453193728493e-06, "loss": 0.0116, "step": 2820 }, { "epoch": 45.44, "grad_norm": 0.20051489770412445, "learning_rate": 1.0718998523113004e-06, "loss": 0.0311, "step": 2840 }, { "epoch": 45.76, "grad_norm": 0.0016040581976994872, "learning_rate": 9.148894247073298e-07, "loss": 0.0283, "step": 2860 }, { "epoch": 46.08, "grad_norm": 0.0015667045954614878, "learning_rate": 7.700936637658779e-07, "loss": 0.0186, "step": 2880 }, { "epoch": 46.4, "grad_norm": 0.04447433352470398, "learning_rate": 6.375860020729541e-07, "loss": 0.0229, "step": 2900 }, { "epoch": 46.72, "grad_norm": 0.20941714942455292, "learning_rate": 5.174336403546226e-07, "loss": 0.0268, "step": 2920 }, { "epoch": 47.04, "grad_norm": 0.0015096565475687385, "learning_rate": 4.096975133963954e-07, "loss": 0.0161, "step": 2940 }, { "epoch": 47.36, "grad_norm": 0.19307351112365723, "learning_rate": 3.144322591404292e-07, "loss": 0.0113, "step": 2960 }, { "epoch": 47.68, "grad_norm": 0.002177381655201316, "learning_rate": 2.316861909760909e-07, "loss": 0.0222, "step": 2980 }, { "epoch": 48.0, "grad_norm": 0.09334852546453476, "learning_rate": 1.6150127323803222e-07, "loss": 0.028, "step": 3000 }, { "epoch": 48.32, "grad_norm": 0.09507758915424347, "learning_rate": 1.0391309992413833e-07, "loss": 0.0217, "step": 3020 }, { "epoch": 48.64, "grad_norm": 0.0010601489339023829, "learning_rate": 5.895087664417876e-08, "loss": 0.0191, "step": 3040 }, { "epoch": 48.96, "grad_norm": 0.001158875529654324, "learning_rate": 2.6637405808302428e-08, "loss": 0.0162, "step": 3060 }, { "epoch": 49.28, "grad_norm": 0.0015347091248258948, "learning_rate": 6.989075062879824e-09, "loss": 0.0146, "step": 3080 }, { "epoch": 49.6, "grad_norm": 0.0010455228621140122, "learning_rate": 1.584897958428755e-11, "loss": 0.0217, "step": 3100 }, { "epoch": 49.6, "step": 3100, "total_flos": 1.7939255986343117e+17, "train_loss": 0.172668604437382, "train_runtime": 11274.7291, "train_samples_per_second": 2.217, "train_steps_per_second": 0.275 } ], "logging_steps": 20, "max_steps": 3100, "num_input_tokens_seen": 0, "num_train_epochs": 50, "save_steps": 1000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 1.7939255986343117e+17, "train_batch_size": 2, "trial_name": null, "trial_params": null }