|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 617.283950617284, |
|
"eval_steps": 500, |
|
"global_step": 10000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.06172839506172839, |
|
"learning_rate": 5e-09, |
|
"loss": 8.604, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 1.2345679012345678, |
|
"learning_rate": 1e-07, |
|
"loss": 8.0873, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 2.4691358024691357, |
|
"learning_rate": 2e-07, |
|
"loss": 7.8074, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 3.7037037037037037, |
|
"learning_rate": 3e-07, |
|
"loss": 7.0763, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 4.938271604938271, |
|
"learning_rate": 4e-07, |
|
"loss": 6.2969, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 6.172839506172839, |
|
"learning_rate": 5e-07, |
|
"loss": 5.9334, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 7.407407407407407, |
|
"learning_rate": 6e-07, |
|
"loss": 5.4919, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 8.641975308641975, |
|
"learning_rate": 7e-07, |
|
"loss": 5.0659, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 9.876543209876543, |
|
"learning_rate": 8e-07, |
|
"loss": 4.9089, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 11.11111111111111, |
|
"learning_rate": 9e-07, |
|
"loss": 4.7144, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 12.345679012345679, |
|
"learning_rate": 1e-06, |
|
"loss": 4.7373, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 13.580246913580247, |
|
"learning_rate": 9.997245010407738e-07, |
|
"loss": 4.3848, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 14.814814814814815, |
|
"learning_rate": 9.994490020815477e-07, |
|
"loss": 4.3457, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 16.049382716049383, |
|
"learning_rate": 9.991735031223215e-07, |
|
"loss": 4.2217, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 17.28395061728395, |
|
"learning_rate": 9.988980041630952e-07, |
|
"loss": 4.1273, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 18.51851851851852, |
|
"learning_rate": 9.986225052038692e-07, |
|
"loss": 4.0123, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 19.753086419753085, |
|
"learning_rate": 9.98347006244643e-07, |
|
"loss": 4.0784, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 20.987654320987655, |
|
"learning_rate": 9.98071507285417e-07, |
|
"loss": 3.9415, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 22.22222222222222, |
|
"learning_rate": 9.977960083261906e-07, |
|
"loss": 3.843, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 23.45679012345679, |
|
"learning_rate": 9.975205093669647e-07, |
|
"loss": 3.8174, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 24.691358024691358, |
|
"learning_rate": 9.972450104077383e-07, |
|
"loss": 3.7376, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 25.925925925925927, |
|
"learning_rate": 9.969695114485122e-07, |
|
"loss": 3.7876, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 27.160493827160494, |
|
"learning_rate": 9.96694012489286e-07, |
|
"loss": 3.7585, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 28.395061728395063, |
|
"learning_rate": 9.9641851353006e-07, |
|
"loss": 3.6304, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 29.62962962962963, |
|
"learning_rate": 9.961430145708338e-07, |
|
"loss": 3.6751, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 30.864197530864196, |
|
"learning_rate": 9.958675156116076e-07, |
|
"loss": 3.677, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 32.098765432098766, |
|
"learning_rate": 9.955920166523815e-07, |
|
"loss": 3.6556, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 33.333333333333336, |
|
"learning_rate": 9.953165176931551e-07, |
|
"loss": 3.5897, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 34.5679012345679, |
|
"learning_rate": 9.950410187339292e-07, |
|
"loss": 3.7304, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 35.80246913580247, |
|
"learning_rate": 9.94765519774703e-07, |
|
"loss": 3.568, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 37.03703703703704, |
|
"learning_rate": 9.94490020815477e-07, |
|
"loss": 3.6207, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 38.27160493827161, |
|
"learning_rate": 9.942145218562508e-07, |
|
"loss": 3.4759, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 39.50617283950617, |
|
"learning_rate": 9.939390228970246e-07, |
|
"loss": 3.4513, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 40.74074074074074, |
|
"learning_rate": 9.936635239377985e-07, |
|
"loss": 3.372, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 41.97530864197531, |
|
"learning_rate": 9.933880249785724e-07, |
|
"loss": 3.2601, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 43.20987654320987, |
|
"learning_rate": 9.93112526019346e-07, |
|
"loss": 3.3071, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 44.44444444444444, |
|
"learning_rate": 9.9283702706012e-07, |
|
"loss": 3.2768, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 45.67901234567901, |
|
"learning_rate": 9.925615281008937e-07, |
|
"loss": 3.3103, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 46.91358024691358, |
|
"learning_rate": 9.922860291416678e-07, |
|
"loss": 3.2807, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 48.148148148148145, |
|
"learning_rate": 9.920105301824414e-07, |
|
"loss": 3.1839, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 49.382716049382715, |
|
"learning_rate": 9.917350312232155e-07, |
|
"loss": 3.1689, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 50.617283950617285, |
|
"learning_rate": 9.914595322639892e-07, |
|
"loss": 3.17, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 51.851851851851855, |
|
"learning_rate": 9.91184033304763e-07, |
|
"loss": 3.1598, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 53.08641975308642, |
|
"learning_rate": 9.909085343455369e-07, |
|
"loss": 3.1215, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 54.32098765432099, |
|
"learning_rate": 9.906330353863107e-07, |
|
"loss": 3.102, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 55.55555555555556, |
|
"learning_rate": 9.903575364270846e-07, |
|
"loss": 3.0819, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 56.79012345679013, |
|
"learning_rate": 9.900820374678584e-07, |
|
"loss": 3.0729, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 58.02469135802469, |
|
"learning_rate": 9.898065385086323e-07, |
|
"loss": 3.0639, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 59.25925925925926, |
|
"learning_rate": 9.89531039549406e-07, |
|
"loss": 3.0172, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 60.49382716049383, |
|
"learning_rate": 9.8925554059018e-07, |
|
"loss": 3.0463, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 61.72839506172839, |
|
"learning_rate": 9.889800416309537e-07, |
|
"loss": 2.9424, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 62.96296296296296, |
|
"learning_rate": 9.887045426717277e-07, |
|
"loss": 3.0018, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 64.19753086419753, |
|
"learning_rate": 9.884290437125014e-07, |
|
"loss": 3.0268, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 65.4320987654321, |
|
"learning_rate": 9.881535447532755e-07, |
|
"loss": 3.0157, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 66.66666666666667, |
|
"learning_rate": 9.878780457940491e-07, |
|
"loss": 2.9853, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 67.90123456790124, |
|
"learning_rate": 9.87602546834823e-07, |
|
"loss": 2.8881, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 69.1358024691358, |
|
"learning_rate": 9.873270478755968e-07, |
|
"loss": 2.965, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 70.37037037037037, |
|
"learning_rate": 9.870515489163707e-07, |
|
"loss": 2.9127, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 71.60493827160494, |
|
"learning_rate": 9.867760499571445e-07, |
|
"loss": 2.8881, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 72.8395061728395, |
|
"learning_rate": 9.865005509979184e-07, |
|
"loss": 2.8964, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 74.07407407407408, |
|
"learning_rate": 9.862250520386923e-07, |
|
"loss": 2.932, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 75.30864197530865, |
|
"learning_rate": 9.85949553079466e-07, |
|
"loss": 2.9062, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 76.54320987654322, |
|
"learning_rate": 9.8567405412024e-07, |
|
"loss": 2.9141, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 77.77777777777777, |
|
"learning_rate": 9.853985551610136e-07, |
|
"loss": 2.8072, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 79.01234567901234, |
|
"learning_rate": 9.851230562017877e-07, |
|
"loss": 2.8403, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 80.24691358024691, |
|
"learning_rate": 9.848475572425613e-07, |
|
"loss": 2.8484, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 81.48148148148148, |
|
"learning_rate": 9.845720582833354e-07, |
|
"loss": 2.8565, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 82.71604938271605, |
|
"learning_rate": 9.842965593241093e-07, |
|
"loss": 2.8241, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 83.95061728395062, |
|
"learning_rate": 9.840210603648831e-07, |
|
"loss": 2.8383, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 85.18518518518519, |
|
"learning_rate": 9.837455614056568e-07, |
|
"loss": 2.8188, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 86.41975308641975, |
|
"learning_rate": 9.834700624464309e-07, |
|
"loss": 2.8681, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 87.65432098765432, |
|
"learning_rate": 9.831945634872045e-07, |
|
"loss": 2.7453, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 88.88888888888889, |
|
"learning_rate": 9.829190645279786e-07, |
|
"loss": 2.7488, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 90.12345679012346, |
|
"learning_rate": 9.826435655687522e-07, |
|
"loss": 2.7859, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 91.35802469135803, |
|
"learning_rate": 9.82368066609526e-07, |
|
"loss": 2.8079, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 92.5925925925926, |
|
"learning_rate": 9.820925676503e-07, |
|
"loss": 2.717, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 93.82716049382717, |
|
"learning_rate": 9.818170686910738e-07, |
|
"loss": 2.7502, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 95.06172839506173, |
|
"learning_rate": 9.815415697318477e-07, |
|
"loss": 2.7395, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 96.29629629629629, |
|
"learning_rate": 9.812660707726215e-07, |
|
"loss": 2.7172, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 97.53086419753086, |
|
"learning_rate": 9.809905718133954e-07, |
|
"loss": 2.6894, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 98.76543209876543, |
|
"learning_rate": 9.807150728541692e-07, |
|
"loss": 2.7527, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 100.0, |
|
"learning_rate": 9.80439573894943e-07, |
|
"loss": 2.7461, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 101.23456790123457, |
|
"learning_rate": 9.801640749357167e-07, |
|
"loss": 2.782, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 102.46913580246914, |
|
"learning_rate": 9.798885759764908e-07, |
|
"loss": 2.7753, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 103.70370370370371, |
|
"learning_rate": 9.796130770172645e-07, |
|
"loss": 2.7902, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 104.93827160493827, |
|
"learning_rate": 9.793375780580385e-07, |
|
"loss": 2.7305, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 106.17283950617283, |
|
"learning_rate": 9.790620790988122e-07, |
|
"loss": 2.7476, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 107.4074074074074, |
|
"learning_rate": 9.787865801395862e-07, |
|
"loss": 2.715, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 108.64197530864197, |
|
"learning_rate": 9.785110811803599e-07, |
|
"loss": 2.6883, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 109.87654320987654, |
|
"learning_rate": 9.782355822211337e-07, |
|
"loss": 2.6921, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 111.11111111111111, |
|
"learning_rate": 9.779600832619076e-07, |
|
"loss": 2.7132, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 112.34567901234568, |
|
"learning_rate": 9.776845843026815e-07, |
|
"loss": 2.6779, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 113.58024691358025, |
|
"learning_rate": 9.774090853434553e-07, |
|
"loss": 2.6119, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 114.81481481481481, |
|
"learning_rate": 9.771335863842292e-07, |
|
"loss": 2.6716, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 116.04938271604938, |
|
"learning_rate": 9.76858087425003e-07, |
|
"loss": 2.6666, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 117.28395061728395, |
|
"learning_rate": 9.765825884657767e-07, |
|
"loss": 2.6476, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 118.51851851851852, |
|
"learning_rate": 9.763070895065508e-07, |
|
"loss": 2.6393, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 119.75308641975309, |
|
"learning_rate": 9.760315905473244e-07, |
|
"loss": 2.627, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 120.98765432098766, |
|
"learning_rate": 9.757560915880985e-07, |
|
"loss": 2.6067, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 122.22222222222223, |
|
"learning_rate": 9.754805926288721e-07, |
|
"loss": 2.6015, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 123.45679012345678, |
|
"learning_rate": 9.752050936696462e-07, |
|
"loss": 2.6428, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 124.69135802469135, |
|
"learning_rate": 9.749295947104198e-07, |
|
"loss": 2.6251, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 125.92592592592592, |
|
"learning_rate": 9.746540957511937e-07, |
|
"loss": 2.655, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 127.1604938271605, |
|
"learning_rate": 9.743785967919676e-07, |
|
"loss": 2.5351, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 128.39506172839506, |
|
"learning_rate": 9.741030978327416e-07, |
|
"loss": 2.582, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 129.62962962962962, |
|
"learning_rate": 9.738275988735153e-07, |
|
"loss": 2.5874, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 130.8641975308642, |
|
"learning_rate": 9.735520999142893e-07, |
|
"loss": 2.5603, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 132.09876543209876, |
|
"learning_rate": 9.73276600955063e-07, |
|
"loss": 2.5398, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 133.33333333333334, |
|
"learning_rate": 9.730011019958369e-07, |
|
"loss": 2.5265, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 134.5679012345679, |
|
"learning_rate": 9.727256030366107e-07, |
|
"loss": 2.6271, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 135.80246913580248, |
|
"learning_rate": 9.724501040773846e-07, |
|
"loss": 2.5626, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 137.03703703703704, |
|
"learning_rate": 9.721746051181584e-07, |
|
"loss": 2.5425, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 138.2716049382716, |
|
"learning_rate": 9.718991061589323e-07, |
|
"loss": 2.5251, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 139.50617283950618, |
|
"learning_rate": 9.716236071997061e-07, |
|
"loss": 2.5817, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 140.74074074074073, |
|
"learning_rate": 9.7134810824048e-07, |
|
"loss": 2.4888, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 141.97530864197532, |
|
"learning_rate": 9.710726092812539e-07, |
|
"loss": 2.5426, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 143.20987654320987, |
|
"learning_rate": 9.707971103220275e-07, |
|
"loss": 2.5293, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 144.44444444444446, |
|
"learning_rate": 9.705216113628016e-07, |
|
"loss": 2.4874, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 145.679012345679, |
|
"learning_rate": 9.702461124035752e-07, |
|
"loss": 2.5487, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 146.91358024691357, |
|
"learning_rate": 9.699706134443493e-07, |
|
"loss": 2.5153, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 148.14814814814815, |
|
"learning_rate": 9.69695114485123e-07, |
|
"loss": 2.4468, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 149.3827160493827, |
|
"learning_rate": 9.69419615525897e-07, |
|
"loss": 2.4731, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 150.6172839506173, |
|
"learning_rate": 9.691441165666707e-07, |
|
"loss": 2.5007, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 151.85185185185185, |
|
"learning_rate": 9.688686176074445e-07, |
|
"loss": 2.4948, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 153.08641975308643, |
|
"learning_rate": 9.685931186482184e-07, |
|
"loss": 2.4867, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 154.320987654321, |
|
"learning_rate": 9.683176196889922e-07, |
|
"loss": 2.4654, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 155.55555555555554, |
|
"learning_rate": 9.68042120729766e-07, |
|
"loss": 2.4621, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 156.79012345679013, |
|
"learning_rate": 9.6776662177054e-07, |
|
"loss": 2.4738, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 158.02469135802468, |
|
"learning_rate": 9.674911228113138e-07, |
|
"loss": 2.4562, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 159.25925925925927, |
|
"learning_rate": 9.672156238520875e-07, |
|
"loss": 2.4937, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 160.49382716049382, |
|
"learning_rate": 9.669401248928615e-07, |
|
"loss": 2.4574, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 161.7283950617284, |
|
"learning_rate": 9.666646259336352e-07, |
|
"loss": 2.4642, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 162.96296296296296, |
|
"learning_rate": 9.663891269744093e-07, |
|
"loss": 2.4736, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 164.19753086419752, |
|
"learning_rate": 9.66113628015183e-07, |
|
"loss": 2.4926, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 165.4320987654321, |
|
"learning_rate": 9.65838129055957e-07, |
|
"loss": 2.4323, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 166.66666666666666, |
|
"learning_rate": 9.655626300967306e-07, |
|
"loss": 2.4508, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 167.90123456790124, |
|
"learning_rate": 9.652871311375045e-07, |
|
"loss": 2.4295, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 169.1358024691358, |
|
"learning_rate": 9.650116321782783e-07, |
|
"loss": 2.4753, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 170.37037037037038, |
|
"learning_rate": 9.647361332190522e-07, |
|
"loss": 2.4323, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 171.60493827160494, |
|
"learning_rate": 9.64460634259826e-07, |
|
"loss": 2.392, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 172.8395061728395, |
|
"learning_rate": 9.641851353006e-07, |
|
"loss": 2.4579, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 174.07407407407408, |
|
"learning_rate": 9.639096363413738e-07, |
|
"loss": 2.4091, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 175.30864197530863, |
|
"learning_rate": 9.636341373821476e-07, |
|
"loss": 2.3802, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 176.54320987654322, |
|
"learning_rate": 9.633586384229215e-07, |
|
"loss": 2.3483, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 177.77777777777777, |
|
"learning_rate": 9.630831394636953e-07, |
|
"loss": 2.374, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 179.01234567901236, |
|
"learning_rate": 9.628076405044692e-07, |
|
"loss": 2.414, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 180.2469135802469, |
|
"learning_rate": 9.62532141545243e-07, |
|
"loss": 2.3971, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 181.4814814814815, |
|
"learning_rate": 9.62256642586017e-07, |
|
"loss": 2.3508, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 182.71604938271605, |
|
"learning_rate": 9.619811436267908e-07, |
|
"loss": 2.4072, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 183.9506172839506, |
|
"learning_rate": 9.617056446675646e-07, |
|
"loss": 2.3853, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 185.1851851851852, |
|
"learning_rate": 9.614301457083383e-07, |
|
"loss": 2.4256, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 186.41975308641975, |
|
"learning_rate": 9.611546467491124e-07, |
|
"loss": 2.3795, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 187.65432098765433, |
|
"learning_rate": 9.60879147789886e-07, |
|
"loss": 2.3763, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 188.88888888888889, |
|
"learning_rate": 9.6060364883066e-07, |
|
"loss": 2.3474, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 190.12345679012347, |
|
"learning_rate": 9.603281498714337e-07, |
|
"loss": 2.4074, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 191.35802469135803, |
|
"learning_rate": 9.600526509122076e-07, |
|
"loss": 2.3438, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 192.59259259259258, |
|
"learning_rate": 9.597771519529814e-07, |
|
"loss": 2.3209, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 193.82716049382717, |
|
"learning_rate": 9.595016529937553e-07, |
|
"loss": 2.3364, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 195.06172839506172, |
|
"learning_rate": 9.592261540345292e-07, |
|
"loss": 2.3624, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 196.2962962962963, |
|
"learning_rate": 9.58950655075303e-07, |
|
"loss": 2.3004, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 197.53086419753086, |
|
"learning_rate": 9.586751561160769e-07, |
|
"loss": 2.2904, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 198.76543209876544, |
|
"learning_rate": 9.583996571568507e-07, |
|
"loss": 2.2971, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 200.0, |
|
"learning_rate": 9.581241581976246e-07, |
|
"loss": 2.3364, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 201.23456790123456, |
|
"learning_rate": 9.578486592383982e-07, |
|
"loss": 2.3191, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 202.46913580246914, |
|
"learning_rate": 9.575731602791723e-07, |
|
"loss": 2.3663, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 203.7037037037037, |
|
"learning_rate": 9.57297661319946e-07, |
|
"loss": 2.3649, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 204.93827160493828, |
|
"learning_rate": 9.5702216236072e-07, |
|
"loss": 2.2762, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 206.17283950617283, |
|
"learning_rate": 9.567466634014937e-07, |
|
"loss": 2.315, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 207.40740740740742, |
|
"learning_rate": 9.564711644422677e-07, |
|
"loss": 2.2924, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 208.64197530864197, |
|
"learning_rate": 9.561956654830414e-07, |
|
"loss": 2.2383, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 209.87654320987653, |
|
"learning_rate": 9.559201665238153e-07, |
|
"loss": 2.2903, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 211.11111111111111, |
|
"learning_rate": 9.556446675645891e-07, |
|
"loss": 2.3423, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 212.34567901234567, |
|
"learning_rate": 9.55369168605363e-07, |
|
"loss": 2.2747, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 213.58024691358025, |
|
"learning_rate": 9.550936696461368e-07, |
|
"loss": 2.2997, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 214.8148148148148, |
|
"learning_rate": 9.548181706869107e-07, |
|
"loss": 2.2166, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 216.0493827160494, |
|
"learning_rate": 9.545426717276845e-07, |
|
"loss": 2.349, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 217.28395061728395, |
|
"learning_rate": 9.542671727684582e-07, |
|
"loss": 2.2522, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 218.5185185185185, |
|
"learning_rate": 9.539916738092323e-07, |
|
"loss": 2.2854, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 219.7530864197531, |
|
"learning_rate": 9.537161748500061e-07, |
|
"loss": 2.2806, |
|
"step": 3560 |
|
}, |
|
{ |
|
"epoch": 220.98765432098764, |
|
"learning_rate": 9.5344067589078e-07, |
|
"loss": 2.233, |
|
"step": 3580 |
|
}, |
|
{ |
|
"epoch": 222.22222222222223, |
|
"learning_rate": 9.531651769315538e-07, |
|
"loss": 2.2588, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 223.45679012345678, |
|
"learning_rate": 9.528896779723277e-07, |
|
"loss": 2.2729, |
|
"step": 3620 |
|
}, |
|
{ |
|
"epoch": 224.69135802469137, |
|
"learning_rate": 9.526141790131015e-07, |
|
"loss": 2.2818, |
|
"step": 3640 |
|
}, |
|
{ |
|
"epoch": 225.92592592592592, |
|
"learning_rate": 9.523386800538753e-07, |
|
"loss": 2.2291, |
|
"step": 3660 |
|
}, |
|
{ |
|
"epoch": 227.1604938271605, |
|
"learning_rate": 9.520631810946491e-07, |
|
"loss": 2.2552, |
|
"step": 3680 |
|
}, |
|
{ |
|
"epoch": 228.39506172839506, |
|
"learning_rate": 9.51787682135423e-07, |
|
"loss": 2.2028, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 229.62962962962962, |
|
"learning_rate": 9.515121831761969e-07, |
|
"loss": 2.1948, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 230.8641975308642, |
|
"learning_rate": 9.512366842169707e-07, |
|
"loss": 2.2981, |
|
"step": 3740 |
|
}, |
|
{ |
|
"epoch": 232.09876543209876, |
|
"learning_rate": 9.509611852577446e-07, |
|
"loss": 2.2519, |
|
"step": 3760 |
|
}, |
|
{ |
|
"epoch": 233.33333333333334, |
|
"learning_rate": 9.506856862985184e-07, |
|
"loss": 2.2159, |
|
"step": 3780 |
|
}, |
|
{ |
|
"epoch": 234.5679012345679, |
|
"learning_rate": 9.504101873392922e-07, |
|
"loss": 2.2122, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 235.80246913580248, |
|
"learning_rate": 9.501346883800661e-07, |
|
"loss": 2.2165, |
|
"step": 3820 |
|
}, |
|
{ |
|
"epoch": 237.03703703703704, |
|
"learning_rate": 9.498591894208399e-07, |
|
"loss": 2.2362, |
|
"step": 3840 |
|
}, |
|
{ |
|
"epoch": 238.2716049382716, |
|
"learning_rate": 9.495836904616138e-07, |
|
"loss": 2.1995, |
|
"step": 3860 |
|
}, |
|
{ |
|
"epoch": 239.50617283950618, |
|
"learning_rate": 9.493081915023877e-07, |
|
"loss": 2.248, |
|
"step": 3880 |
|
}, |
|
{ |
|
"epoch": 240.74074074074073, |
|
"learning_rate": 9.490326925431615e-07, |
|
"loss": 2.1703, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 241.97530864197532, |
|
"learning_rate": 9.487571935839353e-07, |
|
"loss": 2.1987, |
|
"step": 3920 |
|
}, |
|
{ |
|
"epoch": 243.20987654320987, |
|
"learning_rate": 9.484816946247091e-07, |
|
"loss": 2.2023, |
|
"step": 3940 |
|
}, |
|
{ |
|
"epoch": 244.44444444444446, |
|
"learning_rate": 9.48206195665483e-07, |
|
"loss": 2.2292, |
|
"step": 3960 |
|
}, |
|
{ |
|
"epoch": 245.679012345679, |
|
"learning_rate": 9.479306967062568e-07, |
|
"loss": 2.1746, |
|
"step": 3980 |
|
}, |
|
{ |
|
"epoch": 246.91358024691357, |
|
"learning_rate": 9.476551977470307e-07, |
|
"loss": 2.1809, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 248.14814814814815, |
|
"learning_rate": 9.473796987878046e-07, |
|
"loss": 2.1631, |
|
"step": 4020 |
|
}, |
|
{ |
|
"epoch": 249.3827160493827, |
|
"learning_rate": 9.471041998285784e-07, |
|
"loss": 2.1437, |
|
"step": 4040 |
|
}, |
|
{ |
|
"epoch": 250.6172839506173, |
|
"learning_rate": 9.468287008693522e-07, |
|
"loss": 2.1719, |
|
"step": 4060 |
|
}, |
|
{ |
|
"epoch": 251.85185185185185, |
|
"learning_rate": 9.46553201910126e-07, |
|
"loss": 2.1754, |
|
"step": 4080 |
|
}, |
|
{ |
|
"epoch": 253.08641975308643, |
|
"learning_rate": 9.462777029508999e-07, |
|
"loss": 2.172, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 254.320987654321, |
|
"learning_rate": 9.460022039916737e-07, |
|
"loss": 2.2135, |
|
"step": 4120 |
|
}, |
|
{ |
|
"epoch": 255.55555555555554, |
|
"learning_rate": 9.457267050324476e-07, |
|
"loss": 2.1143, |
|
"step": 4140 |
|
}, |
|
{ |
|
"epoch": 256.7901234567901, |
|
"learning_rate": 9.454512060732215e-07, |
|
"loss": 2.1804, |
|
"step": 4160 |
|
}, |
|
{ |
|
"epoch": 258.0246913580247, |
|
"learning_rate": 9.451757071139952e-07, |
|
"loss": 2.147, |
|
"step": 4180 |
|
}, |
|
{ |
|
"epoch": 259.25925925925924, |
|
"learning_rate": 9.449002081547691e-07, |
|
"loss": 2.1618, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 260.4938271604938, |
|
"learning_rate": 9.446247091955429e-07, |
|
"loss": 2.1434, |
|
"step": 4220 |
|
}, |
|
{ |
|
"epoch": 261.7283950617284, |
|
"learning_rate": 9.443492102363168e-07, |
|
"loss": 2.1535, |
|
"step": 4240 |
|
}, |
|
{ |
|
"epoch": 262.962962962963, |
|
"learning_rate": 9.440737112770907e-07, |
|
"loss": 2.2116, |
|
"step": 4260 |
|
}, |
|
{ |
|
"epoch": 264.1975308641975, |
|
"learning_rate": 9.437982123178645e-07, |
|
"loss": 2.1857, |
|
"step": 4280 |
|
}, |
|
{ |
|
"epoch": 265.4320987654321, |
|
"learning_rate": 9.435227133586384e-07, |
|
"loss": 2.1337, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 266.6666666666667, |
|
"learning_rate": 9.432472143994122e-07, |
|
"loss": 2.1554, |
|
"step": 4320 |
|
}, |
|
{ |
|
"epoch": 267.9012345679012, |
|
"learning_rate": 9.429717154401861e-07, |
|
"loss": 2.1565, |
|
"step": 4340 |
|
}, |
|
{ |
|
"epoch": 269.1358024691358, |
|
"learning_rate": 9.4269621648096e-07, |
|
"loss": 2.168, |
|
"step": 4360 |
|
}, |
|
{ |
|
"epoch": 270.3703703703704, |
|
"learning_rate": 9.424207175217338e-07, |
|
"loss": 2.1392, |
|
"step": 4380 |
|
}, |
|
{ |
|
"epoch": 271.60493827160496, |
|
"learning_rate": 9.421452185625077e-07, |
|
"loss": 2.1726, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 272.8395061728395, |
|
"learning_rate": 9.418697196032815e-07, |
|
"loss": 2.135, |
|
"step": 4420 |
|
}, |
|
{ |
|
"epoch": 274.0740740740741, |
|
"learning_rate": 9.415942206440553e-07, |
|
"loss": 2.1329, |
|
"step": 4440 |
|
}, |
|
{ |
|
"epoch": 275.30864197530866, |
|
"learning_rate": 9.413187216848291e-07, |
|
"loss": 2.1334, |
|
"step": 4460 |
|
}, |
|
{ |
|
"epoch": 276.5432098765432, |
|
"learning_rate": 9.41043222725603e-07, |
|
"loss": 2.1309, |
|
"step": 4480 |
|
}, |
|
{ |
|
"epoch": 277.77777777777777, |
|
"learning_rate": 9.407677237663769e-07, |
|
"loss": 2.0872, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 279.01234567901236, |
|
"learning_rate": 9.404922248071507e-07, |
|
"loss": 2.1426, |
|
"step": 4520 |
|
}, |
|
{ |
|
"epoch": 280.24691358024694, |
|
"learning_rate": 9.402167258479246e-07, |
|
"loss": 2.1331, |
|
"step": 4540 |
|
}, |
|
{ |
|
"epoch": 281.48148148148147, |
|
"learning_rate": 9.399412268886984e-07, |
|
"loss": 2.0859, |
|
"step": 4560 |
|
}, |
|
{ |
|
"epoch": 282.71604938271605, |
|
"learning_rate": 9.396657279294723e-07, |
|
"loss": 2.0755, |
|
"step": 4580 |
|
}, |
|
{ |
|
"epoch": 283.95061728395063, |
|
"learning_rate": 9.39390228970246e-07, |
|
"loss": 2.1203, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 285.18518518518516, |
|
"learning_rate": 9.391147300110199e-07, |
|
"loss": 2.0846, |
|
"step": 4620 |
|
}, |
|
{ |
|
"epoch": 286.41975308641975, |
|
"learning_rate": 9.388392310517938e-07, |
|
"loss": 2.1284, |
|
"step": 4640 |
|
}, |
|
{ |
|
"epoch": 287.65432098765433, |
|
"learning_rate": 9.385637320925677e-07, |
|
"loss": 2.107, |
|
"step": 4660 |
|
}, |
|
{ |
|
"epoch": 288.8888888888889, |
|
"learning_rate": 9.382882331333415e-07, |
|
"loss": 2.2206, |
|
"step": 4680 |
|
}, |
|
{ |
|
"epoch": 290.12345679012344, |
|
"learning_rate": 9.380127341741153e-07, |
|
"loss": 2.2475, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 291.358024691358, |
|
"learning_rate": 9.377372352148891e-07, |
|
"loss": 2.1887, |
|
"step": 4720 |
|
}, |
|
{ |
|
"epoch": 292.5925925925926, |
|
"learning_rate": 9.37461736255663e-07, |
|
"loss": 2.1352, |
|
"step": 4740 |
|
}, |
|
{ |
|
"epoch": 293.82716049382714, |
|
"learning_rate": 9.371862372964368e-07, |
|
"loss": 2.1565, |
|
"step": 4760 |
|
}, |
|
{ |
|
"epoch": 295.0617283950617, |
|
"learning_rate": 9.369107383372107e-07, |
|
"loss": 2.1574, |
|
"step": 4780 |
|
}, |
|
{ |
|
"epoch": 296.2962962962963, |
|
"learning_rate": 9.366352393779845e-07, |
|
"loss": 2.1384, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 297.5308641975309, |
|
"learning_rate": 9.363597404187584e-07, |
|
"loss": 2.1534, |
|
"step": 4820 |
|
}, |
|
{ |
|
"epoch": 298.7654320987654, |
|
"learning_rate": 9.360842414595322e-07, |
|
"loss": 2.0941, |
|
"step": 4840 |
|
}, |
|
{ |
|
"epoch": 300.0, |
|
"learning_rate": 9.35808742500306e-07, |
|
"loss": 2.101, |
|
"step": 4860 |
|
}, |
|
{ |
|
"epoch": 301.2345679012346, |
|
"learning_rate": 9.355332435410799e-07, |
|
"loss": 2.1181, |
|
"step": 4880 |
|
}, |
|
{ |
|
"epoch": 302.4691358024691, |
|
"learning_rate": 9.352577445818537e-07, |
|
"loss": 2.0771, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 303.7037037037037, |
|
"learning_rate": 9.349822456226276e-07, |
|
"loss": 2.0886, |
|
"step": 4920 |
|
}, |
|
{ |
|
"epoch": 304.9382716049383, |
|
"learning_rate": 9.347067466634014e-07, |
|
"loss": 2.0715, |
|
"step": 4940 |
|
}, |
|
{ |
|
"epoch": 306.17283950617286, |
|
"learning_rate": 9.344312477041753e-07, |
|
"loss": 2.0883, |
|
"step": 4960 |
|
}, |
|
{ |
|
"epoch": 307.4074074074074, |
|
"learning_rate": 9.34155748744949e-07, |
|
"loss": 2.0767, |
|
"step": 4980 |
|
}, |
|
{ |
|
"epoch": 308.641975308642, |
|
"learning_rate": 9.338802497857229e-07, |
|
"loss": 2.0757, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 309.87654320987656, |
|
"learning_rate": 9.336047508264968e-07, |
|
"loss": 2.0775, |
|
"step": 5020 |
|
}, |
|
{ |
|
"epoch": 311.1111111111111, |
|
"learning_rate": 9.333292518672707e-07, |
|
"loss": 2.0763, |
|
"step": 5040 |
|
}, |
|
{ |
|
"epoch": 312.34567901234567, |
|
"learning_rate": 9.330537529080446e-07, |
|
"loss": 2.065, |
|
"step": 5060 |
|
}, |
|
{ |
|
"epoch": 313.58024691358025, |
|
"learning_rate": 9.327782539488184e-07, |
|
"loss": 2.0996, |
|
"step": 5080 |
|
}, |
|
{ |
|
"epoch": 314.81481481481484, |
|
"learning_rate": 9.325027549895923e-07, |
|
"loss": 2.0837, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 316.04938271604937, |
|
"learning_rate": 9.322272560303662e-07, |
|
"loss": 2.0985, |
|
"step": 5120 |
|
}, |
|
{ |
|
"epoch": 317.28395061728395, |
|
"learning_rate": 9.319517570711399e-07, |
|
"loss": 2.0662, |
|
"step": 5140 |
|
}, |
|
{ |
|
"epoch": 318.51851851851853, |
|
"learning_rate": 9.316762581119138e-07, |
|
"loss": 2.1054, |
|
"step": 5160 |
|
}, |
|
{ |
|
"epoch": 319.75308641975306, |
|
"learning_rate": 9.314007591526876e-07, |
|
"loss": 2.0688, |
|
"step": 5180 |
|
}, |
|
{ |
|
"epoch": 320.98765432098764, |
|
"learning_rate": 9.311252601934615e-07, |
|
"loss": 2.06, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 322.22222222222223, |
|
"learning_rate": 9.308497612342353e-07, |
|
"loss": 2.0608, |
|
"step": 5220 |
|
}, |
|
{ |
|
"epoch": 323.4567901234568, |
|
"learning_rate": 9.305742622750092e-07, |
|
"loss": 2.0238, |
|
"step": 5240 |
|
}, |
|
{ |
|
"epoch": 324.69135802469134, |
|
"learning_rate": 9.30298763315783e-07, |
|
"loss": 2.0672, |
|
"step": 5260 |
|
}, |
|
{ |
|
"epoch": 325.9259259259259, |
|
"learning_rate": 9.300232643565568e-07, |
|
"loss": 2.0045, |
|
"step": 5280 |
|
}, |
|
{ |
|
"epoch": 327.1604938271605, |
|
"learning_rate": 9.297477653973307e-07, |
|
"loss": 2.0297, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 328.39506172839504, |
|
"learning_rate": 9.294722664381045e-07, |
|
"loss": 2.0939, |
|
"step": 5320 |
|
}, |
|
{ |
|
"epoch": 329.6296296296296, |
|
"learning_rate": 9.291967674788784e-07, |
|
"loss": 2.0309, |
|
"step": 5340 |
|
}, |
|
{ |
|
"epoch": 330.8641975308642, |
|
"learning_rate": 9.289212685196523e-07, |
|
"loss": 2.0221, |
|
"step": 5360 |
|
}, |
|
{ |
|
"epoch": 332.0987654320988, |
|
"learning_rate": 9.286457695604261e-07, |
|
"loss": 2.0629, |
|
"step": 5380 |
|
}, |
|
{ |
|
"epoch": 333.3333333333333, |
|
"learning_rate": 9.283702706011999e-07, |
|
"loss": 2.0113, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 334.5679012345679, |
|
"learning_rate": 9.280947716419737e-07, |
|
"loss": 2.0351, |
|
"step": 5420 |
|
}, |
|
{ |
|
"epoch": 335.8024691358025, |
|
"learning_rate": 9.278192726827476e-07, |
|
"loss": 2.0535, |
|
"step": 5440 |
|
}, |
|
{ |
|
"epoch": 337.037037037037, |
|
"learning_rate": 9.275437737235214e-07, |
|
"loss": 2.0653, |
|
"step": 5460 |
|
}, |
|
{ |
|
"epoch": 338.2716049382716, |
|
"learning_rate": 9.272682747642953e-07, |
|
"loss": 2.0103, |
|
"step": 5480 |
|
}, |
|
{ |
|
"epoch": 339.5061728395062, |
|
"learning_rate": 9.269927758050692e-07, |
|
"loss": 2.0367, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 340.74074074074076, |
|
"learning_rate": 9.267172768458429e-07, |
|
"loss": 1.9869, |
|
"step": 5520 |
|
}, |
|
{ |
|
"epoch": 341.9753086419753, |
|
"learning_rate": 9.264417778866168e-07, |
|
"loss": 2.0831, |
|
"step": 5540 |
|
}, |
|
{ |
|
"epoch": 343.2098765432099, |
|
"learning_rate": 9.261662789273906e-07, |
|
"loss": 2.0198, |
|
"step": 5560 |
|
}, |
|
{ |
|
"epoch": 344.44444444444446, |
|
"learning_rate": 9.258907799681645e-07, |
|
"loss": 2.0057, |
|
"step": 5580 |
|
}, |
|
{ |
|
"epoch": 345.679012345679, |
|
"learning_rate": 9.256152810089383e-07, |
|
"loss": 1.9973, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 346.91358024691357, |
|
"learning_rate": 9.253397820497122e-07, |
|
"loss": 2.0431, |
|
"step": 5620 |
|
}, |
|
{ |
|
"epoch": 348.14814814814815, |
|
"learning_rate": 9.250642830904861e-07, |
|
"loss": 2.0132, |
|
"step": 5640 |
|
}, |
|
{ |
|
"epoch": 349.38271604938274, |
|
"learning_rate": 9.247887841312599e-07, |
|
"loss": 2.0589, |
|
"step": 5660 |
|
}, |
|
{ |
|
"epoch": 350.61728395061726, |
|
"learning_rate": 9.245132851720337e-07, |
|
"loss": 1.9975, |
|
"step": 5680 |
|
}, |
|
{ |
|
"epoch": 351.85185185185185, |
|
"learning_rate": 9.242377862128075e-07, |
|
"loss": 2.0196, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 353.08641975308643, |
|
"learning_rate": 9.239622872535814e-07, |
|
"loss": 1.9605, |
|
"step": 5720 |
|
}, |
|
{ |
|
"epoch": 354.320987654321, |
|
"learning_rate": 9.236867882943551e-07, |
|
"loss": 2.0043, |
|
"step": 5740 |
|
}, |
|
{ |
|
"epoch": 355.55555555555554, |
|
"learning_rate": 9.234112893351291e-07, |
|
"loss": 1.9835, |
|
"step": 5760 |
|
}, |
|
{ |
|
"epoch": 356.7901234567901, |
|
"learning_rate": 9.23135790375903e-07, |
|
"loss": 2.0274, |
|
"step": 5780 |
|
}, |
|
{ |
|
"epoch": 358.0246913580247, |
|
"learning_rate": 9.228602914166768e-07, |
|
"loss": 2.0303, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 359.25925925925924, |
|
"learning_rate": 9.225847924574507e-07, |
|
"loss": 1.9691, |
|
"step": 5820 |
|
}, |
|
{ |
|
"epoch": 360.4938271604938, |
|
"learning_rate": 9.223092934982245e-07, |
|
"loss": 2.0049, |
|
"step": 5840 |
|
}, |
|
{ |
|
"epoch": 361.7283950617284, |
|
"learning_rate": 9.220337945389984e-07, |
|
"loss": 1.9652, |
|
"step": 5860 |
|
}, |
|
{ |
|
"epoch": 362.962962962963, |
|
"learning_rate": 9.217582955797723e-07, |
|
"loss": 2.019, |
|
"step": 5880 |
|
}, |
|
{ |
|
"epoch": 364.1975308641975, |
|
"learning_rate": 9.214827966205461e-07, |
|
"loss": 1.9819, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 365.4320987654321, |
|
"learning_rate": 9.2120729766132e-07, |
|
"loss": 2.0201, |
|
"step": 5920 |
|
}, |
|
{ |
|
"epoch": 366.6666666666667, |
|
"learning_rate": 9.209317987020937e-07, |
|
"loss": 1.9934, |
|
"step": 5940 |
|
}, |
|
{ |
|
"epoch": 367.9012345679012, |
|
"learning_rate": 9.206562997428676e-07, |
|
"loss": 1.9566, |
|
"step": 5960 |
|
}, |
|
{ |
|
"epoch": 369.1358024691358, |
|
"learning_rate": 9.203808007836415e-07, |
|
"loss": 1.9902, |
|
"step": 5980 |
|
}, |
|
{ |
|
"epoch": 370.3703703703704, |
|
"learning_rate": 9.201053018244152e-07, |
|
"loss": 2.0059, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 371.60493827160496, |
|
"learning_rate": 9.198298028651892e-07, |
|
"loss": 1.988, |
|
"step": 6020 |
|
}, |
|
{ |
|
"epoch": 372.8395061728395, |
|
"learning_rate": 9.19554303905963e-07, |
|
"loss": 1.9659, |
|
"step": 6040 |
|
}, |
|
{ |
|
"epoch": 374.0740740740741, |
|
"learning_rate": 9.192788049467368e-07, |
|
"loss": 2.0319, |
|
"step": 6060 |
|
}, |
|
{ |
|
"epoch": 375.30864197530866, |
|
"learning_rate": 9.190033059875106e-07, |
|
"loss": 1.9308, |
|
"step": 6080 |
|
}, |
|
{ |
|
"epoch": 376.5432098765432, |
|
"learning_rate": 9.187278070282845e-07, |
|
"loss": 1.967, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 377.77777777777777, |
|
"learning_rate": 9.184523080690584e-07, |
|
"loss": 2.0184, |
|
"step": 6120 |
|
}, |
|
{ |
|
"epoch": 379.01234567901236, |
|
"learning_rate": 9.181768091098322e-07, |
|
"loss": 1.9865, |
|
"step": 6140 |
|
}, |
|
{ |
|
"epoch": 380.24691358024694, |
|
"learning_rate": 9.179013101506061e-07, |
|
"loss": 1.9415, |
|
"step": 6160 |
|
}, |
|
{ |
|
"epoch": 381.48148148148147, |
|
"learning_rate": 9.176258111913799e-07, |
|
"loss": 1.9619, |
|
"step": 6180 |
|
}, |
|
{ |
|
"epoch": 382.71604938271605, |
|
"learning_rate": 9.173503122321538e-07, |
|
"loss": 1.949, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 383.95061728395063, |
|
"learning_rate": 9.170748132729275e-07, |
|
"loss": 1.9467, |
|
"step": 6220 |
|
}, |
|
{ |
|
"epoch": 385.18518518518516, |
|
"learning_rate": 9.167993143137014e-07, |
|
"loss": 1.9691, |
|
"step": 6240 |
|
}, |
|
{ |
|
"epoch": 386.41975308641975, |
|
"learning_rate": 9.165238153544753e-07, |
|
"loss": 1.9643, |
|
"step": 6260 |
|
}, |
|
{ |
|
"epoch": 387.65432098765433, |
|
"learning_rate": 9.162483163952491e-07, |
|
"loss": 1.942, |
|
"step": 6280 |
|
}, |
|
{ |
|
"epoch": 388.8888888888889, |
|
"learning_rate": 9.15972817436023e-07, |
|
"loss": 1.9002, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 390.12345679012344, |
|
"learning_rate": 9.156973184767968e-07, |
|
"loss": 1.9965, |
|
"step": 6320 |
|
}, |
|
{ |
|
"epoch": 391.358024691358, |
|
"learning_rate": 9.154218195175706e-07, |
|
"loss": 1.9343, |
|
"step": 6340 |
|
}, |
|
{ |
|
"epoch": 392.5925925925926, |
|
"learning_rate": 9.151463205583446e-07, |
|
"loss": 1.931, |
|
"step": 6360 |
|
}, |
|
{ |
|
"epoch": 393.82716049382714, |
|
"learning_rate": 9.148708215991183e-07, |
|
"loss": 1.9628, |
|
"step": 6380 |
|
}, |
|
{ |
|
"epoch": 395.0617283950617, |
|
"learning_rate": 9.145953226398922e-07, |
|
"loss": 1.9752, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 396.2962962962963, |
|
"learning_rate": 9.14319823680666e-07, |
|
"loss": 1.9499, |
|
"step": 6420 |
|
}, |
|
{ |
|
"epoch": 397.5308641975309, |
|
"learning_rate": 9.140443247214399e-07, |
|
"loss": 1.9546, |
|
"step": 6440 |
|
}, |
|
{ |
|
"epoch": 398.7654320987654, |
|
"learning_rate": 9.137688257622137e-07, |
|
"loss": 1.8583, |
|
"step": 6460 |
|
}, |
|
{ |
|
"epoch": 400.0, |
|
"learning_rate": 9.134933268029875e-07, |
|
"loss": 1.9405, |
|
"step": 6480 |
|
}, |
|
{ |
|
"epoch": 401.2345679012346, |
|
"learning_rate": 9.132178278437614e-07, |
|
"loss": 1.941, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 402.4691358024691, |
|
"learning_rate": 9.129423288845352e-07, |
|
"loss": 1.9432, |
|
"step": 6520 |
|
}, |
|
{ |
|
"epoch": 403.7037037037037, |
|
"learning_rate": 9.126668299253092e-07, |
|
"loss": 1.9603, |
|
"step": 6540 |
|
}, |
|
{ |
|
"epoch": 404.9382716049383, |
|
"learning_rate": 9.12391330966083e-07, |
|
"loss": 1.9456, |
|
"step": 6560 |
|
}, |
|
{ |
|
"epoch": 406.17283950617286, |
|
"learning_rate": 9.121158320068569e-07, |
|
"loss": 1.9488, |
|
"step": 6580 |
|
}, |
|
{ |
|
"epoch": 407.4074074074074, |
|
"learning_rate": 9.118403330476307e-07, |
|
"loss": 1.9248, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 408.641975308642, |
|
"learning_rate": 9.115648340884045e-07, |
|
"loss": 1.8897, |
|
"step": 6620 |
|
}, |
|
{ |
|
"epoch": 409.87654320987656, |
|
"learning_rate": 9.112893351291784e-07, |
|
"loss": 1.9315, |
|
"step": 6640 |
|
}, |
|
{ |
|
"epoch": 411.1111111111111, |
|
"learning_rate": 9.110138361699522e-07, |
|
"loss": 1.9599, |
|
"step": 6660 |
|
}, |
|
{ |
|
"epoch": 412.34567901234567, |
|
"learning_rate": 9.107383372107261e-07, |
|
"loss": 1.9441, |
|
"step": 6680 |
|
}, |
|
{ |
|
"epoch": 413.58024691358025, |
|
"learning_rate": 9.104628382515e-07, |
|
"loss": 1.8974, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 414.81481481481484, |
|
"learning_rate": 9.101873392922738e-07, |
|
"loss": 1.9078, |
|
"step": 6720 |
|
}, |
|
{ |
|
"epoch": 416.04938271604937, |
|
"learning_rate": 9.099118403330477e-07, |
|
"loss": 1.9587, |
|
"step": 6740 |
|
}, |
|
{ |
|
"epoch": 417.28395061728395, |
|
"learning_rate": 9.096363413738214e-07, |
|
"loss": 2.0016, |
|
"step": 6760 |
|
}, |
|
{ |
|
"epoch": 418.51851851851853, |
|
"learning_rate": 9.093608424145953e-07, |
|
"loss": 1.9557, |
|
"step": 6780 |
|
}, |
|
{ |
|
"epoch": 419.75308641975306, |
|
"learning_rate": 9.090853434553691e-07, |
|
"loss": 1.9719, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 420.98765432098764, |
|
"learning_rate": 9.08809844496143e-07, |
|
"loss": 1.9779, |
|
"step": 6820 |
|
}, |
|
{ |
|
"epoch": 422.22222222222223, |
|
"learning_rate": 9.085343455369169e-07, |
|
"loss": 1.9754, |
|
"step": 6840 |
|
}, |
|
{ |
|
"epoch": 423.4567901234568, |
|
"learning_rate": 9.082588465776907e-07, |
|
"loss": 2.0078, |
|
"step": 6860 |
|
}, |
|
{ |
|
"epoch": 424.69135802469134, |
|
"learning_rate": 9.079833476184645e-07, |
|
"loss": 1.9856, |
|
"step": 6880 |
|
}, |
|
{ |
|
"epoch": 425.9259259259259, |
|
"learning_rate": 9.077078486592383e-07, |
|
"loss": 1.9698, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 427.1604938271605, |
|
"learning_rate": 9.074323497000122e-07, |
|
"loss": 1.9826, |
|
"step": 6920 |
|
}, |
|
{ |
|
"epoch": 428.39506172839504, |
|
"learning_rate": 9.07156850740786e-07, |
|
"loss": 1.9513, |
|
"step": 6940 |
|
}, |
|
{ |
|
"epoch": 429.6296296296296, |
|
"learning_rate": 9.068813517815599e-07, |
|
"loss": 1.93, |
|
"step": 6960 |
|
}, |
|
{ |
|
"epoch": 430.8641975308642, |
|
"learning_rate": 9.066058528223339e-07, |
|
"loss": 1.9779, |
|
"step": 6980 |
|
}, |
|
{ |
|
"epoch": 432.0987654320988, |
|
"learning_rate": 9.063303538631076e-07, |
|
"loss": 1.9937, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 433.3333333333333, |
|
"learning_rate": 9.060548549038814e-07, |
|
"loss": 1.955, |
|
"step": 7020 |
|
}, |
|
{ |
|
"epoch": 434.5679012345679, |
|
"learning_rate": 9.057793559446552e-07, |
|
"loss": 1.9652, |
|
"step": 7040 |
|
}, |
|
{ |
|
"epoch": 435.8024691358025, |
|
"learning_rate": 9.055038569854291e-07, |
|
"loss": 1.9568, |
|
"step": 7060 |
|
}, |
|
{ |
|
"epoch": 437.037037037037, |
|
"learning_rate": 9.05228358026203e-07, |
|
"loss": 1.9787, |
|
"step": 7080 |
|
}, |
|
{ |
|
"epoch": 438.2716049382716, |
|
"learning_rate": 9.049528590669768e-07, |
|
"loss": 1.9137, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 439.5061728395062, |
|
"learning_rate": 9.046773601077507e-07, |
|
"loss": 1.9314, |
|
"step": 7120 |
|
}, |
|
{ |
|
"epoch": 440.74074074074076, |
|
"learning_rate": 9.044018611485244e-07, |
|
"loss": 1.9961, |
|
"step": 7140 |
|
}, |
|
{ |
|
"epoch": 441.9753086419753, |
|
"learning_rate": 9.041263621892983e-07, |
|
"loss": 1.9303, |
|
"step": 7160 |
|
}, |
|
{ |
|
"epoch": 443.2098765432099, |
|
"learning_rate": 9.038508632300721e-07, |
|
"loss": 1.982, |
|
"step": 7180 |
|
}, |
|
{ |
|
"epoch": 444.44444444444446, |
|
"learning_rate": 9.03575364270846e-07, |
|
"loss": 1.9133, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 445.679012345679, |
|
"learning_rate": 9.032998653116199e-07, |
|
"loss": 1.9376, |
|
"step": 7220 |
|
}, |
|
{ |
|
"epoch": 446.91358024691357, |
|
"learning_rate": 9.030243663523937e-07, |
|
"loss": 1.9254, |
|
"step": 7240 |
|
}, |
|
{ |
|
"epoch": 448.14814814814815, |
|
"learning_rate": 9.027488673931676e-07, |
|
"loss": 1.9491, |
|
"step": 7260 |
|
}, |
|
{ |
|
"epoch": 449.38271604938274, |
|
"learning_rate": 9.024733684339414e-07, |
|
"loss": 1.9455, |
|
"step": 7280 |
|
}, |
|
{ |
|
"epoch": 450.61728395061726, |
|
"learning_rate": 9.021978694747153e-07, |
|
"loss": 1.9128, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 451.85185185185185, |
|
"learning_rate": 9.019223705154892e-07, |
|
"loss": 1.9006, |
|
"step": 7320 |
|
}, |
|
{ |
|
"epoch": 453.08641975308643, |
|
"learning_rate": 9.016468715562629e-07, |
|
"loss": 1.9172, |
|
"step": 7340 |
|
}, |
|
{ |
|
"epoch": 454.320987654321, |
|
"learning_rate": 9.013713725970369e-07, |
|
"loss": 1.9754, |
|
"step": 7360 |
|
}, |
|
{ |
|
"epoch": 455.55555555555554, |
|
"learning_rate": 9.010958736378107e-07, |
|
"loss": 1.9759, |
|
"step": 7380 |
|
}, |
|
{ |
|
"epoch": 456.7901234567901, |
|
"learning_rate": 9.008203746785846e-07, |
|
"loss": 1.9452, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 458.0246913580247, |
|
"learning_rate": 9.005448757193583e-07, |
|
"loss": 1.9618, |
|
"step": 7420 |
|
}, |
|
{ |
|
"epoch": 459.25925925925924, |
|
"learning_rate": 9.002693767601322e-07, |
|
"loss": 1.9364, |
|
"step": 7440 |
|
}, |
|
{ |
|
"epoch": 460.4938271604938, |
|
"learning_rate": 8.999938778009061e-07, |
|
"loss": 1.9513, |
|
"step": 7460 |
|
}, |
|
{ |
|
"epoch": 461.7283950617284, |
|
"learning_rate": 8.997183788416799e-07, |
|
"loss": 1.9473, |
|
"step": 7480 |
|
}, |
|
{ |
|
"epoch": 462.962962962963, |
|
"learning_rate": 8.994428798824538e-07, |
|
"loss": 1.9159, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 464.1975308641975, |
|
"learning_rate": 8.991673809232276e-07, |
|
"loss": 1.9195, |
|
"step": 7520 |
|
}, |
|
{ |
|
"epoch": 465.4320987654321, |
|
"learning_rate": 8.988918819640015e-07, |
|
"loss": 1.9647, |
|
"step": 7540 |
|
}, |
|
{ |
|
"epoch": 466.6666666666667, |
|
"learning_rate": 8.986163830047752e-07, |
|
"loss": 1.895, |
|
"step": 7560 |
|
}, |
|
{ |
|
"epoch": 467.9012345679012, |
|
"learning_rate": 8.983408840455491e-07, |
|
"loss": 1.9408, |
|
"step": 7580 |
|
}, |
|
{ |
|
"epoch": 469.1358024691358, |
|
"learning_rate": 8.98065385086323e-07, |
|
"loss": 1.978, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 470.3703703703704, |
|
"learning_rate": 8.977898861270968e-07, |
|
"loss": 1.9085, |
|
"step": 7620 |
|
}, |
|
{ |
|
"epoch": 471.60493827160496, |
|
"learning_rate": 8.975143871678707e-07, |
|
"loss": 1.9833, |
|
"step": 7640 |
|
}, |
|
{ |
|
"epoch": 472.8395061728395, |
|
"learning_rate": 8.972388882086445e-07, |
|
"loss": 1.9559, |
|
"step": 7660 |
|
}, |
|
{ |
|
"epoch": 474.0740740740741, |
|
"learning_rate": 8.969633892494183e-07, |
|
"loss": 1.9333, |
|
"step": 7680 |
|
}, |
|
{ |
|
"epoch": 475.30864197530866, |
|
"learning_rate": 8.966878902901923e-07, |
|
"loss": 1.9075, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 476.5432098765432, |
|
"learning_rate": 8.96412391330966e-07, |
|
"loss": 1.9343, |
|
"step": 7720 |
|
}, |
|
{ |
|
"epoch": 477.77777777777777, |
|
"learning_rate": 8.961368923717399e-07, |
|
"loss": 1.9107, |
|
"step": 7740 |
|
}, |
|
{ |
|
"epoch": 479.01234567901236, |
|
"learning_rate": 8.958613934125137e-07, |
|
"loss": 1.9344, |
|
"step": 7760 |
|
}, |
|
{ |
|
"epoch": 480.24691358024694, |
|
"learning_rate": 8.955858944532876e-07, |
|
"loss": 1.9044, |
|
"step": 7780 |
|
}, |
|
{ |
|
"epoch": 481.48148148148147, |
|
"learning_rate": 8.953103954940614e-07, |
|
"loss": 1.974, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 482.71604938271605, |
|
"learning_rate": 8.950348965348353e-07, |
|
"loss": 1.9241, |
|
"step": 7820 |
|
}, |
|
{ |
|
"epoch": 483.95061728395063, |
|
"learning_rate": 8.947593975756091e-07, |
|
"loss": 1.9377, |
|
"step": 7840 |
|
}, |
|
{ |
|
"epoch": 485.18518518518516, |
|
"learning_rate": 8.944838986163829e-07, |
|
"loss": 1.9326, |
|
"step": 7860 |
|
}, |
|
{ |
|
"epoch": 486.41975308641975, |
|
"learning_rate": 8.942083996571568e-07, |
|
"loss": 1.9028, |
|
"step": 7880 |
|
}, |
|
{ |
|
"epoch": 487.65432098765433, |
|
"learning_rate": 8.939329006979306e-07, |
|
"loss": 1.8872, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 488.8888888888889, |
|
"learning_rate": 8.936574017387045e-07, |
|
"loss": 1.904, |
|
"step": 7920 |
|
}, |
|
{ |
|
"epoch": 490.12345679012344, |
|
"learning_rate": 8.933819027794784e-07, |
|
"loss": 1.8897, |
|
"step": 7940 |
|
}, |
|
{ |
|
"epoch": 491.358024691358, |
|
"learning_rate": 8.931064038202521e-07, |
|
"loss": 1.9026, |
|
"step": 7960 |
|
}, |
|
{ |
|
"epoch": 492.5925925925926, |
|
"learning_rate": 8.92830904861026e-07, |
|
"loss": 1.9287, |
|
"step": 7980 |
|
}, |
|
{ |
|
"epoch": 493.82716049382714, |
|
"learning_rate": 8.925554059017998e-07, |
|
"loss": 1.8277, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 495.0617283950617, |
|
"learning_rate": 8.922799069425738e-07, |
|
"loss": 1.8716, |
|
"step": 8020 |
|
}, |
|
{ |
|
"epoch": 496.2962962962963, |
|
"learning_rate": 8.920044079833476e-07, |
|
"loss": 1.8907, |
|
"step": 8040 |
|
}, |
|
{ |
|
"epoch": 497.5308641975309, |
|
"learning_rate": 8.917289090241215e-07, |
|
"loss": 1.8754, |
|
"step": 8060 |
|
}, |
|
{ |
|
"epoch": 498.7654320987654, |
|
"learning_rate": 8.914534100648954e-07, |
|
"loss": 1.8713, |
|
"step": 8080 |
|
}, |
|
{ |
|
"epoch": 500.0, |
|
"learning_rate": 8.911779111056691e-07, |
|
"loss": 1.8645, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 501.2345679012346, |
|
"learning_rate": 8.90902412146443e-07, |
|
"loss": 1.896, |
|
"step": 8120 |
|
}, |
|
{ |
|
"epoch": 502.4691358024691, |
|
"learning_rate": 8.906269131872168e-07, |
|
"loss": 1.8824, |
|
"step": 8140 |
|
}, |
|
{ |
|
"epoch": 503.7037037037037, |
|
"learning_rate": 8.903514142279907e-07, |
|
"loss": 1.8612, |
|
"step": 8160 |
|
}, |
|
{ |
|
"epoch": 504.9382716049383, |
|
"learning_rate": 8.900759152687646e-07, |
|
"loss": 1.8747, |
|
"step": 8180 |
|
}, |
|
{ |
|
"epoch": 506.17283950617286, |
|
"learning_rate": 8.898004163095384e-07, |
|
"loss": 1.8882, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 507.4074074074074, |
|
"learning_rate": 8.895249173503122e-07, |
|
"loss": 1.8858, |
|
"step": 8220 |
|
}, |
|
{ |
|
"epoch": 508.641975308642, |
|
"learning_rate": 8.89249418391086e-07, |
|
"loss": 1.8599, |
|
"step": 8240 |
|
}, |
|
{ |
|
"epoch": 509.87654320987656, |
|
"learning_rate": 8.889739194318599e-07, |
|
"loss": 1.9073, |
|
"step": 8260 |
|
}, |
|
{ |
|
"epoch": 511.1111111111111, |
|
"learning_rate": 8.886984204726337e-07, |
|
"loss": 1.9222, |
|
"step": 8280 |
|
}, |
|
{ |
|
"epoch": 512.3456790123457, |
|
"learning_rate": 8.884229215134076e-07, |
|
"loss": 1.904, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 513.5802469135803, |
|
"learning_rate": 8.881474225541814e-07, |
|
"loss": 1.8662, |
|
"step": 8320 |
|
}, |
|
{ |
|
"epoch": 514.8148148148148, |
|
"learning_rate": 8.878719235949553e-07, |
|
"loss": 1.8384, |
|
"step": 8340 |
|
}, |
|
{ |
|
"epoch": 516.0493827160494, |
|
"learning_rate": 8.875964246357292e-07, |
|
"loss": 1.9238, |
|
"step": 8360 |
|
}, |
|
{ |
|
"epoch": 517.283950617284, |
|
"learning_rate": 8.873209256765029e-07, |
|
"loss": 1.8626, |
|
"step": 8380 |
|
}, |
|
{ |
|
"epoch": 518.5185185185185, |
|
"learning_rate": 8.870454267172768e-07, |
|
"loss": 1.9077, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 519.7530864197531, |
|
"learning_rate": 8.867699277580506e-07, |
|
"loss": 1.8783, |
|
"step": 8420 |
|
}, |
|
{ |
|
"epoch": 520.9876543209876, |
|
"learning_rate": 8.864944287988245e-07, |
|
"loss": 1.8721, |
|
"step": 8440 |
|
}, |
|
{ |
|
"epoch": 522.2222222222222, |
|
"learning_rate": 8.862189298395984e-07, |
|
"loss": 1.8961, |
|
"step": 8460 |
|
}, |
|
{ |
|
"epoch": 523.4567901234568, |
|
"learning_rate": 8.859434308803721e-07, |
|
"loss": 1.8552, |
|
"step": 8480 |
|
}, |
|
{ |
|
"epoch": 524.6913580246913, |
|
"learning_rate": 8.85667931921146e-07, |
|
"loss": 1.9263, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 525.925925925926, |
|
"learning_rate": 8.853924329619198e-07, |
|
"loss": 1.8674, |
|
"step": 8520 |
|
}, |
|
{ |
|
"epoch": 527.1604938271605, |
|
"learning_rate": 8.851169340026937e-07, |
|
"loss": 1.8876, |
|
"step": 8540 |
|
}, |
|
{ |
|
"epoch": 528.395061728395, |
|
"learning_rate": 8.848414350434676e-07, |
|
"loss": 1.8834, |
|
"step": 8560 |
|
}, |
|
{ |
|
"epoch": 529.6296296296297, |
|
"learning_rate": 8.845659360842414e-07, |
|
"loss": 1.8467, |
|
"step": 8580 |
|
}, |
|
{ |
|
"epoch": 530.8641975308642, |
|
"learning_rate": 8.842904371250153e-07, |
|
"loss": 1.8957, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 532.0987654320987, |
|
"learning_rate": 8.840149381657891e-07, |
|
"loss": 1.8911, |
|
"step": 8620 |
|
}, |
|
{ |
|
"epoch": 533.3333333333334, |
|
"learning_rate": 8.837394392065629e-07, |
|
"loss": 1.8994, |
|
"step": 8640 |
|
}, |
|
{ |
|
"epoch": 534.5679012345679, |
|
"learning_rate": 8.834639402473367e-07, |
|
"loss": 1.9276, |
|
"step": 8660 |
|
}, |
|
{ |
|
"epoch": 535.8024691358024, |
|
"learning_rate": 8.831884412881107e-07, |
|
"loss": 1.8605, |
|
"step": 8680 |
|
}, |
|
{ |
|
"epoch": 537.0370370370371, |
|
"learning_rate": 8.829129423288845e-07, |
|
"loss": 1.8733, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 538.2716049382716, |
|
"learning_rate": 8.826374433696583e-07, |
|
"loss": 1.9128, |
|
"step": 8720 |
|
}, |
|
{ |
|
"epoch": 539.5061728395061, |
|
"learning_rate": 8.823619444104322e-07, |
|
"loss": 1.8563, |
|
"step": 8740 |
|
}, |
|
{ |
|
"epoch": 540.7407407407408, |
|
"learning_rate": 8.82086445451206e-07, |
|
"loss": 1.8647, |
|
"step": 8760 |
|
}, |
|
{ |
|
"epoch": 541.9753086419753, |
|
"learning_rate": 8.818109464919799e-07, |
|
"loss": 1.8478, |
|
"step": 8780 |
|
}, |
|
{ |
|
"epoch": 543.2098765432099, |
|
"learning_rate": 8.815354475327538e-07, |
|
"loss": 1.877, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 544.4444444444445, |
|
"learning_rate": 8.812599485735276e-07, |
|
"loss": 1.8432, |
|
"step": 8820 |
|
}, |
|
{ |
|
"epoch": 545.679012345679, |
|
"learning_rate": 8.809844496143015e-07, |
|
"loss": 1.8245, |
|
"step": 8840 |
|
}, |
|
{ |
|
"epoch": 546.9135802469136, |
|
"learning_rate": 8.807089506550753e-07, |
|
"loss": 1.841, |
|
"step": 8860 |
|
}, |
|
{ |
|
"epoch": 548.1481481481482, |
|
"learning_rate": 8.804334516958492e-07, |
|
"loss": 1.8609, |
|
"step": 8880 |
|
}, |
|
{ |
|
"epoch": 549.3827160493827, |
|
"learning_rate": 8.80157952736623e-07, |
|
"loss": 1.8815, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 550.6172839506173, |
|
"learning_rate": 8.798824537773968e-07, |
|
"loss": 1.8866, |
|
"step": 8920 |
|
}, |
|
{ |
|
"epoch": 551.8518518518518, |
|
"learning_rate": 8.796069548181707e-07, |
|
"loss": 1.8573, |
|
"step": 8940 |
|
}, |
|
{ |
|
"epoch": 553.0864197530864, |
|
"learning_rate": 8.793314558589445e-07, |
|
"loss": 1.8837, |
|
"step": 8960 |
|
}, |
|
{ |
|
"epoch": 554.320987654321, |
|
"learning_rate": 8.790559568997184e-07, |
|
"loss": 1.8875, |
|
"step": 8980 |
|
}, |
|
{ |
|
"epoch": 555.5555555555555, |
|
"learning_rate": 8.787804579404922e-07, |
|
"loss": 1.934, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 556.7901234567901, |
|
"learning_rate": 8.78504958981266e-07, |
|
"loss": 1.9294, |
|
"step": 9020 |
|
}, |
|
{ |
|
"epoch": 558.0246913580247, |
|
"learning_rate": 8.782294600220397e-07, |
|
"loss": 2.0228, |
|
"step": 9040 |
|
}, |
|
{ |
|
"epoch": 559.2592592592592, |
|
"learning_rate": 8.779539610628137e-07, |
|
"loss": 2.0074, |
|
"step": 9060 |
|
}, |
|
{ |
|
"epoch": 560.4938271604939, |
|
"learning_rate": 8.776784621035876e-07, |
|
"loss": 1.8819, |
|
"step": 9080 |
|
}, |
|
{ |
|
"epoch": 561.7283950617284, |
|
"learning_rate": 8.774029631443614e-07, |
|
"loss": 1.8354, |
|
"step": 9100 |
|
}, |
|
{ |
|
"epoch": 562.9629629629629, |
|
"learning_rate": 8.771274641851353e-07, |
|
"loss": 1.8683, |
|
"step": 9120 |
|
}, |
|
{ |
|
"epoch": 564.1975308641976, |
|
"learning_rate": 8.768519652259091e-07, |
|
"loss": 1.8644, |
|
"step": 9140 |
|
}, |
|
{ |
|
"epoch": 565.4320987654321, |
|
"learning_rate": 8.76576466266683e-07, |
|
"loss": 1.8822, |
|
"step": 9160 |
|
}, |
|
{ |
|
"epoch": 566.6666666666666, |
|
"learning_rate": 8.763009673074568e-07, |
|
"loss": 1.8544, |
|
"step": 9180 |
|
}, |
|
{ |
|
"epoch": 567.9012345679013, |
|
"learning_rate": 8.760254683482306e-07, |
|
"loss": 1.8507, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 569.1358024691358, |
|
"learning_rate": 8.757499693890045e-07, |
|
"loss": 1.8816, |
|
"step": 9220 |
|
}, |
|
{ |
|
"epoch": 570.3703703703703, |
|
"learning_rate": 8.754744704297783e-07, |
|
"loss": 1.7993, |
|
"step": 9240 |
|
}, |
|
{ |
|
"epoch": 571.604938271605, |
|
"learning_rate": 8.751989714705522e-07, |
|
"loss": 1.8231, |
|
"step": 9260 |
|
}, |
|
{ |
|
"epoch": 572.8395061728395, |
|
"learning_rate": 8.74923472511326e-07, |
|
"loss": 1.8354, |
|
"step": 9280 |
|
}, |
|
{ |
|
"epoch": 574.074074074074, |
|
"learning_rate": 8.746479735520998e-07, |
|
"loss": 1.8731, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 575.3086419753087, |
|
"learning_rate": 8.743724745928737e-07, |
|
"loss": 1.8377, |
|
"step": 9320 |
|
}, |
|
{ |
|
"epoch": 576.5432098765432, |
|
"learning_rate": 8.740969756336475e-07, |
|
"loss": 1.8211, |
|
"step": 9340 |
|
}, |
|
{ |
|
"epoch": 577.7777777777778, |
|
"learning_rate": 8.738214766744214e-07, |
|
"loss": 1.8321, |
|
"step": 9360 |
|
}, |
|
{ |
|
"epoch": 579.0123456790124, |
|
"learning_rate": 8.735459777151952e-07, |
|
"loss": 1.8091, |
|
"step": 9380 |
|
}, |
|
{ |
|
"epoch": 580.2469135802469, |
|
"learning_rate": 8.732704787559691e-07, |
|
"loss": 1.7772, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 581.4814814814815, |
|
"learning_rate": 8.72994979796743e-07, |
|
"loss": 1.7844, |
|
"step": 9420 |
|
}, |
|
{ |
|
"epoch": 582.716049382716, |
|
"learning_rate": 8.727194808375168e-07, |
|
"loss": 1.7728, |
|
"step": 9440 |
|
}, |
|
{ |
|
"epoch": 583.9506172839506, |
|
"learning_rate": 8.724439818782906e-07, |
|
"loss": 1.7777, |
|
"step": 9460 |
|
}, |
|
{ |
|
"epoch": 585.1851851851852, |
|
"learning_rate": 8.721684829190644e-07, |
|
"loss": 1.7895, |
|
"step": 9480 |
|
}, |
|
{ |
|
"epoch": 586.4197530864197, |
|
"learning_rate": 8.718929839598383e-07, |
|
"loss": 1.7949, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 587.6543209876543, |
|
"learning_rate": 8.716174850006122e-07, |
|
"loss": 1.7314, |
|
"step": 9520 |
|
}, |
|
{ |
|
"epoch": 588.8888888888889, |
|
"learning_rate": 8.713419860413861e-07, |
|
"loss": 1.7573, |
|
"step": 9540 |
|
}, |
|
{ |
|
"epoch": 590.1234567901234, |
|
"learning_rate": 8.710664870821599e-07, |
|
"loss": 1.7663, |
|
"step": 9560 |
|
}, |
|
{ |
|
"epoch": 591.358024691358, |
|
"learning_rate": 8.707909881229337e-07, |
|
"loss": 1.8203, |
|
"step": 9580 |
|
}, |
|
{ |
|
"epoch": 592.5925925925926, |
|
"learning_rate": 8.705154891637076e-07, |
|
"loss": 1.8092, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 593.8271604938271, |
|
"learning_rate": 8.702399902044814e-07, |
|
"loss": 1.7717, |
|
"step": 9620 |
|
}, |
|
{ |
|
"epoch": 595.0617283950618, |
|
"learning_rate": 8.699644912452553e-07, |
|
"loss": 1.7992, |
|
"step": 9640 |
|
}, |
|
{ |
|
"epoch": 596.2962962962963, |
|
"learning_rate": 8.69688992286029e-07, |
|
"loss": 1.7772, |
|
"step": 9660 |
|
}, |
|
{ |
|
"epoch": 597.5308641975308, |
|
"learning_rate": 8.69413493326803e-07, |
|
"loss": 1.7884, |
|
"step": 9680 |
|
}, |
|
{ |
|
"epoch": 598.7654320987655, |
|
"learning_rate": 8.691379943675769e-07, |
|
"loss": 1.7618, |
|
"step": 9700 |
|
}, |
|
{ |
|
"epoch": 600.0, |
|
"learning_rate": 8.688624954083506e-07, |
|
"loss": 1.7789, |
|
"step": 9720 |
|
}, |
|
{ |
|
"epoch": 601.2345679012345, |
|
"learning_rate": 8.685869964491245e-07, |
|
"loss": 1.7624, |
|
"step": 9740 |
|
}, |
|
{ |
|
"epoch": 602.4691358024692, |
|
"learning_rate": 8.683114974898983e-07, |
|
"loss": 1.7588, |
|
"step": 9760 |
|
}, |
|
{ |
|
"epoch": 603.7037037037037, |
|
"learning_rate": 8.680359985306722e-07, |
|
"loss": 1.793, |
|
"step": 9780 |
|
}, |
|
{ |
|
"epoch": 604.9382716049382, |
|
"learning_rate": 8.677604995714461e-07, |
|
"loss": 1.7527, |
|
"step": 9800 |
|
}, |
|
{ |
|
"epoch": 606.1728395061729, |
|
"learning_rate": 8.674850006122198e-07, |
|
"loss": 1.7472, |
|
"step": 9820 |
|
}, |
|
{ |
|
"epoch": 607.4074074074074, |
|
"learning_rate": 8.672095016529937e-07, |
|
"loss": 1.8087, |
|
"step": 9840 |
|
}, |
|
{ |
|
"epoch": 608.641975308642, |
|
"learning_rate": 8.669340026937675e-07, |
|
"loss": 1.7727, |
|
"step": 9860 |
|
}, |
|
{ |
|
"epoch": 609.8765432098766, |
|
"learning_rate": 8.666585037345414e-07, |
|
"loss": 1.7672, |
|
"step": 9880 |
|
}, |
|
{ |
|
"epoch": 611.1111111111111, |
|
"learning_rate": 8.663830047753152e-07, |
|
"loss": 1.7614, |
|
"step": 9900 |
|
}, |
|
{ |
|
"epoch": 612.3456790123457, |
|
"learning_rate": 8.661075058160891e-07, |
|
"loss": 1.7775, |
|
"step": 9920 |
|
}, |
|
{ |
|
"epoch": 613.5802469135803, |
|
"learning_rate": 8.65832006856863e-07, |
|
"loss": 1.7512, |
|
"step": 9940 |
|
}, |
|
{ |
|
"epoch": 614.8148148148148, |
|
"learning_rate": 8.655565078976368e-07, |
|
"loss": 1.7858, |
|
"step": 9960 |
|
}, |
|
{ |
|
"epoch": 616.0493827160494, |
|
"learning_rate": 8.652810089384107e-07, |
|
"loss": 1.8064, |
|
"step": 9980 |
|
}, |
|
{ |
|
"epoch": 617.283950617284, |
|
"learning_rate": 8.650055099791844e-07, |
|
"loss": 1.7649, |
|
"step": 10000 |
|
} |
|
], |
|
"logging_steps": 20, |
|
"max_steps": 65536, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 4096, |
|
"save_steps": 10000, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 2.923583979036672e+16, |
|
"train_batch_size": 5, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|