|
{ |
|
"best_metric": 0.22935771942138672, |
|
"best_model_checkpoint": "output/output__lora/checkpoint-600", |
|
"epoch": 1.7010935601458081, |
|
"eval_steps": 100, |
|
"global_step": 700, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.002430133657351154, |
|
"grad_norm": 2.0297858715057373, |
|
"learning_rate": 0.0, |
|
"loss": 0.3878, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.004860267314702308, |
|
"grad_norm": 2.0333988666534424, |
|
"learning_rate": 8.859191006777897e-06, |
|
"loss": 0.4074, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.007290400972053463, |
|
"grad_norm": 2.1134026050567627, |
|
"learning_rate": 1.4041485532469073e-05, |
|
"loss": 0.3097, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.009720534629404616, |
|
"grad_norm": 2.1784684658050537, |
|
"learning_rate": 1.7718382013555794e-05, |
|
"loss": 0.4578, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.012150668286755772, |
|
"grad_norm": 1.0290217399597168, |
|
"learning_rate": 2.0570404496611053e-05, |
|
"loss": 0.3389, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.014580801944106925, |
|
"grad_norm": 3.486976385116577, |
|
"learning_rate": 2.2900676539246968e-05, |
|
"loss": 0.2617, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.01701093560145808, |
|
"grad_norm": 1.7177977561950684, |
|
"learning_rate": 2.4870893478326387e-05, |
|
"loss": 0.5574, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.019441069258809233, |
|
"grad_norm": 0.8118329048156738, |
|
"learning_rate": 2.6577573020333684e-05, |
|
"loss": 0.4099, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.02187120291616039, |
|
"grad_norm": 2.2652881145477295, |
|
"learning_rate": 2.8082971064938146e-05, |
|
"loss": 0.3207, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.024301336573511544, |
|
"grad_norm": 1.2344919443130493, |
|
"learning_rate": 2.9429595503388953e-05, |
|
"loss": 0.2961, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.026731470230862697, |
|
"grad_norm": 1.2169722318649292, |
|
"learning_rate": 3.064776548439465e-05, |
|
"loss": 0.3689, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.02916160388821385, |
|
"grad_norm": 1.162546992301941, |
|
"learning_rate": 3.1759867546024865e-05, |
|
"loss": 0.351, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.031591737545565005, |
|
"grad_norm": 2.5353000164031982, |
|
"learning_rate": 3.2782902272079295e-05, |
|
"loss": 0.2424, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.03402187120291616, |
|
"grad_norm": 1.6772927045822144, |
|
"learning_rate": 3.373008448510428e-05, |
|
"loss": 0.2366, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.03645200486026731, |
|
"grad_norm": 1.5499287843704224, |
|
"learning_rate": 3.4611890029080124e-05, |
|
"loss": 0.3358, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.038882138517618466, |
|
"grad_norm": 1.383988618850708, |
|
"learning_rate": 3.543676402711159e-05, |
|
"loss": 0.3495, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.041312272174969626, |
|
"grad_norm": 1.4687221050262451, |
|
"learning_rate": 3.621161404374383e-05, |
|
"loss": 0.3973, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.04374240583232078, |
|
"grad_norm": 1.1256321668624878, |
|
"learning_rate": 3.694216207171603e-05, |
|
"loss": 0.2556, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.046172539489671933, |
|
"grad_norm": 0.9001079797744751, |
|
"learning_rate": 3.76332012245438e-05, |
|
"loss": 0.2283, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.04860267314702309, |
|
"grad_norm": 2.164158582687378, |
|
"learning_rate": 3.8288786510166846e-05, |
|
"loss": 0.2065, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.05103280680437424, |
|
"grad_norm": 0.6704838871955872, |
|
"learning_rate": 3.8912379010795455e-05, |
|
"loss": 0.3622, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.053462940461725394, |
|
"grad_norm": 1.380202293395996, |
|
"learning_rate": 3.9506956491172545e-05, |
|
"loss": 0.2949, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.05589307411907655, |
|
"grad_norm": 1.281702995300293, |
|
"learning_rate": 4.007509939970292e-05, |
|
"loss": 0.1766, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.0583232077764277, |
|
"grad_norm": 1.581787347793579, |
|
"learning_rate": 4.061905855280276e-05, |
|
"loss": 0.2676, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.060753341433778855, |
|
"grad_norm": 1.8878554105758667, |
|
"learning_rate": 4.1140808993222106e-05, |
|
"loss": 0.2572, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.06318347509113001, |
|
"grad_norm": 0.6384598612785339, |
|
"learning_rate": 4.164209327885719e-05, |
|
"loss": 0.231, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.06561360874848117, |
|
"grad_norm": 0.9673507809638977, |
|
"learning_rate": 4.2124456597407214e-05, |
|
"loss": 0.1772, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.06804374240583232, |
|
"grad_norm": 1.0678174495697021, |
|
"learning_rate": 4.258927549188218e-05, |
|
"loss": 0.2025, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.07047387606318348, |
|
"grad_norm": 0.9189561009407043, |
|
"learning_rate": 4.303778154313212e-05, |
|
"loss": 0.2345, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.07290400972053462, |
|
"grad_norm": 1.3609440326690674, |
|
"learning_rate": 4.347108103585803e-05, |
|
"loss": 0.2201, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.07533414337788578, |
|
"grad_norm": 1.3813296556472778, |
|
"learning_rate": 4.389017139879164e-05, |
|
"loss": 0.2438, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.07776427703523693, |
|
"grad_norm": 1.1185611486434937, |
|
"learning_rate": 4.429595503388948e-05, |
|
"loss": 0.1972, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.08019441069258809, |
|
"grad_norm": 0.876745343208313, |
|
"learning_rate": 4.468925101686371e-05, |
|
"loss": 0.2754, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.08262454434993925, |
|
"grad_norm": 0.7372068762779236, |
|
"learning_rate": 4.507080505052173e-05, |
|
"loss": 0.2719, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.0850546780072904, |
|
"grad_norm": 1.4229792356491089, |
|
"learning_rate": 4.544129797493744e-05, |
|
"loss": 0.3219, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.08748481166464156, |
|
"grad_norm": 1.4566582441329956, |
|
"learning_rate": 4.5801353078493936e-05, |
|
"loss": 0.1257, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.0899149453219927, |
|
"grad_norm": 1.217944860458374, |
|
"learning_rate": 4.615154240700883e-05, |
|
"loss": 0.3138, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.09234507897934387, |
|
"grad_norm": 1.882925033569336, |
|
"learning_rate": 4.6492392231321696e-05, |
|
"loss": 0.219, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.09477521263669501, |
|
"grad_norm": 0.8372556567192078, |
|
"learning_rate": 4.682438780454837e-05, |
|
"loss": 0.27, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.09720534629404617, |
|
"grad_norm": 0.9709153771400452, |
|
"learning_rate": 4.714797751694474e-05, |
|
"loss": 0.2759, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.09963547995139732, |
|
"grad_norm": 1.1802537441253662, |
|
"learning_rate": 4.7463576537657414e-05, |
|
"loss": 0.2411, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.10206561360874848, |
|
"grad_norm": 0.7289960980415344, |
|
"learning_rate": 4.777157001757336e-05, |
|
"loss": 0.2755, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.10449574726609964, |
|
"grad_norm": 0.8493523001670837, |
|
"learning_rate": 4.8072315915252694e-05, |
|
"loss": 0.3177, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.10692588092345079, |
|
"grad_norm": 1.0613563060760498, |
|
"learning_rate": 4.8366147497950435e-05, |
|
"loss": 0.2056, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.10935601458080195, |
|
"grad_norm": 1.1137558221817017, |
|
"learning_rate": 4.8653375561549195e-05, |
|
"loss": 0.293, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.1117861482381531, |
|
"grad_norm": 0.6552504301071167, |
|
"learning_rate": 4.8934290406480814e-05, |
|
"loss": 0.2478, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.11421628189550426, |
|
"grad_norm": 0.9905387759208679, |
|
"learning_rate": 4.920916360113129e-05, |
|
"loss": 0.2488, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.1166464155528554, |
|
"grad_norm": 1.101691484451294, |
|
"learning_rate": 4.947824955958066e-05, |
|
"loss": 0.1583, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.11907654921020656, |
|
"grad_norm": 1.3675787448883057, |
|
"learning_rate": 4.9741786956652774e-05, |
|
"loss": 0.3629, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.12150668286755771, |
|
"grad_norm": 2.6356711387634277, |
|
"learning_rate": 5e-05, |
|
"loss": 0.285, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.12393681652490887, |
|
"grad_norm": 1.5550183057785034, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1688, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.12636695018226002, |
|
"grad_norm": 1.7116848230361938, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2495, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.12879708383961117, |
|
"grad_norm": 1.3964793682098389, |
|
"learning_rate": 5e-05, |
|
"loss": 0.276, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.13122721749696234, |
|
"grad_norm": 0.8001569509506226, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2976, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.1336573511543135, |
|
"grad_norm": 1.331112027168274, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2725, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.13608748481166463, |
|
"grad_norm": 1.0783005952835083, |
|
"learning_rate": 5e-05, |
|
"loss": 0.188, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.1385176184690158, |
|
"grad_norm": 0.9507465958595276, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2917, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.14094775212636695, |
|
"grad_norm": 0.8285686373710632, |
|
"learning_rate": 5e-05, |
|
"loss": 0.3322, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.1433778857837181, |
|
"grad_norm": 0.8201387524604797, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2296, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.14580801944106925, |
|
"grad_norm": 0.743626594543457, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1785, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.14823815309842042, |
|
"grad_norm": 0.9876251816749573, |
|
"learning_rate": 5e-05, |
|
"loss": 0.3416, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.15066828675577157, |
|
"grad_norm": 1.6088169813156128, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2513, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.15309842041312272, |
|
"grad_norm": 1.2406659126281738, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2856, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.15552855407047386, |
|
"grad_norm": 1.0257083177566528, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1392, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.15795868772782504, |
|
"grad_norm": 0.7829582691192627, |
|
"learning_rate": 5e-05, |
|
"loss": 0.248, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.16038882138517618, |
|
"grad_norm": 0.7451558113098145, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2826, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.16281895504252733, |
|
"grad_norm": 0.8115066885948181, |
|
"learning_rate": 5e-05, |
|
"loss": 0.3322, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.1652490886998785, |
|
"grad_norm": 0.8748694658279419, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2057, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.16767922235722965, |
|
"grad_norm": 0.6844836473464966, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2062, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.1701093560145808, |
|
"grad_norm": 0.9419387578964233, |
|
"learning_rate": 5e-05, |
|
"loss": 0.169, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.17253948967193194, |
|
"grad_norm": 0.8529036045074463, |
|
"learning_rate": 5e-05, |
|
"loss": 0.28, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.17496962332928312, |
|
"grad_norm": 0.8096457719802856, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1699, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.17739975698663427, |
|
"grad_norm": 0.7595834136009216, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2321, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.1798298906439854, |
|
"grad_norm": 0.6918485164642334, |
|
"learning_rate": 5e-05, |
|
"loss": 0.12, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.1822600243013366, |
|
"grad_norm": 0.8465882539749146, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2344, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.18469015795868773, |
|
"grad_norm": 1.27568519115448, |
|
"learning_rate": 5e-05, |
|
"loss": 0.3018, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.18712029161603888, |
|
"grad_norm": 0.907692015171051, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2154, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.18955042527339003, |
|
"grad_norm": 0.9362420439720154, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2024, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.1919805589307412, |
|
"grad_norm": 1.1268069744110107, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1509, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.19441069258809235, |
|
"grad_norm": 2.944066047668457, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1669, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.1968408262454435, |
|
"grad_norm": 0.8719034790992737, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2229, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.19927095990279464, |
|
"grad_norm": 0.7963365912437439, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1485, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.20170109356014582, |
|
"grad_norm": 1.1861976385116577, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1774, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.20413122721749696, |
|
"grad_norm": 1.029300570487976, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2762, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.2065613608748481, |
|
"grad_norm": 0.7859634757041931, |
|
"learning_rate": 5e-05, |
|
"loss": 0.263, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.20899149453219928, |
|
"grad_norm": 1.2088333368301392, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2248, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.21142162818955043, |
|
"grad_norm": 0.9491986036300659, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1388, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.21385176184690158, |
|
"grad_norm": 1.3867732286453247, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1932, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.21628189550425272, |
|
"grad_norm": 1.2418168783187866, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2703, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.2187120291616039, |
|
"grad_norm": 1.514169692993164, |
|
"learning_rate": 5e-05, |
|
"loss": 0.3437, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.22114216281895505, |
|
"grad_norm": 1.1384562253952026, |
|
"learning_rate": 5e-05, |
|
"loss": 0.3157, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.2235722964763062, |
|
"grad_norm": 1.0910252332687378, |
|
"learning_rate": 5e-05, |
|
"loss": 0.3296, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.22600243013365734, |
|
"grad_norm": 0.8576286435127258, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2572, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.2284325637910085, |
|
"grad_norm": 1.0798954963684082, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2669, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.23086269744835966, |
|
"grad_norm": 1.0142358541488647, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2607, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.2332928311057108, |
|
"grad_norm": 0.8707209825515747, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2206, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.23572296476306198, |
|
"grad_norm": 1.038823127746582, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1648, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.23815309842041313, |
|
"grad_norm": 1.3246694803237915, |
|
"learning_rate": 5e-05, |
|
"loss": 0.317, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.24058323207776428, |
|
"grad_norm": 1.0180795192718506, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2653, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.24301336573511542, |
|
"grad_norm": 1.803346037864685, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1221, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.24301336573511542, |
|
"eval_loss": 0.2808912396430969, |
|
"eval_runtime": 508.6685, |
|
"eval_samples_per_second": 5.402, |
|
"eval_steps_per_second": 0.676, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.2454434993924666, |
|
"grad_norm": 1.4270009994506836, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1513, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.24787363304981774, |
|
"grad_norm": 1.5230388641357422, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2563, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.2503037667071689, |
|
"grad_norm": 1.0142149925231934, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2021, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.25273390036452004, |
|
"grad_norm": 1.224928617477417, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2227, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.2551640340218712, |
|
"grad_norm": 0.8966079950332642, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1416, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.25759416767922233, |
|
"grad_norm": 4.392244815826416, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1297, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.2600243013365735, |
|
"grad_norm": 0.7403039932250977, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1383, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.2624544349939247, |
|
"grad_norm": 0.7826500535011292, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1933, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.2648845686512758, |
|
"grad_norm": 1.2112810611724854, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2118, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.267314702308627, |
|
"grad_norm": 1.2826062440872192, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2028, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.26974483596597815, |
|
"grad_norm": 0.9675102829933167, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1481, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.27217496962332927, |
|
"grad_norm": 1.2551299333572388, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2172, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.27460510328068044, |
|
"grad_norm": 1.2526607513427734, |
|
"learning_rate": 5e-05, |
|
"loss": 0.213, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.2770352369380316, |
|
"grad_norm": 1.102669358253479, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2387, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.27946537059538273, |
|
"grad_norm": 1.1452562808990479, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2224, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.2818955042527339, |
|
"grad_norm": 1.4839582443237305, |
|
"learning_rate": 5e-05, |
|
"loss": 0.3572, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.284325637910085, |
|
"grad_norm": 0.7019845247268677, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1531, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.2867557715674362, |
|
"grad_norm": 1.1344879865646362, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1234, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.2891859052247874, |
|
"grad_norm": 0.7753707766532898, |
|
"learning_rate": 5e-05, |
|
"loss": 0.3054, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.2916160388821385, |
|
"grad_norm": 1.2467390298843384, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0979, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.29404617253948967, |
|
"grad_norm": 0.8829241991043091, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1251, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.29647630619684084, |
|
"grad_norm": 0.9316434860229492, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1723, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.29890643985419196, |
|
"grad_norm": 0.9735895395278931, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1992, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.30133657351154314, |
|
"grad_norm": 1.0356009006500244, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2912, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.3037667071688943, |
|
"grad_norm": 0.8671007752418518, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2204, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.30619684082624543, |
|
"grad_norm": 0.9986138939857483, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1615, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.3086269744835966, |
|
"grad_norm": 0.8567970991134644, |
|
"learning_rate": 5e-05, |
|
"loss": 0.3127, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.3110571081409477, |
|
"grad_norm": 1.497078537940979, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1461, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.3134872417982989, |
|
"grad_norm": 1.9854991436004639, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1588, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.3159173754556501, |
|
"grad_norm": 0.9955502152442932, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2703, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.3183475091130012, |
|
"grad_norm": 0.8443557620048523, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1432, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.32077764277035237, |
|
"grad_norm": 0.8821367025375366, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1113, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.32320777642770354, |
|
"grad_norm": 0.997931957244873, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1363, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.32563791008505466, |
|
"grad_norm": 1.2628774642944336, |
|
"learning_rate": 5e-05, |
|
"loss": 0.134, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.32806804374240583, |
|
"grad_norm": 0.7694957852363586, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1425, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.330498177399757, |
|
"grad_norm": 1.5881726741790771, |
|
"learning_rate": 5e-05, |
|
"loss": 0.17, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.33292831105710813, |
|
"grad_norm": 1.2904351949691772, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1735, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.3353584447144593, |
|
"grad_norm": 1.3187663555145264, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2035, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.3377885783718105, |
|
"grad_norm": 1.1753528118133545, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2147, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.3402187120291616, |
|
"grad_norm": 1.6974895000457764, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1303, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.34264884568651277, |
|
"grad_norm": 0.9945745468139648, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1169, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.3450789793438639, |
|
"grad_norm": 1.0099682807922363, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1446, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.34750911300121506, |
|
"grad_norm": 1.223264217376709, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2118, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.34993924665856624, |
|
"grad_norm": 0.8575140833854675, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1109, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.35236938031591736, |
|
"grad_norm": 1.0145286321640015, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1664, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.35479951397326853, |
|
"grad_norm": 0.670891284942627, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1166, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.3572296476306197, |
|
"grad_norm": 1.4332047700881958, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1381, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.3596597812879708, |
|
"grad_norm": 1.0511322021484375, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1279, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.362089914945322, |
|
"grad_norm": 0.751565158367157, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1533, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.3645200486026732, |
|
"grad_norm": 1.612855076789856, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1633, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.3669501822600243, |
|
"grad_norm": 1.2449802160263062, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2037, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.36938031591737547, |
|
"grad_norm": 1.0060147047042847, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0902, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.3718104495747266, |
|
"grad_norm": 0.9183961153030396, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1748, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.37424058323207776, |
|
"grad_norm": 1.2824058532714844, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1903, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.37667071688942894, |
|
"grad_norm": 0.524626612663269, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1014, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.37910085054678005, |
|
"grad_norm": 0.946535587310791, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2512, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.38153098420413123, |
|
"grad_norm": 0.8985329270362854, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1247, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.3839611178614824, |
|
"grad_norm": 0.6025483012199402, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0677, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.3863912515188335, |
|
"grad_norm": 0.9210860729217529, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1361, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.3888213851761847, |
|
"grad_norm": 1.7067134380340576, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2008, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.39125151883353587, |
|
"grad_norm": 0.8162620663642883, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1108, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.393681652490887, |
|
"grad_norm": 0.535029947757721, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1578, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.39611178614823817, |
|
"grad_norm": 0.7219657897949219, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1012, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.3985419198055893, |
|
"grad_norm": 1.0636670589447021, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1648, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.40097205346294046, |
|
"grad_norm": 1.1116470098495483, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1191, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.40340218712029163, |
|
"grad_norm": 0.7233144044876099, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1206, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.40583232077764275, |
|
"grad_norm": 1.7054411172866821, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1466, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.4082624544349939, |
|
"grad_norm": 0.9031124114990234, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1107, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.4106925880923451, |
|
"grad_norm": 0.8140673637390137, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1503, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.4131227217496962, |
|
"grad_norm": 1.2189490795135498, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1081, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.4155528554070474, |
|
"grad_norm": 0.8403185606002808, |
|
"learning_rate": 5e-05, |
|
"loss": 0.098, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.41798298906439857, |
|
"grad_norm": 0.8939706683158875, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0961, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.4204131227217497, |
|
"grad_norm": 0.8737717270851135, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1165, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.42284325637910086, |
|
"grad_norm": 0.9247780442237854, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1787, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.425273390036452, |
|
"grad_norm": 0.9532423615455627, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1676, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.42770352369380316, |
|
"grad_norm": 0.5811410546302795, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1221, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.43013365735115433, |
|
"grad_norm": 0.6600822806358337, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1809, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.43256379100850545, |
|
"grad_norm": 0.5638179183006287, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0668, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.4349939246658566, |
|
"grad_norm": 0.5698412656784058, |
|
"learning_rate": 5e-05, |
|
"loss": 0.164, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.4374240583232078, |
|
"grad_norm": 1.2823748588562012, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2111, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.4398541919805589, |
|
"grad_norm": 1.1044129133224487, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1188, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.4422843256379101, |
|
"grad_norm": 0.9815372824668884, |
|
"learning_rate": 5e-05, |
|
"loss": 0.186, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.44471445929526127, |
|
"grad_norm": 0.8543856739997864, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1237, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.4471445929526124, |
|
"grad_norm": 1.0749317407608032, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1385, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.44957472660996356, |
|
"grad_norm": 1.4067716598510742, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1269, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.4520048602673147, |
|
"grad_norm": 0.933139443397522, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0476, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.45443499392466585, |
|
"grad_norm": 1.2160463333129883, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0823, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.456865127582017, |
|
"grad_norm": 0.8305476903915405, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2482, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.45929526123936815, |
|
"grad_norm": 0.8701503872871399, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2007, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.4617253948967193, |
|
"grad_norm": 1.1040858030319214, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1031, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.4641555285540705, |
|
"grad_norm": 1.149579644203186, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1384, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.4665856622114216, |
|
"grad_norm": 1.0616861581802368, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1961, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.4690157958687728, |
|
"grad_norm": 1.3416138887405396, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2223, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.47144592952612396, |
|
"grad_norm": 1.2578601837158203, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0833, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.4738760631834751, |
|
"grad_norm": 0.9381983876228333, |
|
"learning_rate": 5e-05, |
|
"loss": 0.135, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.47630619684082626, |
|
"grad_norm": 1.0591320991516113, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1658, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.4787363304981774, |
|
"grad_norm": 0.9011105895042419, |
|
"learning_rate": 5e-05, |
|
"loss": 0.189, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.48116646415552855, |
|
"grad_norm": 0.9011105895042419, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1729, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.4835965978128797, |
|
"grad_norm": 1.0387153625488281, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1534, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.48602673147023084, |
|
"grad_norm": 1.7026396989822388, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1534, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.48602673147023084, |
|
"eval_loss": 0.2694580554962158, |
|
"eval_runtime": 510.8356, |
|
"eval_samples_per_second": 5.379, |
|
"eval_steps_per_second": 0.673, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.488456865127582, |
|
"grad_norm": 0.7090218663215637, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1022, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.4908869987849332, |
|
"grad_norm": 0.6188836693763733, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1839, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.4933171324422843, |
|
"grad_norm": 0.8001598715782166, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1525, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.4957472660996355, |
|
"grad_norm": 1.006425142288208, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1417, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.49817739975698666, |
|
"grad_norm": 0.9688583612442017, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1504, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.5006075334143378, |
|
"grad_norm": 0.9707381129264832, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0538, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 0.503037667071689, |
|
"grad_norm": 0.9181280136108398, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2319, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.5054678007290401, |
|
"grad_norm": 1.059377670288086, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2679, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.5078979343863913, |
|
"grad_norm": 0.6514914035797119, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0623, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 0.5103280680437424, |
|
"grad_norm": 0.7265043258666992, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1034, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.5127582017010935, |
|
"grad_norm": 1.5795567035675049, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1236, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 0.5151883353584447, |
|
"grad_norm": 0.8533775806427002, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0702, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.5176184690157959, |
|
"grad_norm": 0.7225982546806335, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2005, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.520048602673147, |
|
"grad_norm": 0.9016891121864319, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1329, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 0.5224787363304981, |
|
"grad_norm": 1.1623356342315674, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1728, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.5249088699878494, |
|
"grad_norm": 1.1869186162948608, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2517, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.5273390036452005, |
|
"grad_norm": 1.2271298170089722, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1402, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 0.5297691373025516, |
|
"grad_norm": 1.182732105255127, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1341, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 0.5321992709599028, |
|
"grad_norm": 0.8633968234062195, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1469, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.534629404617254, |
|
"grad_norm": 0.9293995499610901, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1121, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.5370595382746051, |
|
"grad_norm": 1.3438748121261597, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1718, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 0.5394896719319563, |
|
"grad_norm": 1.2411259412765503, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1304, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.5419198055893074, |
|
"grad_norm": 0.8153496384620667, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1334, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 0.5443499392466585, |
|
"grad_norm": 0.7595478892326355, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0994, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 0.5467800729040098, |
|
"grad_norm": 1.0710370540618896, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0351, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.5492102065613609, |
|
"grad_norm": 1.2209466695785522, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1049, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 0.551640340218712, |
|
"grad_norm": 0.5865522623062134, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0948, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 0.5540704738760632, |
|
"grad_norm": 0.8379694223403931, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1269, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.5565006075334143, |
|
"grad_norm": 0.7321844100952148, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0912, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 0.5589307411907655, |
|
"grad_norm": 1.0253808498382568, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0811, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.5613608748481167, |
|
"grad_norm": 0.9015201330184937, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0725, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 0.5637910085054678, |
|
"grad_norm": 0.8760166168212891, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1727, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 0.5662211421628189, |
|
"grad_norm": 0.892677903175354, |
|
"learning_rate": 5e-05, |
|
"loss": 0.084, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 0.56865127582017, |
|
"grad_norm": 0.6722489595413208, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0845, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 0.5710814094775213, |
|
"grad_norm": 0.6187300682067871, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1269, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.5735115431348724, |
|
"grad_norm": 0.8674693703651428, |
|
"learning_rate": 5e-05, |
|
"loss": 0.09, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 0.5759416767922235, |
|
"grad_norm": 0.9376964569091797, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1916, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 0.5783718104495748, |
|
"grad_norm": 0.6777618527412415, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1321, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 0.5808019441069259, |
|
"grad_norm": 1.1223134994506836, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1811, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 0.583232077764277, |
|
"grad_norm": 0.8348031044006348, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1508, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.5856622114216282, |
|
"grad_norm": 0.7387763857841492, |
|
"learning_rate": 5e-05, |
|
"loss": 0.058, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 0.5880923450789793, |
|
"grad_norm": 0.6327199339866638, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1096, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 0.5905224787363305, |
|
"grad_norm": 0.7040870785713196, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1281, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 0.5929526123936817, |
|
"grad_norm": 1.2110347747802734, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1959, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 0.5953827460510328, |
|
"grad_norm": 0.8480271100997925, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1234, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.5978128797083839, |
|
"grad_norm": 1.0009665489196777, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1223, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 0.6002430133657352, |
|
"grad_norm": 0.7057176828384399, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1747, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 0.6026731470230863, |
|
"grad_norm": 0.9945976138114929, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1766, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 0.6051032806804374, |
|
"grad_norm": 0.7461872696876526, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1224, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 0.6075334143377886, |
|
"grad_norm": 0.5109842419624329, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1105, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.6099635479951397, |
|
"grad_norm": 0.7398750185966492, |
|
"learning_rate": 5e-05, |
|
"loss": 0.139, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 0.6123936816524909, |
|
"grad_norm": 0.7645841836929321, |
|
"learning_rate": 5e-05, |
|
"loss": 0.11, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 0.6148238153098421, |
|
"grad_norm": 0.8627939224243164, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1277, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 0.6172539489671932, |
|
"grad_norm": 0.6933022737503052, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1283, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 0.6196840826245443, |
|
"grad_norm": 0.8254914283752441, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1087, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.6221142162818954, |
|
"grad_norm": 0.9561761617660522, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1019, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 0.6245443499392467, |
|
"grad_norm": 1.0569480657577515, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1235, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 0.6269744835965978, |
|
"grad_norm": 1.0708096027374268, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1797, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 0.6294046172539489, |
|
"grad_norm": 0.5817265510559082, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0787, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 0.6318347509113001, |
|
"grad_norm": 0.8255194425582886, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1754, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.6342648845686513, |
|
"grad_norm": 0.8279481530189514, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1668, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 0.6366950182260024, |
|
"grad_norm": 0.6393010020256042, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0756, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 0.6391251518833536, |
|
"grad_norm": 0.49387434124946594, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0757, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 0.6415552855407047, |
|
"grad_norm": 0.5705033540725708, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0949, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 0.6439854191980559, |
|
"grad_norm": 0.9022333025932312, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1716, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.6464155528554071, |
|
"grad_norm": 0.5299041271209717, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0799, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 0.6488456865127582, |
|
"grad_norm": 0.7919138669967651, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1476, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 0.6512758201701093, |
|
"grad_norm": 1.394972324371338, |
|
"learning_rate": 5e-05, |
|
"loss": 0.112, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 0.6537059538274606, |
|
"grad_norm": 0.9046385884284973, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1683, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 0.6561360874848117, |
|
"grad_norm": 0.7190728783607483, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1434, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.6585662211421628, |
|
"grad_norm": 1.0634692907333374, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1244, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 0.660996354799514, |
|
"grad_norm": 0.569398045539856, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0728, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 0.6634264884568651, |
|
"grad_norm": 0.9083871841430664, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1467, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 0.6658566221142163, |
|
"grad_norm": 0.7927613854408264, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1832, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 0.6682867557715675, |
|
"grad_norm": 1.0088517665863037, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1422, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.6707168894289186, |
|
"grad_norm": 0.5374720692634583, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0971, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 0.6731470230862697, |
|
"grad_norm": 0.7200155854225159, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0722, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 0.675577156743621, |
|
"grad_norm": 0.7518504858016968, |
|
"learning_rate": 5e-05, |
|
"loss": 0.126, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 0.6780072904009721, |
|
"grad_norm": 0.8544089794158936, |
|
"learning_rate": 5e-05, |
|
"loss": 0.113, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 0.6804374240583232, |
|
"grad_norm": 0.7537454962730408, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0615, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.6828675577156743, |
|
"grad_norm": 0.8440731763839722, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0933, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 0.6852976913730255, |
|
"grad_norm": 1.7643004655838013, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1807, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 0.6877278250303767, |
|
"grad_norm": 0.6502506732940674, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1189, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 0.6901579586877278, |
|
"grad_norm": 0.8678569197654724, |
|
"learning_rate": 5e-05, |
|
"loss": 0.094, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 0.692588092345079, |
|
"grad_norm": 0.8592473864555359, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1197, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.6950182260024301, |
|
"grad_norm": 0.8664891719818115, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0575, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 0.6974483596597812, |
|
"grad_norm": 0.5715927481651306, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0668, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 0.6998784933171325, |
|
"grad_norm": 1.0751949548721313, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1457, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 0.7023086269744836, |
|
"grad_norm": 0.8039306402206421, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1206, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 0.7047387606318347, |
|
"grad_norm": 1.5916086435317993, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1471, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.707168894289186, |
|
"grad_norm": 0.7829633355140686, |
|
"learning_rate": 5e-05, |
|
"loss": 0.106, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 0.7095990279465371, |
|
"grad_norm": 0.8277871608734131, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0347, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 0.7120291616038882, |
|
"grad_norm": 0.9798584580421448, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1231, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 0.7144592952612394, |
|
"grad_norm": 0.6273385286331177, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0838, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 0.7168894289185905, |
|
"grad_norm": 1.2111278772354126, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1971, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.7193195625759417, |
|
"grad_norm": 0.5927327275276184, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1198, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 0.7217496962332929, |
|
"grad_norm": 0.672064483165741, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1295, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 0.724179829890644, |
|
"grad_norm": 0.6398962736129761, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0933, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 0.7266099635479951, |
|
"grad_norm": 0.7458367943763733, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1407, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 0.7290400972053463, |
|
"grad_norm": 0.6593539118766785, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1055, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.7290400972053463, |
|
"eval_loss": 0.24557192623615265, |
|
"eval_runtime": 513.4098, |
|
"eval_samples_per_second": 5.352, |
|
"eval_steps_per_second": 0.67, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.7314702308626975, |
|
"grad_norm": 0.5555610656738281, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1641, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 0.7339003645200486, |
|
"grad_norm": 0.49084270000457764, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0786, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 0.7363304981773997, |
|
"grad_norm": 0.9267096519470215, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2053, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 0.7387606318347509, |
|
"grad_norm": 0.440594345331192, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1224, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 0.741190765492102, |
|
"grad_norm": 1.0237218141555786, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1263, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.7436208991494532, |
|
"grad_norm": 1.0087871551513672, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1218, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 0.7460510328068044, |
|
"grad_norm": 0.8280074000358582, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1851, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 0.7484811664641555, |
|
"grad_norm": 0.681713342666626, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1757, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 0.7509113001215066, |
|
"grad_norm": 4.372740268707275, |
|
"learning_rate": 5e-05, |
|
"loss": 0.104, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 0.7533414337788579, |
|
"grad_norm": 0.8002400398254395, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0922, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.755771567436209, |
|
"grad_norm": 0.7685486674308777, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1391, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 0.7582017010935601, |
|
"grad_norm": 1.282363772392273, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1141, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 0.7606318347509113, |
|
"grad_norm": 0.9099876284599304, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1396, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 0.7630619684082625, |
|
"grad_norm": 0.5911856293678284, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0631, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 0.7654921020656136, |
|
"grad_norm": 0.7148822546005249, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1714, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.7679222357229648, |
|
"grad_norm": 0.6756410002708435, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0894, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 0.7703523693803159, |
|
"grad_norm": 0.8938581347465515, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1538, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 0.772782503037667, |
|
"grad_norm": 1.061012625694275, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1069, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 0.7752126366950183, |
|
"grad_norm": 0.8262349963188171, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0745, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 0.7776427703523694, |
|
"grad_norm": 1.3609825372695923, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0689, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.7800729040097205, |
|
"grad_norm": 1.2151660919189453, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0863, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 0.7825030376670717, |
|
"grad_norm": 0.7165310978889465, |
|
"learning_rate": 5e-05, |
|
"loss": 0.084, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 0.7849331713244229, |
|
"grad_norm": 1.2151451110839844, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1229, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 0.787363304981774, |
|
"grad_norm": 0.9600043296813965, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0524, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 0.7897934386391251, |
|
"grad_norm": 1.1785893440246582, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1635, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.7922235722964763, |
|
"grad_norm": 0.6502353549003601, |
|
"learning_rate": 5e-05, |
|
"loss": 0.103, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 0.7946537059538274, |
|
"grad_norm": 0.6894869208335876, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0824, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 0.7970838396111786, |
|
"grad_norm": 0.8827742338180542, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0899, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 0.7995139732685298, |
|
"grad_norm": 0.82513827085495, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0642, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 0.8019441069258809, |
|
"grad_norm": 0.7896872758865356, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1869, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.804374240583232, |
|
"grad_norm": 1.421380639076233, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0668, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 0.8068043742405833, |
|
"grad_norm": 0.826987087726593, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1543, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 0.8092345078979344, |
|
"grad_norm": 0.7504023909568787, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1167, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 0.8116646415552855, |
|
"grad_norm": 0.9217634797096252, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0801, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 0.8140947752126367, |
|
"grad_norm": 0.6423681974411011, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0445, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.8165249088699879, |
|
"grad_norm": 0.3945442736148834, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0737, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 0.818955042527339, |
|
"grad_norm": 0.6396763324737549, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1118, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 0.8213851761846902, |
|
"grad_norm": 1.31553053855896, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0989, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 0.8238153098420413, |
|
"grad_norm": 0.962189793586731, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1043, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 0.8262454434993924, |
|
"grad_norm": 0.9305171370506287, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1333, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.8286755771567437, |
|
"grad_norm": 0.6102299690246582, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0743, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 0.8311057108140948, |
|
"grad_norm": 0.9117159247398376, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0399, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 0.8335358444714459, |
|
"grad_norm": 0.614378035068512, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0767, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 0.8359659781287971, |
|
"grad_norm": 0.934684157371521, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0814, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 0.8383961117861483, |
|
"grad_norm": 0.8567538261413574, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0866, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.8408262454434994, |
|
"grad_norm": 1.083308219909668, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0534, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 0.8432563791008505, |
|
"grad_norm": 0.8664528727531433, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1348, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 0.8456865127582017, |
|
"grad_norm": 0.7151762843132019, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0615, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 0.8481166464155528, |
|
"grad_norm": 0.784515917301178, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1014, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 0.850546780072904, |
|
"grad_norm": 0.7713536620140076, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0684, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.8529769137302552, |
|
"grad_norm": 0.8839780688285828, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2673, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 0.8554070473876063, |
|
"grad_norm": 0.6095230579376221, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1667, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 0.8578371810449574, |
|
"grad_norm": 1.1133002042770386, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1746, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 0.8602673147023087, |
|
"grad_norm": 0.5206813812255859, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1507, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 0.8626974483596598, |
|
"grad_norm": 0.6527222990989685, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1045, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.8651275820170109, |
|
"grad_norm": 0.6029911637306213, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1157, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 0.8675577156743621, |
|
"grad_norm": 0.7916355729103088, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0817, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 0.8699878493317132, |
|
"grad_norm": 0.6410179734230042, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0707, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 0.8724179829890644, |
|
"grad_norm": 1.0566637516021729, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0774, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 0.8748481166464156, |
|
"grad_norm": 0.6340796947479248, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1003, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.8772782503037667, |
|
"grad_norm": 0.6086530089378357, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0441, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 0.8797083839611178, |
|
"grad_norm": 0.8955773115158081, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1389, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 0.8821385176184691, |
|
"grad_norm": 0.4679882228374481, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0708, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 0.8845686512758202, |
|
"grad_norm": 0.6539987921714783, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1078, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 0.8869987849331713, |
|
"grad_norm": 0.8322189450263977, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0477, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.8894289185905225, |
|
"grad_norm": 0.5329148769378662, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0638, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 0.8918590522478737, |
|
"grad_norm": 0.7909948825836182, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0941, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 0.8942891859052248, |
|
"grad_norm": 0.6703280806541443, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0989, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 0.8967193195625759, |
|
"grad_norm": 0.6218695640563965, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0712, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 0.8991494532199271, |
|
"grad_norm": 1.6482758522033691, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1537, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.9015795868772782, |
|
"grad_norm": 0.7616936564445496, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1234, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 0.9040097205346294, |
|
"grad_norm": 0.7617467045783997, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0895, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 0.9064398541919806, |
|
"grad_norm": 0.7406112551689148, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0559, |
|
"step": 373 |
|
}, |
|
{ |
|
"epoch": 0.9088699878493317, |
|
"grad_norm": 1.1627577543258667, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0807, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 0.9113001215066828, |
|
"grad_norm": 0.6091077923774719, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0649, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.913730255164034, |
|
"grad_norm": 0.8935027122497559, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1563, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 0.9161603888213852, |
|
"grad_norm": 0.8287168741226196, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0725, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 0.9185905224787363, |
|
"grad_norm": 0.5430625677108765, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0984, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 0.9210206561360875, |
|
"grad_norm": 0.7538096308708191, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1231, |
|
"step": 379 |
|
}, |
|
{ |
|
"epoch": 0.9234507897934386, |
|
"grad_norm": 0.9483656287193298, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1341, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.9258809234507898, |
|
"grad_norm": 0.5689654350280762, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1116, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 0.928311057108141, |
|
"grad_norm": 0.987369179725647, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1815, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 0.9307411907654921, |
|
"grad_norm": 0.872460126876831, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0994, |
|
"step": 383 |
|
}, |
|
{ |
|
"epoch": 0.9331713244228432, |
|
"grad_norm": 0.785040020942688, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1006, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 0.9356014580801945, |
|
"grad_norm": 0.6228112578392029, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0805, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.9380315917375456, |
|
"grad_norm": 0.746306836605072, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0672, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 0.9404617253948967, |
|
"grad_norm": 0.44906482100486755, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0831, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 0.9428918590522479, |
|
"grad_norm": 0.7880069613456726, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0871, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 0.945321992709599, |
|
"grad_norm": 0.7579517960548401, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1306, |
|
"step": 389 |
|
}, |
|
{ |
|
"epoch": 0.9477521263669502, |
|
"grad_norm": 0.5668421387672424, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0738, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.9501822600243013, |
|
"grad_norm": 0.8249059319496155, |
|
"learning_rate": 5e-05, |
|
"loss": 0.069, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 0.9526123936816525, |
|
"grad_norm": 0.7063718438148499, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0488, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 0.9550425273390036, |
|
"grad_norm": 0.9006249904632568, |
|
"learning_rate": 5e-05, |
|
"loss": 0.029, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 0.9574726609963548, |
|
"grad_norm": 0.5778633952140808, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0859, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 0.959902794653706, |
|
"grad_norm": 0.9541341066360474, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0698, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.9623329283110571, |
|
"grad_norm": 0.7825549244880676, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0714, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 0.9647630619684082, |
|
"grad_norm": 0.729721188545227, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0651, |
|
"step": 397 |
|
}, |
|
{ |
|
"epoch": 0.9671931956257594, |
|
"grad_norm": 0.7052625417709351, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1518, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 0.9696233292831106, |
|
"grad_norm": 2.3800599575042725, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0919, |
|
"step": 399 |
|
}, |
|
{ |
|
"epoch": 0.9720534629404617, |
|
"grad_norm": 0.5317090749740601, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0828, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.9720534629404617, |
|
"eval_loss": 0.2335556447505951, |
|
"eval_runtime": 513.3338, |
|
"eval_samples_per_second": 5.353, |
|
"eval_steps_per_second": 0.67, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.9744835965978129, |
|
"grad_norm": 0.6768686771392822, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1675, |
|
"step": 401 |
|
}, |
|
{ |
|
"epoch": 0.976913730255164, |
|
"grad_norm": 0.5037873983383179, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0199, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 0.9793438639125152, |
|
"grad_norm": 1.7766000032424927, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0846, |
|
"step": 403 |
|
}, |
|
{ |
|
"epoch": 0.9817739975698664, |
|
"grad_norm": 0.978345513343811, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0593, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 0.9842041312272175, |
|
"grad_norm": 1.4286214113235474, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1963, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.9866342648845686, |
|
"grad_norm": 1.1499220132827759, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0786, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 0.9890643985419199, |
|
"grad_norm": 1.4914474487304688, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1152, |
|
"step": 407 |
|
}, |
|
{ |
|
"epoch": 0.991494532199271, |
|
"grad_norm": 0.7232803106307983, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1338, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 0.9939246658566221, |
|
"grad_norm": 0.8456845283508301, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0631, |
|
"step": 409 |
|
}, |
|
{ |
|
"epoch": 0.9963547995139733, |
|
"grad_norm": 0.5165804028511047, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0725, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.9987849331713244, |
|
"grad_norm": 0.8364010453224182, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1215, |
|
"step": 411 |
|
}, |
|
{ |
|
"epoch": 1.0012150668286757, |
|
"grad_norm": 0.6961699724197388, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0327, |
|
"step": 412 |
|
}, |
|
{ |
|
"epoch": 1.0036452004860268, |
|
"grad_norm": 0.9358980655670166, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0724, |
|
"step": 413 |
|
}, |
|
{ |
|
"epoch": 1.006075334143378, |
|
"grad_norm": 0.641076922416687, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0824, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 1.008505467800729, |
|
"grad_norm": 0.6918608546257019, |
|
"learning_rate": 5e-05, |
|
"loss": 0.083, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 1.0109356014580801, |
|
"grad_norm": 0.7481106519699097, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0887, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 1.0133657351154313, |
|
"grad_norm": 0.7577409148216248, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0684, |
|
"step": 417 |
|
}, |
|
{ |
|
"epoch": 1.0157958687727826, |
|
"grad_norm": 1.0815515518188477, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1167, |
|
"step": 418 |
|
}, |
|
{ |
|
"epoch": 1.0182260024301337, |
|
"grad_norm": 1.248225450515747, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0644, |
|
"step": 419 |
|
}, |
|
{ |
|
"epoch": 1.0206561360874848, |
|
"grad_norm": 0.7970739006996155, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1712, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 1.023086269744836, |
|
"grad_norm": 1.343601942062378, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1359, |
|
"step": 421 |
|
}, |
|
{ |
|
"epoch": 1.025516403402187, |
|
"grad_norm": 0.705987274646759, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0961, |
|
"step": 422 |
|
}, |
|
{ |
|
"epoch": 1.0279465370595382, |
|
"grad_norm": 0.5520442724227905, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0839, |
|
"step": 423 |
|
}, |
|
{ |
|
"epoch": 1.0303766707168893, |
|
"grad_norm": 0.5799084305763245, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0759, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 1.0328068043742407, |
|
"grad_norm": 0.610801637172699, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0593, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 1.0352369380315918, |
|
"grad_norm": 1.035436987876892, |
|
"learning_rate": 5e-05, |
|
"loss": 0.194, |
|
"step": 426 |
|
}, |
|
{ |
|
"epoch": 1.037667071688943, |
|
"grad_norm": 0.9973961114883423, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0954, |
|
"step": 427 |
|
}, |
|
{ |
|
"epoch": 1.040097205346294, |
|
"grad_norm": 0.8617413640022278, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0917, |
|
"step": 428 |
|
}, |
|
{ |
|
"epoch": 1.0425273390036451, |
|
"grad_norm": 0.9682510495185852, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1391, |
|
"step": 429 |
|
}, |
|
{ |
|
"epoch": 1.0449574726609963, |
|
"grad_norm": 0.3044584393501282, |
|
"learning_rate": 5e-05, |
|
"loss": 0.025, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 1.0473876063183476, |
|
"grad_norm": 0.6183155179023743, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0347, |
|
"step": 431 |
|
}, |
|
{ |
|
"epoch": 1.0498177399756987, |
|
"grad_norm": 0.37442171573638916, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0452, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 1.0522478736330498, |
|
"grad_norm": 0.565359890460968, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0943, |
|
"step": 433 |
|
}, |
|
{ |
|
"epoch": 1.054678007290401, |
|
"grad_norm": 0.6165122389793396, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0851, |
|
"step": 434 |
|
}, |
|
{ |
|
"epoch": 1.057108140947752, |
|
"grad_norm": 0.7150765061378479, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0704, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 1.0595382746051032, |
|
"grad_norm": 0.7983779311180115, |
|
"learning_rate": 5e-05, |
|
"loss": 0.121, |
|
"step": 436 |
|
}, |
|
{ |
|
"epoch": 1.0619684082624545, |
|
"grad_norm": 0.5900336503982544, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0938, |
|
"step": 437 |
|
}, |
|
{ |
|
"epoch": 1.0643985419198057, |
|
"grad_norm": 0.6962295174598694, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1227, |
|
"step": 438 |
|
}, |
|
{ |
|
"epoch": 1.0668286755771568, |
|
"grad_norm": 1.0119210481643677, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0737, |
|
"step": 439 |
|
}, |
|
{ |
|
"epoch": 1.069258809234508, |
|
"grad_norm": 0.4835350811481476, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0644, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 1.071688942891859, |
|
"grad_norm": 0.6343538761138916, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0796, |
|
"step": 441 |
|
}, |
|
{ |
|
"epoch": 1.0741190765492101, |
|
"grad_norm": 0.41634368896484375, |
|
"learning_rate": 5e-05, |
|
"loss": 0.087, |
|
"step": 442 |
|
}, |
|
{ |
|
"epoch": 1.0765492102065615, |
|
"grad_norm": 0.7903019189834595, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0564, |
|
"step": 443 |
|
}, |
|
{ |
|
"epoch": 1.0789793438639126, |
|
"grad_norm": 0.45615118741989136, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0928, |
|
"step": 444 |
|
}, |
|
{ |
|
"epoch": 1.0814094775212637, |
|
"grad_norm": 0.5224602222442627, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1134, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 1.0838396111786148, |
|
"grad_norm": 0.7974500060081482, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1592, |
|
"step": 446 |
|
}, |
|
{ |
|
"epoch": 1.086269744835966, |
|
"grad_norm": 0.545687735080719, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0892, |
|
"step": 447 |
|
}, |
|
{ |
|
"epoch": 1.088699878493317, |
|
"grad_norm": 0.88418048620224, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0406, |
|
"step": 448 |
|
}, |
|
{ |
|
"epoch": 1.0911300121506682, |
|
"grad_norm": 0.6175111532211304, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0536, |
|
"step": 449 |
|
}, |
|
{ |
|
"epoch": 1.0935601458080195, |
|
"grad_norm": 0.5780444741249084, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0806, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 1.0959902794653706, |
|
"grad_norm": 2.7306559085845947, |
|
"learning_rate": 5e-05, |
|
"loss": 0.102, |
|
"step": 451 |
|
}, |
|
{ |
|
"epoch": 1.0984204131227218, |
|
"grad_norm": 0.8727555871009827, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0723, |
|
"step": 452 |
|
}, |
|
{ |
|
"epoch": 1.1008505467800729, |
|
"grad_norm": 0.49243301153182983, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0589, |
|
"step": 453 |
|
}, |
|
{ |
|
"epoch": 1.103280680437424, |
|
"grad_norm": 1.0444567203521729, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1075, |
|
"step": 454 |
|
}, |
|
{ |
|
"epoch": 1.1057108140947751, |
|
"grad_norm": 0.5982443690299988, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0646, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 1.1081409477521265, |
|
"grad_norm": 0.4428943395614624, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0421, |
|
"step": 456 |
|
}, |
|
{ |
|
"epoch": 1.1105710814094776, |
|
"grad_norm": 0.612250566482544, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0694, |
|
"step": 457 |
|
}, |
|
{ |
|
"epoch": 1.1130012150668287, |
|
"grad_norm": 0.8445409536361694, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1089, |
|
"step": 458 |
|
}, |
|
{ |
|
"epoch": 1.1154313487241798, |
|
"grad_norm": 0.5859739184379578, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1143, |
|
"step": 459 |
|
}, |
|
{ |
|
"epoch": 1.117861482381531, |
|
"grad_norm": 0.38835424184799194, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0866, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 1.120291616038882, |
|
"grad_norm": 0.7463281750679016, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0815, |
|
"step": 461 |
|
}, |
|
{ |
|
"epoch": 1.1227217496962334, |
|
"grad_norm": 0.6839048266410828, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1349, |
|
"step": 462 |
|
}, |
|
{ |
|
"epoch": 1.1251518833535845, |
|
"grad_norm": 1.4645401239395142, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0799, |
|
"step": 463 |
|
}, |
|
{ |
|
"epoch": 1.1275820170109356, |
|
"grad_norm": 0.7507435083389282, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0598, |
|
"step": 464 |
|
}, |
|
{ |
|
"epoch": 1.1300121506682868, |
|
"grad_norm": 0.7818620800971985, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0938, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 1.1324422843256379, |
|
"grad_norm": 0.6816830635070801, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0787, |
|
"step": 466 |
|
}, |
|
{ |
|
"epoch": 1.134872417982989, |
|
"grad_norm": 0.575626015663147, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1461, |
|
"step": 467 |
|
}, |
|
{ |
|
"epoch": 1.13730255164034, |
|
"grad_norm": 0.73714679479599, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0761, |
|
"step": 468 |
|
}, |
|
{ |
|
"epoch": 1.1397326852976915, |
|
"grad_norm": 0.7288960814476013, |
|
"learning_rate": 5e-05, |
|
"loss": 0.074, |
|
"step": 469 |
|
}, |
|
{ |
|
"epoch": 1.1421628189550426, |
|
"grad_norm": 0.5384907126426697, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0732, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 1.1445929526123937, |
|
"grad_norm": 0.6071632504463196, |
|
"learning_rate": 5e-05, |
|
"loss": 0.03, |
|
"step": 471 |
|
}, |
|
{ |
|
"epoch": 1.1470230862697448, |
|
"grad_norm": 0.5744608044624329, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0912, |
|
"step": 472 |
|
}, |
|
{ |
|
"epoch": 1.149453219927096, |
|
"grad_norm": 0.9392772912979126, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1024, |
|
"step": 473 |
|
}, |
|
{ |
|
"epoch": 1.1518833535844473, |
|
"grad_norm": 1.170997142791748, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1027, |
|
"step": 474 |
|
}, |
|
{ |
|
"epoch": 1.1543134872417984, |
|
"grad_norm": 0.791221559047699, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1305, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 1.1567436208991495, |
|
"grad_norm": 0.5443055033683777, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0749, |
|
"step": 476 |
|
}, |
|
{ |
|
"epoch": 1.1591737545565006, |
|
"grad_norm": 0.8006246089935303, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0976, |
|
"step": 477 |
|
}, |
|
{ |
|
"epoch": 1.1616038882138517, |
|
"grad_norm": 0.47016945481300354, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0386, |
|
"step": 478 |
|
}, |
|
{ |
|
"epoch": 1.1640340218712029, |
|
"grad_norm": 0.7536148428916931, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1385, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 1.166464155528554, |
|
"grad_norm": 0.6575655341148376, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0851, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 1.1688942891859053, |
|
"grad_norm": 0.484667032957077, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0301, |
|
"step": 481 |
|
}, |
|
{ |
|
"epoch": 1.1713244228432564, |
|
"grad_norm": 0.8194088339805603, |
|
"learning_rate": 5e-05, |
|
"loss": 0.129, |
|
"step": 482 |
|
}, |
|
{ |
|
"epoch": 1.1737545565006076, |
|
"grad_norm": 0.6559669375419617, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0667, |
|
"step": 483 |
|
}, |
|
{ |
|
"epoch": 1.1761846901579587, |
|
"grad_norm": 0.8752015829086304, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0637, |
|
"step": 484 |
|
}, |
|
{ |
|
"epoch": 1.1786148238153098, |
|
"grad_norm": 0.6029524207115173, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0557, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 1.181044957472661, |
|
"grad_norm": 0.5431787371635437, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0338, |
|
"step": 486 |
|
}, |
|
{ |
|
"epoch": 1.183475091130012, |
|
"grad_norm": 0.5199941992759705, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0603, |
|
"step": 487 |
|
}, |
|
{ |
|
"epoch": 1.1859052247873634, |
|
"grad_norm": 0.520455539226532, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0221, |
|
"step": 488 |
|
}, |
|
{ |
|
"epoch": 1.1883353584447145, |
|
"grad_norm": 0.5124850273132324, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0306, |
|
"step": 489 |
|
}, |
|
{ |
|
"epoch": 1.1907654921020656, |
|
"grad_norm": 0.6515238881111145, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0612, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 1.1931956257594167, |
|
"grad_norm": 0.816235363483429, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1585, |
|
"step": 491 |
|
}, |
|
{ |
|
"epoch": 1.1956257594167679, |
|
"grad_norm": 0.6307040452957153, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0551, |
|
"step": 492 |
|
}, |
|
{ |
|
"epoch": 1.1980558930741192, |
|
"grad_norm": 0.46716415882110596, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0457, |
|
"step": 493 |
|
}, |
|
{ |
|
"epoch": 1.2004860267314703, |
|
"grad_norm": 0.5497251749038696, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0874, |
|
"step": 494 |
|
}, |
|
{ |
|
"epoch": 1.2029161603888214, |
|
"grad_norm": 1.0956586599349976, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0796, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 1.2053462940461726, |
|
"grad_norm": 0.6908737421035767, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0406, |
|
"step": 496 |
|
}, |
|
{ |
|
"epoch": 1.2077764277035237, |
|
"grad_norm": 0.8598071932792664, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1157, |
|
"step": 497 |
|
}, |
|
{ |
|
"epoch": 1.2102065613608748, |
|
"grad_norm": 0.6752752065658569, |
|
"learning_rate": 5e-05, |
|
"loss": 0.068, |
|
"step": 498 |
|
}, |
|
{ |
|
"epoch": 1.212636695018226, |
|
"grad_norm": 0.5425324440002441, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0976, |
|
"step": 499 |
|
}, |
|
{ |
|
"epoch": 1.2150668286755772, |
|
"grad_norm": 0.7367557883262634, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1028, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 1.2150668286755772, |
|
"eval_loss": 0.23548956215381622, |
|
"eval_runtime": 512.2983, |
|
"eval_samples_per_second": 5.364, |
|
"eval_steps_per_second": 0.671, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 1.2174969623329284, |
|
"grad_norm": 0.8351901769638062, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0927, |
|
"step": 501 |
|
}, |
|
{ |
|
"epoch": 1.2199270959902795, |
|
"grad_norm": 0.5388638377189636, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0956, |
|
"step": 502 |
|
}, |
|
{ |
|
"epoch": 1.2223572296476306, |
|
"grad_norm": 0.9194818139076233, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0533, |
|
"step": 503 |
|
}, |
|
{ |
|
"epoch": 1.2247873633049817, |
|
"grad_norm": 0.7008382081985474, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0554, |
|
"step": 504 |
|
}, |
|
{ |
|
"epoch": 1.2272174969623328, |
|
"grad_norm": 0.5537183284759521, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1277, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 1.229647630619684, |
|
"grad_norm": 0.570819079875946, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1037, |
|
"step": 506 |
|
}, |
|
{ |
|
"epoch": 1.2320777642770353, |
|
"grad_norm": 0.5867807269096375, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0578, |
|
"step": 507 |
|
}, |
|
{ |
|
"epoch": 1.2345078979343864, |
|
"grad_norm": 0.7660566568374634, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0856, |
|
"step": 508 |
|
}, |
|
{ |
|
"epoch": 1.2369380315917375, |
|
"grad_norm": 0.7493465542793274, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0567, |
|
"step": 509 |
|
}, |
|
{ |
|
"epoch": 1.2393681652490887, |
|
"grad_norm": 0.6718665957450867, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1113, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 1.2417982989064398, |
|
"grad_norm": 0.7018153667449951, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0966, |
|
"step": 511 |
|
}, |
|
{ |
|
"epoch": 1.2442284325637911, |
|
"grad_norm": 1.3940482139587402, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1352, |
|
"step": 512 |
|
}, |
|
{ |
|
"epoch": 1.2466585662211422, |
|
"grad_norm": 0.663314938545227, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0556, |
|
"step": 513 |
|
}, |
|
{ |
|
"epoch": 1.2490886998784934, |
|
"grad_norm": 1.0090134143829346, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0352, |
|
"step": 514 |
|
}, |
|
{ |
|
"epoch": 1.2515188335358445, |
|
"grad_norm": 0.5742064714431763, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0669, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 1.2539489671931956, |
|
"grad_norm": 0.726700484752655, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0533, |
|
"step": 516 |
|
}, |
|
{ |
|
"epoch": 1.2563791008505467, |
|
"grad_norm": 0.6788687109947205, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0567, |
|
"step": 517 |
|
}, |
|
{ |
|
"epoch": 1.2588092345078978, |
|
"grad_norm": 0.48673132061958313, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0406, |
|
"step": 518 |
|
}, |
|
{ |
|
"epoch": 1.2612393681652492, |
|
"grad_norm": 1.1016216278076172, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0668, |
|
"step": 519 |
|
}, |
|
{ |
|
"epoch": 1.2636695018226003, |
|
"grad_norm": 1.2345398664474487, |
|
"learning_rate": 5e-05, |
|
"loss": 0.029, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 1.2660996354799514, |
|
"grad_norm": 0.7507830858230591, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0991, |
|
"step": 521 |
|
}, |
|
{ |
|
"epoch": 1.2685297691373025, |
|
"grad_norm": 0.819266140460968, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0723, |
|
"step": 522 |
|
}, |
|
{ |
|
"epoch": 1.2709599027946537, |
|
"grad_norm": 0.9876227974891663, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1555, |
|
"step": 523 |
|
}, |
|
{ |
|
"epoch": 1.273390036452005, |
|
"grad_norm": 0.7902403473854065, |
|
"learning_rate": 5e-05, |
|
"loss": 0.156, |
|
"step": 524 |
|
}, |
|
{ |
|
"epoch": 1.2758201701093559, |
|
"grad_norm": 0.6342129111289978, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1161, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 1.2782503037667072, |
|
"grad_norm": 0.9039121866226196, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1078, |
|
"step": 526 |
|
}, |
|
{ |
|
"epoch": 1.2806804374240583, |
|
"grad_norm": 0.9820278882980347, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1182, |
|
"step": 527 |
|
}, |
|
{ |
|
"epoch": 1.2831105710814095, |
|
"grad_norm": 0.45999497175216675, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0469, |
|
"step": 528 |
|
}, |
|
{ |
|
"epoch": 1.2855407047387606, |
|
"grad_norm": 0.6754646301269531, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0875, |
|
"step": 529 |
|
}, |
|
{ |
|
"epoch": 1.2879708383961117, |
|
"grad_norm": 0.7348721027374268, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0639, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 1.290400972053463, |
|
"grad_norm": 0.7348721027374268, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0764, |
|
"step": 531 |
|
}, |
|
{ |
|
"epoch": 1.2928311057108142, |
|
"grad_norm": 0.4331410229206085, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0278, |
|
"step": 532 |
|
}, |
|
{ |
|
"epoch": 1.2952612393681653, |
|
"grad_norm": 0.7986178994178772, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0893, |
|
"step": 533 |
|
}, |
|
{ |
|
"epoch": 1.2976913730255164, |
|
"grad_norm": 0.7255426049232483, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0802, |
|
"step": 534 |
|
}, |
|
{ |
|
"epoch": 1.3001215066828675, |
|
"grad_norm": 0.590578556060791, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0853, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 1.3025516403402186, |
|
"grad_norm": 0.6147257685661316, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0462, |
|
"step": 536 |
|
}, |
|
{ |
|
"epoch": 1.3049817739975698, |
|
"grad_norm": 0.6174589991569519, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0758, |
|
"step": 537 |
|
}, |
|
{ |
|
"epoch": 1.307411907654921, |
|
"grad_norm": 0.5428439378738403, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1452, |
|
"step": 538 |
|
}, |
|
{ |
|
"epoch": 1.3098420413122722, |
|
"grad_norm": 0.554800271987915, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0266, |
|
"step": 539 |
|
}, |
|
{ |
|
"epoch": 1.3122721749696233, |
|
"grad_norm": 0.627275824546814, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0604, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 1.3147023086269745, |
|
"grad_norm": 0.5268173813819885, |
|
"learning_rate": 5e-05, |
|
"loss": 0.036, |
|
"step": 541 |
|
}, |
|
{ |
|
"epoch": 1.3171324422843256, |
|
"grad_norm": 0.4525201916694641, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0746, |
|
"step": 542 |
|
}, |
|
{ |
|
"epoch": 1.319562575941677, |
|
"grad_norm": 1.299209475517273, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0573, |
|
"step": 543 |
|
}, |
|
{ |
|
"epoch": 1.3219927095990278, |
|
"grad_norm": 0.5899752378463745, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0373, |
|
"step": 544 |
|
}, |
|
{ |
|
"epoch": 1.3244228432563792, |
|
"grad_norm": 0.5044320225715637, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0415, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 1.3268529769137303, |
|
"grad_norm": 0.4812248945236206, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0898, |
|
"step": 546 |
|
}, |
|
{ |
|
"epoch": 1.3292831105710814, |
|
"grad_norm": 0.47461068630218506, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0248, |
|
"step": 547 |
|
}, |
|
{ |
|
"epoch": 1.3317132442284325, |
|
"grad_norm": 0.5620236396789551, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0418, |
|
"step": 548 |
|
}, |
|
{ |
|
"epoch": 1.3341433778857836, |
|
"grad_norm": 0.8658341765403748, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0208, |
|
"step": 549 |
|
}, |
|
{ |
|
"epoch": 1.336573511543135, |
|
"grad_norm": 0.6640876531600952, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0652, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 1.339003645200486, |
|
"grad_norm": 0.3873005509376526, |
|
"learning_rate": 5e-05, |
|
"loss": 0.018, |
|
"step": 551 |
|
}, |
|
{ |
|
"epoch": 1.3414337788578372, |
|
"grad_norm": 0.4546399414539337, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0823, |
|
"step": 552 |
|
}, |
|
{ |
|
"epoch": 1.3438639125151883, |
|
"grad_norm": 0.5188842415809631, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0604, |
|
"step": 553 |
|
}, |
|
{ |
|
"epoch": 1.3462940461725394, |
|
"grad_norm": 0.41030353307724, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0112, |
|
"step": 554 |
|
}, |
|
{ |
|
"epoch": 1.3487241798298906, |
|
"grad_norm": 0.7154520750045776, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0687, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 1.3511543134872417, |
|
"grad_norm": 0.6743410229682922, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1262, |
|
"step": 556 |
|
}, |
|
{ |
|
"epoch": 1.353584447144593, |
|
"grad_norm": 0.38846057653427124, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0371, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 1.3560145808019441, |
|
"grad_norm": 0.5022516846656799, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0159, |
|
"step": 558 |
|
}, |
|
{ |
|
"epoch": 1.3584447144592953, |
|
"grad_norm": 0.5445972084999084, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0652, |
|
"step": 559 |
|
}, |
|
{ |
|
"epoch": 1.3608748481166464, |
|
"grad_norm": 0.6346146464347839, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1128, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 1.3633049817739975, |
|
"grad_norm": 0.7129392623901367, |
|
"learning_rate": 5e-05, |
|
"loss": 0.059, |
|
"step": 561 |
|
}, |
|
{ |
|
"epoch": 1.3657351154313488, |
|
"grad_norm": 0.4905102252960205, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1114, |
|
"step": 562 |
|
}, |
|
{ |
|
"epoch": 1.3681652490886997, |
|
"grad_norm": 0.7290562391281128, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1029, |
|
"step": 563 |
|
}, |
|
{ |
|
"epoch": 1.370595382746051, |
|
"grad_norm": 0.5576648116111755, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0383, |
|
"step": 564 |
|
}, |
|
{ |
|
"epoch": 1.3730255164034022, |
|
"grad_norm": 0.595310389995575, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0906, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 1.3754556500607533, |
|
"grad_norm": 0.7409107089042664, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0239, |
|
"step": 566 |
|
}, |
|
{ |
|
"epoch": 1.3778857837181044, |
|
"grad_norm": 1.0220900774002075, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0537, |
|
"step": 567 |
|
}, |
|
{ |
|
"epoch": 1.3803159173754556, |
|
"grad_norm": 1.0970425605773926, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0615, |
|
"step": 568 |
|
}, |
|
{ |
|
"epoch": 1.382746051032807, |
|
"grad_norm": 0.45453736186027527, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1, |
|
"step": 569 |
|
}, |
|
{ |
|
"epoch": 1.385176184690158, |
|
"grad_norm": 0.61972576379776, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1029, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 1.3876063183475091, |
|
"grad_norm": 0.47734493017196655, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0683, |
|
"step": 571 |
|
}, |
|
{ |
|
"epoch": 1.3900364520048603, |
|
"grad_norm": 0.3732990026473999, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0706, |
|
"step": 572 |
|
}, |
|
{ |
|
"epoch": 1.3924665856622114, |
|
"grad_norm": 0.49241572618484497, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0667, |
|
"step": 573 |
|
}, |
|
{ |
|
"epoch": 1.3948967193195625, |
|
"grad_norm": 0.6907554268836975, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0602, |
|
"step": 574 |
|
}, |
|
{ |
|
"epoch": 1.3973268529769136, |
|
"grad_norm": 0.8935093879699707, |
|
"learning_rate": 5e-05, |
|
"loss": 0.2114, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 1.399756986634265, |
|
"grad_norm": 0.2839941680431366, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0257, |
|
"step": 576 |
|
}, |
|
{ |
|
"epoch": 1.402187120291616, |
|
"grad_norm": 0.7202839255332947, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0878, |
|
"step": 577 |
|
}, |
|
{ |
|
"epoch": 1.4046172539489672, |
|
"grad_norm": 0.6520135402679443, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0295, |
|
"step": 578 |
|
}, |
|
{ |
|
"epoch": 1.4070473876063183, |
|
"grad_norm": 0.7360886335372925, |
|
"learning_rate": 5e-05, |
|
"loss": 0.065, |
|
"step": 579 |
|
}, |
|
{ |
|
"epoch": 1.4094775212636694, |
|
"grad_norm": 0.5855164527893066, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0752, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 1.4119076549210208, |
|
"grad_norm": 0.38474756479263306, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0541, |
|
"step": 581 |
|
}, |
|
{ |
|
"epoch": 1.414337788578372, |
|
"grad_norm": 0.5105370283126831, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0749, |
|
"step": 582 |
|
}, |
|
{ |
|
"epoch": 1.416767922235723, |
|
"grad_norm": 0.5076673030853271, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0528, |
|
"step": 583 |
|
}, |
|
{ |
|
"epoch": 1.4191980558930741, |
|
"grad_norm": 0.5034826397895813, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0505, |
|
"step": 584 |
|
}, |
|
{ |
|
"epoch": 1.4216281895504252, |
|
"grad_norm": 0.6378771066665649, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0676, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 1.4240583232077764, |
|
"grad_norm": 0.5122814774513245, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0515, |
|
"step": 586 |
|
}, |
|
{ |
|
"epoch": 1.4264884568651275, |
|
"grad_norm": 0.5807358026504517, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0507, |
|
"step": 587 |
|
}, |
|
{ |
|
"epoch": 1.4289185905224788, |
|
"grad_norm": 0.5156981348991394, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0503, |
|
"step": 588 |
|
}, |
|
{ |
|
"epoch": 1.43134872417983, |
|
"grad_norm": 0.9181635975837708, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0964, |
|
"step": 589 |
|
}, |
|
{ |
|
"epoch": 1.433778857837181, |
|
"grad_norm": 0.38104763627052307, |
|
"learning_rate": 5e-05, |
|
"loss": 0.055, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 1.4362089914945322, |
|
"grad_norm": 0.4987882673740387, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0347, |
|
"step": 591 |
|
}, |
|
{ |
|
"epoch": 1.4386391251518833, |
|
"grad_norm": 0.5201525688171387, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0904, |
|
"step": 592 |
|
}, |
|
{ |
|
"epoch": 1.4410692588092346, |
|
"grad_norm": 0.5238099694252014, |
|
"learning_rate": 5e-05, |
|
"loss": 0.077, |
|
"step": 593 |
|
}, |
|
{ |
|
"epoch": 1.4434993924665855, |
|
"grad_norm": 0.924893319606781, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1248, |
|
"step": 594 |
|
}, |
|
{ |
|
"epoch": 1.4459295261239369, |
|
"grad_norm": 0.6318195462226868, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0267, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 1.448359659781288, |
|
"grad_norm": 0.9913945198059082, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0917, |
|
"step": 596 |
|
}, |
|
{ |
|
"epoch": 1.4507897934386391, |
|
"grad_norm": 0.4909703731536865, |
|
"learning_rate": 5e-05, |
|
"loss": 0.036, |
|
"step": 597 |
|
}, |
|
{ |
|
"epoch": 1.4532199270959902, |
|
"grad_norm": 0.5913828015327454, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0353, |
|
"step": 598 |
|
}, |
|
{ |
|
"epoch": 1.4556500607533414, |
|
"grad_norm": 0.3615020215511322, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0517, |
|
"step": 599 |
|
}, |
|
{ |
|
"epoch": 1.4580801944106927, |
|
"grad_norm": 0.5106006264686584, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0444, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 1.4580801944106927, |
|
"eval_loss": 0.22935771942138672, |
|
"eval_runtime": 507.4275, |
|
"eval_samples_per_second": 5.416, |
|
"eval_steps_per_second": 0.678, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 1.4605103280680438, |
|
"grad_norm": 0.7134807705879211, |
|
"learning_rate": 5e-05, |
|
"loss": 0.056, |
|
"step": 601 |
|
}, |
|
{ |
|
"epoch": 1.462940461725395, |
|
"grad_norm": 0.2919784188270569, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0146, |
|
"step": 602 |
|
}, |
|
{ |
|
"epoch": 1.465370595382746, |
|
"grad_norm": 0.6526296734809875, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1207, |
|
"step": 603 |
|
}, |
|
{ |
|
"epoch": 1.4678007290400972, |
|
"grad_norm": 0.5833859443664551, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0657, |
|
"step": 604 |
|
}, |
|
{ |
|
"epoch": 1.4702308626974483, |
|
"grad_norm": 0.6107626557350159, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0655, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 1.4726609963547994, |
|
"grad_norm": 0.5227603912353516, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0504, |
|
"step": 606 |
|
}, |
|
{ |
|
"epoch": 1.4750911300121508, |
|
"grad_norm": 0.6677300333976746, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0465, |
|
"step": 607 |
|
}, |
|
{ |
|
"epoch": 1.4775212636695019, |
|
"grad_norm": 0.7898095846176147, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0932, |
|
"step": 608 |
|
}, |
|
{ |
|
"epoch": 1.479951397326853, |
|
"grad_norm": 0.9366335868835449, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0819, |
|
"step": 609 |
|
}, |
|
{ |
|
"epoch": 1.482381530984204, |
|
"grad_norm": 0.6056090593338013, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0638, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 1.4848116646415552, |
|
"grad_norm": 0.7640931606292725, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1152, |
|
"step": 611 |
|
}, |
|
{ |
|
"epoch": 1.4872417982989066, |
|
"grad_norm": 0.7496662735939026, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0461, |
|
"step": 612 |
|
}, |
|
{ |
|
"epoch": 1.4896719319562575, |
|
"grad_norm": 0.6870647072792053, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0812, |
|
"step": 613 |
|
}, |
|
{ |
|
"epoch": 1.4921020656136088, |
|
"grad_norm": 0.7589767575263977, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0846, |
|
"step": 614 |
|
}, |
|
{ |
|
"epoch": 1.49453219927096, |
|
"grad_norm": 0.681777834892273, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0407, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 1.496962332928311, |
|
"grad_norm": 0.7369627952575684, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0955, |
|
"step": 616 |
|
}, |
|
{ |
|
"epoch": 1.4993924665856622, |
|
"grad_norm": 0.7186546921730042, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0745, |
|
"step": 617 |
|
}, |
|
{ |
|
"epoch": 1.5018226002430133, |
|
"grad_norm": 0.6271875500679016, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0533, |
|
"step": 618 |
|
}, |
|
{ |
|
"epoch": 1.5042527339003646, |
|
"grad_norm": 0.5378838777542114, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0688, |
|
"step": 619 |
|
}, |
|
{ |
|
"epoch": 1.5066828675577155, |
|
"grad_norm": 0.7019214630126953, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0369, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 1.5091130012150669, |
|
"grad_norm": 0.9618707299232483, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0469, |
|
"step": 621 |
|
}, |
|
{ |
|
"epoch": 1.511543134872418, |
|
"grad_norm": 0.4820462167263031, |
|
"learning_rate": 5e-05, |
|
"loss": 0.074, |
|
"step": 622 |
|
}, |
|
{ |
|
"epoch": 1.513973268529769, |
|
"grad_norm": 0.411123126745224, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0307, |
|
"step": 623 |
|
}, |
|
{ |
|
"epoch": 1.5164034021871204, |
|
"grad_norm": 0.5718781352043152, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0276, |
|
"step": 624 |
|
}, |
|
{ |
|
"epoch": 1.5188335358444713, |
|
"grad_norm": 0.4341799020767212, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0651, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 1.5212636695018227, |
|
"grad_norm": 0.515855073928833, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0413, |
|
"step": 626 |
|
}, |
|
{ |
|
"epoch": 1.5236938031591738, |
|
"grad_norm": 0.6174566149711609, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0307, |
|
"step": 627 |
|
}, |
|
{ |
|
"epoch": 1.526123936816525, |
|
"grad_norm": 0.5931458473205566, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0852, |
|
"step": 628 |
|
}, |
|
{ |
|
"epoch": 1.528554070473876, |
|
"grad_norm": 0.46777036786079407, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0438, |
|
"step": 629 |
|
}, |
|
{ |
|
"epoch": 1.5309842041312272, |
|
"grad_norm": 0.4254392087459564, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0606, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 1.5334143377885785, |
|
"grad_norm": 0.5670008063316345, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0514, |
|
"step": 631 |
|
}, |
|
{ |
|
"epoch": 1.5358444714459294, |
|
"grad_norm": 1.1566509008407593, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1133, |
|
"step": 632 |
|
}, |
|
{ |
|
"epoch": 1.5382746051032807, |
|
"grad_norm": 0.5660587549209595, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0544, |
|
"step": 633 |
|
}, |
|
{ |
|
"epoch": 1.5407047387606319, |
|
"grad_norm": 0.5406026840209961, |
|
"learning_rate": 5e-05, |
|
"loss": 0.046, |
|
"step": 634 |
|
}, |
|
{ |
|
"epoch": 1.543134872417983, |
|
"grad_norm": 0.49723613262176514, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0177, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 1.545565006075334, |
|
"grad_norm": 1.263848900794983, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0659, |
|
"step": 636 |
|
}, |
|
{ |
|
"epoch": 1.5479951397326852, |
|
"grad_norm": 0.6971784830093384, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0747, |
|
"step": 637 |
|
}, |
|
{ |
|
"epoch": 1.5504252733900366, |
|
"grad_norm": 0.5171895623207092, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0512, |
|
"step": 638 |
|
}, |
|
{ |
|
"epoch": 1.5528554070473874, |
|
"grad_norm": 0.7318109273910522, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0655, |
|
"step": 639 |
|
}, |
|
{ |
|
"epoch": 1.5552855407047388, |
|
"grad_norm": 0.46023720502853394, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0568, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 1.55771567436209, |
|
"grad_norm": 0.7200696468353271, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0511, |
|
"step": 641 |
|
}, |
|
{ |
|
"epoch": 1.560145808019441, |
|
"grad_norm": 1.0338624715805054, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0822, |
|
"step": 642 |
|
}, |
|
{ |
|
"epoch": 1.5625759416767924, |
|
"grad_norm": 0.7976377010345459, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0623, |
|
"step": 643 |
|
}, |
|
{ |
|
"epoch": 1.5650060753341433, |
|
"grad_norm": 0.6757969260215759, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0737, |
|
"step": 644 |
|
}, |
|
{ |
|
"epoch": 1.5674362089914946, |
|
"grad_norm": 1.4083917140960693, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0374, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 1.5698663426488457, |
|
"grad_norm": 0.348634272813797, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0249, |
|
"step": 646 |
|
}, |
|
{ |
|
"epoch": 1.5722964763061968, |
|
"grad_norm": 0.7224584817886353, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1013, |
|
"step": 647 |
|
}, |
|
{ |
|
"epoch": 1.574726609963548, |
|
"grad_norm": 0.919998288154602, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1093, |
|
"step": 648 |
|
}, |
|
{ |
|
"epoch": 1.577156743620899, |
|
"grad_norm": 0.7341464161872864, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1234, |
|
"step": 649 |
|
}, |
|
{ |
|
"epoch": 1.5795868772782504, |
|
"grad_norm": 0.8548154234886169, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1628, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 1.5820170109356013, |
|
"grad_norm": 0.7931060194969177, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1439, |
|
"step": 651 |
|
}, |
|
{ |
|
"epoch": 1.5844471445929527, |
|
"grad_norm": 0.9627839922904968, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0327, |
|
"step": 652 |
|
}, |
|
{ |
|
"epoch": 1.5868772782503038, |
|
"grad_norm": 0.8406124114990234, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0574, |
|
"step": 653 |
|
}, |
|
{ |
|
"epoch": 1.589307411907655, |
|
"grad_norm": 0.7667601108551025, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0551, |
|
"step": 654 |
|
}, |
|
{ |
|
"epoch": 1.5917375455650062, |
|
"grad_norm": 0.4747026264667511, |
|
"learning_rate": 5e-05, |
|
"loss": 0.052, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 1.5941676792223571, |
|
"grad_norm": 0.632870078086853, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0587, |
|
"step": 656 |
|
}, |
|
{ |
|
"epoch": 1.5965978128797085, |
|
"grad_norm": 0.7700726985931396, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0986, |
|
"step": 657 |
|
}, |
|
{ |
|
"epoch": 1.5990279465370594, |
|
"grad_norm": 1.121201992034912, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0318, |
|
"step": 658 |
|
}, |
|
{ |
|
"epoch": 1.6014580801944107, |
|
"grad_norm": 0.6213607788085938, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0932, |
|
"step": 659 |
|
}, |
|
{ |
|
"epoch": 1.6038882138517618, |
|
"grad_norm": 0.30946460366249084, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0333, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 1.606318347509113, |
|
"grad_norm": 0.5171956419944763, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0617, |
|
"step": 661 |
|
}, |
|
{ |
|
"epoch": 1.6087484811664643, |
|
"grad_norm": 0.6288155317306519, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0506, |
|
"step": 662 |
|
}, |
|
{ |
|
"epoch": 1.6111786148238152, |
|
"grad_norm": 0.6943643689155579, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0285, |
|
"step": 663 |
|
}, |
|
{ |
|
"epoch": 1.6136087484811665, |
|
"grad_norm": 0.8144752383232117, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0484, |
|
"step": 664 |
|
}, |
|
{ |
|
"epoch": 1.6160388821385177, |
|
"grad_norm": 0.4773041307926178, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0993, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 1.6184690157958688, |
|
"grad_norm": 0.7703022956848145, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0735, |
|
"step": 666 |
|
}, |
|
{ |
|
"epoch": 1.62089914945322, |
|
"grad_norm": 0.6951743960380554, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0446, |
|
"step": 667 |
|
}, |
|
{ |
|
"epoch": 1.623329283110571, |
|
"grad_norm": 0.4781995415687561, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0178, |
|
"step": 668 |
|
}, |
|
{ |
|
"epoch": 1.6257594167679223, |
|
"grad_norm": 0.5066447854042053, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0969, |
|
"step": 669 |
|
}, |
|
{ |
|
"epoch": 1.6281895504252732, |
|
"grad_norm": 0.4288381040096283, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0317, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 1.6306196840826246, |
|
"grad_norm": 0.5544601678848267, |
|
"learning_rate": 5e-05, |
|
"loss": 0.035, |
|
"step": 671 |
|
}, |
|
{ |
|
"epoch": 1.6330498177399757, |
|
"grad_norm": 0.48424339294433594, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0231, |
|
"step": 672 |
|
}, |
|
{ |
|
"epoch": 1.6354799513973268, |
|
"grad_norm": 0.770346462726593, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0767, |
|
"step": 673 |
|
}, |
|
{ |
|
"epoch": 1.6379100850546782, |
|
"grad_norm": 0.7101414203643799, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0247, |
|
"step": 674 |
|
}, |
|
{ |
|
"epoch": 1.640340218712029, |
|
"grad_norm": 0.9238234758377075, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0528, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 1.6427703523693804, |
|
"grad_norm": 0.7192332148551941, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0738, |
|
"step": 676 |
|
}, |
|
{ |
|
"epoch": 1.6452004860267315, |
|
"grad_norm": 0.5393381118774414, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0337, |
|
"step": 677 |
|
}, |
|
{ |
|
"epoch": 1.6476306196840826, |
|
"grad_norm": 0.2713688015937805, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0159, |
|
"step": 678 |
|
}, |
|
{ |
|
"epoch": 1.6500607533414338, |
|
"grad_norm": 0.3296976089477539, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0608, |
|
"step": 679 |
|
}, |
|
{ |
|
"epoch": 1.6524908869987849, |
|
"grad_norm": 0.4995911717414856, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0119, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 1.6549210206561362, |
|
"grad_norm": 0.5983487367630005, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0188, |
|
"step": 681 |
|
}, |
|
{ |
|
"epoch": 1.6573511543134871, |
|
"grad_norm": 0.4184721112251282, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0207, |
|
"step": 682 |
|
}, |
|
{ |
|
"epoch": 1.6597812879708385, |
|
"grad_norm": 0.6056463122367859, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0321, |
|
"step": 683 |
|
}, |
|
{ |
|
"epoch": 1.6622114216281896, |
|
"grad_norm": 0.8600037693977356, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0716, |
|
"step": 684 |
|
}, |
|
{ |
|
"epoch": 1.6646415552855407, |
|
"grad_norm": 0.7863733768463135, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1251, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 1.6670716889428918, |
|
"grad_norm": 0.4304426908493042, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0824, |
|
"step": 686 |
|
}, |
|
{ |
|
"epoch": 1.669501822600243, |
|
"grad_norm": 0.6984259486198425, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0332, |
|
"step": 687 |
|
}, |
|
{ |
|
"epoch": 1.6719319562575943, |
|
"grad_norm": 0.400698184967041, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0444, |
|
"step": 688 |
|
}, |
|
{ |
|
"epoch": 1.6743620899149452, |
|
"grad_norm": 0.5730535387992859, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0488, |
|
"step": 689 |
|
}, |
|
{ |
|
"epoch": 1.6767922235722965, |
|
"grad_norm": 0.4522067606449127, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0596, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 1.6792223572296476, |
|
"grad_norm": 0.44992417097091675, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0311, |
|
"step": 691 |
|
}, |
|
{ |
|
"epoch": 1.6816524908869988, |
|
"grad_norm": 0.8305267095565796, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0933, |
|
"step": 692 |
|
}, |
|
{ |
|
"epoch": 1.68408262454435, |
|
"grad_norm": 0.7188434600830078, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0443, |
|
"step": 693 |
|
}, |
|
{ |
|
"epoch": 1.686512758201701, |
|
"grad_norm": 0.5052474141120911, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1067, |
|
"step": 694 |
|
}, |
|
{ |
|
"epoch": 1.6889428918590523, |
|
"grad_norm": 0.686128556728363, |
|
"learning_rate": 5e-05, |
|
"loss": 0.055, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 1.6913730255164034, |
|
"grad_norm": 0.4116796851158142, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0728, |
|
"step": 696 |
|
}, |
|
{ |
|
"epoch": 1.6938031591737546, |
|
"grad_norm": 0.44450122117996216, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0359, |
|
"step": 697 |
|
}, |
|
{ |
|
"epoch": 1.6962332928311057, |
|
"grad_norm": 0.3920440375804901, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0654, |
|
"step": 698 |
|
}, |
|
{ |
|
"epoch": 1.6986634264884568, |
|
"grad_norm": 0.7238947749137878, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0593, |
|
"step": 699 |
|
}, |
|
{ |
|
"epoch": 1.7010935601458081, |
|
"grad_norm": 0.3644774854183197, |
|
"learning_rate": 5e-05, |
|
"loss": 0.054, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 1.7010935601458081, |
|
"eval_loss": 0.23474913835525513, |
|
"eval_runtime": 505.0402, |
|
"eval_samples_per_second": 5.441, |
|
"eval_steps_per_second": 0.681, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 1.7010935601458081, |
|
"step": 700, |
|
"total_flos": 2.1333404857820774e+18, |
|
"train_loss": 0.12394342508699213, |
|
"train_runtime": 14191.792, |
|
"train_samples_per_second": 2.255, |
|
"train_steps_per_second": 0.07 |
|
} |
|
], |
|
"logging_steps": 1.0, |
|
"max_steps": 1000, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 100, |
|
"total_flos": 2.1333404857820774e+18, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|