|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 2.0, |
|
"eval_steps": 25, |
|
"global_step": 812, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"grad_norm": 33.25565995239216, |
|
"learning_rate": 2.439024390243903e-07, |
|
"loss": 1.8579, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 20.62131418379264, |
|
"learning_rate": 1.2195121951219514e-06, |
|
"loss": 1.8594, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 9.116745598504249, |
|
"learning_rate": 2.4390243902439027e-06, |
|
"loss": 1.6368, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 4.845477457562938, |
|
"learning_rate": 3.6585365853658537e-06, |
|
"loss": 1.5273, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 4.064982240065612, |
|
"learning_rate": 4.8780487804878055e-06, |
|
"loss": 1.4879, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 3.842051647757176, |
|
"learning_rate": 6.0975609756097564e-06, |
|
"loss": 1.4854, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"eval_loss": 1.4539871215820312, |
|
"eval_runtime": 37.3786, |
|
"eval_samples_per_second": 35.047, |
|
"eval_steps_per_second": 1.097, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 3.5736405142429235, |
|
"learning_rate": 7.317073170731707e-06, |
|
"loss": 1.4689, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 3.601301646883329, |
|
"learning_rate": 8.536585365853658e-06, |
|
"loss": 1.4492, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 3.5795153985197454, |
|
"learning_rate": 9.756097560975611e-06, |
|
"loss": 1.4464, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 3.547211967993725, |
|
"learning_rate": 1.0975609756097562e-05, |
|
"loss": 1.4118, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 3.8241967984903127, |
|
"learning_rate": 1.2195121951219513e-05, |
|
"loss": 1.43, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"eval_loss": 1.4291119575500488, |
|
"eval_runtime": 37.1035, |
|
"eval_samples_per_second": 35.307, |
|
"eval_steps_per_second": 1.105, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 6.65980179405396, |
|
"learning_rate": 1.3414634146341466e-05, |
|
"loss": 1.4254, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 4.155175626685716, |
|
"learning_rate": 1.4634146341463415e-05, |
|
"loss": 1.4337, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 3.6031251352197695, |
|
"learning_rate": 1.585365853658537e-05, |
|
"loss": 1.4482, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 3.9000471804568297, |
|
"learning_rate": 1.7073170731707317e-05, |
|
"loss": 1.452, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 3.6748440024674487, |
|
"learning_rate": 1.829268292682927e-05, |
|
"loss": 1.4188, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"eval_loss": 1.4451020956039429, |
|
"eval_runtime": 36.3992, |
|
"eval_samples_per_second": 35.99, |
|
"eval_steps_per_second": 1.126, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 4.784684166408002, |
|
"learning_rate": 1.9512195121951222e-05, |
|
"loss": 1.422, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 3.777533605838261, |
|
"learning_rate": 1.9999166586547382e-05, |
|
"loss": 1.4567, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 3.551321168481626, |
|
"learning_rate": 1.999407400739705e-05, |
|
"loss": 1.4674, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 3.625908306378223, |
|
"learning_rate": 1.9984354211555646e-05, |
|
"loss": 1.4554, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 4.243745995028721, |
|
"learning_rate": 1.997001169925015e-05, |
|
"loss": 1.4465, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"eval_loss": 1.4564708471298218, |
|
"eval_runtime": 35.8982, |
|
"eval_samples_per_second": 36.492, |
|
"eval_steps_per_second": 1.142, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 3.7056910011856434, |
|
"learning_rate": 1.995105311100698e-05, |
|
"loss": 1.435, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 4.028708939881987, |
|
"learning_rate": 1.9927487224577402e-05, |
|
"loss": 1.4554, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 3.4936872675120205, |
|
"learning_rate": 1.9899324950873532e-05, |
|
"loss": 1.4371, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 5.446138884571553, |
|
"learning_rate": 1.986657932891657e-05, |
|
"loss": 1.4213, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 3.5762471286009836, |
|
"learning_rate": 1.9829265519799824e-05, |
|
"loss": 1.4499, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"eval_loss": 1.4525413513183594, |
|
"eval_runtime": 36.7696, |
|
"eval_samples_per_second": 35.627, |
|
"eval_steps_per_second": 1.115, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 3.651881589784425, |
|
"learning_rate": 1.9787400799669155e-05, |
|
"loss": 1.4406, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 3.5390077509830626, |
|
"learning_rate": 1.9741004551724208e-05, |
|
"loss": 1.4683, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 3.654802825210285, |
|
"learning_rate": 1.9690098257244063e-05, |
|
"loss": 1.46, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 4.828779092366187, |
|
"learning_rate": 1.963470548564149e-05, |
|
"loss": 1.457, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 3.221793591458839, |
|
"learning_rate": 1.9574851883550395e-05, |
|
"loss": 1.4387, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"eval_loss": 1.4409288167953491, |
|
"eval_runtime": 36.7378, |
|
"eval_samples_per_second": 35.658, |
|
"eval_steps_per_second": 1.116, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 4.429483358899047, |
|
"learning_rate": 1.9510565162951538e-05, |
|
"loss": 1.4227, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 3.3883915102781494, |
|
"learning_rate": 1.9441875088342e-05, |
|
"loss": 1.4335, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 3.4947168397325368, |
|
"learning_rate": 1.9368813462954316e-05, |
|
"loss": 1.4165, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 3.498600082458413, |
|
"learning_rate": 1.9291414114031744e-05, |
|
"loss": 1.411, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 3.540671355408263, |
|
"learning_rate": 1.9209712877166347e-05, |
|
"loss": 1.4368, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"eval_loss": 1.4226408004760742, |
|
"eval_runtime": 36.4019, |
|
"eval_samples_per_second": 35.987, |
|
"eval_steps_per_second": 1.126, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 324.64947934206117, |
|
"learning_rate": 1.9123747579707275e-05, |
|
"loss": 1.3938, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 9.509687913297793, |
|
"learning_rate": 1.9033558023246845e-05, |
|
"loss": 1.4201, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 5.133252955012539, |
|
"learning_rate": 1.8939185965192572e-05, |
|
"loss": 1.413, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 5.023875370016508, |
|
"learning_rate": 1.8840675099433637e-05, |
|
"loss": 1.4426, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 4.144939646556894, |
|
"learning_rate": 1.873807103611081e-05, |
|
"loss": 1.4095, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"eval_loss": 1.422728419303894, |
|
"eval_runtime": 36.3022, |
|
"eval_samples_per_second": 36.086, |
|
"eval_steps_per_second": 1.129, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 3.8094743377142, |
|
"learning_rate": 1.8631421280499114e-05, |
|
"loss": 1.4219, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 3.6484435225782885, |
|
"learning_rate": 1.8520775211013094e-05, |
|
"loss": 1.399, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 3.8418704373836454, |
|
"learning_rate": 1.8406184056344784e-05, |
|
"loss": 1.424, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 3.5369039139594354, |
|
"learning_rate": 1.8287700871745036e-05, |
|
"loss": 1.3809, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 4.306679118424655, |
|
"learning_rate": 1.816538051445916e-05, |
|
"loss": 1.4065, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"eval_loss": 1.4043676853179932, |
|
"eval_runtime": 36.1499, |
|
"eval_samples_per_second": 36.238, |
|
"eval_steps_per_second": 1.134, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 3.793497887600387, |
|
"learning_rate": 1.8039279618328215e-05, |
|
"loss": 1.4393, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 3.5806655996080217, |
|
"learning_rate": 1.7909456567567774e-05, |
|
"loss": 1.3877, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 4.335034005215255, |
|
"learning_rate": 1.777597146973627e-05, |
|
"loss": 1.3992, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 3.5829112355976407, |
|
"learning_rate": 1.7638886127905427e-05, |
|
"loss": 1.3966, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 3.380252848613958, |
|
"learning_rate": 1.7498264012045686e-05, |
|
"loss": 1.3747, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"eval_loss": 1.3881479501724243, |
|
"eval_runtime": 36.7171, |
|
"eval_samples_per_second": 35.678, |
|
"eval_steps_per_second": 1.117, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 3.610102942308515, |
|
"learning_rate": 1.7354170229639856e-05, |
|
"loss": 1.3726, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 3.3439020430777155, |
|
"learning_rate": 1.720667149553861e-05, |
|
"loss": 1.3682, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 3.4956412955508402, |
|
"learning_rate": 1.705583610107178e-05, |
|
"loss": 1.3552, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 3.4985126593068165, |
|
"learning_rate": 1.690173388242972e-05, |
|
"loss": 1.3895, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 3.7382334626759746, |
|
"learning_rate": 1.6744436188329455e-05, |
|
"loss": 1.3847, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"eval_loss": 1.3719505071640015, |
|
"eval_runtime": 36.7391, |
|
"eval_samples_per_second": 35.657, |
|
"eval_steps_per_second": 1.116, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 3.7522691462559363, |
|
"learning_rate": 1.658401584698049e-05, |
|
"loss": 1.3954, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 3.599034550869163, |
|
"learning_rate": 1.6420547132365637e-05, |
|
"loss": 1.3558, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 3.780355027101858, |
|
"learning_rate": 1.6254105729852466e-05, |
|
"loss": 1.3836, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 3.723697013821863, |
|
"learning_rate": 1.6084768701151263e-05, |
|
"loss": 1.3753, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 3.285217439584574, |
|
"learning_rate": 1.5912614448635784e-05, |
|
"loss": 1.3526, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"eval_loss": 1.358110785484314, |
|
"eval_runtime": 36.5535, |
|
"eval_samples_per_second": 35.838, |
|
"eval_steps_per_second": 1.122, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 4.389266580261243, |
|
"learning_rate": 1.573772267904325e-05, |
|
"loss": 1.3432, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 3.562805344105124, |
|
"learning_rate": 1.5560174366570448e-05, |
|
"loss": 1.3479, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 3.557282541178432, |
|
"learning_rate": 1.5380051715383e-05, |
|
"loss": 1.3482, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 3.171518329728872, |
|
"learning_rate": 1.5197438121555159e-05, |
|
"loss": 1.3627, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 3.4241008103077006, |
|
"learning_rate": 1.5012418134457756e-05, |
|
"loss": 1.3072, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"eval_loss": 1.3410331010818481, |
|
"eval_runtime": 36.5162, |
|
"eval_samples_per_second": 35.874, |
|
"eval_steps_per_second": 1.123, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 3.4417136894356077, |
|
"learning_rate": 1.4825077417612187e-05, |
|
"loss": 1.3713, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 3.55225659028301, |
|
"learning_rate": 1.4635502709028511e-05, |
|
"loss": 1.3492, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 4.864004021362561, |
|
"learning_rate": 1.4443781781046135e-05, |
|
"loss": 1.3297, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 3.714231728878515, |
|
"learning_rate": 1.425000339969554e-05, |
|
"loss": 1.292, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 3.499541537796859, |
|
"learning_rate": 1.4054257283599974e-05, |
|
"loss": 1.3146, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"eval_loss": 1.3238551616668701, |
|
"eval_runtime": 35.9276, |
|
"eval_samples_per_second": 36.462, |
|
"eval_steps_per_second": 1.141, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 3.3766322552695374, |
|
"learning_rate": 1.3856634062436073e-05, |
|
"loss": 1.2782, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 3.430094949177884, |
|
"learning_rate": 1.3657225234972695e-05, |
|
"loss": 1.291, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 3.7053693958311964, |
|
"learning_rate": 1.3456123126707334e-05, |
|
"loss": 1.3275, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 3.528099228472523, |
|
"learning_rate": 1.3253420847119804e-05, |
|
"loss": 1.2867, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 3.6078436890974706, |
|
"learning_rate": 1.304921224656289e-05, |
|
"loss": 1.2783, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"eval_loss": 1.310630440711975, |
|
"eval_runtime": 36.5504, |
|
"eval_samples_per_second": 35.841, |
|
"eval_steps_per_second": 1.122, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 3.523684230098565, |
|
"learning_rate": 1.2843591872810039e-05, |
|
"loss": 1.2782, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 3.7267322056062144, |
|
"learning_rate": 1.2636654927280074e-05, |
|
"loss": 1.2761, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 3.65799573792828, |
|
"learning_rate": 1.2428497220959359e-05, |
|
"loss": 1.2864, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 3.520887899492278, |
|
"learning_rate": 1.2219215130041656e-05, |
|
"loss": 1.2936, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 3.878015265725218, |
|
"learning_rate": 1.2008905551306356e-05, |
|
"loss": 1.2773, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"eval_loss": 1.2895601987838745, |
|
"eval_runtime": 36.4176, |
|
"eval_samples_per_second": 35.972, |
|
"eval_steps_per_second": 1.126, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 6.1260498556243155, |
|
"learning_rate": 1.1797665857255622e-05, |
|
"loss": 1.2861, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"grad_norm": 4.466084749533314, |
|
"learning_rate": 1.1585593851031346e-05, |
|
"loss": 0.8568, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"grad_norm": 4.230541766753645, |
|
"learning_rate": 1.1372787721132648e-05, |
|
"loss": 0.7304, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"grad_norm": 4.050955020609926, |
|
"learning_rate": 1.1159345995955007e-05, |
|
"loss": 0.7438, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"grad_norm": 3.8978479248153, |
|
"learning_rate": 1.0945367498171992e-05, |
|
"loss": 0.7303, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"eval_loss": 1.3326823711395264, |
|
"eval_runtime": 37.9893, |
|
"eval_samples_per_second": 34.483, |
|
"eval_steps_per_second": 1.079, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"grad_norm": 3.7185440445464963, |
|
"learning_rate": 1.0730951298980776e-05, |
|
"loss": 0.7407, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"grad_norm": 3.7568671101121303, |
|
"learning_rate": 1.0516196672232538e-05, |
|
"loss": 0.7147, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"grad_norm": 4.175956955241419, |
|
"learning_rate": 1.0301203048469084e-05, |
|
"loss": 0.7187, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"grad_norm": 3.5713973562394794, |
|
"learning_rate": 1.0086069968886885e-05, |
|
"loss": 0.6899, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"grad_norm": 3.926229796886718, |
|
"learning_rate": 9.87089703924991e-06, |
|
"loss": 0.6911, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"eval_loss": 1.3080617189407349, |
|
"eval_runtime": 36.6218, |
|
"eval_samples_per_second": 35.771, |
|
"eval_steps_per_second": 1.12, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"grad_norm": 4.041765832088674, |
|
"learning_rate": 9.655783883772546e-06, |
|
"loss": 0.6783, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"grad_norm": 3.822198052373504, |
|
"learning_rate": 9.440830098993969e-06, |
|
"loss": 0.7047, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"grad_norm": 3.74103783619035, |
|
"learning_rate": 9.22613520766537e-06, |
|
"loss": 0.7148, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"grad_norm": 3.863229052811563, |
|
"learning_rate": 9.011798612671286e-06, |
|
"loss": 0.6882, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"grad_norm": 3.7184780256426233, |
|
"learning_rate": 8.797919551006477e-06, |
|
"loss": 0.7044, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"eval_loss": 1.3011983633041382, |
|
"eval_runtime": 36.1296, |
|
"eval_samples_per_second": 36.258, |
|
"eval_steps_per_second": 1.135, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"grad_norm": 3.61644498812297, |
|
"learning_rate": 8.58459704782957e-06, |
|
"loss": 0.7014, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"grad_norm": 3.7252005213479427, |
|
"learning_rate": 8.371929870614834e-06, |
|
"loss": 0.689, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"grad_norm": 3.6457105180525105, |
|
"learning_rate": 8.1600164834232e-06, |
|
"loss": 0.6913, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"grad_norm": 7.856843188311362, |
|
"learning_rate": 7.948955001313812e-06, |
|
"loss": 0.6851, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"grad_norm": 3.583455067918794, |
|
"learning_rate": 7.738843144917119e-06, |
|
"loss": 0.6597, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"eval_loss": 1.2844963073730469, |
|
"eval_runtime": 36.2065, |
|
"eval_samples_per_second": 36.181, |
|
"eval_steps_per_second": 1.132, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"grad_norm": 6.459020130486219, |
|
"learning_rate": 7.529778195190644e-06, |
|
"loss": 0.6643, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"grad_norm": 3.8211583260556905, |
|
"learning_rate": 7.321856948378259e-06, |
|
"loss": 0.6848, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"grad_norm": 3.7126647718977095, |
|
"learning_rate": 7.115175671193913e-06, |
|
"loss": 0.6894, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"grad_norm": 3.659148206814036, |
|
"learning_rate": 6.909830056250527e-06, |
|
"loss": 0.6846, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"grad_norm": 3.5912318784335757, |
|
"learning_rate": 6.7059151777547e-06, |
|
"loss": 0.6717, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"eval_loss": 1.2701396942138672, |
|
"eval_runtime": 36.3576, |
|
"eval_samples_per_second": 36.031, |
|
"eval_steps_per_second": 1.128, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"grad_norm": 4.43495688718006, |
|
"learning_rate": 6.503525447487717e-06, |
|
"loss": 0.6644, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"grad_norm": 3.754517876550191, |
|
"learning_rate": 6.3027545710932715e-06, |
|
"loss": 0.6531, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"grad_norm": 3.745749569347864, |
|
"learning_rate": 6.103695504692122e-06, |
|
"loss": 0.6528, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"grad_norm": 3.724963177872727, |
|
"learning_rate": 5.906440411843787e-06, |
|
"loss": 0.6679, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"grad_norm": 3.5172682214444926, |
|
"learning_rate": 5.711080620875165e-06, |
|
"loss": 0.6538, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"eval_loss": 1.2666361331939697, |
|
"eval_runtime": 36.5896, |
|
"eval_samples_per_second": 35.802, |
|
"eval_steps_per_second": 1.121, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"grad_norm": 3.6339406360777096, |
|
"learning_rate": 5.517706582595896e-06, |
|
"loss": 0.6794, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"grad_norm": 3.7173322180156116, |
|
"learning_rate": 5.32640782841998e-06, |
|
"loss": 0.6479, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"grad_norm": 3.661315456496335, |
|
"learning_rate": 5.137272928913097e-06, |
|
"loss": 0.6582, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"grad_norm": 4.487184920335217, |
|
"learning_rate": 4.950389452784796e-06, |
|
"loss": 0.6697, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"grad_norm": 3.5318063600935776, |
|
"learning_rate": 4.7658439263445015e-06, |
|
"loss": 0.6484, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"eval_loss": 1.2480911016464233, |
|
"eval_runtime": 36.7469, |
|
"eval_samples_per_second": 35.649, |
|
"eval_steps_per_second": 1.116, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"grad_norm": 3.6935613568961463, |
|
"learning_rate": 4.583721793440188e-06, |
|
"loss": 0.6621, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"grad_norm": 3.6221958065559203, |
|
"learning_rate": 4.404107375898234e-06, |
|
"loss": 0.6633, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"grad_norm": 3.9034034514689533, |
|
"learning_rate": 4.2270838344827285e-06, |
|
"loss": 0.6561, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"grad_norm": 3.7752277109004733, |
|
"learning_rate": 4.052733130392367e-06, |
|
"loss": 0.65, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"grad_norm": 3.63792060813426, |
|
"learning_rate": 3.881135987312758e-06, |
|
"loss": 0.633, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"eval_loss": 1.2404245138168335, |
|
"eval_runtime": 36.9, |
|
"eval_samples_per_second": 35.501, |
|
"eval_steps_per_second": 1.111, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"grad_norm": 3.629431521288044, |
|
"learning_rate": 3.712371854041654e-06, |
|
"loss": 0.6252, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"grad_norm": 3.7538748748698962, |
|
"learning_rate": 3.546518867704499e-06, |
|
"loss": 0.6549, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"grad_norm": 3.5943200372390676, |
|
"learning_rate": 3.383653817577216e-06, |
|
"loss": 0.6365, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"grad_norm": 3.7988035352394673, |
|
"learning_rate": 3.223852109533112e-06, |
|
"loss": 0.6157, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"grad_norm": 3.7396973266303886, |
|
"learning_rate": 3.0671877311302247e-06, |
|
"loss": 0.6252, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"eval_loss": 1.2332385778427124, |
|
"eval_runtime": 36.3217, |
|
"eval_samples_per_second": 36.067, |
|
"eval_steps_per_second": 1.129, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"grad_norm": 3.6119019240279258, |
|
"learning_rate": 2.9137332173554043e-06, |
|
"loss": 0.6215, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"grad_norm": 3.561248255506443, |
|
"learning_rate": 2.763559617040876e-06, |
|
"loss": 0.6364, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"grad_norm": 4.068465368690645, |
|
"learning_rate": 2.616736459968936e-06, |
|
"loss": 0.6193, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"grad_norm": 3.808522974597512, |
|
"learning_rate": 2.473331724679917e-06, |
|
"loss": 0.6622, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"grad_norm": 3.743484183877061, |
|
"learning_rate": 2.3334118069984102e-06, |
|
"loss": 0.6423, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"eval_loss": 1.2212291955947876, |
|
"eval_runtime": 36.995, |
|
"eval_samples_per_second": 35.41, |
|
"eval_steps_per_second": 1.108, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"grad_norm": 3.7762899770508453, |
|
"learning_rate": 2.197041489292244e-06, |
|
"loss": 0.64, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"grad_norm": 3.7440789835416575, |
|
"learning_rate": 2.064283910478527e-06, |
|
"loss": 0.6627, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"grad_norm": 3.53490892439879, |
|
"learning_rate": 1.9352005367905536e-06, |
|
"loss": 0.6253, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"grad_norm": 3.743044691468231, |
|
"learning_rate": 1.8098511333192026e-06, |
|
"loss": 0.6248, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"grad_norm": 3.508178112989742, |
|
"learning_rate": 1.6882937363419205e-06, |
|
"loss": 0.6182, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"eval_loss": 1.2115447521209717, |
|
"eval_runtime": 36.7342, |
|
"eval_samples_per_second": 35.662, |
|
"eval_steps_per_second": 1.116, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"grad_norm": 3.82584115413825, |
|
"learning_rate": 1.5705846264521728e-06, |
|
"loss": 0.6159, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"grad_norm": 3.805908154584629, |
|
"learning_rate": 1.45677830250173e-06, |
|
"loss": 0.5992, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"grad_norm": 3.6691587173463174, |
|
"learning_rate": 1.3469274563679401e-06, |
|
"loss": 0.6379, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"grad_norm": 3.6099414977809534, |
|
"learning_rate": 1.2410829485575703e-06, |
|
"loss": 0.6217, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"grad_norm": 3.6351646415567833, |
|
"learning_rate": 1.1392937846586216e-06, |
|
"loss": 0.6218, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"eval_loss": 1.2099106311798096, |
|
"eval_runtime": 37.2149, |
|
"eval_samples_per_second": 35.201, |
|
"eval_steps_per_second": 1.102, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"grad_norm": 3.5759017601516625, |
|
"learning_rate": 1.0416070926509114e-06, |
|
"loss": 0.625, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"grad_norm": 4.079216714693347, |
|
"learning_rate": 9.48068101086026e-07, |
|
"loss": 0.6503, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"grad_norm": 3.690374175674403, |
|
"learning_rate": 8.58720118146662e-07, |
|
"loss": 0.6279, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"grad_norm": 3.653451951807952, |
|
"learning_rate": 7.736045115951252e-07, |
|
"loss": 0.6214, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"grad_norm": 3.9843182779544253, |
|
"learning_rate": 6.927606896202066e-07, |
|
"loss": 0.6293, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"eval_loss": 1.2038028240203857, |
|
"eval_runtime": 36.8911, |
|
"eval_samples_per_second": 35.51, |
|
"eval_steps_per_second": 1.111, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"grad_norm": 3.625232494017856, |
|
"learning_rate": 6.162260825913591e-07, |
|
"loss": 0.6138, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"grad_norm": 3.438461239374458, |
|
"learning_rate": 5.440361257285742e-07, |
|
"loss": 0.6035, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"grad_norm": 3.468914334826838, |
|
"learning_rate": 4.762242426960262e-07, |
|
"loss": 0.6123, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"grad_norm": 3.4298511965711214, |
|
"learning_rate": 4.128218301270359e-07, |
|
"loss": 0.5911, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"grad_norm": 3.648217807793934, |
|
"learning_rate": 3.538582430875659e-07, |
|
"loss": 0.5978, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"eval_loss": 1.20143461227417, |
|
"eval_runtime": 36.7633, |
|
"eval_samples_per_second": 35.633, |
|
"eval_steps_per_second": 1.115, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"grad_norm": 3.6847453856191597, |
|
"learning_rate": 2.9936078148492973e-07, |
|
"loss": 0.6013, |
|
"step": 755 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"grad_norm": 3.6349536920444874, |
|
"learning_rate": 2.493546774280531e-07, |
|
"loss": 0.6004, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"grad_norm": 3.5351570751944594, |
|
"learning_rate": 2.0386308354509944e-07, |
|
"loss": 0.5957, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"grad_norm": 3.623279846120391, |
|
"learning_rate": 1.6290706226390286e-07, |
|
"loss": 0.5953, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"grad_norm": 3.6727168784725897, |
|
"learning_rate": 1.2650557606013635e-07, |
|
"loss": 0.6176, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"eval_loss": 1.19950532913208, |
|
"eval_runtime": 36.6969, |
|
"eval_samples_per_second": 35.698, |
|
"eval_steps_per_second": 1.117, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"grad_norm": 3.873849616801828, |
|
"learning_rate": 9.467547867777261e-08, |
|
"loss": 0.5817, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"grad_norm": 3.554084353387301, |
|
"learning_rate": 6.74315073258569e-08, |
|
"loss": 0.6145, |
|
"step": 785 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"grad_norm": 7.330372861717701, |
|
"learning_rate": 4.478627585524753e-08, |
|
"loss": 0.6256, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"grad_norm": 3.617776301851038, |
|
"learning_rate": 2.675026891844512e-08, |
|
"loss": 0.6224, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"grad_norm": 3.5902020548965763, |
|
"learning_rate": 1.3331837115241331e-08, |
|
"loss": 0.6475, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"eval_loss": 1.199601411819458, |
|
"eval_runtime": 36.7572, |
|
"eval_samples_per_second": 35.639, |
|
"eval_steps_per_second": 1.115, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"grad_norm": 3.5573953568277306, |
|
"learning_rate": 4.537193126427086e-09, |
|
"loss": 0.5941, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 3.684445766409529, |
|
"learning_rate": 3.7040883734462685e-10, |
|
"loss": 0.6021, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"step": 812, |
|
"total_flos": 42504070103040.0, |
|
"train_loss": 1.0263223388218528, |
|
"train_runtime": 5267.6103, |
|
"train_samples_per_second": 4.93, |
|
"train_steps_per_second": 0.154 |
|
} |
|
], |
|
"logging_steps": 5, |
|
"max_steps": 812, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 2, |
|
"save_steps": 100, |
|
"total_flos": 42504070103040.0, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|