|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 2.0, |
|
"eval_steps": 500, |
|
"global_step": 152, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.013157894736842105, |
|
"grad_norm": 13.235902064816667, |
|
"learning_rate": 6.25e-07, |
|
"loss": 0.3431, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.02631578947368421, |
|
"grad_norm": 13.138228566029483, |
|
"learning_rate": 1.25e-06, |
|
"loss": 0.3429, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.039473684210526314, |
|
"grad_norm": 11.866560617710881, |
|
"learning_rate": 1.8750000000000003e-06, |
|
"loss": 0.2846, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.05263157894736842, |
|
"grad_norm": 11.872638763731226, |
|
"learning_rate": 2.5e-06, |
|
"loss": 0.3455, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.06578947368421052, |
|
"grad_norm": 7.262421698320824, |
|
"learning_rate": 3.125e-06, |
|
"loss": 0.2313, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.07894736842105263, |
|
"grad_norm": 5.162108989306741, |
|
"learning_rate": 3.7500000000000005e-06, |
|
"loss": 0.1935, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.09210526315789473, |
|
"grad_norm": 4.989963293072479, |
|
"learning_rate": 4.3750000000000005e-06, |
|
"loss": 0.1607, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.10526315789473684, |
|
"grad_norm": 3.2196850103315575, |
|
"learning_rate": 5e-06, |
|
"loss": 0.1746, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.11842105263157894, |
|
"grad_norm": 9.267917781353754, |
|
"learning_rate": 5.625e-06, |
|
"loss": 0.2141, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.13157894736842105, |
|
"grad_norm": 2.9550126149670968, |
|
"learning_rate": 6.25e-06, |
|
"loss": 0.1466, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.14473684210526316, |
|
"grad_norm": 3.2305434151019665, |
|
"learning_rate": 6.875e-06, |
|
"loss": 0.1559, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.15789473684210525, |
|
"grad_norm": 2.303728369668543, |
|
"learning_rate": 7.500000000000001e-06, |
|
"loss": 0.1459, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.17105263157894737, |
|
"grad_norm": 2.1650046636974225, |
|
"learning_rate": 8.125000000000001e-06, |
|
"loss": 0.1158, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.18421052631578946, |
|
"grad_norm": 2.43429839604823, |
|
"learning_rate": 8.750000000000001e-06, |
|
"loss": 0.1429, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.19736842105263158, |
|
"grad_norm": 2.256199461941204, |
|
"learning_rate": 9.375000000000001e-06, |
|
"loss": 0.1569, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.21052631578947367, |
|
"grad_norm": 2.540675164259285, |
|
"learning_rate": 1e-05, |
|
"loss": 0.1389, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.2236842105263158, |
|
"grad_norm": 3.0603124517729934, |
|
"learning_rate": 9.998666040558187e-06, |
|
"loss": 0.1481, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.23684210526315788, |
|
"grad_norm": 3.15679704599401, |
|
"learning_rate": 9.994664874011864e-06, |
|
"loss": 0.1543, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 1.2073031998301942, |
|
"learning_rate": 9.987998635318586e-06, |
|
"loss": 0.0912, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.2631578947368421, |
|
"grad_norm": 1.5485358071033657, |
|
"learning_rate": 9.978670881475173e-06, |
|
"loss": 0.1454, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.27631578947368424, |
|
"grad_norm": 1.837098464284506, |
|
"learning_rate": 9.96668658961975e-06, |
|
"loss": 0.1263, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.2894736842105263, |
|
"grad_norm": 2.195521777946691, |
|
"learning_rate": 9.952052154376027e-06, |
|
"loss": 0.1347, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.3026315789473684, |
|
"grad_norm": 1.4759607642496113, |
|
"learning_rate": 9.93477538444123e-06, |
|
"loss": 0.1036, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.3157894736842105, |
|
"grad_norm": 3.4617468292311653, |
|
"learning_rate": 9.91486549841951e-06, |
|
"loss": 0.1401, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.32894736842105265, |
|
"grad_norm": 1.8290234660376485, |
|
"learning_rate": 9.892333119903045e-06, |
|
"loss": 0.1191, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.34210526315789475, |
|
"grad_norm": 1.9446124807007885, |
|
"learning_rate": 9.867190271803466e-06, |
|
"loss": 0.1156, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.35526315789473684, |
|
"grad_norm": 1.924481905745065, |
|
"learning_rate": 9.839450369936615e-06, |
|
"loss": 0.1203, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.3684210526315789, |
|
"grad_norm": 0.9472794741039898, |
|
"learning_rate": 9.809128215864096e-06, |
|
"loss": 0.084, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.3815789473684211, |
|
"grad_norm": 1.8614087289828243, |
|
"learning_rate": 9.776239988995401e-06, |
|
"loss": 0.1237, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.39473684210526316, |
|
"grad_norm": 1.8725601311672047, |
|
"learning_rate": 9.74080323795483e-06, |
|
"loss": 0.109, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.40789473684210525, |
|
"grad_norm": 1.6989866208387698, |
|
"learning_rate": 9.702836871217838e-06, |
|
"loss": 0.1078, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.42105263157894735, |
|
"grad_norm": 1.4620493708990816, |
|
"learning_rate": 9.66236114702178e-06, |
|
"loss": 0.0993, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.4342105263157895, |
|
"grad_norm": 1.4349318310245505, |
|
"learning_rate": 9.619397662556434e-06, |
|
"loss": 0.1168, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.4473684210526316, |
|
"grad_norm": 1.2971799391770196, |
|
"learning_rate": 9.573969342440107e-06, |
|
"loss": 0.0989, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.4605263157894737, |
|
"grad_norm": 1.3284875209748295, |
|
"learning_rate": 9.52610042648741e-06, |
|
"loss": 0.1052, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.47368421052631576, |
|
"grad_norm": 1.517294107060111, |
|
"learning_rate": 9.475816456775313e-06, |
|
"loss": 0.1095, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.4868421052631579, |
|
"grad_norm": 1.8550537994393244, |
|
"learning_rate": 9.423144264014278e-06, |
|
"loss": 0.0975, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 2.2490574894196733, |
|
"learning_rate": 9.368111953231849e-06, |
|
"loss": 0.1171, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.5131578947368421, |
|
"grad_norm": 1.839240097339726, |
|
"learning_rate": 9.310748888776254e-06, |
|
"loss": 0.0916, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.5263157894736842, |
|
"grad_norm": 1.0701426736883897, |
|
"learning_rate": 9.251085678648072e-06, |
|
"loss": 0.1004, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.5394736842105263, |
|
"grad_norm": 1.3170647678764655, |
|
"learning_rate": 9.189154158168293e-06, |
|
"loss": 0.0873, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.5526315789473685, |
|
"grad_norm": 1.2646530838963848, |
|
"learning_rate": 9.124987372991512e-06, |
|
"loss": 0.1062, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.5657894736842105, |
|
"grad_norm": 2.0376323047774947, |
|
"learning_rate": 9.058619561473308e-06, |
|
"loss": 0.1446, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.5789473684210527, |
|
"grad_norm": 1.4525149631147896, |
|
"learning_rate": 8.990086136401199e-06, |
|
"loss": 0.1063, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.5921052631578947, |
|
"grad_norm": 1.7227367203376753, |
|
"learning_rate": 8.91942366609897e-06, |
|
"loss": 0.1288, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.6052631578947368, |
|
"grad_norm": 1.6374368378542756, |
|
"learning_rate": 8.846669854914395e-06, |
|
"loss": 0.1101, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.618421052631579, |
|
"grad_norm": 1.776661892531926, |
|
"learning_rate": 8.771863523100821e-06, |
|
"loss": 0.0863, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.631578947368421, |
|
"grad_norm": 2.101189002973872, |
|
"learning_rate": 8.695044586103297e-06, |
|
"loss": 0.1155, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.6447368421052632, |
|
"grad_norm": 1.4990143401140024, |
|
"learning_rate": 8.616254033260351e-06, |
|
"loss": 0.1233, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.6578947368421053, |
|
"grad_norm": 1.300205397968391, |
|
"learning_rate": 8.535533905932739e-06, |
|
"loss": 0.0722, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.6710526315789473, |
|
"grad_norm": 1.2976123144962795, |
|
"learning_rate": 8.452927275070858e-06, |
|
"loss": 0.0955, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.6842105263157895, |
|
"grad_norm": 3.1993999412027807, |
|
"learning_rate": 8.368478218232787e-06, |
|
"loss": 0.0992, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.6973684210526315, |
|
"grad_norm": 1.511040367087983, |
|
"learning_rate": 8.282231796065215e-06, |
|
"loss": 0.1239, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.7105263157894737, |
|
"grad_norm": 1.735863282573467, |
|
"learning_rate": 8.194234028259806e-06, |
|
"loss": 0.1114, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.7236842105263158, |
|
"grad_norm": 1.062538506538223, |
|
"learning_rate": 8.104531868997858e-06, |
|
"loss": 0.0833, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.7368421052631579, |
|
"grad_norm": 1.4057184361489892, |
|
"learning_rate": 8.013173181896283e-06, |
|
"loss": 0.1348, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 1.0217071440957173, |
|
"learning_rate": 7.920206714468383e-06, |
|
"loss": 0.0869, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.7631578947368421, |
|
"grad_norm": 0.9647935123156808, |
|
"learning_rate": 7.82568207211296e-06, |
|
"loss": 0.0774, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.7763157894736842, |
|
"grad_norm": 1.2119765768642377, |
|
"learning_rate": 7.729649691645673e-06, |
|
"loss": 0.0867, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.7894736842105263, |
|
"grad_norm": 0.684404887958961, |
|
"learning_rate": 7.63216081438678e-06, |
|
"loss": 0.077, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.8026315789473685, |
|
"grad_norm": 1.202460612915356, |
|
"learning_rate": 7.533267458819597e-06, |
|
"loss": 0.0823, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.8157894736842105, |
|
"grad_norm": 1.307993207575058, |
|
"learning_rate": 7.4330223928342814e-06, |
|
"loss": 0.0902, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.8289473684210527, |
|
"grad_norm": 1.6640855566750032, |
|
"learning_rate": 7.33147910557174e-06, |
|
"loss": 0.141, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.8421052631578947, |
|
"grad_norm": 1.3895288949997349, |
|
"learning_rate": 7.2286917788826926e-06, |
|
"loss": 0.0992, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.8552631578947368, |
|
"grad_norm": 0.86233230188121, |
|
"learning_rate": 7.124715258417111e-06, |
|
"loss": 0.0663, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.868421052631579, |
|
"grad_norm": 1.3734677592094817, |
|
"learning_rate": 7.019605024359475e-06, |
|
"loss": 0.0848, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.881578947368421, |
|
"grad_norm": 1.2016915391705267, |
|
"learning_rate": 6.913417161825449e-06, |
|
"loss": 0.1137, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.8947368421052632, |
|
"grad_norm": 1.1962233631272159, |
|
"learning_rate": 6.806208330935766e-06, |
|
"loss": 0.0866, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.9078947368421053, |
|
"grad_norm": 1.3774507325980982, |
|
"learning_rate": 6.698035736583307e-06, |
|
"loss": 0.0975, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.9210526315789473, |
|
"grad_norm": 1.0910714190292217, |
|
"learning_rate": 6.588957097909509e-06, |
|
"loss": 0.0804, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.9342105263157895, |
|
"grad_norm": 0.7559368152288192, |
|
"learning_rate": 6.4790306175063535e-06, |
|
"loss": 0.0748, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.9473684210526315, |
|
"grad_norm": 1.320278904610911, |
|
"learning_rate": 6.368314950360416e-06, |
|
"loss": 0.1092, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.9605263157894737, |
|
"grad_norm": 0.9978117274759357, |
|
"learning_rate": 6.2568691725555144e-06, |
|
"loss": 0.0669, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.9736842105263158, |
|
"grad_norm": 1.184512537053634, |
|
"learning_rate": 6.144752749750671e-06, |
|
"loss": 0.0791, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.9868421052631579, |
|
"grad_norm": 1.3888364419849502, |
|
"learning_rate": 6.0320255054501985e-06, |
|
"loss": 0.0791, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 1.0696792385978997, |
|
"learning_rate": 5.918747589082853e-06, |
|
"loss": 0.1352, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 0.09702859073877335, |
|
"eval_runtime": 143.4285, |
|
"eval_samples_per_second": 35.593, |
|
"eval_steps_per_second": 1.116, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 1.013157894736842, |
|
"grad_norm": 1.130795788902584, |
|
"learning_rate": 5.804979443907065e-06, |
|
"loss": 0.0565, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 1.0263157894736843, |
|
"grad_norm": 0.9742675703939438, |
|
"learning_rate": 5.690781774759412e-06, |
|
"loss": 0.0826, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 1.0394736842105263, |
|
"grad_norm": 1.4529332894640457, |
|
"learning_rate": 5.576215515663489e-06, |
|
"loss": 0.109, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 1.0526315789473684, |
|
"grad_norm": 0.8745980190537407, |
|
"learning_rate": 5.46134179731651e-06, |
|
"loss": 0.0815, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 1.0657894736842106, |
|
"grad_norm": 0.965352409023328, |
|
"learning_rate": 5.346221914470959e-06, |
|
"loss": 0.0753, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 1.0789473684210527, |
|
"grad_norm": 0.9281912350813449, |
|
"learning_rate": 5.230917293228699e-06, |
|
"loss": 0.0604, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 1.0921052631578947, |
|
"grad_norm": 0.7117517424840752, |
|
"learning_rate": 5.115489458265006e-06, |
|
"loss": 0.0548, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 1.1052631578947367, |
|
"grad_norm": 0.8174226361772433, |
|
"learning_rate": 5e-06, |
|
"loss": 0.0798, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 1.118421052631579, |
|
"grad_norm": 1.091499890906096, |
|
"learning_rate": 4.8845105417349955e-06, |
|
"loss": 0.0669, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 1.131578947368421, |
|
"grad_norm": 0.8489893677105097, |
|
"learning_rate": 4.7690827067713035e-06, |
|
"loss": 0.0799, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 1.1447368421052633, |
|
"grad_norm": 1.0103580405572923, |
|
"learning_rate": 4.653778085529043e-06, |
|
"loss": 0.0697, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 1.1578947368421053, |
|
"grad_norm": 0.8915773944210619, |
|
"learning_rate": 4.53865820268349e-06, |
|
"loss": 0.0786, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 1.1710526315789473, |
|
"grad_norm": 0.9795278903306394, |
|
"learning_rate": 4.4237844843365126e-06, |
|
"loss": 0.0948, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 1.1842105263157894, |
|
"grad_norm": 0.6718237305799917, |
|
"learning_rate": 4.309218225240591e-06, |
|
"loss": 0.0751, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 1.1973684210526316, |
|
"grad_norm": 0.9809953018950844, |
|
"learning_rate": 4.195020556092935e-06, |
|
"loss": 0.0717, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 1.2105263157894737, |
|
"grad_norm": 1.3739884608587827, |
|
"learning_rate": 4.081252410917148e-06, |
|
"loss": 0.0834, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 1.2236842105263157, |
|
"grad_norm": 0.7354951934416454, |
|
"learning_rate": 3.967974494549803e-06, |
|
"loss": 0.0497, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 1.236842105263158, |
|
"grad_norm": 0.9321611181418619, |
|
"learning_rate": 3.855247250249331e-06, |
|
"loss": 0.1091, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"grad_norm": 0.8153028496714768, |
|
"learning_rate": 3.743130827444487e-06, |
|
"loss": 0.0719, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 1.263157894736842, |
|
"grad_norm": 1.7763781332034168, |
|
"learning_rate": 3.6316850496395863e-06, |
|
"loss": 0.0995, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 1.2763157894736843, |
|
"grad_norm": 1.0403921146171855, |
|
"learning_rate": 3.5209693824936486e-06, |
|
"loss": 0.1035, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 1.2894736842105263, |
|
"grad_norm": 0.9375966336686881, |
|
"learning_rate": 3.4110429020904924e-06, |
|
"loss": 0.0854, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 1.3026315789473684, |
|
"grad_norm": 0.8044967231814447, |
|
"learning_rate": 3.301964263416693e-06, |
|
"loss": 0.0545, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 1.3157894736842106, |
|
"grad_norm": 1.057615646119616, |
|
"learning_rate": 3.1937916690642356e-06, |
|
"loss": 0.0721, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 1.3289473684210527, |
|
"grad_norm": 0.8460218568489392, |
|
"learning_rate": 3.0865828381745515e-06, |
|
"loss": 0.0756, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 1.3421052631578947, |
|
"grad_norm": 0.7973010176418315, |
|
"learning_rate": 2.980394975640526e-06, |
|
"loss": 0.0701, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 1.3552631578947367, |
|
"grad_norm": 0.43625526231871226, |
|
"learning_rate": 2.8752847415828923e-06, |
|
"loss": 0.0439, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 1.368421052631579, |
|
"grad_norm": 0.7846073292481729, |
|
"learning_rate": 2.771308221117309e-06, |
|
"loss": 0.068, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 1.381578947368421, |
|
"grad_norm": 0.7894886291152677, |
|
"learning_rate": 2.668520894428259e-06, |
|
"loss": 0.0521, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 1.3947368421052633, |
|
"grad_norm": 0.8898713056015292, |
|
"learning_rate": 2.5669776071657194e-06, |
|
"loss": 0.0832, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 1.4078947368421053, |
|
"grad_norm": 0.7239448528037634, |
|
"learning_rate": 2.466732541180404e-06, |
|
"loss": 0.0651, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 1.4210526315789473, |
|
"grad_norm": 1.1924698653865555, |
|
"learning_rate": 2.3678391856132203e-06, |
|
"loss": 0.0871, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 1.4342105263157894, |
|
"grad_norm": 0.796957419192979, |
|
"learning_rate": 2.2703503083543288e-06, |
|
"loss": 0.0632, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 1.4473684210526316, |
|
"grad_norm": 0.8317972188965687, |
|
"learning_rate": 2.174317927887041e-06, |
|
"loss": 0.0627, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 1.4605263157894737, |
|
"grad_norm": 0.6383766307808258, |
|
"learning_rate": 2.0797932855316183e-06, |
|
"loss": 0.0685, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 1.4736842105263157, |
|
"grad_norm": 0.8167329309674938, |
|
"learning_rate": 1.9868268181037186e-06, |
|
"loss": 0.0674, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 1.486842105263158, |
|
"grad_norm": 0.7471966235227357, |
|
"learning_rate": 1.8954681310021434e-06, |
|
"loss": 0.0729, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"grad_norm": 0.6587846660979715, |
|
"learning_rate": 1.8057659717401948e-06, |
|
"loss": 0.0481, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 1.513157894736842, |
|
"grad_norm": 0.8623712101197185, |
|
"learning_rate": 1.7177682039347875e-06, |
|
"loss": 0.0751, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 1.526315789473684, |
|
"grad_norm": 0.7600610777742871, |
|
"learning_rate": 1.6315217817672142e-06, |
|
"loss": 0.0754, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 1.5394736842105263, |
|
"grad_norm": 0.7028388956696653, |
|
"learning_rate": 1.5470727249291423e-06, |
|
"loss": 0.0591, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 1.5526315789473686, |
|
"grad_norm": 0.9080662531327958, |
|
"learning_rate": 1.4644660940672628e-06, |
|
"loss": 0.0665, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 1.5657894736842106, |
|
"grad_norm": 0.8002081473664321, |
|
"learning_rate": 1.383745966739652e-06, |
|
"loss": 0.0567, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 1.5789473684210527, |
|
"grad_norm": 0.8070376846944848, |
|
"learning_rate": 1.3049554138967052e-06, |
|
"loss": 0.074, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 1.5921052631578947, |
|
"grad_norm": 0.7133857988769062, |
|
"learning_rate": 1.2281364768991804e-06, |
|
"loss": 0.082, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 1.6052631578947367, |
|
"grad_norm": 0.7063230781754961, |
|
"learning_rate": 1.1533301450856054e-06, |
|
"loss": 0.0793, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 1.618421052631579, |
|
"grad_norm": 0.5118303753159236, |
|
"learning_rate": 1.0805763339010329e-06, |
|
"loss": 0.0606, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 1.631578947368421, |
|
"grad_norm": 0.6255028097581671, |
|
"learning_rate": 1.0099138635988026e-06, |
|
"loss": 0.0624, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 1.6447368421052633, |
|
"grad_norm": 1.2565545443199133, |
|
"learning_rate": 9.41380438526694e-07, |
|
"loss": 0.0834, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 1.6578947368421053, |
|
"grad_norm": 0.6949398501709794, |
|
"learning_rate": 8.750126270084891e-07, |
|
"loss": 0.0554, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 1.6710526315789473, |
|
"grad_norm": 0.6169370398639232, |
|
"learning_rate": 8.108458418317089e-07, |
|
"loss": 0.0756, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 1.6842105263157894, |
|
"grad_norm": 0.5783403647508548, |
|
"learning_rate": 7.489143213519301e-07, |
|
"loss": 0.0484, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 1.6973684210526314, |
|
"grad_norm": 0.613206478877195, |
|
"learning_rate": 6.892511112237472e-07, |
|
"loss": 0.0674, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 1.7105263157894737, |
|
"grad_norm": 0.7610207582376373, |
|
"learning_rate": 6.318880467681527e-07, |
|
"loss": 0.0802, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 1.723684210526316, |
|
"grad_norm": 0.6484932637313764, |
|
"learning_rate": 5.768557359857241e-07, |
|
"loss": 0.0799, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 1.736842105263158, |
|
"grad_norm": 0.801045963046033, |
|
"learning_rate": 5.241835432246888e-07, |
|
"loss": 0.0631, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"grad_norm": 0.5006567755599417, |
|
"learning_rate": 4.738995735125895e-07, |
|
"loss": 0.0602, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 1.763157894736842, |
|
"grad_norm": 0.7321825355306667, |
|
"learning_rate": 4.2603065755989493e-07, |
|
"loss": 0.0574, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 1.776315789473684, |
|
"grad_norm": 0.5455281345073466, |
|
"learning_rate": 3.8060233744356634e-07, |
|
"loss": 0.0585, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 1.7894736842105263, |
|
"grad_norm": 0.6941675384816356, |
|
"learning_rate": 3.3763885297822153e-07, |
|
"loss": 0.0593, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 1.8026315789473686, |
|
"grad_norm": 0.607277109667664, |
|
"learning_rate": 2.9716312878216194e-07, |
|
"loss": 0.0596, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 1.8157894736842106, |
|
"grad_norm": 0.5526534850437881, |
|
"learning_rate": 2.5919676204517073e-07, |
|
"loss": 0.0574, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 1.8289473684210527, |
|
"grad_norm": 0.7027887496493381, |
|
"learning_rate": 2.237600110046001e-07, |
|
"loss": 0.0883, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 1.8421052631578947, |
|
"grad_norm": 0.5658475866263035, |
|
"learning_rate": 1.908717841359048e-07, |
|
"loss": 0.0593, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 1.8552631578947367, |
|
"grad_norm": 0.6440458312241749, |
|
"learning_rate": 1.6054963006338742e-07, |
|
"loss": 0.0645, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 1.868421052631579, |
|
"grad_norm": 0.8251777255090132, |
|
"learning_rate": 1.328097281965357e-07, |
|
"loss": 0.091, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 1.881578947368421, |
|
"grad_norm": 0.584859727969167, |
|
"learning_rate": 1.0766688009695548e-07, |
|
"loss": 0.0624, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 1.8947368421052633, |
|
"grad_norm": 0.6478070234261503, |
|
"learning_rate": 8.513450158049109e-08, |
|
"loss": 0.0662, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 1.9078947368421053, |
|
"grad_norm": 0.5481010295457636, |
|
"learning_rate": 6.522461555877213e-08, |
|
"loss": 0.0717, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 1.9210526315789473, |
|
"grad_norm": 0.5866004853500736, |
|
"learning_rate": 4.794784562397459e-08, |
|
"loss": 0.0729, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 1.9342105263157894, |
|
"grad_norm": 0.7549384445552164, |
|
"learning_rate": 3.3313410380250157e-08, |
|
"loss": 0.0936, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 1.9473684210526314, |
|
"grad_norm": 0.8887389895072535, |
|
"learning_rate": 2.1329118524827662e-08, |
|
"loss": 0.1033, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 1.9605263157894737, |
|
"grad_norm": 0.6459019124164774, |
|
"learning_rate": 1.200136468141544e-08, |
|
"loss": 0.0951, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 1.973684210526316, |
|
"grad_norm": 0.5993524993653697, |
|
"learning_rate": 5.3351259881379016e-09, |
|
"loss": 0.063, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 1.986842105263158, |
|
"grad_norm": 0.6183094577100168, |
|
"learning_rate": 1.3339594418138036e-09, |
|
"loss": 0.0715, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 0.47166327201031033, |
|
"learning_rate": 0.0, |
|
"loss": 0.051, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 0.08257210999727249, |
|
"eval_runtime": 143.687, |
|
"eval_samples_per_second": 35.529, |
|
"eval_steps_per_second": 1.114, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"step": 152, |
|
"total_flos": 4.516827336435302e+16, |
|
"train_loss": 0.09894711428665016, |
|
"train_runtime": 2051.5561, |
|
"train_samples_per_second": 9.454, |
|
"train_steps_per_second": 0.074 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 152, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 2, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 4.516827336435302e+16, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|