|
{ |
|
"best_metric": 0.9354831576347351, |
|
"best_model_checkpoint": "data/Llama-31-8B_task-1_60-samples_config-4_full/checkpoint-287", |
|
"epoch": 104.34782608695652, |
|
"eval_steps": 500, |
|
"global_step": 300, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.34782608695652173, |
|
"grad_norm": 2.058788776397705, |
|
"learning_rate": 3.3333333333333335e-07, |
|
"loss": 2.5057, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.6956521739130435, |
|
"grad_norm": 2.0099380016326904, |
|
"learning_rate": 6.666666666666667e-07, |
|
"loss": 2.5391, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.6956521739130435, |
|
"eval_loss": 2.416827917098999, |
|
"eval_runtime": 4.8263, |
|
"eval_samples_per_second": 2.486, |
|
"eval_steps_per_second": 2.486, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 1.391304347826087, |
|
"grad_norm": 2.0508503913879395, |
|
"learning_rate": 1.3333333333333334e-06, |
|
"loss": 2.5182, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 1.7391304347826086, |
|
"eval_loss": 2.4065160751342773, |
|
"eval_runtime": 4.8317, |
|
"eval_samples_per_second": 2.484, |
|
"eval_steps_per_second": 2.484, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 2.0869565217391304, |
|
"grad_norm": 1.9739177227020264, |
|
"learning_rate": 2.0000000000000003e-06, |
|
"loss": 2.5073, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 2.782608695652174, |
|
"grad_norm": 1.7424209117889404, |
|
"learning_rate": 2.666666666666667e-06, |
|
"loss": 2.4879, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 2.782608695652174, |
|
"eval_loss": 2.3912572860717773, |
|
"eval_runtime": 4.8313, |
|
"eval_samples_per_second": 2.484, |
|
"eval_steps_per_second": 2.484, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 3.4782608695652173, |
|
"grad_norm": 1.7085793018341064, |
|
"learning_rate": 3.3333333333333333e-06, |
|
"loss": 2.4947, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 3.8260869565217392, |
|
"eval_loss": 2.372018575668335, |
|
"eval_runtime": 4.8224, |
|
"eval_samples_per_second": 2.488, |
|
"eval_steps_per_second": 2.488, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 4.173913043478261, |
|
"grad_norm": 1.7649006843566895, |
|
"learning_rate": 4.000000000000001e-06, |
|
"loss": 2.4911, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 4.869565217391305, |
|
"grad_norm": 1.5436650514602661, |
|
"learning_rate": 4.666666666666667e-06, |
|
"loss": 2.4335, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 4.869565217391305, |
|
"eval_loss": 2.3479087352752686, |
|
"eval_runtime": 4.8287, |
|
"eval_samples_per_second": 2.485, |
|
"eval_steps_per_second": 2.485, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 5.565217391304348, |
|
"grad_norm": 1.5427696704864502, |
|
"learning_rate": 5.333333333333334e-06, |
|
"loss": 2.4424, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 5.913043478260869, |
|
"eval_loss": 2.310939073562622, |
|
"eval_runtime": 4.8306, |
|
"eval_samples_per_second": 2.484, |
|
"eval_steps_per_second": 2.484, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 6.260869565217392, |
|
"grad_norm": 1.6715421676635742, |
|
"learning_rate": 6e-06, |
|
"loss": 2.3843, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 6.956521739130435, |
|
"grad_norm": 1.8118067979812622, |
|
"learning_rate": 6.666666666666667e-06, |
|
"loss": 2.3698, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 6.956521739130435, |
|
"eval_loss": 2.2672276496887207, |
|
"eval_runtime": 4.8229, |
|
"eval_samples_per_second": 2.488, |
|
"eval_steps_per_second": 2.488, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 7.6521739130434785, |
|
"grad_norm": 1.769518494606018, |
|
"learning_rate": 7.333333333333333e-06, |
|
"loss": 2.3512, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_loss": 2.2129247188568115, |
|
"eval_runtime": 4.8207, |
|
"eval_samples_per_second": 2.489, |
|
"eval_steps_per_second": 2.489, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 8.347826086956522, |
|
"grad_norm": 1.4598925113677979, |
|
"learning_rate": 8.000000000000001e-06, |
|
"loss": 2.32, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 8.695652173913043, |
|
"eval_loss": 2.182978391647339, |
|
"eval_runtime": 4.8234, |
|
"eval_samples_per_second": 2.488, |
|
"eval_steps_per_second": 2.488, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 9.043478260869565, |
|
"grad_norm": 1.3492408990859985, |
|
"learning_rate": 8.666666666666668e-06, |
|
"loss": 2.2635, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 9.73913043478261, |
|
"grad_norm": 1.289293646812439, |
|
"learning_rate": 9.333333333333334e-06, |
|
"loss": 2.2555, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 9.73913043478261, |
|
"eval_loss": 2.12662935256958, |
|
"eval_runtime": 4.8292, |
|
"eval_samples_per_second": 2.485, |
|
"eval_steps_per_second": 2.485, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 10.434782608695652, |
|
"grad_norm": 1.3162202835083008, |
|
"learning_rate": 1e-05, |
|
"loss": 2.1681, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 10.782608695652174, |
|
"eval_loss": 2.0537397861480713, |
|
"eval_runtime": 4.8349, |
|
"eval_samples_per_second": 2.482, |
|
"eval_steps_per_second": 2.482, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 11.130434782608695, |
|
"grad_norm": 1.3937517404556274, |
|
"learning_rate": 9.99864620589731e-06, |
|
"loss": 2.1476, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 11.826086956521738, |
|
"grad_norm": 1.0678613185882568, |
|
"learning_rate": 9.994585556692624e-06, |
|
"loss": 2.0737, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 11.826086956521738, |
|
"eval_loss": 1.9880471229553223, |
|
"eval_runtime": 4.8264, |
|
"eval_samples_per_second": 2.486, |
|
"eval_steps_per_second": 2.486, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 12.521739130434783, |
|
"grad_norm": 1.7729694843292236, |
|
"learning_rate": 9.987820251299121e-06, |
|
"loss": 2.0403, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 12.869565217391305, |
|
"eval_loss": 1.9276618957519531, |
|
"eval_runtime": 4.8258, |
|
"eval_samples_per_second": 2.487, |
|
"eval_steps_per_second": 2.487, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 13.217391304347826, |
|
"grad_norm": 1.1538618803024292, |
|
"learning_rate": 9.978353953249023e-06, |
|
"loss": 2.005, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 13.91304347826087, |
|
"grad_norm": 0.9281337261199951, |
|
"learning_rate": 9.966191788709716e-06, |
|
"loss": 1.9476, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 13.91304347826087, |
|
"eval_loss": 1.871132254600525, |
|
"eval_runtime": 4.8266, |
|
"eval_samples_per_second": 2.486, |
|
"eval_steps_per_second": 2.486, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 14.608695652173914, |
|
"grad_norm": 0.8572821617126465, |
|
"learning_rate": 9.951340343707852e-06, |
|
"loss": 1.9204, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 14.956521739130435, |
|
"eval_loss": 1.8154711723327637, |
|
"eval_runtime": 4.8222, |
|
"eval_samples_per_second": 2.489, |
|
"eval_steps_per_second": 2.489, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 15.304347826086957, |
|
"grad_norm": 0.8229424953460693, |
|
"learning_rate": 9.933807660562898e-06, |
|
"loss": 1.8583, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"grad_norm": 0.8312278985977173, |
|
"learning_rate": 9.913603233532067e-06, |
|
"loss": 1.8461, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"eval_loss": 1.7615426778793335, |
|
"eval_runtime": 4.8193, |
|
"eval_samples_per_second": 2.49, |
|
"eval_steps_per_second": 2.49, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 16.695652173913043, |
|
"grad_norm": 0.8826591372489929, |
|
"learning_rate": 9.890738003669029e-06, |
|
"loss": 1.8095, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 16.695652173913043, |
|
"eval_loss": 1.7235876321792603, |
|
"eval_runtime": 4.8227, |
|
"eval_samples_per_second": 2.488, |
|
"eval_steps_per_second": 2.488, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 17.391304347826086, |
|
"grad_norm": 0.9071628451347351, |
|
"learning_rate": 9.86522435289912e-06, |
|
"loss": 1.7597, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 17.73913043478261, |
|
"eval_loss": 1.6579965353012085, |
|
"eval_runtime": 4.8233, |
|
"eval_samples_per_second": 2.488, |
|
"eval_steps_per_second": 2.488, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 18.08695652173913, |
|
"grad_norm": 0.9293504357337952, |
|
"learning_rate": 9.83707609731432e-06, |
|
"loss": 1.6805, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 18.782608695652176, |
|
"grad_norm": 0.8878646492958069, |
|
"learning_rate": 9.806308479691595e-06, |
|
"loss": 1.6484, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 18.782608695652176, |
|
"eval_loss": 1.5918583869934082, |
|
"eval_runtime": 4.8278, |
|
"eval_samples_per_second": 2.486, |
|
"eval_steps_per_second": 2.486, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 19.47826086956522, |
|
"grad_norm": 1.0276228189468384, |
|
"learning_rate": 9.77293816123866e-06, |
|
"loss": 1.6443, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 19.82608695652174, |
|
"eval_loss": 1.5261772871017456, |
|
"eval_runtime": 4.8278, |
|
"eval_samples_per_second": 2.486, |
|
"eval_steps_per_second": 2.486, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 20.17391304347826, |
|
"grad_norm": 1.2222720384597778, |
|
"learning_rate": 9.736983212571646e-06, |
|
"loss": 1.5605, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 20.869565217391305, |
|
"grad_norm": 1.040975570678711, |
|
"learning_rate": 9.698463103929542e-06, |
|
"loss": 1.5204, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 20.869565217391305, |
|
"eval_loss": 1.4560518264770508, |
|
"eval_runtime": 4.8273, |
|
"eval_samples_per_second": 2.486, |
|
"eval_steps_per_second": 2.486, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 21.565217391304348, |
|
"grad_norm": 0.8610687255859375, |
|
"learning_rate": 9.657398694630713e-06, |
|
"loss": 1.463, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 21.91304347826087, |
|
"eval_loss": 1.3959513902664185, |
|
"eval_runtime": 4.8279, |
|
"eval_samples_per_second": 2.486, |
|
"eval_steps_per_second": 2.486, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 22.26086956521739, |
|
"grad_norm": 0.9347313642501831, |
|
"learning_rate": 9.613812221777212e-06, |
|
"loss": 1.4249, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 22.956521739130434, |
|
"grad_norm": 0.911578357219696, |
|
"learning_rate": 9.567727288213005e-06, |
|
"loss": 1.3833, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 22.956521739130434, |
|
"eval_loss": 1.3404473066329956, |
|
"eval_runtime": 4.8261, |
|
"eval_samples_per_second": 2.486, |
|
"eval_steps_per_second": 2.486, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 23.652173913043477, |
|
"grad_norm": 0.8542620539665222, |
|
"learning_rate": 9.519168849742603e-06, |
|
"loss": 1.3385, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 24.0, |
|
"eval_loss": 1.2874761819839478, |
|
"eval_runtime": 4.8231, |
|
"eval_samples_per_second": 2.488, |
|
"eval_steps_per_second": 2.488, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 24.347826086956523, |
|
"grad_norm": 0.90443354845047, |
|
"learning_rate": 9.468163201617063e-06, |
|
"loss": 1.3094, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 24.695652173913043, |
|
"eval_loss": 1.2503851652145386, |
|
"eval_runtime": 4.8253, |
|
"eval_samples_per_second": 2.487, |
|
"eval_steps_per_second": 2.487, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 25.043478260869566, |
|
"grad_norm": 0.8763517141342163, |
|
"learning_rate": 9.414737964294636e-06, |
|
"loss": 1.2694, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 25.73913043478261, |
|
"grad_norm": 0.7862639427185059, |
|
"learning_rate": 9.358922068483813e-06, |
|
"loss": 1.2303, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 25.73913043478261, |
|
"eval_loss": 1.200731873512268, |
|
"eval_runtime": 4.8257, |
|
"eval_samples_per_second": 2.487, |
|
"eval_steps_per_second": 2.487, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 26.434782608695652, |
|
"grad_norm": 0.7432467341423035, |
|
"learning_rate": 9.30074573947683e-06, |
|
"loss": 1.1677, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 26.782608695652176, |
|
"eval_loss": 1.1599854230880737, |
|
"eval_runtime": 4.8291, |
|
"eval_samples_per_second": 2.485, |
|
"eval_steps_per_second": 2.485, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 27.130434782608695, |
|
"grad_norm": 0.6284440159797668, |
|
"learning_rate": 9.24024048078213e-06, |
|
"loss": 1.175, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 27.82608695652174, |
|
"grad_norm": 0.5942031145095825, |
|
"learning_rate": 9.177439057064684e-06, |
|
"loss": 1.1674, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 27.82608695652174, |
|
"eval_loss": 1.1331721544265747, |
|
"eval_runtime": 4.8209, |
|
"eval_samples_per_second": 2.489, |
|
"eval_steps_per_second": 2.489, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 28.52173913043478, |
|
"grad_norm": 0.5919143557548523, |
|
"learning_rate": 9.112375476403313e-06, |
|
"loss": 1.1068, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 28.869565217391305, |
|
"eval_loss": 1.1100302934646606, |
|
"eval_runtime": 4.8285, |
|
"eval_samples_per_second": 2.485, |
|
"eval_steps_per_second": 2.485, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 29.217391304347824, |
|
"grad_norm": 0.653171181678772, |
|
"learning_rate": 9.045084971874738e-06, |
|
"loss": 1.0762, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 29.91304347826087, |
|
"grad_norm": 0.6148922443389893, |
|
"learning_rate": 8.97560398247424e-06, |
|
"loss": 1.104, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 29.91304347826087, |
|
"eval_loss": 1.0883766412734985, |
|
"eval_runtime": 4.8276, |
|
"eval_samples_per_second": 2.486, |
|
"eval_steps_per_second": 2.486, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 30.608695652173914, |
|
"grad_norm": 0.5734871029853821, |
|
"learning_rate": 8.903970133383297e-06, |
|
"loss": 1.0617, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 30.956521739130434, |
|
"eval_loss": 1.0717229843139648, |
|
"eval_runtime": 4.8304, |
|
"eval_samples_per_second": 2.484, |
|
"eval_steps_per_second": 2.484, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 31.304347826086957, |
|
"grad_norm": 0.5561685562133789, |
|
"learning_rate": 8.83022221559489e-06, |
|
"loss": 1.0674, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 32.0, |
|
"grad_norm": 0.522966206073761, |
|
"learning_rate": 8.754400164907496e-06, |
|
"loss": 1.0354, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 32.0, |
|
"eval_loss": 1.057722568511963, |
|
"eval_runtime": 4.8513, |
|
"eval_samples_per_second": 2.474, |
|
"eval_steps_per_second": 2.474, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 32.69565217391305, |
|
"grad_norm": 0.46999314427375793, |
|
"learning_rate": 8.676545040299145e-06, |
|
"loss": 1.0195, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 32.69565217391305, |
|
"eval_loss": 1.049912452697754, |
|
"eval_runtime": 4.823, |
|
"eval_samples_per_second": 2.488, |
|
"eval_steps_per_second": 2.488, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 33.391304347826086, |
|
"grad_norm": 0.49655812978744507, |
|
"learning_rate": 8.596699001693257e-06, |
|
"loss": 1.0659, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 33.73913043478261, |
|
"eval_loss": 1.0396418571472168, |
|
"eval_runtime": 4.8306, |
|
"eval_samples_per_second": 2.484, |
|
"eval_steps_per_second": 2.484, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 34.08695652173913, |
|
"grad_norm": 0.44858285784721375, |
|
"learning_rate": 8.51490528712831e-06, |
|
"loss": 0.9917, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 34.78260869565217, |
|
"grad_norm": 0.44484949111938477, |
|
"learning_rate": 8.43120818934367e-06, |
|
"loss": 1.0118, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 34.78260869565217, |
|
"eval_loss": 1.031006932258606, |
|
"eval_runtime": 4.8275, |
|
"eval_samples_per_second": 2.486, |
|
"eval_steps_per_second": 2.486, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 35.47826086956522, |
|
"grad_norm": 0.3746640086174011, |
|
"learning_rate": 8.345653031794292e-06, |
|
"loss": 1.0009, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 35.82608695652174, |
|
"eval_loss": 1.0246673822402954, |
|
"eval_runtime": 4.8365, |
|
"eval_samples_per_second": 2.481, |
|
"eval_steps_per_second": 2.481, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 36.17391304347826, |
|
"grad_norm": 0.4000430107116699, |
|
"learning_rate": 8.258286144107277e-06, |
|
"loss": 0.9993, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 36.869565217391305, |
|
"grad_norm": 0.44598138332366943, |
|
"learning_rate": 8.16915483699355e-06, |
|
"loss": 0.9938, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 36.869565217391305, |
|
"eval_loss": 1.0180528163909912, |
|
"eval_runtime": 4.8232, |
|
"eval_samples_per_second": 2.488, |
|
"eval_steps_per_second": 2.488, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 37.56521739130435, |
|
"grad_norm": 0.40717804431915283, |
|
"learning_rate": 8.078307376628292e-06, |
|
"loss": 0.9736, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 37.91304347826087, |
|
"eval_loss": 1.0123671293258667, |
|
"eval_runtime": 4.8227, |
|
"eval_samples_per_second": 2.488, |
|
"eval_steps_per_second": 2.488, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 38.26086956521739, |
|
"grad_norm": 0.4021812975406647, |
|
"learning_rate": 7.985792958513932e-06, |
|
"loss": 0.9937, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 38.95652173913044, |
|
"grad_norm": 0.3653680086135864, |
|
"learning_rate": 7.891661680839932e-06, |
|
"loss": 0.9888, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 38.95652173913044, |
|
"eval_loss": 1.0076369047164917, |
|
"eval_runtime": 4.8197, |
|
"eval_samples_per_second": 2.49, |
|
"eval_steps_per_second": 2.49, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 39.65217391304348, |
|
"grad_norm": 0.36845943331718445, |
|
"learning_rate": 7.795964517353734e-06, |
|
"loss": 0.9637, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 40.0, |
|
"eval_loss": 1.0018855333328247, |
|
"eval_runtime": 4.8188, |
|
"eval_samples_per_second": 2.49, |
|
"eval_steps_per_second": 2.49, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 40.34782608695652, |
|
"grad_norm": 0.3739126920700073, |
|
"learning_rate": 7.698753289757565e-06, |
|
"loss": 0.9769, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 40.69565217391305, |
|
"eval_loss": 0.998671293258667, |
|
"eval_runtime": 4.8236, |
|
"eval_samples_per_second": 2.488, |
|
"eval_steps_per_second": 2.488, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 41.04347826086956, |
|
"grad_norm": 0.36004525423049927, |
|
"learning_rate": 7.600080639646077e-06, |
|
"loss": 0.9648, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 41.73913043478261, |
|
"grad_norm": 0.3854582607746124, |
|
"learning_rate": 7.500000000000001e-06, |
|
"loss": 0.936, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 41.73913043478261, |
|
"eval_loss": 0.9939180016517639, |
|
"eval_runtime": 4.8226, |
|
"eval_samples_per_second": 2.488, |
|
"eval_steps_per_second": 2.488, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 42.43478260869565, |
|
"grad_norm": 0.3872157335281372, |
|
"learning_rate": 7.398565566251232e-06, |
|
"loss": 0.9863, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 42.78260869565217, |
|
"eval_loss": 0.9906224608421326, |
|
"eval_runtime": 4.8249, |
|
"eval_samples_per_second": 2.487, |
|
"eval_steps_per_second": 2.487, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 43.130434782608695, |
|
"grad_norm": 0.45275992155075073, |
|
"learning_rate": 7.295832266935059e-06, |
|
"loss": 0.9381, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 43.82608695652174, |
|
"grad_norm": 0.3777986466884613, |
|
"learning_rate": 7.191855733945388e-06, |
|
"loss": 0.9626, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 43.82608695652174, |
|
"eval_loss": 0.9862757325172424, |
|
"eval_runtime": 4.8256, |
|
"eval_samples_per_second": 2.487, |
|
"eval_steps_per_second": 2.487, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 44.52173913043478, |
|
"grad_norm": 0.3554269075393677, |
|
"learning_rate": 7.08669227240909e-06, |
|
"loss": 0.9438, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 44.869565217391305, |
|
"eval_loss": 0.9825396537780762, |
|
"eval_runtime": 4.8278, |
|
"eval_samples_per_second": 2.486, |
|
"eval_steps_per_second": 2.486, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 45.21739130434783, |
|
"grad_norm": 0.4316815435886383, |
|
"learning_rate": 6.980398830195785e-06, |
|
"loss": 0.9621, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 45.91304347826087, |
|
"grad_norm": 0.3970458209514618, |
|
"learning_rate": 6.873032967079562e-06, |
|
"loss": 0.9034, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 45.91304347826087, |
|
"eval_loss": 0.980421781539917, |
|
"eval_runtime": 4.8261, |
|
"eval_samples_per_second": 2.486, |
|
"eval_steps_per_second": 2.486, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 46.608695652173914, |
|
"grad_norm": 0.36759161949157715, |
|
"learning_rate": 6.7646528235693445e-06, |
|
"loss": 0.9398, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 46.95652173913044, |
|
"eval_loss": 0.976270854473114, |
|
"eval_runtime": 4.8284, |
|
"eval_samples_per_second": 2.485, |
|
"eval_steps_per_second": 2.485, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 47.30434782608695, |
|
"grad_norm": 0.4169844388961792, |
|
"learning_rate": 6.655317089424791e-06, |
|
"loss": 0.9681, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 48.0, |
|
"grad_norm": 0.3838629126548767, |
|
"learning_rate": 6.545084971874738e-06, |
|
"loss": 0.9206, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 48.0, |
|
"eval_loss": 0.9739949703216553, |
|
"eval_runtime": 4.8201, |
|
"eval_samples_per_second": 2.49, |
|
"eval_steps_per_second": 2.49, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 48.69565217391305, |
|
"grad_norm": 0.43010860681533813, |
|
"learning_rate": 6.434016163555452e-06, |
|
"loss": 0.9251, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 48.69565217391305, |
|
"eval_loss": 0.9728464484214783, |
|
"eval_runtime": 4.8239, |
|
"eval_samples_per_second": 2.488, |
|
"eval_steps_per_second": 2.488, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 49.391304347826086, |
|
"grad_norm": 0.3847574293613434, |
|
"learning_rate": 6.322170810186013e-06, |
|
"loss": 0.9245, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 49.73913043478261, |
|
"eval_loss": 0.9704022407531738, |
|
"eval_runtime": 4.8259, |
|
"eval_samples_per_second": 2.487, |
|
"eval_steps_per_second": 2.487, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 50.08695652173913, |
|
"grad_norm": 0.519916832447052, |
|
"learning_rate": 6.209609477998339e-06, |
|
"loss": 0.9363, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 50.78260869565217, |
|
"grad_norm": 0.46963444352149963, |
|
"learning_rate": 6.0963931209395165e-06, |
|
"loss": 0.9332, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 50.78260869565217, |
|
"eval_loss": 0.9670591354370117, |
|
"eval_runtime": 4.8259, |
|
"eval_samples_per_second": 2.487, |
|
"eval_steps_per_second": 2.487, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 51.47826086956522, |
|
"grad_norm": 0.39990323781967163, |
|
"learning_rate": 5.982583047664151e-06, |
|
"loss": 0.9012, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 51.82608695652174, |
|
"eval_loss": 0.9651336669921875, |
|
"eval_runtime": 4.822, |
|
"eval_samples_per_second": 2.489, |
|
"eval_steps_per_second": 2.489, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 52.17391304347826, |
|
"grad_norm": 0.6043806076049805, |
|
"learning_rate": 5.8682408883346535e-06, |
|
"loss": 0.915, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 52.869565217391305, |
|
"grad_norm": 0.4192598760128021, |
|
"learning_rate": 5.753428561247416e-06, |
|
"loss": 0.9075, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 52.869565217391305, |
|
"eval_loss": 0.9627308249473572, |
|
"eval_runtime": 4.8216, |
|
"eval_samples_per_second": 2.489, |
|
"eval_steps_per_second": 2.489, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 53.56521739130435, |
|
"grad_norm": 0.42425036430358887, |
|
"learning_rate": 5.638208239302975e-06, |
|
"loss": 0.9031, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 53.91304347826087, |
|
"eval_loss": 0.9614337086677551, |
|
"eval_runtime": 4.8281, |
|
"eval_samples_per_second": 2.485, |
|
"eval_steps_per_second": 2.485, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 54.26086956521739, |
|
"grad_norm": 0.49185582995414734, |
|
"learning_rate": 5.522642316338268e-06, |
|
"loss": 0.9089, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 54.95652173913044, |
|
"grad_norm": 0.45466744899749756, |
|
"learning_rate": 5.406793373339292e-06, |
|
"loss": 0.8969, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 54.95652173913044, |
|
"eval_loss": 0.9591866135597229, |
|
"eval_runtime": 4.827, |
|
"eval_samples_per_second": 2.486, |
|
"eval_steps_per_second": 2.486, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 55.65217391304348, |
|
"grad_norm": 0.4060223698616028, |
|
"learning_rate": 5.290724144552379e-06, |
|
"loss": 0.9102, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 56.0, |
|
"eval_loss": 0.9583068490028381, |
|
"eval_runtime": 4.8182, |
|
"eval_samples_per_second": 2.491, |
|
"eval_steps_per_second": 2.491, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 56.34782608695652, |
|
"grad_norm": 0.46647414565086365, |
|
"learning_rate": 5.174497483512506e-06, |
|
"loss": 0.8955, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 56.69565217391305, |
|
"eval_loss": 0.9563028216362, |
|
"eval_runtime": 4.8258, |
|
"eval_samples_per_second": 2.487, |
|
"eval_steps_per_second": 2.487, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 57.04347826086956, |
|
"grad_norm": 0.5661187767982483, |
|
"learning_rate": 5.0581763290069865e-06, |
|
"loss": 0.9039, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 57.73913043478261, |
|
"grad_norm": 0.4606133699417114, |
|
"learning_rate": 4.941823670993016e-06, |
|
"loss": 0.8775, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 57.73913043478261, |
|
"eval_loss": 0.9546887278556824, |
|
"eval_runtime": 4.8258, |
|
"eval_samples_per_second": 2.487, |
|
"eval_steps_per_second": 2.487, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 58.43478260869565, |
|
"grad_norm": 0.4152405261993408, |
|
"learning_rate": 4.825502516487497e-06, |
|
"loss": 0.8879, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 58.78260869565217, |
|
"eval_loss": 0.9539990425109863, |
|
"eval_runtime": 4.8323, |
|
"eval_samples_per_second": 2.483, |
|
"eval_steps_per_second": 2.483, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 59.130434782608695, |
|
"grad_norm": 0.46841880679130554, |
|
"learning_rate": 4.7092758554476215e-06, |
|
"loss": 0.9008, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 59.82608695652174, |
|
"grad_norm": 0.5221750736236572, |
|
"learning_rate": 4.59320662666071e-06, |
|
"loss": 0.8805, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 59.82608695652174, |
|
"eval_loss": 0.9510453343391418, |
|
"eval_runtime": 4.8267, |
|
"eval_samples_per_second": 2.486, |
|
"eval_steps_per_second": 2.486, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 60.52173913043478, |
|
"grad_norm": 0.45134589076042175, |
|
"learning_rate": 4.477357683661734e-06, |
|
"loss": 0.8982, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 60.869565217391305, |
|
"eval_loss": 0.9505268931388855, |
|
"eval_runtime": 4.8236, |
|
"eval_samples_per_second": 2.488, |
|
"eval_steps_per_second": 2.488, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 61.21739130434783, |
|
"grad_norm": 0.5287215113639832, |
|
"learning_rate": 4.361791760697027e-06, |
|
"loss": 0.8613, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 61.91304347826087, |
|
"grad_norm": 0.4524611532688141, |
|
"learning_rate": 4.246571438752585e-06, |
|
"loss": 0.8897, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 61.91304347826087, |
|
"eval_loss": 0.9494467377662659, |
|
"eval_runtime": 4.8277, |
|
"eval_samples_per_second": 2.486, |
|
"eval_steps_per_second": 2.486, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 62.608695652173914, |
|
"grad_norm": 0.4399959146976471, |
|
"learning_rate": 4.131759111665349e-06, |
|
"loss": 0.8515, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 62.95652173913044, |
|
"eval_loss": 0.9478859305381775, |
|
"eval_runtime": 4.8225, |
|
"eval_samples_per_second": 2.488, |
|
"eval_steps_per_second": 2.488, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 63.30434782608695, |
|
"grad_norm": 0.5381987690925598, |
|
"learning_rate": 4.017416952335849e-06, |
|
"loss": 0.9205, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 64.0, |
|
"grad_norm": 0.5035319924354553, |
|
"learning_rate": 3.903606879060483e-06, |
|
"loss": 0.8637, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 64.0, |
|
"eval_loss": 0.9468648433685303, |
|
"eval_runtime": 4.8249, |
|
"eval_samples_per_second": 2.487, |
|
"eval_steps_per_second": 2.487, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 64.69565217391305, |
|
"grad_norm": 0.4583827555179596, |
|
"learning_rate": 3.790390522001662e-06, |
|
"loss": 0.8719, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 64.69565217391305, |
|
"eval_loss": 0.9470892548561096, |
|
"eval_runtime": 4.824, |
|
"eval_samples_per_second": 2.488, |
|
"eval_steps_per_second": 2.488, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 65.3913043478261, |
|
"grad_norm": 0.5146140456199646, |
|
"learning_rate": 3.6778291898139907e-06, |
|
"loss": 0.8635, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 65.73913043478261, |
|
"eval_loss": 0.9452293515205383, |
|
"eval_runtime": 4.8297, |
|
"eval_samples_per_second": 2.485, |
|
"eval_steps_per_second": 2.485, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 66.08695652173913, |
|
"grad_norm": 0.5240075588226318, |
|
"learning_rate": 3.5659838364445505e-06, |
|
"loss": 0.8809, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 66.78260869565217, |
|
"grad_norm": 0.4892469048500061, |
|
"learning_rate": 3.4549150281252635e-06, |
|
"loss": 0.8579, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 66.78260869565217, |
|
"eval_loss": 0.9444965720176697, |
|
"eval_runtime": 4.8242, |
|
"eval_samples_per_second": 2.487, |
|
"eval_steps_per_second": 2.487, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 67.47826086956522, |
|
"grad_norm": 0.5614312291145325, |
|
"learning_rate": 3.3446829105752103e-06, |
|
"loss": 0.8465, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 67.82608695652173, |
|
"eval_loss": 0.9434299468994141, |
|
"eval_runtime": 4.8252, |
|
"eval_samples_per_second": 2.487, |
|
"eval_steps_per_second": 2.487, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 68.17391304347827, |
|
"grad_norm": 0.5231534242630005, |
|
"learning_rate": 3.2353471764306567e-06, |
|
"loss": 0.8791, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 68.8695652173913, |
|
"grad_norm": 0.5704526305198669, |
|
"learning_rate": 3.12696703292044e-06, |
|
"loss": 0.8588, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 68.8695652173913, |
|
"eval_loss": 0.9436232447624207, |
|
"eval_runtime": 4.8241, |
|
"eval_samples_per_second": 2.488, |
|
"eval_steps_per_second": 2.488, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 69.56521739130434, |
|
"grad_norm": 0.5715120434761047, |
|
"learning_rate": 3.019601169804216e-06, |
|
"loss": 0.868, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 69.91304347826087, |
|
"eval_loss": 0.9421423077583313, |
|
"eval_runtime": 4.8217, |
|
"eval_samples_per_second": 2.489, |
|
"eval_steps_per_second": 2.489, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 70.26086956521739, |
|
"grad_norm": 0.7542048692703247, |
|
"learning_rate": 2.9133077275909112e-06, |
|
"loss": 0.8691, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 70.95652173913044, |
|
"grad_norm": 0.5780174732208252, |
|
"learning_rate": 2.8081442660546126e-06, |
|
"loss": 0.8523, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 70.95652173913044, |
|
"eval_loss": 0.9417908191680908, |
|
"eval_runtime": 4.8236, |
|
"eval_samples_per_second": 2.488, |
|
"eval_steps_per_second": 2.488, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 71.65217391304348, |
|
"grad_norm": 0.5677995681762695, |
|
"learning_rate": 2.7041677330649408e-06, |
|
"loss": 0.8654, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 72.0, |
|
"eval_loss": 0.9404106140136719, |
|
"eval_runtime": 4.8254, |
|
"eval_samples_per_second": 2.487, |
|
"eval_steps_per_second": 2.487, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 72.34782608695652, |
|
"grad_norm": 0.5781915187835693, |
|
"learning_rate": 2.601434433748771e-06, |
|
"loss": 0.8525, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 72.69565217391305, |
|
"eval_loss": 0.9405445456504822, |
|
"eval_runtime": 4.8227, |
|
"eval_samples_per_second": 2.488, |
|
"eval_steps_per_second": 2.488, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 73.04347826086956, |
|
"grad_norm": 0.5282586812973022, |
|
"learning_rate": 2.5000000000000015e-06, |
|
"loss": 0.8419, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 73.73913043478261, |
|
"grad_norm": 0.7152076363563538, |
|
"learning_rate": 2.3999193603539234e-06, |
|
"loss": 0.8565, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 73.73913043478261, |
|
"eval_loss": 0.9400214552879333, |
|
"eval_runtime": 4.8268, |
|
"eval_samples_per_second": 2.486, |
|
"eval_steps_per_second": 2.486, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 74.43478260869566, |
|
"grad_norm": 0.5548727512359619, |
|
"learning_rate": 2.3012467102424373e-06, |
|
"loss": 0.8424, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 74.78260869565217, |
|
"eval_loss": 0.9406850337982178, |
|
"eval_runtime": 4.8319, |
|
"eval_samples_per_second": 2.483, |
|
"eval_steps_per_second": 2.483, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 75.1304347826087, |
|
"grad_norm": 0.6737377047538757, |
|
"learning_rate": 2.204035482646267e-06, |
|
"loss": 0.8541, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 75.82608695652173, |
|
"grad_norm": 0.5421648621559143, |
|
"learning_rate": 2.1083383191600676e-06, |
|
"loss": 0.8342, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 75.82608695652173, |
|
"eval_loss": 0.9395294189453125, |
|
"eval_runtime": 4.8312, |
|
"eval_samples_per_second": 2.484, |
|
"eval_steps_per_second": 2.484, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 76.52173913043478, |
|
"grad_norm": 0.6473485827445984, |
|
"learning_rate": 2.0142070414860704e-06, |
|
"loss": 0.8539, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 76.8695652173913, |
|
"eval_loss": 0.9393353462219238, |
|
"eval_runtime": 4.8273, |
|
"eval_samples_per_second": 2.486, |
|
"eval_steps_per_second": 2.486, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 77.21739130434783, |
|
"grad_norm": 0.6593698263168335, |
|
"learning_rate": 1.9216926233717087e-06, |
|
"loss": 0.8531, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 77.91304347826087, |
|
"grad_norm": 0.612883448600769, |
|
"learning_rate": 1.8308451630064484e-06, |
|
"loss": 0.8413, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 77.91304347826087, |
|
"eval_loss": 0.9383335113525391, |
|
"eval_runtime": 4.8256, |
|
"eval_samples_per_second": 2.487, |
|
"eval_steps_per_second": 2.487, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 78.6086956521739, |
|
"grad_norm": 0.600571870803833, |
|
"learning_rate": 1.7417138558927244e-06, |
|
"loss": 0.8488, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 78.95652173913044, |
|
"eval_loss": 0.938234806060791, |
|
"eval_runtime": 4.8279, |
|
"eval_samples_per_second": 2.486, |
|
"eval_steps_per_second": 2.486, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 79.30434782608695, |
|
"grad_norm": 0.5200878977775574, |
|
"learning_rate": 1.6543469682057105e-06, |
|
"loss": 0.8469, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 80.0, |
|
"grad_norm": 0.5332892537117004, |
|
"learning_rate": 1.5687918106563326e-06, |
|
"loss": 0.8319, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 80.0, |
|
"eval_loss": 0.9395381808280945, |
|
"eval_runtime": 4.8185, |
|
"eval_samples_per_second": 2.49, |
|
"eval_steps_per_second": 2.49, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 80.69565217391305, |
|
"grad_norm": 0.5492921471595764, |
|
"learning_rate": 1.4850947128716914e-06, |
|
"loss": 0.8402, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 80.69565217391305, |
|
"eval_loss": 0.9382343292236328, |
|
"eval_runtime": 4.8345, |
|
"eval_samples_per_second": 2.482, |
|
"eval_steps_per_second": 2.482, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 81.3913043478261, |
|
"grad_norm": 0.6471025943756104, |
|
"learning_rate": 1.4033009983067454e-06, |
|
"loss": 0.8604, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 81.73913043478261, |
|
"eval_loss": 0.9375772476196289, |
|
"eval_runtime": 4.8261, |
|
"eval_samples_per_second": 2.487, |
|
"eval_steps_per_second": 2.487, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 82.08695652173913, |
|
"grad_norm": 0.6458178758621216, |
|
"learning_rate": 1.3234549597008572e-06, |
|
"loss": 0.8182, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 82.78260869565217, |
|
"grad_norm": 0.639596164226532, |
|
"learning_rate": 1.2455998350925042e-06, |
|
"loss": 0.8516, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 82.78260869565217, |
|
"eval_loss": 0.9374153017997742, |
|
"eval_runtime": 4.8221, |
|
"eval_samples_per_second": 2.489, |
|
"eval_steps_per_second": 2.489, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 83.47826086956522, |
|
"grad_norm": 0.5777437686920166, |
|
"learning_rate": 1.1697777844051105e-06, |
|
"loss": 0.8195, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 83.82608695652173, |
|
"eval_loss": 0.9378395676612854, |
|
"eval_runtime": 4.822, |
|
"eval_samples_per_second": 2.489, |
|
"eval_steps_per_second": 2.489, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 84.17391304347827, |
|
"grad_norm": 0.6369026303291321, |
|
"learning_rate": 1.096029866616704e-06, |
|
"loss": 0.8491, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 84.8695652173913, |
|
"grad_norm": 0.6049113273620605, |
|
"learning_rate": 1.0243960175257605e-06, |
|
"loss": 0.8456, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 84.8695652173913, |
|
"eval_loss": 0.9380698204040527, |
|
"eval_runtime": 4.8285, |
|
"eval_samples_per_second": 2.485, |
|
"eval_steps_per_second": 2.485, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 85.56521739130434, |
|
"grad_norm": 0.6273989677429199, |
|
"learning_rate": 9.549150281252633e-07, |
|
"loss": 0.8313, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 85.91304347826087, |
|
"eval_loss": 0.9373624324798584, |
|
"eval_runtime": 4.8293, |
|
"eval_samples_per_second": 2.485, |
|
"eval_steps_per_second": 2.485, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 86.26086956521739, |
|
"grad_norm": 0.6432866454124451, |
|
"learning_rate": 8.876245235966884e-07, |
|
"loss": 0.8236, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 86.95652173913044, |
|
"grad_norm": 0.6188175082206726, |
|
"learning_rate": 8.225609429353187e-07, |
|
"loss": 0.8415, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 86.95652173913044, |
|
"eval_loss": 0.9369013905525208, |
|
"eval_runtime": 4.8226, |
|
"eval_samples_per_second": 2.488, |
|
"eval_steps_per_second": 2.488, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 87.65217391304348, |
|
"grad_norm": 0.5841935276985168, |
|
"learning_rate": 7.597595192178702e-07, |
|
"loss": 0.8318, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 88.0, |
|
"eval_loss": 0.9364583492279053, |
|
"eval_runtime": 4.8193, |
|
"eval_samples_per_second": 2.49, |
|
"eval_steps_per_second": 2.49, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 88.34782608695652, |
|
"grad_norm": 0.6125827431678772, |
|
"learning_rate": 6.992542605231739e-07, |
|
"loss": 0.8271, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 88.69565217391305, |
|
"eval_loss": 0.9369893670082092, |
|
"eval_runtime": 4.8228, |
|
"eval_samples_per_second": 2.488, |
|
"eval_steps_per_second": 2.488, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 89.04347826086956, |
|
"grad_norm": 0.5938271284103394, |
|
"learning_rate": 6.410779315161885e-07, |
|
"loss": 0.8479, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 89.73913043478261, |
|
"grad_norm": 0.6235251426696777, |
|
"learning_rate": 5.852620357053651e-07, |
|
"loss": 0.8361, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 89.73913043478261, |
|
"eval_loss": 0.9364290237426758, |
|
"eval_runtime": 4.8251, |
|
"eval_samples_per_second": 2.487, |
|
"eval_steps_per_second": 2.487, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 90.43478260869566, |
|
"grad_norm": 0.6185177564620972, |
|
"learning_rate": 5.318367983829393e-07, |
|
"loss": 0.8216, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 90.78260869565217, |
|
"eval_loss": 0.9365130066871643, |
|
"eval_runtime": 4.8285, |
|
"eval_samples_per_second": 2.485, |
|
"eval_steps_per_second": 2.485, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 91.1304347826087, |
|
"grad_norm": 0.6606243848800659, |
|
"learning_rate": 4.808311502573976e-07, |
|
"loss": 0.8269, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 91.82608695652173, |
|
"grad_norm": 0.6094285845756531, |
|
"learning_rate": 4.322727117869951e-07, |
|
"loss": 0.8387, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 91.82608695652173, |
|
"eval_loss": 0.936617910861969, |
|
"eval_runtime": 4.8322, |
|
"eval_samples_per_second": 2.483, |
|
"eval_steps_per_second": 2.483, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 92.52173913043478, |
|
"grad_norm": 0.6140238642692566, |
|
"learning_rate": 3.8618777822278854e-07, |
|
"loss": 0.8457, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 92.8695652173913, |
|
"eval_loss": 0.9366480708122253, |
|
"eval_runtime": 4.826, |
|
"eval_samples_per_second": 2.487, |
|
"eval_steps_per_second": 2.487, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 93.21739130434783, |
|
"grad_norm": 0.7491313219070435, |
|
"learning_rate": 3.426013053692878e-07, |
|
"loss": 0.822, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 93.91304347826087, |
|
"grad_norm": 0.6457171440124512, |
|
"learning_rate": 3.015368960704584e-07, |
|
"loss": 0.8491, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 93.91304347826087, |
|
"eval_loss": 0.9367270469665527, |
|
"eval_runtime": 4.8289, |
|
"eval_samples_per_second": 2.485, |
|
"eval_steps_per_second": 2.485, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 94.6086956521739, |
|
"grad_norm": 0.6033779382705688, |
|
"learning_rate": 2.63016787428354e-07, |
|
"loss": 0.8171, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 94.95652173913044, |
|
"eval_loss": 0.9357192516326904, |
|
"eval_runtime": 4.8259, |
|
"eval_samples_per_second": 2.487, |
|
"eval_steps_per_second": 2.487, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 95.30434782608695, |
|
"grad_norm": 0.6807251572608948, |
|
"learning_rate": 2.2706183876134047e-07, |
|
"loss": 0.8514, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 96.0, |
|
"grad_norm": 0.6259418725967407, |
|
"learning_rate": 1.9369152030840553e-07, |
|
"loss": 0.8168, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 96.0, |
|
"eval_loss": 0.9367218017578125, |
|
"eval_runtime": 4.8235, |
|
"eval_samples_per_second": 2.488, |
|
"eval_steps_per_second": 2.488, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 96.69565217391305, |
|
"grad_norm": 0.673136293888092, |
|
"learning_rate": 1.6292390268568103e-07, |
|
"loss": 0.8161, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 96.69565217391305, |
|
"eval_loss": 0.9363929629325867, |
|
"eval_runtime": 4.8214, |
|
"eval_samples_per_second": 2.489, |
|
"eval_steps_per_second": 2.489, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 97.3913043478261, |
|
"grad_norm": 0.5515984296798706, |
|
"learning_rate": 1.3477564710088097e-07, |
|
"loss": 0.8442, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 97.73913043478261, |
|
"eval_loss": 0.9356262683868408, |
|
"eval_runtime": 4.8255, |
|
"eval_samples_per_second": 2.487, |
|
"eval_steps_per_second": 2.487, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 98.08695652173913, |
|
"grad_norm": 0.6790144443511963, |
|
"learning_rate": 1.0926199633097156e-07, |
|
"loss": 0.8332, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 98.78260869565217, |
|
"grad_norm": 0.5827218890190125, |
|
"learning_rate": 8.639676646793382e-08, |
|
"loss": 0.8388, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 98.78260869565217, |
|
"eval_loss": 0.936260461807251, |
|
"eval_runtime": 4.8222, |
|
"eval_samples_per_second": 2.489, |
|
"eval_steps_per_second": 2.489, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 99.47826086956522, |
|
"grad_norm": 0.5916593670845032, |
|
"learning_rate": 6.61923394371039e-08, |
|
"loss": 0.8365, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 99.82608695652173, |
|
"eval_loss": 0.9354831576347351, |
|
"eval_runtime": 4.831, |
|
"eval_samples_per_second": 2.484, |
|
"eval_steps_per_second": 2.484, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 100.17391304347827, |
|
"grad_norm": 0.9281457662582397, |
|
"learning_rate": 4.865965629214819e-08, |
|
"loss": 0.8107, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 100.8695652173913, |
|
"grad_norm": 0.6146421432495117, |
|
"learning_rate": 3.3808211290284886e-08, |
|
"loss": 0.8493, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 100.8695652173913, |
|
"eval_loss": 0.9359912276268005, |
|
"eval_runtime": 4.826, |
|
"eval_samples_per_second": 2.487, |
|
"eval_steps_per_second": 2.487, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 101.56521739130434, |
|
"grad_norm": 0.6010165214538574, |
|
"learning_rate": 2.1646046750978255e-08, |
|
"loss": 0.8267, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 101.91304347826087, |
|
"eval_loss": 0.9354932904243469, |
|
"eval_runtime": 4.8205, |
|
"eval_samples_per_second": 2.489, |
|
"eval_steps_per_second": 2.489, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 102.26086956521739, |
|
"grad_norm": 0.6717774868011475, |
|
"learning_rate": 1.2179748700879013e-08, |
|
"loss": 0.8293, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 102.95652173913044, |
|
"grad_norm": 0.5756195783615112, |
|
"learning_rate": 5.414443307377171e-09, |
|
"loss": 0.8304, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 102.95652173913044, |
|
"eval_loss": 0.9361304640769958, |
|
"eval_runtime": 4.8264, |
|
"eval_samples_per_second": 2.486, |
|
"eval_steps_per_second": 2.486, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 103.65217391304348, |
|
"grad_norm": 0.6329156160354614, |
|
"learning_rate": 1.3537941026914302e-09, |
|
"loss": 0.8216, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 104.0, |
|
"eval_loss": 0.9361245632171631, |
|
"eval_runtime": 4.8193, |
|
"eval_samples_per_second": 2.49, |
|
"eval_steps_per_second": 2.49, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 104.34782608695652, |
|
"grad_norm": 0.6687307953834534, |
|
"learning_rate": 0.0, |
|
"loss": 0.8436, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 104.34782608695652, |
|
"eval_loss": 0.9358415007591248, |
|
"eval_runtime": 4.8451, |
|
"eval_samples_per_second": 2.477, |
|
"eval_steps_per_second": 2.477, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 104.34782608695652, |
|
"step": 300, |
|
"total_flos": 7.184388637222502e+16, |
|
"train_loss": 1.1654362523555755, |
|
"train_runtime": 5630.1569, |
|
"train_samples_per_second": 1.226, |
|
"train_steps_per_second": 0.053 |
|
} |
|
], |
|
"logging_steps": 2, |
|
"max_steps": 300, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 150, |
|
"save_steps": 25, |
|
"stateful_callbacks": { |
|
"EarlyStoppingCallback": { |
|
"args": { |
|
"early_stopping_patience": 7, |
|
"early_stopping_threshold": 0.0 |
|
}, |
|
"attributes": { |
|
"early_stopping_patience_counter": 0 |
|
} |
|
}, |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 7.184388637222502e+16, |
|
"train_batch_size": 1, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|