{ "best_metric": null, "best_model_checkpoint": null, "epoch": 39.388888888888886, "eval_steps": 500, "global_step": 280, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.17, "learning_rate": 8.333333333333334e-06, "loss": 1.727, "step": 3 }, { "epoch": 0.33, "learning_rate": 1.6666666666666667e-05, "loss": 1.9644, "step": 6 }, { "epoch": 1.11, "learning_rate": 2.5e-05, "loss": 1.8722, "step": 9 }, { "epoch": 1.28, "learning_rate": 3.3333333333333335e-05, "loss": 1.6195, "step": 12 }, { "epoch": 2.06, "learning_rate": 4.166666666666667e-05, "loss": 1.8716, "step": 15 }, { "epoch": 2.22, "learning_rate": 5e-05, "loss": 1.5027, "step": 18 }, { "epoch": 2.39, "learning_rate": 5.833333333333334e-05, "loss": 1.6891, "step": 21 }, { "epoch": 3.17, "learning_rate": 6.666666666666667e-05, "loss": 1.5746, "step": 24 }, { "epoch": 3.33, "learning_rate": 7.500000000000001e-05, "loss": 1.3677, "step": 27 }, { "epoch": 4.11, "learning_rate": 8.333333333333334e-05, "loss": 1.5316, "step": 30 }, { "epoch": 4.28, "learning_rate": 9.166666666666667e-05, "loss": 1.361, "step": 33 }, { "epoch": 5.06, "learning_rate": 0.0001, "loss": 1.042, "step": 36 }, { "epoch": 5.22, "learning_rate": 0.00010833333333333333, "loss": 1.2362, "step": 39 }, { "epoch": 5.39, "learning_rate": 0.00011666666666666668, "loss": 1.1618, "step": 42 }, { "epoch": 6.17, "learning_rate": 0.000125, "loss": 1.0433, "step": 45 }, { "epoch": 6.33, "learning_rate": 0.00013333333333333334, "loss": 1.0156, "step": 48 }, { "epoch": 7.11, "learning_rate": 0.00014166666666666668, "loss": 0.794, "step": 51 }, { "epoch": 7.28, "learning_rate": 0.00015000000000000001, "loss": 0.8204, "step": 54 }, { "epoch": 8.06, "learning_rate": 0.00015833333333333332, "loss": 0.6787, "step": 57 }, { "epoch": 8.22, "learning_rate": 0.0001666666666666667, "loss": 0.5017, "step": 60 }, { "epoch": 8.39, "learning_rate": 0.000175, "loss": 0.4836, "step": 63 }, { "epoch": 9.17, "learning_rate": 0.00018333333333333334, "loss": 0.3545, "step": 66 }, { "epoch": 9.33, "learning_rate": 0.00019166666666666667, "loss": 0.5009, "step": 69 }, { "epoch": 10.11, "learning_rate": 0.0002, "loss": 0.3405, "step": 72 }, { "epoch": 10.28, "learning_rate": 0.0001990740740740741, "loss": 0.3549, "step": 75 }, { "epoch": 11.06, "learning_rate": 0.00019814814814814814, "loss": 0.2224, "step": 78 }, { "epoch": 11.22, "learning_rate": 0.00019722222222222225, "loss": 0.2505, "step": 81 }, { "epoch": 11.39, "learning_rate": 0.0001962962962962963, "loss": 0.2139, "step": 84 }, { "epoch": 12.17, "learning_rate": 0.00019537037037037038, "loss": 0.1912, "step": 87 }, { "epoch": 12.33, "learning_rate": 0.00019444444444444446, "loss": 0.1754, "step": 90 }, { "epoch": 13.11, "learning_rate": 0.0001935185185185185, "loss": 0.1442, "step": 93 }, { "epoch": 13.28, "learning_rate": 0.0001925925925925926, "loss": 0.1333, "step": 96 }, { "epoch": 14.06, "learning_rate": 0.00019166666666666667, "loss": 0.1117, "step": 99 }, { "epoch": 14.22, "learning_rate": 0.00019074074074074075, "loss": 0.0805, "step": 102 }, { "epoch": 14.39, "learning_rate": 0.00018981481481481483, "loss": 0.0944, "step": 105 }, { "epoch": 15.17, "learning_rate": 0.00018888888888888888, "loss": 0.0615, "step": 108 }, { "epoch": 15.33, "learning_rate": 0.00018796296296296296, "loss": 0.0458, "step": 111 }, { "epoch": 16.11, "learning_rate": 0.00018703703703703704, "loss": 0.0527, "step": 114 }, { "epoch": 16.28, "learning_rate": 0.00018611111111111112, "loss": 0.0355, "step": 117 }, { "epoch": 17.06, "learning_rate": 0.0001851851851851852, "loss": 0.0407, "step": 120 }, { "epoch": 17.22, "learning_rate": 0.00018425925925925926, "loss": 0.0283, "step": 123 }, { "epoch": 17.39, "learning_rate": 0.00018333333333333334, "loss": 0.0329, "step": 126 }, { "epoch": 18.17, "learning_rate": 0.00018240740740740742, "loss": 0.0239, "step": 129 }, { "epoch": 18.33, "learning_rate": 0.0001814814814814815, "loss": 0.0241, "step": 132 }, { "epoch": 19.11, "learning_rate": 0.00018055555555555557, "loss": 0.0138, "step": 135 }, { "epoch": 19.28, "learning_rate": 0.00017962962962962963, "loss": 0.0215, "step": 138 }, { "epoch": 20.06, "learning_rate": 0.0001787037037037037, "loss": 0.0198, "step": 141 }, { "epoch": 20.22, "learning_rate": 0.00017777777777777779, "loss": 0.0164, "step": 144 }, { "epoch": 20.39, "learning_rate": 0.00017685185185185187, "loss": 0.0218, "step": 147 }, { "epoch": 21.17, "learning_rate": 0.00017592592592592595, "loss": 0.0139, "step": 150 }, { "epoch": 21.33, "learning_rate": 0.000175, "loss": 0.0191, "step": 153 }, { "epoch": 22.11, "learning_rate": 0.00017407407407407408, "loss": 0.0111, "step": 156 }, { "epoch": 22.28, "learning_rate": 0.00017314814814814816, "loss": 0.0134, "step": 159 }, { "epoch": 23.06, "learning_rate": 0.00017222222222222224, "loss": 0.0163, "step": 162 }, { "epoch": 23.22, "learning_rate": 0.00017129629629629632, "loss": 0.0097, "step": 165 }, { "epoch": 23.39, "learning_rate": 0.00017037037037037037, "loss": 0.0141, "step": 168 }, { "epoch": 24.17, "learning_rate": 0.00016944444444444445, "loss": 0.0122, "step": 171 }, { "epoch": 24.33, "learning_rate": 0.00016851851851851853, "loss": 0.0111, "step": 174 }, { "epoch": 25.11, "learning_rate": 0.00016759259259259258, "loss": 0.0105, "step": 177 }, { "epoch": 25.28, "learning_rate": 0.0001666666666666667, "loss": 0.0087, "step": 180 }, { "epoch": 26.06, "learning_rate": 0.00016574074074074074, "loss": 0.0097, "step": 183 }, { "epoch": 26.22, "learning_rate": 0.00016481481481481482, "loss": 0.0099, "step": 186 }, { "epoch": 26.39, "learning_rate": 0.0001638888888888889, "loss": 0.009, "step": 189 }, { "epoch": 27.17, "learning_rate": 0.00016296296296296295, "loss": 0.0097, "step": 192 }, { "epoch": 27.33, "learning_rate": 0.00016203703703703706, "loss": 0.008, "step": 195 }, { "epoch": 28.11, "learning_rate": 0.0001611111111111111, "loss": 0.008, "step": 198 }, { "epoch": 28.28, "learning_rate": 0.0001601851851851852, "loss": 0.0055, "step": 201 }, { "epoch": 29.06, "learning_rate": 0.00015925925925925927, "loss": 0.0094, "step": 204 }, { "epoch": 29.22, "learning_rate": 0.00015833333333333332, "loss": 0.0052, "step": 207 }, { "epoch": 29.39, "learning_rate": 0.00015740740740740743, "loss": 0.0072, "step": 210 }, { "epoch": 30.17, "learning_rate": 0.00015648148148148148, "loss": 0.0066, "step": 213 }, { "epoch": 30.33, "learning_rate": 0.00015555555555555556, "loss": 0.006, "step": 216 }, { "epoch": 31.11, "learning_rate": 0.00015462962962962964, "loss": 0.0078, "step": 219 }, { "epoch": 31.28, "learning_rate": 0.0001537037037037037, "loss": 0.0051, "step": 222 }, { "epoch": 32.06, "learning_rate": 0.00015277777777777777, "loss": 0.0042, "step": 225 }, { "epoch": 32.22, "learning_rate": 0.00015185185185185185, "loss": 0.0058, "step": 228 }, { "epoch": 32.39, "learning_rate": 0.00015092592592592593, "loss": 0.0059, "step": 231 }, { "epoch": 33.17, "learning_rate": 0.00015000000000000001, "loss": 0.0065, "step": 234 }, { "epoch": 33.33, "learning_rate": 0.00014907407407407407, "loss": 0.0047, "step": 237 }, { "epoch": 34.11, "learning_rate": 0.00014814814814814815, "loss": 0.0061, "step": 240 }, { "epoch": 34.28, "learning_rate": 0.00014722222222222223, "loss": 0.0031, "step": 243 }, { "epoch": 35.06, "learning_rate": 0.0001462962962962963, "loss": 0.0066, "step": 246 }, { "epoch": 35.22, "learning_rate": 0.00014537037037037039, "loss": 0.003, "step": 249 }, { "epoch": 35.39, "learning_rate": 0.00014444444444444444, "loss": 0.0054, "step": 252 }, { "epoch": 36.17, "learning_rate": 0.00014351851851851852, "loss": 0.0029, "step": 255 }, { "epoch": 36.33, "learning_rate": 0.0001425925925925926, "loss": 0.0063, "step": 258 }, { "epoch": 37.11, "learning_rate": 0.00014166666666666668, "loss": 0.0039, "step": 261 }, { "epoch": 37.28, "learning_rate": 0.00014074074074074076, "loss": 0.0043, "step": 264 }, { "epoch": 38.06, "learning_rate": 0.0001398148148148148, "loss": 0.0063, "step": 267 }, { "epoch": 38.22, "learning_rate": 0.0001388888888888889, "loss": 0.003, "step": 270 }, { "epoch": 38.39, "learning_rate": 0.00013796296296296297, "loss": 0.005, "step": 273 }, { "epoch": 39.17, "learning_rate": 0.00013703703703703705, "loss": 0.005, "step": 276 }, { "epoch": 39.33, "learning_rate": 0.00013611111111111113, "loss": 0.0041, "step": 279 } ], "logging_steps": 3, "max_steps": 720, "num_train_epochs": 40, "save_steps": 500, "total_flos": 2.27623670120448e+16, "trial_name": null, "trial_params": null }