llama2-testv8 / checkpoint-280 /trainer_state.json
Flyfer's picture
Upload folder using huggingface_hub
9ae1154
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 39.388888888888886,
"eval_steps": 500,
"global_step": 280,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.17,
"learning_rate": 8.333333333333334e-06,
"loss": 1.727,
"step": 3
},
{
"epoch": 0.33,
"learning_rate": 1.6666666666666667e-05,
"loss": 1.9644,
"step": 6
},
{
"epoch": 1.11,
"learning_rate": 2.5e-05,
"loss": 1.8722,
"step": 9
},
{
"epoch": 1.28,
"learning_rate": 3.3333333333333335e-05,
"loss": 1.6195,
"step": 12
},
{
"epoch": 2.06,
"learning_rate": 4.166666666666667e-05,
"loss": 1.8716,
"step": 15
},
{
"epoch": 2.22,
"learning_rate": 5e-05,
"loss": 1.5027,
"step": 18
},
{
"epoch": 2.39,
"learning_rate": 5.833333333333334e-05,
"loss": 1.6891,
"step": 21
},
{
"epoch": 3.17,
"learning_rate": 6.666666666666667e-05,
"loss": 1.5746,
"step": 24
},
{
"epoch": 3.33,
"learning_rate": 7.500000000000001e-05,
"loss": 1.3677,
"step": 27
},
{
"epoch": 4.11,
"learning_rate": 8.333333333333334e-05,
"loss": 1.5316,
"step": 30
},
{
"epoch": 4.28,
"learning_rate": 9.166666666666667e-05,
"loss": 1.361,
"step": 33
},
{
"epoch": 5.06,
"learning_rate": 0.0001,
"loss": 1.042,
"step": 36
},
{
"epoch": 5.22,
"learning_rate": 0.00010833333333333333,
"loss": 1.2362,
"step": 39
},
{
"epoch": 5.39,
"learning_rate": 0.00011666666666666668,
"loss": 1.1618,
"step": 42
},
{
"epoch": 6.17,
"learning_rate": 0.000125,
"loss": 1.0433,
"step": 45
},
{
"epoch": 6.33,
"learning_rate": 0.00013333333333333334,
"loss": 1.0156,
"step": 48
},
{
"epoch": 7.11,
"learning_rate": 0.00014166666666666668,
"loss": 0.794,
"step": 51
},
{
"epoch": 7.28,
"learning_rate": 0.00015000000000000001,
"loss": 0.8204,
"step": 54
},
{
"epoch": 8.06,
"learning_rate": 0.00015833333333333332,
"loss": 0.6787,
"step": 57
},
{
"epoch": 8.22,
"learning_rate": 0.0001666666666666667,
"loss": 0.5017,
"step": 60
},
{
"epoch": 8.39,
"learning_rate": 0.000175,
"loss": 0.4836,
"step": 63
},
{
"epoch": 9.17,
"learning_rate": 0.00018333333333333334,
"loss": 0.3545,
"step": 66
},
{
"epoch": 9.33,
"learning_rate": 0.00019166666666666667,
"loss": 0.5009,
"step": 69
},
{
"epoch": 10.11,
"learning_rate": 0.0002,
"loss": 0.3405,
"step": 72
},
{
"epoch": 10.28,
"learning_rate": 0.0001990740740740741,
"loss": 0.3549,
"step": 75
},
{
"epoch": 11.06,
"learning_rate": 0.00019814814814814814,
"loss": 0.2224,
"step": 78
},
{
"epoch": 11.22,
"learning_rate": 0.00019722222222222225,
"loss": 0.2505,
"step": 81
},
{
"epoch": 11.39,
"learning_rate": 0.0001962962962962963,
"loss": 0.2139,
"step": 84
},
{
"epoch": 12.17,
"learning_rate": 0.00019537037037037038,
"loss": 0.1912,
"step": 87
},
{
"epoch": 12.33,
"learning_rate": 0.00019444444444444446,
"loss": 0.1754,
"step": 90
},
{
"epoch": 13.11,
"learning_rate": 0.0001935185185185185,
"loss": 0.1442,
"step": 93
},
{
"epoch": 13.28,
"learning_rate": 0.0001925925925925926,
"loss": 0.1333,
"step": 96
},
{
"epoch": 14.06,
"learning_rate": 0.00019166666666666667,
"loss": 0.1117,
"step": 99
},
{
"epoch": 14.22,
"learning_rate": 0.00019074074074074075,
"loss": 0.0805,
"step": 102
},
{
"epoch": 14.39,
"learning_rate": 0.00018981481481481483,
"loss": 0.0944,
"step": 105
},
{
"epoch": 15.17,
"learning_rate": 0.00018888888888888888,
"loss": 0.0615,
"step": 108
},
{
"epoch": 15.33,
"learning_rate": 0.00018796296296296296,
"loss": 0.0458,
"step": 111
},
{
"epoch": 16.11,
"learning_rate": 0.00018703703703703704,
"loss": 0.0527,
"step": 114
},
{
"epoch": 16.28,
"learning_rate": 0.00018611111111111112,
"loss": 0.0355,
"step": 117
},
{
"epoch": 17.06,
"learning_rate": 0.0001851851851851852,
"loss": 0.0407,
"step": 120
},
{
"epoch": 17.22,
"learning_rate": 0.00018425925925925926,
"loss": 0.0283,
"step": 123
},
{
"epoch": 17.39,
"learning_rate": 0.00018333333333333334,
"loss": 0.0329,
"step": 126
},
{
"epoch": 18.17,
"learning_rate": 0.00018240740740740742,
"loss": 0.0239,
"step": 129
},
{
"epoch": 18.33,
"learning_rate": 0.0001814814814814815,
"loss": 0.0241,
"step": 132
},
{
"epoch": 19.11,
"learning_rate": 0.00018055555555555557,
"loss": 0.0138,
"step": 135
},
{
"epoch": 19.28,
"learning_rate": 0.00017962962962962963,
"loss": 0.0215,
"step": 138
},
{
"epoch": 20.06,
"learning_rate": 0.0001787037037037037,
"loss": 0.0198,
"step": 141
},
{
"epoch": 20.22,
"learning_rate": 0.00017777777777777779,
"loss": 0.0164,
"step": 144
},
{
"epoch": 20.39,
"learning_rate": 0.00017685185185185187,
"loss": 0.0218,
"step": 147
},
{
"epoch": 21.17,
"learning_rate": 0.00017592592592592595,
"loss": 0.0139,
"step": 150
},
{
"epoch": 21.33,
"learning_rate": 0.000175,
"loss": 0.0191,
"step": 153
},
{
"epoch": 22.11,
"learning_rate": 0.00017407407407407408,
"loss": 0.0111,
"step": 156
},
{
"epoch": 22.28,
"learning_rate": 0.00017314814814814816,
"loss": 0.0134,
"step": 159
},
{
"epoch": 23.06,
"learning_rate": 0.00017222222222222224,
"loss": 0.0163,
"step": 162
},
{
"epoch": 23.22,
"learning_rate": 0.00017129629629629632,
"loss": 0.0097,
"step": 165
},
{
"epoch": 23.39,
"learning_rate": 0.00017037037037037037,
"loss": 0.0141,
"step": 168
},
{
"epoch": 24.17,
"learning_rate": 0.00016944444444444445,
"loss": 0.0122,
"step": 171
},
{
"epoch": 24.33,
"learning_rate": 0.00016851851851851853,
"loss": 0.0111,
"step": 174
},
{
"epoch": 25.11,
"learning_rate": 0.00016759259259259258,
"loss": 0.0105,
"step": 177
},
{
"epoch": 25.28,
"learning_rate": 0.0001666666666666667,
"loss": 0.0087,
"step": 180
},
{
"epoch": 26.06,
"learning_rate": 0.00016574074074074074,
"loss": 0.0097,
"step": 183
},
{
"epoch": 26.22,
"learning_rate": 0.00016481481481481482,
"loss": 0.0099,
"step": 186
},
{
"epoch": 26.39,
"learning_rate": 0.0001638888888888889,
"loss": 0.009,
"step": 189
},
{
"epoch": 27.17,
"learning_rate": 0.00016296296296296295,
"loss": 0.0097,
"step": 192
},
{
"epoch": 27.33,
"learning_rate": 0.00016203703703703706,
"loss": 0.008,
"step": 195
},
{
"epoch": 28.11,
"learning_rate": 0.0001611111111111111,
"loss": 0.008,
"step": 198
},
{
"epoch": 28.28,
"learning_rate": 0.0001601851851851852,
"loss": 0.0055,
"step": 201
},
{
"epoch": 29.06,
"learning_rate": 0.00015925925925925927,
"loss": 0.0094,
"step": 204
},
{
"epoch": 29.22,
"learning_rate": 0.00015833333333333332,
"loss": 0.0052,
"step": 207
},
{
"epoch": 29.39,
"learning_rate": 0.00015740740740740743,
"loss": 0.0072,
"step": 210
},
{
"epoch": 30.17,
"learning_rate": 0.00015648148148148148,
"loss": 0.0066,
"step": 213
},
{
"epoch": 30.33,
"learning_rate": 0.00015555555555555556,
"loss": 0.006,
"step": 216
},
{
"epoch": 31.11,
"learning_rate": 0.00015462962962962964,
"loss": 0.0078,
"step": 219
},
{
"epoch": 31.28,
"learning_rate": 0.0001537037037037037,
"loss": 0.0051,
"step": 222
},
{
"epoch": 32.06,
"learning_rate": 0.00015277777777777777,
"loss": 0.0042,
"step": 225
},
{
"epoch": 32.22,
"learning_rate": 0.00015185185185185185,
"loss": 0.0058,
"step": 228
},
{
"epoch": 32.39,
"learning_rate": 0.00015092592592592593,
"loss": 0.0059,
"step": 231
},
{
"epoch": 33.17,
"learning_rate": 0.00015000000000000001,
"loss": 0.0065,
"step": 234
},
{
"epoch": 33.33,
"learning_rate": 0.00014907407407407407,
"loss": 0.0047,
"step": 237
},
{
"epoch": 34.11,
"learning_rate": 0.00014814814814814815,
"loss": 0.0061,
"step": 240
},
{
"epoch": 34.28,
"learning_rate": 0.00014722222222222223,
"loss": 0.0031,
"step": 243
},
{
"epoch": 35.06,
"learning_rate": 0.0001462962962962963,
"loss": 0.0066,
"step": 246
},
{
"epoch": 35.22,
"learning_rate": 0.00014537037037037039,
"loss": 0.003,
"step": 249
},
{
"epoch": 35.39,
"learning_rate": 0.00014444444444444444,
"loss": 0.0054,
"step": 252
},
{
"epoch": 36.17,
"learning_rate": 0.00014351851851851852,
"loss": 0.0029,
"step": 255
},
{
"epoch": 36.33,
"learning_rate": 0.0001425925925925926,
"loss": 0.0063,
"step": 258
},
{
"epoch": 37.11,
"learning_rate": 0.00014166666666666668,
"loss": 0.0039,
"step": 261
},
{
"epoch": 37.28,
"learning_rate": 0.00014074074074074076,
"loss": 0.0043,
"step": 264
},
{
"epoch": 38.06,
"learning_rate": 0.0001398148148148148,
"loss": 0.0063,
"step": 267
},
{
"epoch": 38.22,
"learning_rate": 0.0001388888888888889,
"loss": 0.003,
"step": 270
},
{
"epoch": 38.39,
"learning_rate": 0.00013796296296296297,
"loss": 0.005,
"step": 273
},
{
"epoch": 39.17,
"learning_rate": 0.00013703703703703705,
"loss": 0.005,
"step": 276
},
{
"epoch": 39.33,
"learning_rate": 0.00013611111111111113,
"loss": 0.0041,
"step": 279
}
],
"logging_steps": 3,
"max_steps": 720,
"num_train_epochs": 40,
"save_steps": 500,
"total_flos": 2.27623670120448e+16,
"trial_name": null,
"trial_params": null
}