|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 375.83892617449663, |
|
"global_step": 56000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 3.36, |
|
"learning_rate": 4.956305928411633e-05, |
|
"loss": 6.8613, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 6.71, |
|
"learning_rate": 4.912611856823266e-05, |
|
"loss": 6.0182, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 10.07, |
|
"learning_rate": 4.868917785234899e-05, |
|
"loss": 5.8106, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 13.42, |
|
"learning_rate": 4.825223713646533e-05, |
|
"loss": 5.6038, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 16.78, |
|
"learning_rate": 4.781529642058166e-05, |
|
"loss": 5.3695, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 20.13, |
|
"learning_rate": 4.737835570469799e-05, |
|
"loss": 5.1081, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 23.49, |
|
"learning_rate": 4.694141498881432e-05, |
|
"loss": 4.84, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 26.85, |
|
"learning_rate": 4.6504474272930655e-05, |
|
"loss": 4.6172, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 30.2, |
|
"learning_rate": 4.6067533557046985e-05, |
|
"loss": 4.4057, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 33.56, |
|
"learning_rate": 4.563059284116331e-05, |
|
"loss": 4.2367, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 36.91, |
|
"learning_rate": 4.5193652125279645e-05, |
|
"loss": 4.0491, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 40.27, |
|
"learning_rate": 4.4756711409395975e-05, |
|
"loss": 3.9008, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 43.62, |
|
"learning_rate": 4.4319770693512305e-05, |
|
"loss": 3.7661, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 46.98, |
|
"learning_rate": 4.3882829977628635e-05, |
|
"loss": 3.6335, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 50.34, |
|
"learning_rate": 4.344588926174497e-05, |
|
"loss": 3.4885, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 53.69, |
|
"learning_rate": 4.30089485458613e-05, |
|
"loss": 3.377, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 57.05, |
|
"learning_rate": 4.257200782997763e-05, |
|
"loss": 3.275, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 60.4, |
|
"learning_rate": 4.213506711409396e-05, |
|
"loss": 3.1363, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 63.76, |
|
"learning_rate": 4.169812639821029e-05, |
|
"loss": 3.0257, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 67.11, |
|
"learning_rate": 4.126118568232663e-05, |
|
"loss": 2.9229, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 70.47, |
|
"learning_rate": 4.082424496644295e-05, |
|
"loss": 2.8369, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 73.83, |
|
"learning_rate": 4.038730425055929e-05, |
|
"loss": 2.7189, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 77.18, |
|
"learning_rate": 3.995036353467562e-05, |
|
"loss": 2.6285, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 80.54, |
|
"learning_rate": 3.951342281879195e-05, |
|
"loss": 2.5286, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 83.89, |
|
"learning_rate": 3.907648210290828e-05, |
|
"loss": 2.4418, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 87.25, |
|
"learning_rate": 3.863954138702461e-05, |
|
"loss": 2.3471, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 90.6, |
|
"learning_rate": 3.8202600671140944e-05, |
|
"loss": 2.2438, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 93.96, |
|
"learning_rate": 3.7765659955257274e-05, |
|
"loss": 2.1722, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 97.32, |
|
"learning_rate": 3.7328719239373604e-05, |
|
"loss": 2.0995, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 100.67, |
|
"learning_rate": 3.6891778523489934e-05, |
|
"loss": 2.0343, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 104.03, |
|
"learning_rate": 3.645483780760627e-05, |
|
"loss": 1.9535, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 107.38, |
|
"learning_rate": 3.6017897091722594e-05, |
|
"loss": 1.8755, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 110.74, |
|
"learning_rate": 3.5580956375838924e-05, |
|
"loss": 1.8003, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 114.09, |
|
"learning_rate": 3.514401565995526e-05, |
|
"loss": 1.7354, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 117.45, |
|
"learning_rate": 3.470707494407159e-05, |
|
"loss": 1.666, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 120.81, |
|
"learning_rate": 3.427013422818792e-05, |
|
"loss": 1.6116, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 124.16, |
|
"learning_rate": 3.383319351230425e-05, |
|
"loss": 1.5483, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 127.52, |
|
"learning_rate": 3.339625279642059e-05, |
|
"loss": 1.4874, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 130.87, |
|
"learning_rate": 3.295931208053692e-05, |
|
"loss": 1.4294, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 134.23, |
|
"learning_rate": 3.252237136465325e-05, |
|
"loss": 1.3839, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 137.58, |
|
"learning_rate": 3.2085430648769577e-05, |
|
"loss": 1.32, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 140.94, |
|
"learning_rate": 3.1648489932885906e-05, |
|
"loss": 1.2635, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 144.3, |
|
"learning_rate": 3.1211549217002236e-05, |
|
"loss": 1.2201, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 147.65, |
|
"learning_rate": 3.0774608501118566e-05, |
|
"loss": 1.1751, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 151.01, |
|
"learning_rate": 3.0337667785234903e-05, |
|
"loss": 1.1269, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 154.36, |
|
"learning_rate": 2.990072706935123e-05, |
|
"loss": 1.087, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 157.72, |
|
"learning_rate": 2.9463786353467566e-05, |
|
"loss": 1.0491, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 161.07, |
|
"learning_rate": 2.9026845637583893e-05, |
|
"loss": 1.0008, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 164.43, |
|
"learning_rate": 2.8589904921700223e-05, |
|
"loss": 0.9718, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 167.79, |
|
"learning_rate": 2.8152964205816556e-05, |
|
"loss": 0.9386, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 171.14, |
|
"learning_rate": 2.7716023489932886e-05, |
|
"loss": 0.8923, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 174.5, |
|
"learning_rate": 2.727908277404922e-05, |
|
"loss": 0.8642, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 177.85, |
|
"learning_rate": 2.684214205816555e-05, |
|
"loss": 0.8361, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 181.21, |
|
"learning_rate": 2.6405201342281882e-05, |
|
"loss": 0.8075, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 184.56, |
|
"learning_rate": 2.5968260626398212e-05, |
|
"loss": 0.7729, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 187.92, |
|
"learning_rate": 2.5531319910514546e-05, |
|
"loss": 0.7429, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 191.28, |
|
"learning_rate": 2.5094379194630872e-05, |
|
"loss": 0.7253, |
|
"step": 28500 |
|
}, |
|
{ |
|
"epoch": 194.63, |
|
"learning_rate": 2.4657438478747206e-05, |
|
"loss": 0.693, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 197.99, |
|
"learning_rate": 2.4220497762863535e-05, |
|
"loss": 0.6717, |
|
"step": 29500 |
|
}, |
|
{ |
|
"epoch": 201.34, |
|
"learning_rate": 2.378355704697987e-05, |
|
"loss": 0.648, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 204.7, |
|
"learning_rate": 2.3346616331096195e-05, |
|
"loss": 0.617, |
|
"step": 30500 |
|
}, |
|
{ |
|
"epoch": 208.05, |
|
"learning_rate": 2.290967561521253e-05, |
|
"loss": 0.6145, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 211.41, |
|
"learning_rate": 2.247273489932886e-05, |
|
"loss": 0.5925, |
|
"step": 31500 |
|
}, |
|
{ |
|
"epoch": 214.77, |
|
"learning_rate": 2.2035794183445192e-05, |
|
"loss": 0.574, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 218.12, |
|
"learning_rate": 2.1598853467561522e-05, |
|
"loss": 0.5549, |
|
"step": 32500 |
|
}, |
|
{ |
|
"epoch": 221.48, |
|
"learning_rate": 2.1161912751677855e-05, |
|
"loss": 0.543, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 224.83, |
|
"learning_rate": 2.0724972035794185e-05, |
|
"loss": 0.5261, |
|
"step": 33500 |
|
}, |
|
{ |
|
"epoch": 228.19, |
|
"learning_rate": 2.0288031319910515e-05, |
|
"loss": 0.5083, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 231.54, |
|
"learning_rate": 1.9851090604026848e-05, |
|
"loss": 0.4944, |
|
"step": 34500 |
|
}, |
|
{ |
|
"epoch": 234.9, |
|
"learning_rate": 1.9414149888143178e-05, |
|
"loss": 0.4789, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 238.26, |
|
"learning_rate": 1.8977209172259508e-05, |
|
"loss": 0.4702, |
|
"step": 35500 |
|
}, |
|
{ |
|
"epoch": 241.61, |
|
"learning_rate": 1.8540268456375838e-05, |
|
"loss": 0.4567, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 244.97, |
|
"learning_rate": 1.810332774049217e-05, |
|
"loss": 0.4417, |
|
"step": 36500 |
|
}, |
|
{ |
|
"epoch": 248.32, |
|
"learning_rate": 1.76663870246085e-05, |
|
"loss": 0.4414, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 251.68, |
|
"learning_rate": 1.7229446308724834e-05, |
|
"loss": 0.431, |
|
"step": 37500 |
|
}, |
|
{ |
|
"epoch": 255.03, |
|
"learning_rate": 1.6792505592841164e-05, |
|
"loss": 0.4134, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 258.39, |
|
"learning_rate": 1.6355564876957498e-05, |
|
"loss": 0.4047, |
|
"step": 38500 |
|
}, |
|
{ |
|
"epoch": 261.74, |
|
"learning_rate": 1.5918624161073824e-05, |
|
"loss": 0.3894, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 265.1, |
|
"learning_rate": 1.5481683445190158e-05, |
|
"loss": 0.3924, |
|
"step": 39500 |
|
}, |
|
{ |
|
"epoch": 268.46, |
|
"learning_rate": 1.5044742729306487e-05, |
|
"loss": 0.3782, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 271.81, |
|
"learning_rate": 1.4607802013422819e-05, |
|
"loss": 0.3786, |
|
"step": 40500 |
|
}, |
|
{ |
|
"epoch": 275.17, |
|
"learning_rate": 1.417086129753915e-05, |
|
"loss": 0.361, |
|
"step": 41000 |
|
}, |
|
{ |
|
"epoch": 278.52, |
|
"learning_rate": 1.3733920581655482e-05, |
|
"loss": 0.3588, |
|
"step": 41500 |
|
}, |
|
{ |
|
"epoch": 281.88, |
|
"learning_rate": 1.3296979865771814e-05, |
|
"loss": 0.35, |
|
"step": 42000 |
|
}, |
|
{ |
|
"epoch": 285.23, |
|
"learning_rate": 1.2860039149888146e-05, |
|
"loss": 0.349, |
|
"step": 42500 |
|
}, |
|
{ |
|
"epoch": 288.59, |
|
"learning_rate": 1.2423098434004475e-05, |
|
"loss": 0.344, |
|
"step": 43000 |
|
}, |
|
{ |
|
"epoch": 291.95, |
|
"learning_rate": 1.1986157718120807e-05, |
|
"loss": 0.3314, |
|
"step": 43500 |
|
}, |
|
{ |
|
"epoch": 295.3, |
|
"learning_rate": 1.1549217002237137e-05, |
|
"loss": 0.3288, |
|
"step": 44000 |
|
}, |
|
{ |
|
"epoch": 298.66, |
|
"learning_rate": 1.1112276286353469e-05, |
|
"loss": 0.3259, |
|
"step": 44500 |
|
}, |
|
{ |
|
"epoch": 302.01, |
|
"learning_rate": 1.06753355704698e-05, |
|
"loss": 0.3161, |
|
"step": 45000 |
|
}, |
|
{ |
|
"epoch": 305.37, |
|
"learning_rate": 1.023839485458613e-05, |
|
"loss": 0.3108, |
|
"step": 45500 |
|
}, |
|
{ |
|
"epoch": 308.72, |
|
"learning_rate": 9.80145413870246e-06, |
|
"loss": 0.3141, |
|
"step": 46000 |
|
}, |
|
{ |
|
"epoch": 312.08, |
|
"learning_rate": 9.364513422818792e-06, |
|
"loss": 0.3022, |
|
"step": 46500 |
|
}, |
|
{ |
|
"epoch": 315.44, |
|
"learning_rate": 8.927572706935123e-06, |
|
"loss": 0.2995, |
|
"step": 47000 |
|
}, |
|
{ |
|
"epoch": 318.79, |
|
"learning_rate": 8.490631991051455e-06, |
|
"loss": 0.294, |
|
"step": 47500 |
|
}, |
|
{ |
|
"epoch": 322.15, |
|
"learning_rate": 8.053691275167785e-06, |
|
"loss": 0.2935, |
|
"step": 48000 |
|
}, |
|
{ |
|
"epoch": 325.5, |
|
"learning_rate": 7.6167505592841164e-06, |
|
"loss": 0.2906, |
|
"step": 48500 |
|
}, |
|
{ |
|
"epoch": 328.86, |
|
"learning_rate": 7.179809843400448e-06, |
|
"loss": 0.2854, |
|
"step": 49000 |
|
}, |
|
{ |
|
"epoch": 332.21, |
|
"learning_rate": 6.74286912751678e-06, |
|
"loss": 0.2768, |
|
"step": 49500 |
|
}, |
|
{ |
|
"epoch": 335.57, |
|
"learning_rate": 6.3059284116331096e-06, |
|
"loss": 0.2799, |
|
"step": 50000 |
|
}, |
|
{ |
|
"epoch": 338.93, |
|
"learning_rate": 5.868987695749441e-06, |
|
"loss": 0.2696, |
|
"step": 50500 |
|
}, |
|
{ |
|
"epoch": 342.28, |
|
"learning_rate": 5.432046979865772e-06, |
|
"loss": 0.2713, |
|
"step": 51000 |
|
}, |
|
{ |
|
"epoch": 345.64, |
|
"learning_rate": 4.995106263982103e-06, |
|
"loss": 0.2651, |
|
"step": 51500 |
|
}, |
|
{ |
|
"epoch": 348.99, |
|
"learning_rate": 4.558165548098434e-06, |
|
"loss": 0.2651, |
|
"step": 52000 |
|
}, |
|
{ |
|
"epoch": 352.35, |
|
"learning_rate": 4.121224832214765e-06, |
|
"loss": 0.2621, |
|
"step": 52500 |
|
}, |
|
{ |
|
"epoch": 355.7, |
|
"learning_rate": 3.6842841163310963e-06, |
|
"loss": 0.2671, |
|
"step": 53000 |
|
}, |
|
{ |
|
"epoch": 359.06, |
|
"learning_rate": 3.247343400447427e-06, |
|
"loss": 0.2538, |
|
"step": 53500 |
|
}, |
|
{ |
|
"epoch": 362.42, |
|
"learning_rate": 2.8104026845637587e-06, |
|
"loss": 0.2588, |
|
"step": 54000 |
|
}, |
|
{ |
|
"epoch": 365.77, |
|
"learning_rate": 2.37346196868009e-06, |
|
"loss": 0.2522, |
|
"step": 54500 |
|
}, |
|
{ |
|
"epoch": 369.13, |
|
"learning_rate": 1.9365212527964206e-06, |
|
"loss": 0.2577, |
|
"step": 55000 |
|
}, |
|
{ |
|
"epoch": 372.48, |
|
"learning_rate": 1.4995805369127516e-06, |
|
"loss": 0.244, |
|
"step": 55500 |
|
}, |
|
{ |
|
"epoch": 375.84, |
|
"learning_rate": 1.062639821029083e-06, |
|
"loss": 0.2492, |
|
"step": 56000 |
|
} |
|
], |
|
"max_steps": 57216, |
|
"num_train_epochs": 384, |
|
"total_flos": 5.557730674767398e+16, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|