|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 3.0, |
|
"eval_steps": 500, |
|
"global_step": 672, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.004464285714285714, |
|
"grad_norm": 118.73274230957031, |
|
"learning_rate": 0.0, |
|
"loss": 3.3178, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.008928571428571428, |
|
"grad_norm": 120.67337036132812, |
|
"learning_rate": 4.6275642631951835e-06, |
|
"loss": 3.3568, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.013392857142857142, |
|
"grad_norm": 115.98211669921875, |
|
"learning_rate": 7.3345158268416935e-06, |
|
"loss": 3.2911, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.017857142857142856, |
|
"grad_norm": 19.323978424072266, |
|
"learning_rate": 9.255128526390367e-06, |
|
"loss": 2.589, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.022321428571428572, |
|
"grad_norm": 11.930215835571289, |
|
"learning_rate": 1.0744871473609633e-05, |
|
"loss": 2.1897, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.026785714285714284, |
|
"grad_norm": 11.268843650817871, |
|
"learning_rate": 1.1962080090036879e-05, |
|
"loss": 1.6538, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.03125, |
|
"grad_norm": 17.312570571899414, |
|
"learning_rate": 1.299121531141887e-05, |
|
"loss": 1.1579, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.03571428571428571, |
|
"grad_norm": 11.755829811096191, |
|
"learning_rate": 1.388269278958555e-05, |
|
"loss": 0.6901, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.04017857142857143, |
|
"grad_norm": 9.42724609375, |
|
"learning_rate": 1.4669031653683387e-05, |
|
"loss": 0.442, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.044642857142857144, |
|
"grad_norm": 7.048767566680908, |
|
"learning_rate": 1.537243573680482e-05, |
|
"loss": 0.2483, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.049107142857142856, |
|
"grad_norm": 11.449494361877441, |
|
"learning_rate": 1.600874212937343e-05, |
|
"loss": 0.184, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.05357142857142857, |
|
"grad_norm": 4.351762771606445, |
|
"learning_rate": 1.6589644353232063e-05, |
|
"loss": 0.0867, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.05803571428571429, |
|
"grad_norm": 8.21219253540039, |
|
"learning_rate": 1.712402259777778e-05, |
|
"loss": 0.079, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.0625, |
|
"grad_norm": 2.505552053451538, |
|
"learning_rate": 1.7618779574614054e-05, |
|
"loss": 0.1795, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.06696428571428571, |
|
"grad_norm": 1.9123979806900024, |
|
"learning_rate": 1.8079387300451327e-05, |
|
"loss": 0.0686, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.07142857142857142, |
|
"grad_norm": 1.5148472785949707, |
|
"learning_rate": 1.8510257052780734e-05, |
|
"loss": 0.0645, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.07589285714285714, |
|
"grad_norm": 1.7530207633972168, |
|
"learning_rate": 1.891499697130832e-05, |
|
"loss": 0.0598, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.08035714285714286, |
|
"grad_norm": 1.9798786640167236, |
|
"learning_rate": 1.929659591687857e-05, |
|
"loss": 0.0519, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.08482142857142858, |
|
"grad_norm": 1.4390745162963867, |
|
"learning_rate": 1.9657557553855117e-05, |
|
"loss": 0.0359, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.08928571428571429, |
|
"grad_norm": 0.8786221742630005, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0746, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.09375, |
|
"grad_norm": 0.8981115818023682, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0331, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.09821428571428571, |
|
"grad_norm": 2.092228889465332, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0339, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.10267857142857142, |
|
"grad_norm": 0.6631488800048828, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0339, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.10714285714285714, |
|
"grad_norm": 0.7038796544075012, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0381, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.11160714285714286, |
|
"grad_norm": 1.3783732652664185, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0934, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.11607142857142858, |
|
"grad_norm": 2.0931007862091064, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0409, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.12053571428571429, |
|
"grad_norm": 3.667280435562134, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0325, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.125, |
|
"grad_norm": 1.4841816425323486, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0242, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.12946428571428573, |
|
"grad_norm": 1.6250470876693726, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0337, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.13392857142857142, |
|
"grad_norm": 0.6866446137428284, |
|
"learning_rate": 2e-05, |
|
"loss": 0.031, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.13839285714285715, |
|
"grad_norm": 1.0550711154937744, |
|
"learning_rate": 2e-05, |
|
"loss": 0.05, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.14285714285714285, |
|
"grad_norm": 0.4552890360355377, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0329, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.14732142857142858, |
|
"grad_norm": 1.899696946144104, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0191, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.15178571428571427, |
|
"grad_norm": 31.991363525390625, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0252, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.15625, |
|
"grad_norm": 7.586791038513184, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0318, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.16071428571428573, |
|
"grad_norm": 0.9316713809967041, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0275, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.16517857142857142, |
|
"grad_norm": 0.8239340782165527, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0175, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.16964285714285715, |
|
"grad_norm": 0.2483951449394226, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0102, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.17410714285714285, |
|
"grad_norm": 0.41918331384658813, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0235, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.17857142857142858, |
|
"grad_norm": 0.5702436566352844, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0344, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.18303571428571427, |
|
"grad_norm": 0.31215420365333557, |
|
"learning_rate": 2e-05, |
|
"loss": 0.018, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.1875, |
|
"grad_norm": 0.36833837628364563, |
|
"learning_rate": 2e-05, |
|
"loss": 0.024, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.19196428571428573, |
|
"grad_norm": 0.14042280614376068, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0075, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.19642857142857142, |
|
"grad_norm": 0.25796693563461304, |
|
"learning_rate": 2e-05, |
|
"loss": 0.016, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.20089285714285715, |
|
"grad_norm": 0.32930517196655273, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0217, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.20535714285714285, |
|
"grad_norm": 0.5433669090270996, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0204, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.20982142857142858, |
|
"grad_norm": 0.2668748199939728, |
|
"learning_rate": 2e-05, |
|
"loss": 0.012, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.21428571428571427, |
|
"grad_norm": 0.12396270036697388, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0071, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.21875, |
|
"grad_norm": 0.3738039433956146, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0172, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.22321428571428573, |
|
"grad_norm": 0.2019537091255188, |
|
"learning_rate": 2e-05, |
|
"loss": 0.009, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.22767857142857142, |
|
"grad_norm": 0.2405676245689392, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0124, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.23214285714285715, |
|
"grad_norm": 0.3147808611392975, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0106, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.23660714285714285, |
|
"grad_norm": 0.19005292654037476, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0086, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.24107142857142858, |
|
"grad_norm": 0.5401691198348999, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0202, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.24553571428571427, |
|
"grad_norm": 0.10109489411115646, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0061, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.5185668468475342, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0275, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.2544642857142857, |
|
"grad_norm": 0.3303714990615845, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0154, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.25892857142857145, |
|
"grad_norm": 0.31375738978385925, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0169, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.26339285714285715, |
|
"grad_norm": 0.33531704545021057, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0173, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.26785714285714285, |
|
"grad_norm": 0.34593161940574646, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0152, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.27232142857142855, |
|
"grad_norm": 0.339955598115921, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0179, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.2767857142857143, |
|
"grad_norm": 0.40512269735336304, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0181, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.28125, |
|
"grad_norm": 0.39926639199256897, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0229, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.2857142857142857, |
|
"grad_norm": 0.49879220128059387, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0182, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.29017857142857145, |
|
"grad_norm": 0.6592809557914734, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0791, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.29464285714285715, |
|
"grad_norm": 1.4562678337097168, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0391, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.29910714285714285, |
|
"grad_norm": 0.2671646773815155, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0122, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.30357142857142855, |
|
"grad_norm": 0.398265540599823, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0113, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.3080357142857143, |
|
"grad_norm": 0.4379926323890686, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0557, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.3125, |
|
"grad_norm": 0.18577145040035248, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0105, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.3169642857142857, |
|
"grad_norm": 0.27026066184043884, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0097, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.32142857142857145, |
|
"grad_norm": 0.35076332092285156, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0122, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.32589285714285715, |
|
"grad_norm": 0.16470995545387268, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0094, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.33035714285714285, |
|
"grad_norm": 0.2735673785209656, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0131, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.33482142857142855, |
|
"grad_norm": 0.4593258798122406, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0074, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.3392857142857143, |
|
"grad_norm": 0.285243421792984, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0074, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.34375, |
|
"grad_norm": 0.10123898833990097, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0073, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.3482142857142857, |
|
"grad_norm": 0.17487011849880219, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0086, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.35267857142857145, |
|
"grad_norm": 0.3148305416107178, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0318, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.35714285714285715, |
|
"grad_norm": 0.4343680143356323, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0109, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.36160714285714285, |
|
"grad_norm": 0.24763138592243195, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0096, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.36607142857142855, |
|
"grad_norm": 0.12427603453397751, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0084, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.3705357142857143, |
|
"grad_norm": 0.37291279435157776, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0322, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.375, |
|
"grad_norm": 0.1884593665599823, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0066, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.3794642857142857, |
|
"grad_norm": 0.22058643400669098, |
|
"learning_rate": 2e-05, |
|
"loss": 0.01, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.38392857142857145, |
|
"grad_norm": 0.209506556391716, |
|
"learning_rate": 2e-05, |
|
"loss": 0.01, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.38839285714285715, |
|
"grad_norm": 0.15874722599983215, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0096, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.39285714285714285, |
|
"grad_norm": 0.24815426766872406, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0126, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.39732142857142855, |
|
"grad_norm": 0.17406296730041504, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0089, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.4017857142857143, |
|
"grad_norm": 0.2863009572029114, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0168, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.40625, |
|
"grad_norm": 0.3468080759048462, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0317, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.4107142857142857, |
|
"grad_norm": 0.3998833894729614, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0398, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.41517857142857145, |
|
"grad_norm": 0.16211910545825958, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0093, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.41964285714285715, |
|
"grad_norm": 0.27588722109794617, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0119, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.42410714285714285, |
|
"grad_norm": 0.17133073508739471, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0108, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.42857142857142855, |
|
"grad_norm": 0.22069743275642395, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0113, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.4330357142857143, |
|
"grad_norm": 0.14985281229019165, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0056, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.4375, |
|
"grad_norm": 0.2660472095012665, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0092, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.4419642857142857, |
|
"grad_norm": 0.15048441290855408, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0107, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.44642857142857145, |
|
"grad_norm": 0.09269213676452637, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0065, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.45089285714285715, |
|
"grad_norm": 0.08095856755971909, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0073, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.45535714285714285, |
|
"grad_norm": 0.22052565217018127, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0077, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.45982142857142855, |
|
"grad_norm": 0.15201696753501892, |
|
"learning_rate": 2e-05, |
|
"loss": 0.011, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.4642857142857143, |
|
"grad_norm": 0.2241186946630478, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0112, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.46875, |
|
"grad_norm": 0.2035360485315323, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0086, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.4732142857142857, |
|
"grad_norm": 0.16850705444812775, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0081, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.47767857142857145, |
|
"grad_norm": 0.265540212392807, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0116, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.48214285714285715, |
|
"grad_norm": 0.15132322907447815, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0087, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.48660714285714285, |
|
"grad_norm": 0.23885183036327362, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0087, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.49107142857142855, |
|
"grad_norm": 0.16768532991409302, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0093, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.4955357142857143, |
|
"grad_norm": 0.156480073928833, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0048, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.31494247913360596, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0171, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.5044642857142857, |
|
"grad_norm": 0.3249429166316986, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0136, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.5089285714285714, |
|
"grad_norm": 0.18467825651168823, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0113, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.5133928571428571, |
|
"grad_norm": 0.11450862139463425, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0058, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.5178571428571429, |
|
"grad_norm": 0.12011968344449997, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0069, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.5223214285714286, |
|
"grad_norm": 0.4256938397884369, |
|
"learning_rate": 2e-05, |
|
"loss": 0.03, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.5267857142857143, |
|
"grad_norm": 0.12585486471652985, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0076, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.53125, |
|
"grad_norm": 0.13013215363025665, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0067, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.5357142857142857, |
|
"grad_norm": 0.44324806332588196, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0299, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.5401785714285714, |
|
"grad_norm": 0.09240884333848953, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0081, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.5446428571428571, |
|
"grad_norm": 0.11633799970149994, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0073, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.5491071428571429, |
|
"grad_norm": 0.13074152171611786, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0083, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.5535714285714286, |
|
"grad_norm": 0.08824117481708527, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0066, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.5580357142857143, |
|
"grad_norm": 0.1764586865901947, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0115, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.5625, |
|
"grad_norm": 0.28081825375556946, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0157, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.5669642857142857, |
|
"grad_norm": 0.14706389605998993, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0097, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.5714285714285714, |
|
"grad_norm": 0.13169962167739868, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0075, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.5758928571428571, |
|
"grad_norm": 0.24221737682819366, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0094, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.5803571428571429, |
|
"grad_norm": 0.2343599945306778, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0125, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.5848214285714286, |
|
"grad_norm": 0.24817775189876556, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0169, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.5892857142857143, |
|
"grad_norm": 0.1439865231513977, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0088, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.59375, |
|
"grad_norm": 0.5774053335189819, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0115, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.5982142857142857, |
|
"grad_norm": 0.27249982953071594, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0079, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.6026785714285714, |
|
"grad_norm": 0.21357686817646027, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0052, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.6071428571428571, |
|
"grad_norm": 0.139152392745018, |
|
"learning_rate": 2e-05, |
|
"loss": 0.007, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.6116071428571429, |
|
"grad_norm": 0.3676150143146515, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0181, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.6160714285714286, |
|
"grad_norm": 0.1653941422700882, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0103, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.6205357142857143, |
|
"grad_norm": 0.13131186366081238, |
|
"learning_rate": 2e-05, |
|
"loss": 0.009, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.625, |
|
"grad_norm": 0.1705312728881836, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0127, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.6294642857142857, |
|
"grad_norm": 0.20293530821800232, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0106, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.6339285714285714, |
|
"grad_norm": 0.19262683391571045, |
|
"learning_rate": 2e-05, |
|
"loss": 0.011, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.6383928571428571, |
|
"grad_norm": 0.13301971554756165, |
|
"learning_rate": 2e-05, |
|
"loss": 0.01, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.6428571428571429, |
|
"grad_norm": 0.13495324552059174, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0087, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.6473214285714286, |
|
"grad_norm": 0.16713371872901917, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0062, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.6517857142857143, |
|
"grad_norm": 0.153954416513443, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0092, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.65625, |
|
"grad_norm": 0.13245932757854462, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0098, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.6607142857142857, |
|
"grad_norm": 0.10969486832618713, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0069, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.6651785714285714, |
|
"grad_norm": 0.3466426432132721, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0155, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.6696428571428571, |
|
"grad_norm": 0.17744660377502441, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0093, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.6741071428571429, |
|
"grad_norm": 0.505264163017273, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0188, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.6785714285714286, |
|
"grad_norm": 0.225211963057518, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0124, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.6830357142857143, |
|
"grad_norm": 0.14465954899787903, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0082, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.6875, |
|
"grad_norm": 0.13549940288066864, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0092, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.6919642857142857, |
|
"grad_norm": 0.17627279460430145, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0107, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.6964285714285714, |
|
"grad_norm": 0.1911449432373047, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0111, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.7008928571428571, |
|
"grad_norm": 0.18086941540241241, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0074, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.7053571428571429, |
|
"grad_norm": 0.280097633600235, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0089, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.7098214285714286, |
|
"grad_norm": 0.1896984875202179, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0077, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.7142857142857143, |
|
"grad_norm": 0.16626662015914917, |
|
"learning_rate": 2e-05, |
|
"loss": 0.005, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.71875, |
|
"grad_norm": 0.18548309803009033, |
|
"learning_rate": 2e-05, |
|
"loss": 0.012, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.7232142857142857, |
|
"grad_norm": 0.19919419288635254, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0099, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.7276785714285714, |
|
"grad_norm": 0.2509930729866028, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0076, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.7321428571428571, |
|
"grad_norm": 0.21043901145458221, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0145, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.7366071428571429, |
|
"grad_norm": 0.11943473666906357, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0058, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.7410714285714286, |
|
"grad_norm": 0.2109983116388321, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0062, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.7455357142857143, |
|
"grad_norm": 0.26444128155708313, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0076, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 0.2525187134742737, |
|
"learning_rate": 2e-05, |
|
"loss": 0.01, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.7544642857142857, |
|
"grad_norm": 0.26021668314933777, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0135, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.7589285714285714, |
|
"grad_norm": 0.25887635350227356, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0134, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.7633928571428571, |
|
"grad_norm": 0.4230816960334778, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0087, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.7678571428571429, |
|
"grad_norm": 0.4137205183506012, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0108, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.7723214285714286, |
|
"grad_norm": 0.12480466812849045, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0081, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.7767857142857143, |
|
"grad_norm": 0.2187943309545517, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0062, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.78125, |
|
"grad_norm": 0.09233838319778442, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0075, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.7857142857142857, |
|
"grad_norm": 0.1769775003194809, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0091, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.7901785714285714, |
|
"grad_norm": 0.12783905863761902, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0071, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.7946428571428571, |
|
"grad_norm": 0.3727477788925171, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0103, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.7991071428571429, |
|
"grad_norm": 0.17474934458732605, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0112, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.8035714285714286, |
|
"grad_norm": 0.10964759439229965, |
|
"learning_rate": 2e-05, |
|
"loss": 0.009, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.8080357142857143, |
|
"grad_norm": 0.386709988117218, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0098, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.8125, |
|
"grad_norm": 0.20667214691638947, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0107, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.8169642857142857, |
|
"grad_norm": 0.13784578442573547, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0062, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.8214285714285714, |
|
"grad_norm": 0.11419745534658432, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0068, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.8258928571428571, |
|
"grad_norm": 0.1253211796283722, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0076, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.8303571428571429, |
|
"grad_norm": 0.16259260475635529, |
|
"learning_rate": 2e-05, |
|
"loss": 0.008, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.8348214285714286, |
|
"grad_norm": 0.22637289762496948, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0125, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.8392857142857143, |
|
"grad_norm": 0.11347130686044693, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0063, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.84375, |
|
"grad_norm": 0.23896394670009613, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0069, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.8482142857142857, |
|
"grad_norm": 0.12304235994815826, |
|
"learning_rate": 2e-05, |
|
"loss": 0.006, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.8526785714285714, |
|
"grad_norm": 0.14501796662807465, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0077, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.8571428571428571, |
|
"grad_norm": 0.22769401967525482, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0079, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.8616071428571429, |
|
"grad_norm": 0.20632179081439972, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0109, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.8660714285714286, |
|
"grad_norm": 0.09508738666772842, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0066, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.8705357142857143, |
|
"grad_norm": 0.16956081986427307, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0056, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.875, |
|
"grad_norm": 0.19381490349769592, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0142, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.8794642857142857, |
|
"grad_norm": 0.28475290536880493, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0123, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.8839285714285714, |
|
"grad_norm": 0.13424070179462433, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0091, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.8883928571428571, |
|
"grad_norm": 0.20577555894851685, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0098, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.8928571428571429, |
|
"grad_norm": 0.34151265025138855, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0142, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.8973214285714286, |
|
"grad_norm": 0.211790069937706, |
|
"learning_rate": 2e-05, |
|
"loss": 0.007, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.9017857142857143, |
|
"grad_norm": 0.17058762907981873, |
|
"learning_rate": 2e-05, |
|
"loss": 0.008, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.90625, |
|
"grad_norm": 0.3145332634449005, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0212, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.9107142857142857, |
|
"grad_norm": 0.1913270354270935, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0119, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.9151785714285714, |
|
"grad_norm": 0.17364099621772766, |
|
"learning_rate": 2e-05, |
|
"loss": 0.01, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.9196428571428571, |
|
"grad_norm": 0.20001767575740814, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0101, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 0.9241071428571429, |
|
"grad_norm": 0.13925513625144958, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0061, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.9285714285714286, |
|
"grad_norm": 0.13868604600429535, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0074, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.9330357142857143, |
|
"grad_norm": 0.27624329924583435, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0081, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 0.9375, |
|
"grad_norm": 0.1608811616897583, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0086, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.9419642857142857, |
|
"grad_norm": 0.0851345956325531, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0065, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 0.9464285714285714, |
|
"grad_norm": 0.34213757514953613, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0141, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.9508928571428571, |
|
"grad_norm": 0.09006419777870178, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0053, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.9553571428571429, |
|
"grad_norm": 0.1785372495651245, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0094, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 0.9598214285714286, |
|
"grad_norm": 0.1560213714838028, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0077, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.9642857142857143, |
|
"grad_norm": 0.11964119970798492, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0083, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.96875, |
|
"grad_norm": 0.11234603822231293, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0067, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 0.9732142857142857, |
|
"grad_norm": 0.11584550887346268, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0095, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 0.9776785714285714, |
|
"grad_norm": 0.04019077867269516, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0048, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.9821428571428571, |
|
"grad_norm": 0.20271597802639008, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0068, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.9866071428571429, |
|
"grad_norm": 0.11852418631315231, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0075, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 0.9910714285714286, |
|
"grad_norm": 0.08768610656261444, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0059, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.9955357142857143, |
|
"grad_norm": 0.2536848485469818, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0083, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.19689269363880157, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0095, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 1.0044642857142858, |
|
"grad_norm": 0.10952406376600266, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0075, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 1.0089285714285714, |
|
"grad_norm": 0.19677336513996124, |
|
"learning_rate": 2e-05, |
|
"loss": 0.005, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 1.0133928571428572, |
|
"grad_norm": 0.12416896969079971, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0057, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 1.0178571428571428, |
|
"grad_norm": 0.14256438612937927, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0062, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 1.0223214285714286, |
|
"grad_norm": 0.17529906332492828, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0078, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 1.0267857142857142, |
|
"grad_norm": 0.22035321593284607, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0081, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 1.03125, |
|
"grad_norm": 0.21250584721565247, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0067, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 1.0357142857142858, |
|
"grad_norm": 0.18901652097702026, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0075, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 1.0401785714285714, |
|
"grad_norm": 0.19155669212341309, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0117, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 1.0446428571428572, |
|
"grad_norm": 0.1865350306034088, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0078, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 1.0491071428571428, |
|
"grad_norm": 0.15170250833034515, |
|
"learning_rate": 2e-05, |
|
"loss": 0.008, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 1.0535714285714286, |
|
"grad_norm": 0.10518006980419159, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0062, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 1.0580357142857142, |
|
"grad_norm": 0.2728783190250397, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0084, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 1.0625, |
|
"grad_norm": 0.21080271899700165, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0102, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 1.0669642857142858, |
|
"grad_norm": 0.13723178207874298, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0089, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 1.0714285714285714, |
|
"grad_norm": 0.19935685396194458, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0057, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 1.0758928571428572, |
|
"grad_norm": 0.12787075340747833, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0071, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 1.0803571428571428, |
|
"grad_norm": 0.3154740035533905, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0106, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 1.0848214285714286, |
|
"grad_norm": 0.1168612465262413, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0076, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 1.0892857142857142, |
|
"grad_norm": 0.10894633084535599, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0079, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 1.09375, |
|
"grad_norm": 0.1844814121723175, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0103, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 1.0982142857142858, |
|
"grad_norm": 0.5986953377723694, |
|
"learning_rate": 2e-05, |
|
"loss": 0.008, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 1.1026785714285714, |
|
"grad_norm": 0.23711133003234863, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0079, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 1.1071428571428572, |
|
"grad_norm": 0.28520116209983826, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0071, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 1.1116071428571428, |
|
"grad_norm": 0.2094752937555313, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0077, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 1.1160714285714286, |
|
"grad_norm": 0.17041270434856415, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0085, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 1.1205357142857142, |
|
"grad_norm": 0.26283618807792664, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0078, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 1.125, |
|
"grad_norm": 0.40016844868659973, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0084, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 1.1294642857142858, |
|
"grad_norm": 0.13586895167827606, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0075, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 1.1339285714285714, |
|
"grad_norm": 0.26292648911476135, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0078, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 1.1383928571428572, |
|
"grad_norm": 0.6343429088592529, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0093, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 1.1428571428571428, |
|
"grad_norm": 0.1948457509279251, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0063, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 1.1473214285714286, |
|
"grad_norm": 0.18401798605918884, |
|
"learning_rate": 2e-05, |
|
"loss": 0.008, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 1.1517857142857142, |
|
"grad_norm": 0.4655361771583557, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0138, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 1.15625, |
|
"grad_norm": 0.11816074699163437, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0084, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 1.1607142857142858, |
|
"grad_norm": 0.11838362365961075, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0088, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 1.1651785714285714, |
|
"grad_norm": 0.16618025302886963, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0072, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 1.1696428571428572, |
|
"grad_norm": 0.1511249542236328, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0103, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 1.1741071428571428, |
|
"grad_norm": 0.14847303926944733, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0093, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 1.1785714285714286, |
|
"grad_norm": 0.26684054732322693, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0078, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 1.1830357142857142, |
|
"grad_norm": 0.13314294815063477, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0088, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 1.1875, |
|
"grad_norm": 0.16953180730342865, |
|
"learning_rate": 2e-05, |
|
"loss": 0.008, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 1.1919642857142858, |
|
"grad_norm": 0.1619475930929184, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0058, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 1.1964285714285714, |
|
"grad_norm": 0.19401992857456207, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0103, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 1.2008928571428572, |
|
"grad_norm": 0.16010133922100067, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0072, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 1.2053571428571428, |
|
"grad_norm": 0.20059579610824585, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0073, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 1.2098214285714286, |
|
"grad_norm": 0.07786630094051361, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0059, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 1.2142857142857142, |
|
"grad_norm": 0.0951201319694519, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0079, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 1.21875, |
|
"grad_norm": 0.16951927542686462, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0094, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 1.2232142857142858, |
|
"grad_norm": 0.1350620537996292, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0071, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 1.2276785714285714, |
|
"grad_norm": 0.09529034048318863, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0071, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 1.2321428571428572, |
|
"grad_norm": 0.18849754333496094, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0112, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 1.2366071428571428, |
|
"grad_norm": 0.08065006136894226, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0053, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 1.2410714285714286, |
|
"grad_norm": 0.12004493921995163, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0095, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 1.2455357142857142, |
|
"grad_norm": 0.0798286497592926, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0053, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"grad_norm": 0.08489444851875305, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0062, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 1.2544642857142856, |
|
"grad_norm": 0.17156155407428741, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0072, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 1.2589285714285714, |
|
"grad_norm": 0.17118088901042938, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0102, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 1.2633928571428572, |
|
"grad_norm": 0.224991574883461, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0097, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 1.2678571428571428, |
|
"grad_norm": 0.11658413708209991, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0079, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 1.2723214285714286, |
|
"grad_norm": 0.20273533463478088, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0093, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 1.2767857142857144, |
|
"grad_norm": 0.09004299342632294, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0064, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 1.28125, |
|
"grad_norm": 0.18204613029956818, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0091, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 1.2857142857142856, |
|
"grad_norm": 0.049572091549634933, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0049, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 1.2901785714285714, |
|
"grad_norm": 0.1771448701620102, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0078, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 1.2946428571428572, |
|
"grad_norm": 0.13888761401176453, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0091, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 1.2991071428571428, |
|
"grad_norm": 0.197177454829216, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0089, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 1.3035714285714286, |
|
"grad_norm": 0.243514284491539, |
|
"learning_rate": 2e-05, |
|
"loss": 0.013, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 1.3080357142857144, |
|
"grad_norm": 0.09940832108259201, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0081, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 1.3125, |
|
"grad_norm": 0.299482136964798, |
|
"learning_rate": 2e-05, |
|
"loss": 0.013, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 1.3169642857142856, |
|
"grad_norm": 0.17354537546634674, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0086, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 1.3214285714285714, |
|
"grad_norm": 0.1544426530599594, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0059, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 1.3258928571428572, |
|
"grad_norm": 0.15909938514232635, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0075, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 1.3303571428571428, |
|
"grad_norm": 0.21977274119853973, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0086, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 1.3348214285714286, |
|
"grad_norm": 0.3289298415184021, |
|
"learning_rate": 2e-05, |
|
"loss": 0.005, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 1.3392857142857144, |
|
"grad_norm": 0.22125479578971863, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0101, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 1.34375, |
|
"grad_norm": 0.10905671864748001, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0075, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 1.3482142857142856, |
|
"grad_norm": 0.3605917692184448, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0086, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 1.3526785714285714, |
|
"grad_norm": 0.18675312399864197, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0062, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 1.3571428571428572, |
|
"grad_norm": 0.23379294574260712, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0112, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 1.3616071428571428, |
|
"grad_norm": 0.09766809642314911, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0065, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 1.3660714285714286, |
|
"grad_norm": 0.11652177572250366, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0101, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 1.3705357142857144, |
|
"grad_norm": 0.11609390377998352, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0087, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 1.375, |
|
"grad_norm": 0.3085818290710449, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0075, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 1.3794642857142856, |
|
"grad_norm": 0.30840498208999634, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0078, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 1.3839285714285714, |
|
"grad_norm": 0.12295576184988022, |
|
"learning_rate": 2e-05, |
|
"loss": 0.007, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 1.3883928571428572, |
|
"grad_norm": 0.2735302746295929, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0094, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 1.3928571428571428, |
|
"grad_norm": 0.22401046752929688, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0072, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 1.3973214285714286, |
|
"grad_norm": 0.2643716335296631, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0058, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 1.4017857142857144, |
|
"grad_norm": 0.09197328239679337, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0069, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 1.40625, |
|
"grad_norm": 0.15055829286575317, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0115, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 1.4107142857142856, |
|
"grad_norm": 0.10229172557592392, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0073, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 1.4151785714285714, |
|
"grad_norm": 0.1713918000459671, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0074, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 1.4196428571428572, |
|
"grad_norm": 0.17321155965328217, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0128, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 1.4241071428571428, |
|
"grad_norm": 0.14712217450141907, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0089, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 1.4285714285714286, |
|
"grad_norm": 0.08568955212831497, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0063, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 1.4330357142857144, |
|
"grad_norm": 0.10814769566059113, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0071, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 1.4375, |
|
"grad_norm": 0.11376317590475082, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0088, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 1.4419642857142856, |
|
"grad_norm": 0.09950172901153564, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0055, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 1.4464285714285714, |
|
"grad_norm": 0.10877425968647003, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0071, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 1.4508928571428572, |
|
"grad_norm": 0.16534090042114258, |
|
"learning_rate": 2e-05, |
|
"loss": 0.012, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 1.4553571428571428, |
|
"grad_norm": 0.11280371993780136, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0064, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 1.4598214285714286, |
|
"grad_norm": 0.09798318147659302, |
|
"learning_rate": 2e-05, |
|
"loss": 0.007, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 1.4642857142857144, |
|
"grad_norm": 0.21448449790477753, |
|
"learning_rate": 2e-05, |
|
"loss": 0.005, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 1.46875, |
|
"grad_norm": 0.11371182650327682, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0093, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 1.4732142857142856, |
|
"grad_norm": 0.08604440838098526, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0076, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 1.4776785714285714, |
|
"grad_norm": 0.19020172953605652, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0095, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 1.4821428571428572, |
|
"grad_norm": 0.07287605851888657, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0066, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 1.4866071428571428, |
|
"grad_norm": 0.23569829761981964, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0165, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 1.4910714285714286, |
|
"grad_norm": 0.13371407985687256, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0058, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 1.4955357142857144, |
|
"grad_norm": 0.2280549257993698, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0069, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"grad_norm": 0.09851525723934174, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0076, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 1.5044642857142856, |
|
"grad_norm": 0.17804968357086182, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0102, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 1.5089285714285714, |
|
"grad_norm": 0.08316322416067123, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0077, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 1.5133928571428572, |
|
"grad_norm": 0.09718410670757294, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0073, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 1.5178571428571428, |
|
"grad_norm": 0.20578241348266602, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0076, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 1.5223214285714286, |
|
"grad_norm": 0.22264248132705688, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0082, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 1.5267857142857144, |
|
"grad_norm": 0.13287998735904694, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0081, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 1.53125, |
|
"grad_norm": 0.17487944662570953, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0085, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 1.5357142857142856, |
|
"grad_norm": 0.34332460165023804, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0082, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 1.5401785714285714, |
|
"grad_norm": 0.12836158275604248, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0097, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 1.5446428571428572, |
|
"grad_norm": 0.21668285131454468, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0112, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 1.5491071428571428, |
|
"grad_norm": 0.20517054200172424, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0108, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 1.5535714285714286, |
|
"grad_norm": 0.3114481270313263, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0111, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 1.5580357142857144, |
|
"grad_norm": 0.19148893654346466, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0087, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 1.5625, |
|
"grad_norm": 0.07075007259845734, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0072, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 1.5669642857142856, |
|
"grad_norm": 0.12846648693084717, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0072, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 1.5714285714285714, |
|
"grad_norm": 0.1478446125984192, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0077, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 1.5758928571428572, |
|
"grad_norm": 0.16469308733940125, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0064, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 1.5803571428571428, |
|
"grad_norm": 0.07830236107110977, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0069, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 1.5848214285714286, |
|
"grad_norm": 0.1975899487733841, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0097, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 1.5892857142857144, |
|
"grad_norm": 0.0999147966504097, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0063, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 1.59375, |
|
"grad_norm": 0.13135221600532532, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0089, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 1.5982142857142856, |
|
"grad_norm": 0.15470464527606964, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0082, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 1.6026785714285714, |
|
"grad_norm": 0.12623795866966248, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0055, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 1.6071428571428572, |
|
"grad_norm": 0.14909787476062775, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0088, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 1.6116071428571428, |
|
"grad_norm": 0.16480869054794312, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0083, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 1.6160714285714286, |
|
"grad_norm": 0.2582751512527466, |
|
"learning_rate": 2e-05, |
|
"loss": 0.011, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 1.6205357142857144, |
|
"grad_norm": 0.11021137982606888, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0076, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 1.625, |
|
"grad_norm": 0.07676777243614197, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0082, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 1.6294642857142856, |
|
"grad_norm": 0.08494916558265686, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0059, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 1.6339285714285714, |
|
"grad_norm": 0.10398717224597931, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0072, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 1.6383928571428572, |
|
"grad_norm": 0.09390587359666824, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0079, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 1.6428571428571428, |
|
"grad_norm": 0.13992364704608917, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0085, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 1.6473214285714286, |
|
"grad_norm": 0.0727076306939125, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0064, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 1.6517857142857144, |
|
"grad_norm": 0.0938880518078804, |
|
"learning_rate": 2e-05, |
|
"loss": 0.007, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 1.65625, |
|
"grad_norm": 0.1349678784608841, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0083, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 1.6607142857142856, |
|
"grad_norm": 0.13271740078926086, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0074, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 1.6651785714285714, |
|
"grad_norm": 0.06232460215687752, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0058, |
|
"step": 373 |
|
}, |
|
{ |
|
"epoch": 1.6696428571428572, |
|
"grad_norm": 0.06478999555110931, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0062, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 1.6741071428571428, |
|
"grad_norm": 0.16400626301765442, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0091, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 1.6785714285714286, |
|
"grad_norm": 0.0681205615401268, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0051, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 1.6830357142857144, |
|
"grad_norm": 0.11032702773809433, |
|
"learning_rate": 2e-05, |
|
"loss": 0.008, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 1.6875, |
|
"grad_norm": 0.14635053277015686, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0091, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 1.6919642857142856, |
|
"grad_norm": 0.10891900956630707, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0065, |
|
"step": 379 |
|
}, |
|
{ |
|
"epoch": 1.6964285714285714, |
|
"grad_norm": 0.1412774920463562, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0059, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 1.7008928571428572, |
|
"grad_norm": 0.14827723801136017, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0053, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 1.7053571428571428, |
|
"grad_norm": 0.07940851151943207, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0076, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 1.7098214285714286, |
|
"grad_norm": 0.3140534460544586, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0082, |
|
"step": 383 |
|
}, |
|
{ |
|
"epoch": 1.7142857142857144, |
|
"grad_norm": 0.09956244379281998, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0065, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 1.71875, |
|
"grad_norm": 0.18810197710990906, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0084, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 1.7232142857142856, |
|
"grad_norm": 0.2445448487997055, |
|
"learning_rate": 2e-05, |
|
"loss": 0.013, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 1.7276785714285714, |
|
"grad_norm": 0.19565074145793915, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0106, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 1.7321428571428572, |
|
"grad_norm": 0.07408107817173004, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0054, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 1.7366071428571428, |
|
"grad_norm": 0.06907396763563156, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0055, |
|
"step": 389 |
|
}, |
|
{ |
|
"epoch": 1.7410714285714286, |
|
"grad_norm": 0.12801378965377808, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0092, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 1.7455357142857144, |
|
"grad_norm": 0.06491488218307495, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0065, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"grad_norm": 0.11618059128522873, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0058, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 1.7544642857142856, |
|
"grad_norm": 0.08376041054725647, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0054, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 1.7589285714285714, |
|
"grad_norm": 0.10667941719293594, |
|
"learning_rate": 2e-05, |
|
"loss": 0.01, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 1.7633928571428572, |
|
"grad_norm": 0.13104568421840668, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0073, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 1.7678571428571428, |
|
"grad_norm": 0.09212382137775421, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0076, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 1.7723214285714286, |
|
"grad_norm": 0.23598161339759827, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0093, |
|
"step": 397 |
|
}, |
|
{ |
|
"epoch": 1.7767857142857144, |
|
"grad_norm": 0.0689774751663208, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0054, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 1.78125, |
|
"grad_norm": 0.08331548422574997, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0088, |
|
"step": 399 |
|
}, |
|
{ |
|
"epoch": 1.7857142857142856, |
|
"grad_norm": 0.08793813735246658, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0065, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 1.7901785714285714, |
|
"grad_norm": 0.1980009227991104, |
|
"learning_rate": 2e-05, |
|
"loss": 0.007, |
|
"step": 401 |
|
}, |
|
{ |
|
"epoch": 1.7946428571428572, |
|
"grad_norm": 0.11130912601947784, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0081, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 1.7991071428571428, |
|
"grad_norm": 0.1447134166955948, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0067, |
|
"step": 403 |
|
}, |
|
{ |
|
"epoch": 1.8035714285714286, |
|
"grad_norm": 0.12303169071674347, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0081, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 1.8080357142857144, |
|
"grad_norm": 0.10215016454458237, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0065, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 1.8125, |
|
"grad_norm": 0.08570464700460434, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0072, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 1.8169642857142856, |
|
"grad_norm": 0.0914364829659462, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0067, |
|
"step": 407 |
|
}, |
|
{ |
|
"epoch": 1.8214285714285714, |
|
"grad_norm": 0.22513145208358765, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0089, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 1.8258928571428572, |
|
"grad_norm": 0.11378823965787888, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0071, |
|
"step": 409 |
|
}, |
|
{ |
|
"epoch": 1.8303571428571428, |
|
"grad_norm": 0.09297342598438263, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0091, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 1.8348214285714286, |
|
"grad_norm": 0.19639909267425537, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0113, |
|
"step": 411 |
|
}, |
|
{ |
|
"epoch": 1.8392857142857144, |
|
"grad_norm": 0.19027334451675415, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0086, |
|
"step": 412 |
|
}, |
|
{ |
|
"epoch": 1.84375, |
|
"grad_norm": 0.14595958590507507, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0096, |
|
"step": 413 |
|
}, |
|
{ |
|
"epoch": 1.8482142857142856, |
|
"grad_norm": 0.15500158071517944, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0089, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 1.8526785714285714, |
|
"grad_norm": 0.1046842709183693, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0075, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 1.8571428571428572, |
|
"grad_norm": 0.17138102650642395, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0089, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 1.8616071428571428, |
|
"grad_norm": 0.05790293961763382, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0064, |
|
"step": 417 |
|
}, |
|
{ |
|
"epoch": 1.8660714285714286, |
|
"grad_norm": 0.09619224816560745, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0073, |
|
"step": 418 |
|
}, |
|
{ |
|
"epoch": 1.8705357142857144, |
|
"grad_norm": 0.0759076327085495, |
|
"learning_rate": 2e-05, |
|
"loss": 0.007, |
|
"step": 419 |
|
}, |
|
{ |
|
"epoch": 1.875, |
|
"grad_norm": 0.09200000762939453, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0072, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 1.8794642857142856, |
|
"grad_norm": 0.10770270973443985, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0057, |
|
"step": 421 |
|
}, |
|
{ |
|
"epoch": 1.8839285714285714, |
|
"grad_norm": 0.12750625610351562, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0077, |
|
"step": 422 |
|
}, |
|
{ |
|
"epoch": 1.8883928571428572, |
|
"grad_norm": 0.07994697988033295, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0054, |
|
"step": 423 |
|
}, |
|
{ |
|
"epoch": 1.8928571428571428, |
|
"grad_norm": 0.16716712713241577, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0074, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 1.8973214285714286, |
|
"grad_norm": 0.10230574756860733, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0098, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 1.9017857142857144, |
|
"grad_norm": 0.17903654277324677, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0095, |
|
"step": 426 |
|
}, |
|
{ |
|
"epoch": 1.90625, |
|
"grad_norm": 0.12698078155517578, |
|
"learning_rate": 2e-05, |
|
"loss": 0.008, |
|
"step": 427 |
|
}, |
|
{ |
|
"epoch": 1.9107142857142856, |
|
"grad_norm": 0.12924636900424957, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0092, |
|
"step": 428 |
|
}, |
|
{ |
|
"epoch": 1.9151785714285714, |
|
"grad_norm": 0.07155907899141312, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0067, |
|
"step": 429 |
|
}, |
|
{ |
|
"epoch": 1.9196428571428572, |
|
"grad_norm": 0.10363934934139252, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0073, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 1.9241071428571428, |
|
"grad_norm": 0.11996760219335556, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0059, |
|
"step": 431 |
|
}, |
|
{ |
|
"epoch": 1.9285714285714286, |
|
"grad_norm": 0.10494901239871979, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0063, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 1.9330357142857144, |
|
"grad_norm": 0.18249598145484924, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0076, |
|
"step": 433 |
|
}, |
|
{ |
|
"epoch": 1.9375, |
|
"grad_norm": 0.13870225846767426, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0094, |
|
"step": 434 |
|
}, |
|
{ |
|
"epoch": 1.9419642857142856, |
|
"grad_norm": 0.09797409921884537, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0052, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 1.9464285714285714, |
|
"grad_norm": 0.34259602427482605, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0088, |
|
"step": 436 |
|
}, |
|
{ |
|
"epoch": 1.9508928571428572, |
|
"grad_norm": 0.16433046758174896, |
|
"learning_rate": 2e-05, |
|
"loss": 0.007, |
|
"step": 437 |
|
}, |
|
{ |
|
"epoch": 1.9553571428571428, |
|
"grad_norm": 0.1122380718588829, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0066, |
|
"step": 438 |
|
}, |
|
{ |
|
"epoch": 1.9598214285714286, |
|
"grad_norm": 0.09637662768363953, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0057, |
|
"step": 439 |
|
}, |
|
{ |
|
"epoch": 1.9642857142857144, |
|
"grad_norm": 0.1375151425600052, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0082, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 1.96875, |
|
"grad_norm": 0.1535470336675644, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0084, |
|
"step": 441 |
|
}, |
|
{ |
|
"epoch": 1.9732142857142856, |
|
"grad_norm": 0.08874325454235077, |
|
"learning_rate": 2e-05, |
|
"loss": 0.006, |
|
"step": 442 |
|
}, |
|
{ |
|
"epoch": 1.9776785714285714, |
|
"grad_norm": 0.1507561206817627, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0086, |
|
"step": 443 |
|
}, |
|
{ |
|
"epoch": 1.9821428571428572, |
|
"grad_norm": 0.10507656633853912, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0093, |
|
"step": 444 |
|
}, |
|
{ |
|
"epoch": 1.9866071428571428, |
|
"grad_norm": 0.19875890016555786, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0094, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 1.9910714285714286, |
|
"grad_norm": 0.09340199083089828, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0084, |
|
"step": 446 |
|
}, |
|
{ |
|
"epoch": 1.9955357142857144, |
|
"grad_norm": 0.06378448009490967, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0074, |
|
"step": 447 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 0.07624057680368423, |
|
"learning_rate": 2e-05, |
|
"loss": 0.006, |
|
"step": 448 |
|
}, |
|
{ |
|
"epoch": 2.0044642857142856, |
|
"grad_norm": 0.11424339562654495, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0067, |
|
"step": 449 |
|
}, |
|
{ |
|
"epoch": 2.0089285714285716, |
|
"grad_norm": 0.1132059097290039, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0072, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 2.013392857142857, |
|
"grad_norm": 0.06459866464138031, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0066, |
|
"step": 451 |
|
}, |
|
{ |
|
"epoch": 2.017857142857143, |
|
"grad_norm": 0.05782695114612579, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0052, |
|
"step": 452 |
|
}, |
|
{ |
|
"epoch": 2.0223214285714284, |
|
"grad_norm": 0.09242033213376999, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0074, |
|
"step": 453 |
|
}, |
|
{ |
|
"epoch": 2.0267857142857144, |
|
"grad_norm": 0.07749675959348679, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0079, |
|
"step": 454 |
|
}, |
|
{ |
|
"epoch": 2.03125, |
|
"grad_norm": 0.18795126676559448, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0056, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 2.0357142857142856, |
|
"grad_norm": 0.08309225738048553, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0063, |
|
"step": 456 |
|
}, |
|
{ |
|
"epoch": 2.0401785714285716, |
|
"grad_norm": 0.07069497555494308, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0067, |
|
"step": 457 |
|
}, |
|
{ |
|
"epoch": 2.044642857142857, |
|
"grad_norm": 0.17089472711086273, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0112, |
|
"step": 458 |
|
}, |
|
{ |
|
"epoch": 2.049107142857143, |
|
"grad_norm": 0.15188992023468018, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0067, |
|
"step": 459 |
|
}, |
|
{ |
|
"epoch": 2.0535714285714284, |
|
"grad_norm": 0.16761493682861328, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0067, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 2.0580357142857144, |
|
"grad_norm": 0.06758573651313782, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0071, |
|
"step": 461 |
|
}, |
|
{ |
|
"epoch": 2.0625, |
|
"grad_norm": 0.05804857611656189, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0059, |
|
"step": 462 |
|
}, |
|
{ |
|
"epoch": 2.0669642857142856, |
|
"grad_norm": 0.0982503667473793, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0059, |
|
"step": 463 |
|
}, |
|
{ |
|
"epoch": 2.0714285714285716, |
|
"grad_norm": 0.16003400087356567, |
|
"learning_rate": 2e-05, |
|
"loss": 0.008, |
|
"step": 464 |
|
}, |
|
{ |
|
"epoch": 2.075892857142857, |
|
"grad_norm": 0.0709138959646225, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0065, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 2.080357142857143, |
|
"grad_norm": 0.1403375118970871, |
|
"learning_rate": 2e-05, |
|
"loss": 0.009, |
|
"step": 466 |
|
}, |
|
{ |
|
"epoch": 2.0848214285714284, |
|
"grad_norm": 0.103029765188694, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0068, |
|
"step": 467 |
|
}, |
|
{ |
|
"epoch": 2.0892857142857144, |
|
"grad_norm": 0.06266972422599792, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0064, |
|
"step": 468 |
|
}, |
|
{ |
|
"epoch": 2.09375, |
|
"grad_norm": 0.09076008945703506, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0069, |
|
"step": 469 |
|
}, |
|
{ |
|
"epoch": 2.0982142857142856, |
|
"grad_norm": 0.04648640751838684, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0061, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 2.1026785714285716, |
|
"grad_norm": 0.1504824310541153, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0117, |
|
"step": 471 |
|
}, |
|
{ |
|
"epoch": 2.107142857142857, |
|
"grad_norm": 0.0596964955329895, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0053, |
|
"step": 472 |
|
}, |
|
{ |
|
"epoch": 2.111607142857143, |
|
"grad_norm": 0.11864890903234482, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0072, |
|
"step": 473 |
|
}, |
|
{ |
|
"epoch": 2.1160714285714284, |
|
"grad_norm": 0.12213343381881714, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0082, |
|
"step": 474 |
|
}, |
|
{ |
|
"epoch": 2.1205357142857144, |
|
"grad_norm": 0.11169180274009705, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0082, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 2.125, |
|
"grad_norm": 0.1022031381726265, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0073, |
|
"step": 476 |
|
}, |
|
{ |
|
"epoch": 2.1294642857142856, |
|
"grad_norm": 0.07049895823001862, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0056, |
|
"step": 477 |
|
}, |
|
{ |
|
"epoch": 2.1339285714285716, |
|
"grad_norm": 0.11264973878860474, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0079, |
|
"step": 478 |
|
}, |
|
{ |
|
"epoch": 2.138392857142857, |
|
"grad_norm": 0.07296457141637802, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0063, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 2.142857142857143, |
|
"grad_norm": 0.11134887486696243, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0079, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 2.1473214285714284, |
|
"grad_norm": 0.06486133486032486, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0067, |
|
"step": 481 |
|
}, |
|
{ |
|
"epoch": 2.1517857142857144, |
|
"grad_norm": 0.0773395225405693, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0063, |
|
"step": 482 |
|
}, |
|
{ |
|
"epoch": 2.15625, |
|
"grad_norm": 0.13982515037059784, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0091, |
|
"step": 483 |
|
}, |
|
{ |
|
"epoch": 2.1607142857142856, |
|
"grad_norm": 0.10648379474878311, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0078, |
|
"step": 484 |
|
}, |
|
{ |
|
"epoch": 2.1651785714285716, |
|
"grad_norm": 0.10965216904878616, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0079, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 2.169642857142857, |
|
"grad_norm": 0.1464068591594696, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0066, |
|
"step": 486 |
|
}, |
|
{ |
|
"epoch": 2.174107142857143, |
|
"grad_norm": 0.1061873659491539, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0076, |
|
"step": 487 |
|
}, |
|
{ |
|
"epoch": 2.1785714285714284, |
|
"grad_norm": 0.061869096010923386, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0054, |
|
"step": 488 |
|
}, |
|
{ |
|
"epoch": 2.1830357142857144, |
|
"grad_norm": 0.09782952070236206, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0092, |
|
"step": 489 |
|
}, |
|
{ |
|
"epoch": 2.1875, |
|
"grad_norm": 0.08471392840147018, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0079, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 2.1919642857142856, |
|
"grad_norm": 0.10430457442998886, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0089, |
|
"step": 491 |
|
}, |
|
{ |
|
"epoch": 2.1964285714285716, |
|
"grad_norm": 0.07273336499929428, |
|
"learning_rate": 2e-05, |
|
"loss": 0.008, |
|
"step": 492 |
|
}, |
|
{ |
|
"epoch": 2.200892857142857, |
|
"grad_norm": 0.11176946014165878, |
|
"learning_rate": 2e-05, |
|
"loss": 0.008, |
|
"step": 493 |
|
}, |
|
{ |
|
"epoch": 2.205357142857143, |
|
"grad_norm": 0.06163870543241501, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0064, |
|
"step": 494 |
|
}, |
|
{ |
|
"epoch": 2.2098214285714284, |
|
"grad_norm": 0.10791970789432526, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0072, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 2.2142857142857144, |
|
"grad_norm": 0.20584189891815186, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0058, |
|
"step": 496 |
|
}, |
|
{ |
|
"epoch": 2.21875, |
|
"grad_norm": 0.06841476261615753, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0057, |
|
"step": 497 |
|
}, |
|
{ |
|
"epoch": 2.2232142857142856, |
|
"grad_norm": 0.08570678532123566, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0075, |
|
"step": 498 |
|
}, |
|
{ |
|
"epoch": 2.2276785714285716, |
|
"grad_norm": 0.056776780635118484, |
|
"learning_rate": 2e-05, |
|
"loss": 0.007, |
|
"step": 499 |
|
}, |
|
{ |
|
"epoch": 2.232142857142857, |
|
"grad_norm": 0.09146109968423843, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0078, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 2.236607142857143, |
|
"grad_norm": 0.09654641896486282, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0047, |
|
"step": 501 |
|
}, |
|
{ |
|
"epoch": 2.2410714285714284, |
|
"grad_norm": 0.13811850547790527, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0098, |
|
"step": 502 |
|
}, |
|
{ |
|
"epoch": 2.2455357142857144, |
|
"grad_norm": 0.10810638964176178, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0068, |
|
"step": 503 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"grad_norm": 0.21392522752285004, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0088, |
|
"step": 504 |
|
}, |
|
{ |
|
"epoch": 2.2544642857142856, |
|
"grad_norm": 0.11717508733272552, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0088, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 2.2589285714285716, |
|
"grad_norm": 0.05745083466172218, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0059, |
|
"step": 506 |
|
}, |
|
{ |
|
"epoch": 2.263392857142857, |
|
"grad_norm": 0.13906385004520416, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0077, |
|
"step": 507 |
|
}, |
|
{ |
|
"epoch": 2.267857142857143, |
|
"grad_norm": 0.20656032860279083, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0079, |
|
"step": 508 |
|
}, |
|
{ |
|
"epoch": 2.2723214285714284, |
|
"grad_norm": 0.11236809194087982, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0078, |
|
"step": 509 |
|
}, |
|
{ |
|
"epoch": 2.2767857142857144, |
|
"grad_norm": 0.08433755487203598, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0049, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 2.28125, |
|
"grad_norm": 0.1257009208202362, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0098, |
|
"step": 511 |
|
}, |
|
{ |
|
"epoch": 2.2857142857142856, |
|
"grad_norm": 0.08553122729063034, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0071, |
|
"step": 512 |
|
}, |
|
{ |
|
"epoch": 2.2901785714285716, |
|
"grad_norm": 0.13585498929023743, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0079, |
|
"step": 513 |
|
}, |
|
{ |
|
"epoch": 2.294642857142857, |
|
"grad_norm": 0.08864466845989227, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0077, |
|
"step": 514 |
|
}, |
|
{ |
|
"epoch": 2.299107142857143, |
|
"grad_norm": 0.11119361966848373, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0071, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 2.3035714285714284, |
|
"grad_norm": 0.07501739263534546, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0063, |
|
"step": 516 |
|
}, |
|
{ |
|
"epoch": 2.3080357142857144, |
|
"grad_norm": 0.1268368661403656, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0066, |
|
"step": 517 |
|
}, |
|
{ |
|
"epoch": 2.3125, |
|
"grad_norm": 0.07790718972682953, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0076, |
|
"step": 518 |
|
}, |
|
{ |
|
"epoch": 2.3169642857142856, |
|
"grad_norm": 0.11105362325906754, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0068, |
|
"step": 519 |
|
}, |
|
{ |
|
"epoch": 2.3214285714285716, |
|
"grad_norm": 0.07915301620960236, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0069, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 2.325892857142857, |
|
"grad_norm": 0.08192173391580582, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0063, |
|
"step": 521 |
|
}, |
|
{ |
|
"epoch": 2.330357142857143, |
|
"grad_norm": 0.1169026792049408, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0071, |
|
"step": 522 |
|
}, |
|
{ |
|
"epoch": 2.3348214285714284, |
|
"grad_norm": 0.21077628433704376, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0081, |
|
"step": 523 |
|
}, |
|
{ |
|
"epoch": 2.3392857142857144, |
|
"grad_norm": 0.07853112369775772, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0079, |
|
"step": 524 |
|
}, |
|
{ |
|
"epoch": 2.34375, |
|
"grad_norm": 0.13631939888000488, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0093, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 2.3482142857142856, |
|
"grad_norm": 0.08369535207748413, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0068, |
|
"step": 526 |
|
}, |
|
{ |
|
"epoch": 2.3526785714285716, |
|
"grad_norm": 0.055638108402490616, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0077, |
|
"step": 527 |
|
}, |
|
{ |
|
"epoch": 2.357142857142857, |
|
"grad_norm": 0.06996078044176102, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0062, |
|
"step": 528 |
|
}, |
|
{ |
|
"epoch": 2.361607142857143, |
|
"grad_norm": 0.08065176755189896, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0061, |
|
"step": 529 |
|
}, |
|
{ |
|
"epoch": 2.3660714285714284, |
|
"grad_norm": 0.12515677511692047, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0068, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 2.3705357142857144, |
|
"grad_norm": 0.1157045066356659, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0084, |
|
"step": 531 |
|
}, |
|
{ |
|
"epoch": 2.375, |
|
"grad_norm": 0.11465250700712204, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0071, |
|
"step": 532 |
|
}, |
|
{ |
|
"epoch": 2.3794642857142856, |
|
"grad_norm": 0.07074079662561417, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0043, |
|
"step": 533 |
|
}, |
|
{ |
|
"epoch": 2.3839285714285716, |
|
"grad_norm": 0.09491592645645142, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0053, |
|
"step": 534 |
|
}, |
|
{ |
|
"epoch": 2.388392857142857, |
|
"grad_norm": 0.09677135199308395, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0087, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 2.392857142857143, |
|
"grad_norm": 0.12837085127830505, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0089, |
|
"step": 536 |
|
}, |
|
{ |
|
"epoch": 2.3973214285714284, |
|
"grad_norm": 0.1667841225862503, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0082, |
|
"step": 537 |
|
}, |
|
{ |
|
"epoch": 2.4017857142857144, |
|
"grad_norm": 0.1354755461215973, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0096, |
|
"step": 538 |
|
}, |
|
{ |
|
"epoch": 2.40625, |
|
"grad_norm": 0.17251300811767578, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0085, |
|
"step": 539 |
|
}, |
|
{ |
|
"epoch": 2.4107142857142856, |
|
"grad_norm": 0.12392369657754898, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0079, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 2.4151785714285716, |
|
"grad_norm": 0.10283850133419037, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0074, |
|
"step": 541 |
|
}, |
|
{ |
|
"epoch": 2.419642857142857, |
|
"grad_norm": 0.22526662051677704, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0095, |
|
"step": 542 |
|
}, |
|
{ |
|
"epoch": 2.424107142857143, |
|
"grad_norm": 0.16849340498447418, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0072, |
|
"step": 543 |
|
}, |
|
{ |
|
"epoch": 2.4285714285714284, |
|
"grad_norm": 0.1429499238729477, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0065, |
|
"step": 544 |
|
}, |
|
{ |
|
"epoch": 2.4330357142857144, |
|
"grad_norm": 0.13862621784210205, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0075, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 2.4375, |
|
"grad_norm": 0.1709546595811844, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0072, |
|
"step": 546 |
|
}, |
|
{ |
|
"epoch": 2.4419642857142856, |
|
"grad_norm": 0.08103591948747635, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0076, |
|
"step": 547 |
|
}, |
|
{ |
|
"epoch": 2.4464285714285716, |
|
"grad_norm": 0.09880509972572327, |
|
"learning_rate": 2e-05, |
|
"loss": 0.007, |
|
"step": 548 |
|
}, |
|
{ |
|
"epoch": 2.450892857142857, |
|
"grad_norm": 0.11954550445079803, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0065, |
|
"step": 549 |
|
}, |
|
{ |
|
"epoch": 2.455357142857143, |
|
"grad_norm": 0.20056311786174774, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0111, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 2.4598214285714284, |
|
"grad_norm": 0.08780152350664139, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0087, |
|
"step": 551 |
|
}, |
|
{ |
|
"epoch": 2.4642857142857144, |
|
"grad_norm": 0.526067852973938, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0063, |
|
"step": 552 |
|
}, |
|
{ |
|
"epoch": 2.46875, |
|
"grad_norm": 3.777337074279785, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0112, |
|
"step": 553 |
|
}, |
|
{ |
|
"epoch": 2.4732142857142856, |
|
"grad_norm": 1.3595633506774902, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0283, |
|
"step": 554 |
|
}, |
|
{ |
|
"epoch": 2.4776785714285716, |
|
"grad_norm": 2.552967071533203, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0405, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 2.482142857142857, |
|
"grad_norm": 1.628051996231079, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0271, |
|
"step": 556 |
|
}, |
|
{ |
|
"epoch": 2.486607142857143, |
|
"grad_norm": 1.537842869758606, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0432, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 2.4910714285714284, |
|
"grad_norm": 1.3012171983718872, |
|
"learning_rate": 2e-05, |
|
"loss": 0.025, |
|
"step": 558 |
|
}, |
|
{ |
|
"epoch": 2.4955357142857144, |
|
"grad_norm": 1.285781741142273, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0254, |
|
"step": 559 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"grad_norm": 2.0242481231689453, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0153, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 2.5044642857142856, |
|
"grad_norm": 6.9972357749938965, |
|
"learning_rate": 2e-05, |
|
"loss": 0.1738, |
|
"step": 561 |
|
}, |
|
{ |
|
"epoch": 2.508928571428571, |
|
"grad_norm": 1.1082912683486938, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0221, |
|
"step": 562 |
|
}, |
|
{ |
|
"epoch": 2.513392857142857, |
|
"grad_norm": 3.411374092102051, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0272, |
|
"step": 563 |
|
}, |
|
{ |
|
"epoch": 2.517857142857143, |
|
"grad_norm": 10.052404403686523, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0897, |
|
"step": 564 |
|
}, |
|
{ |
|
"epoch": 2.522321428571429, |
|
"grad_norm": 2.627420663833618, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0292, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 2.5267857142857144, |
|
"grad_norm": 3.180161237716675, |
|
"learning_rate": 2e-05, |
|
"loss": 0.034, |
|
"step": 566 |
|
}, |
|
{ |
|
"epoch": 2.53125, |
|
"grad_norm": 4.870179653167725, |
|
"learning_rate": 2e-05, |
|
"loss": 0.1194, |
|
"step": 567 |
|
}, |
|
{ |
|
"epoch": 2.5357142857142856, |
|
"grad_norm": 4.233872890472412, |
|
"learning_rate": 2e-05, |
|
"loss": 0.1097, |
|
"step": 568 |
|
}, |
|
{ |
|
"epoch": 2.540178571428571, |
|
"grad_norm": 1.2434048652648926, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0258, |
|
"step": 569 |
|
}, |
|
{ |
|
"epoch": 2.544642857142857, |
|
"grad_norm": 2.1450345516204834, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0845, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 2.549107142857143, |
|
"grad_norm": 1.3749841451644897, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0326, |
|
"step": 571 |
|
}, |
|
{ |
|
"epoch": 2.553571428571429, |
|
"grad_norm": 0.5364799499511719, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0141, |
|
"step": 572 |
|
}, |
|
{ |
|
"epoch": 2.5580357142857144, |
|
"grad_norm": 0.25783586502075195, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0102, |
|
"step": 573 |
|
}, |
|
{ |
|
"epoch": 2.5625, |
|
"grad_norm": 0.31643831729888916, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0156, |
|
"step": 574 |
|
}, |
|
{ |
|
"epoch": 2.5669642857142856, |
|
"grad_norm": 0.16987808048725128, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0106, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 2.571428571428571, |
|
"grad_norm": 0.13765154778957367, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0097, |
|
"step": 576 |
|
}, |
|
{ |
|
"epoch": 2.575892857142857, |
|
"grad_norm": 0.1949804574251175, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0122, |
|
"step": 577 |
|
}, |
|
{ |
|
"epoch": 2.580357142857143, |
|
"grad_norm": 0.1344509869813919, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0089, |
|
"step": 578 |
|
}, |
|
{ |
|
"epoch": 2.584821428571429, |
|
"grad_norm": 0.16776832938194275, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0084, |
|
"step": 579 |
|
}, |
|
{ |
|
"epoch": 2.5892857142857144, |
|
"grad_norm": 0.17442850768566132, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0075, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 2.59375, |
|
"grad_norm": 0.30235546827316284, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0195, |
|
"step": 581 |
|
}, |
|
{ |
|
"epoch": 2.5982142857142856, |
|
"grad_norm": 0.14607034623622894, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0078, |
|
"step": 582 |
|
}, |
|
{ |
|
"epoch": 2.602678571428571, |
|
"grad_norm": 0.23565685749053955, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0089, |
|
"step": 583 |
|
}, |
|
{ |
|
"epoch": 2.607142857142857, |
|
"grad_norm": 0.15442737936973572, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0056, |
|
"step": 584 |
|
}, |
|
{ |
|
"epoch": 2.611607142857143, |
|
"grad_norm": 0.08780539035797119, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0066, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 2.616071428571429, |
|
"grad_norm": 0.21069885790348053, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0119, |
|
"step": 586 |
|
}, |
|
{ |
|
"epoch": 2.6205357142857144, |
|
"grad_norm": 0.07902006804943085, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0052, |
|
"step": 587 |
|
}, |
|
{ |
|
"epoch": 2.625, |
|
"grad_norm": 0.18614117801189423, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0075, |
|
"step": 588 |
|
}, |
|
{ |
|
"epoch": 2.6294642857142856, |
|
"grad_norm": 0.21328520774841309, |
|
"learning_rate": 2e-05, |
|
"loss": 0.009, |
|
"step": 589 |
|
}, |
|
{ |
|
"epoch": 2.633928571428571, |
|
"grad_norm": 0.13880658149719238, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0075, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 2.638392857142857, |
|
"grad_norm": 1.2935315370559692, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0163, |
|
"step": 591 |
|
}, |
|
{ |
|
"epoch": 2.642857142857143, |
|
"grad_norm": 0.2837115526199341, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0127, |
|
"step": 592 |
|
}, |
|
{ |
|
"epoch": 2.647321428571429, |
|
"grad_norm": 0.908935010433197, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0073, |
|
"step": 593 |
|
}, |
|
{ |
|
"epoch": 2.6517857142857144, |
|
"grad_norm": 1.0788531303405762, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0163, |
|
"step": 594 |
|
}, |
|
{ |
|
"epoch": 2.65625, |
|
"grad_norm": 3.155200242996216, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0355, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 2.6607142857142856, |
|
"grad_norm": 6.767505168914795, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0595, |
|
"step": 596 |
|
}, |
|
{ |
|
"epoch": 2.665178571428571, |
|
"grad_norm": 1.3904411792755127, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0465, |
|
"step": 597 |
|
}, |
|
{ |
|
"epoch": 2.669642857142857, |
|
"grad_norm": 5.910096645355225, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0229, |
|
"step": 598 |
|
}, |
|
{ |
|
"epoch": 2.674107142857143, |
|
"grad_norm": 1.484846591949463, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0302, |
|
"step": 599 |
|
}, |
|
{ |
|
"epoch": 2.678571428571429, |
|
"grad_norm": 0.35779526829719543, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0131, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 2.6830357142857144, |
|
"grad_norm": 0.7565224766731262, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0135, |
|
"step": 601 |
|
}, |
|
{ |
|
"epoch": 2.6875, |
|
"grad_norm": 0.5273329019546509, |
|
"learning_rate": 2e-05, |
|
"loss": 0.021, |
|
"step": 602 |
|
}, |
|
{ |
|
"epoch": 2.6919642857142856, |
|
"grad_norm": 0.23534627258777618, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0127, |
|
"step": 603 |
|
}, |
|
{ |
|
"epoch": 2.696428571428571, |
|
"grad_norm": 4.416169166564941, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0172, |
|
"step": 604 |
|
}, |
|
{ |
|
"epoch": 2.700892857142857, |
|
"grad_norm": 0.26013216376304626, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0107, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 2.705357142857143, |
|
"grad_norm": 4.598071575164795, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0608, |
|
"step": 606 |
|
}, |
|
{ |
|
"epoch": 2.709821428571429, |
|
"grad_norm": 2.953446865081787, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0273, |
|
"step": 607 |
|
}, |
|
{ |
|
"epoch": 2.7142857142857144, |
|
"grad_norm": 1.1341112852096558, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0165, |
|
"step": 608 |
|
}, |
|
{ |
|
"epoch": 2.71875, |
|
"grad_norm": 0.1598537564277649, |
|
"learning_rate": 2e-05, |
|
"loss": 0.011, |
|
"step": 609 |
|
}, |
|
{ |
|
"epoch": 2.7232142857142856, |
|
"grad_norm": 0.24823372066020966, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0096, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 2.727678571428571, |
|
"grad_norm": 0.775355875492096, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0085, |
|
"step": 611 |
|
}, |
|
{ |
|
"epoch": 2.732142857142857, |
|
"grad_norm": 0.08136750012636185, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0088, |
|
"step": 612 |
|
}, |
|
{ |
|
"epoch": 2.736607142857143, |
|
"grad_norm": 9.017096519470215, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0224, |
|
"step": 613 |
|
}, |
|
{ |
|
"epoch": 2.741071428571429, |
|
"grad_norm": 0.416096955537796, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0108, |
|
"step": 614 |
|
}, |
|
{ |
|
"epoch": 2.7455357142857144, |
|
"grad_norm": 10.557942390441895, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0159, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"grad_norm": 1.6279311180114746, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0752, |
|
"step": 616 |
|
}, |
|
{ |
|
"epoch": 2.7544642857142856, |
|
"grad_norm": 1.5116238594055176, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0344, |
|
"step": 617 |
|
}, |
|
{ |
|
"epoch": 2.758928571428571, |
|
"grad_norm": 0.34003084897994995, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0101, |
|
"step": 618 |
|
}, |
|
{ |
|
"epoch": 2.763392857142857, |
|
"grad_norm": 0.8232696056365967, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0162, |
|
"step": 619 |
|
}, |
|
{ |
|
"epoch": 2.767857142857143, |
|
"grad_norm": 0.1500774621963501, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0104, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 2.772321428571429, |
|
"grad_norm": 7.253946781158447, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0323, |
|
"step": 621 |
|
}, |
|
{ |
|
"epoch": 2.7767857142857144, |
|
"grad_norm": 0.3118407428264618, |
|
"learning_rate": 2e-05, |
|
"loss": 0.014, |
|
"step": 622 |
|
}, |
|
{ |
|
"epoch": 2.78125, |
|
"grad_norm": 7.626855373382568, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0127, |
|
"step": 623 |
|
}, |
|
{ |
|
"epoch": 2.7857142857142856, |
|
"grad_norm": 1.0912909507751465, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0254, |
|
"step": 624 |
|
}, |
|
{ |
|
"epoch": 2.790178571428571, |
|
"grad_norm": 0.22835303843021393, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0145, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 2.794642857142857, |
|
"grad_norm": 6.9012908935546875, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0444, |
|
"step": 626 |
|
}, |
|
{ |
|
"epoch": 2.799107142857143, |
|
"grad_norm": 0.27794891595840454, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0083, |
|
"step": 627 |
|
}, |
|
{ |
|
"epoch": 2.803571428571429, |
|
"grad_norm": 0.15842191874980927, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0075, |
|
"step": 628 |
|
}, |
|
{ |
|
"epoch": 2.8080357142857144, |
|
"grad_norm": 67.39225769042969, |
|
"learning_rate": 2e-05, |
|
"loss": 0.025, |
|
"step": 629 |
|
}, |
|
{ |
|
"epoch": 2.8125, |
|
"grad_norm": 0.18912476301193237, |
|
"learning_rate": 2e-05, |
|
"loss": 0.011, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 2.8169642857142856, |
|
"grad_norm": 0.36825433373451233, |
|
"learning_rate": 2e-05, |
|
"loss": 0.009, |
|
"step": 631 |
|
}, |
|
{ |
|
"epoch": 2.821428571428571, |
|
"grad_norm": 0.22129683196544647, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0093, |
|
"step": 632 |
|
}, |
|
{ |
|
"epoch": 2.825892857142857, |
|
"grad_norm": 0.1257479190826416, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0064, |
|
"step": 633 |
|
}, |
|
{ |
|
"epoch": 2.830357142857143, |
|
"grad_norm": 0.21349452435970306, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0111, |
|
"step": 634 |
|
}, |
|
{ |
|
"epoch": 2.834821428571429, |
|
"grad_norm": 0.09613204002380371, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0078, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 2.8392857142857144, |
|
"grad_norm": 0.1867591142654419, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0072, |
|
"step": 636 |
|
}, |
|
{ |
|
"epoch": 2.84375, |
|
"grad_norm": 0.3712433874607086, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0151, |
|
"step": 637 |
|
}, |
|
{ |
|
"epoch": 2.8482142857142856, |
|
"grad_norm": 0.15406033396720886, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0077, |
|
"step": 638 |
|
}, |
|
{ |
|
"epoch": 2.852678571428571, |
|
"grad_norm": 0.3927120566368103, |
|
"learning_rate": 2e-05, |
|
"loss": 0.01, |
|
"step": 639 |
|
}, |
|
{ |
|
"epoch": 2.857142857142857, |
|
"grad_norm": 0.1449354887008667, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0072, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 2.861607142857143, |
|
"grad_norm": 8.02448558807373, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0109, |
|
"step": 641 |
|
}, |
|
{ |
|
"epoch": 2.866071428571429, |
|
"grad_norm": 0.09204132109880447, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0059, |
|
"step": 642 |
|
}, |
|
{ |
|
"epoch": 2.8705357142857144, |
|
"grad_norm": 0.1463308483362198, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0084, |
|
"step": 643 |
|
}, |
|
{ |
|
"epoch": 2.875, |
|
"grad_norm": 0.11871679872274399, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0072, |
|
"step": 644 |
|
}, |
|
{ |
|
"epoch": 2.8794642857142856, |
|
"grad_norm": 0.21967895328998566, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0086, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 2.883928571428571, |
|
"grad_norm": 1.2042790651321411, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0138, |
|
"step": 646 |
|
}, |
|
{ |
|
"epoch": 2.888392857142857, |
|
"grad_norm": 0.12121184170246124, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0113, |
|
"step": 647 |
|
}, |
|
{ |
|
"epoch": 2.892857142857143, |
|
"grad_norm": 0.11416534334421158, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0099, |
|
"step": 648 |
|
}, |
|
{ |
|
"epoch": 2.897321428571429, |
|
"grad_norm": 0.11109648644924164, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0068, |
|
"step": 649 |
|
}, |
|
{ |
|
"epoch": 2.9017857142857144, |
|
"grad_norm": 0.09570077806711197, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0067, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 2.90625, |
|
"grad_norm": 0.155325785279274, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0079, |
|
"step": 651 |
|
}, |
|
{ |
|
"epoch": 2.9107142857142856, |
|
"grad_norm": 0.11843915283679962, |
|
"learning_rate": 2e-05, |
|
"loss": 0.01, |
|
"step": 652 |
|
}, |
|
{ |
|
"epoch": 2.915178571428571, |
|
"grad_norm": 0.1322368085384369, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0081, |
|
"step": 653 |
|
}, |
|
{ |
|
"epoch": 2.919642857142857, |
|
"grad_norm": 0.11188462376594543, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0078, |
|
"step": 654 |
|
}, |
|
{ |
|
"epoch": 2.924107142857143, |
|
"grad_norm": 0.1739046424627304, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0093, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 2.928571428571429, |
|
"grad_norm": 0.09202327579259872, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0084, |
|
"step": 656 |
|
}, |
|
{ |
|
"epoch": 2.9330357142857144, |
|
"grad_norm": 0.13602964580059052, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0086, |
|
"step": 657 |
|
}, |
|
{ |
|
"epoch": 2.9375, |
|
"grad_norm": 0.1424887776374817, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0074, |
|
"step": 658 |
|
}, |
|
{ |
|
"epoch": 2.9419642857142856, |
|
"grad_norm": 0.11014612764120102, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0083, |
|
"step": 659 |
|
}, |
|
{ |
|
"epoch": 2.946428571428571, |
|
"grad_norm": 0.08697912842035294, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0083, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 2.950892857142857, |
|
"grad_norm": 0.11514980345964432, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0078, |
|
"step": 661 |
|
}, |
|
{ |
|
"epoch": 2.955357142857143, |
|
"grad_norm": 0.11393480747938156, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0074, |
|
"step": 662 |
|
}, |
|
{ |
|
"epoch": 2.959821428571429, |
|
"grad_norm": 0.2073274850845337, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0116, |
|
"step": 663 |
|
}, |
|
{ |
|
"epoch": 2.9642857142857144, |
|
"grad_norm": 0.11760195344686508, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0083, |
|
"step": 664 |
|
}, |
|
{ |
|
"epoch": 2.96875, |
|
"grad_norm": 0.12061617523431778, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0074, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 2.9732142857142856, |
|
"grad_norm": 0.11407007277011871, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0073, |
|
"step": 666 |
|
}, |
|
{ |
|
"epoch": 2.977678571428571, |
|
"grad_norm": 0.06950999796390533, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0057, |
|
"step": 667 |
|
}, |
|
{ |
|
"epoch": 2.982142857142857, |
|
"grad_norm": 0.2019437998533249, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0112, |
|
"step": 668 |
|
}, |
|
{ |
|
"epoch": 2.986607142857143, |
|
"grad_norm": 0.10581085830926895, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0056, |
|
"step": 669 |
|
}, |
|
{ |
|
"epoch": 2.991071428571429, |
|
"grad_norm": 0.20939861238002777, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0091, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 2.9955357142857144, |
|
"grad_norm": 0.09610752016305923, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0093, |
|
"step": 671 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"grad_norm": 0.1326116919517517, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0091, |
|
"step": 672 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"step": 672, |
|
"total_flos": 54480040427520.0, |
|
"train_loss": 0.0410676886753035, |
|
"train_runtime": 14462.764, |
|
"train_samples_per_second": 2.968, |
|
"train_steps_per_second": 0.046 |
|
} |
|
], |
|
"logging_steps": 1.0, |
|
"max_steps": 672, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 10000, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 54480040427520.0, |
|
"train_batch_size": 4, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|