{ "best_metric": null, "best_model_checkpoint": null, "epoch": 3.0, "eval_steps": 500, "global_step": 660, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.004545454545454545, "grad_norm": 3.4694509506225586, "learning_rate": 0.0004992424242424243, "loss": 1.6877, "step": 1 }, { "epoch": 0.00909090909090909, "grad_norm": 2.844703435897827, "learning_rate": 0.0004984848484848485, "loss": 1.7528, "step": 2 }, { "epoch": 0.013636363636363636, "grad_norm": 4.147863388061523, "learning_rate": 0.0004977272727272727, "loss": 2.6111, "step": 3 }, { "epoch": 0.01818181818181818, "grad_norm": 2.755852699279785, "learning_rate": 0.000496969696969697, "loss": 1.9464, "step": 4 }, { "epoch": 0.022727272727272728, "grad_norm": 4.124767780303955, "learning_rate": 0.0004962121212121212, "loss": 2.2121, "step": 5 }, { "epoch": 0.02727272727272727, "grad_norm": 3.923773765563965, "learning_rate": 0.0004954545454545455, "loss": 2.6991, "step": 6 }, { "epoch": 0.031818181818181815, "grad_norm": 4.66182279586792, "learning_rate": 0.0004946969696969697, "loss": 2.176, "step": 7 }, { "epoch": 0.03636363636363636, "grad_norm": 3.7830166816711426, "learning_rate": 0.000493939393939394, "loss": 2.7265, "step": 8 }, { "epoch": 0.04090909090909091, "grad_norm": 3.966615676879883, "learning_rate": 0.0004931818181818182, "loss": 3.1926, "step": 9 }, { "epoch": 0.045454545454545456, "grad_norm": 3.281916618347168, "learning_rate": 0.0004924242424242425, "loss": 2.1706, "step": 10 }, { "epoch": 0.05, "grad_norm": 4.288072109222412, "learning_rate": 0.0004916666666666666, "loss": 3.1348, "step": 11 }, { "epoch": 0.05454545454545454, "grad_norm": 3.2815868854522705, "learning_rate": 0.0004909090909090909, "loss": 2.2711, "step": 12 }, { "epoch": 0.05909090909090909, "grad_norm": 2.75382924079895, "learning_rate": 0.0004901515151515152, "loss": 1.762, "step": 13 }, { "epoch": 0.06363636363636363, "grad_norm": 2.6767005920410156, "learning_rate": 0.0004893939393939393, "loss": 1.591, "step": 14 }, { "epoch": 0.06818181818181818, "grad_norm": 3.613719940185547, "learning_rate": 0.0004886363636363636, "loss": 2.1644, "step": 15 }, { "epoch": 0.07272727272727272, "grad_norm": 3.3433680534362793, "learning_rate": 0.00048787878787878784, "loss": 2.9727, "step": 16 }, { "epoch": 0.07727272727272727, "grad_norm": 3.7183644771575928, "learning_rate": 0.0004871212121212121, "loss": 1.8928, "step": 17 }, { "epoch": 0.08181818181818182, "grad_norm": 4.1484575271606445, "learning_rate": 0.0004863636363636364, "loss": 2.6002, "step": 18 }, { "epoch": 0.08636363636363636, "grad_norm": 3.281487464904785, "learning_rate": 0.0004856060606060606, "loss": 1.9074, "step": 19 }, { "epoch": 0.09090909090909091, "grad_norm": 3.0067665576934814, "learning_rate": 0.0004848484848484849, "loss": 2.0375, "step": 20 }, { "epoch": 0.09545454545454546, "grad_norm": 2.8053739070892334, "learning_rate": 0.00048409090909090906, "loss": 1.7248, "step": 21 }, { "epoch": 0.1, "grad_norm": 2.619422435760498, "learning_rate": 0.00048333333333333334, "loss": 2.6251, "step": 22 }, { "epoch": 0.10454545454545454, "grad_norm": 3.1278717517852783, "learning_rate": 0.0004825757575757576, "loss": 2.7065, "step": 23 }, { "epoch": 0.10909090909090909, "grad_norm": 2.723963975906372, "learning_rate": 0.00048181818181818184, "loss": 1.8838, "step": 24 }, { "epoch": 0.11363636363636363, "grad_norm": 2.6069819927215576, "learning_rate": 0.0004810606060606061, "loss": 1.9516, "step": 25 }, { "epoch": 0.11818181818181818, "grad_norm": 2.426720380783081, "learning_rate": 0.0004803030303030303, "loss": 2.1439, "step": 26 }, { "epoch": 0.12272727272727273, "grad_norm": 2.3633666038513184, "learning_rate": 0.00047954545454545456, "loss": 2.0113, "step": 27 }, { "epoch": 0.12727272727272726, "grad_norm": 2.988654136657715, "learning_rate": 0.0004787878787878788, "loss": 2.1691, "step": 28 }, { "epoch": 0.1318181818181818, "grad_norm": 2.713346481323242, "learning_rate": 0.00047803030303030306, "loss": 2.1206, "step": 29 }, { "epoch": 0.13636363636363635, "grad_norm": 2.9896864891052246, "learning_rate": 0.0004772727272727273, "loss": 2.139, "step": 30 }, { "epoch": 0.1409090909090909, "grad_norm": 2.7176098823547363, "learning_rate": 0.0004765151515151515, "loss": 2.2194, "step": 31 }, { "epoch": 0.14545454545454545, "grad_norm": 2.6730499267578125, "learning_rate": 0.0004757575757575758, "loss": 1.5875, "step": 32 }, { "epoch": 0.15, "grad_norm": 3.921717643737793, "learning_rate": 0.000475, "loss": 1.9037, "step": 33 }, { "epoch": 0.15454545454545454, "grad_norm": 2.800473690032959, "learning_rate": 0.0004742424242424243, "loss": 1.8628, "step": 34 }, { "epoch": 0.1590909090909091, "grad_norm": 2.7188827991485596, "learning_rate": 0.0004734848484848485, "loss": 2.1262, "step": 35 }, { "epoch": 0.16363636363636364, "grad_norm": 2.794339895248413, "learning_rate": 0.0004727272727272727, "loss": 2.3508, "step": 36 }, { "epoch": 0.16818181818181818, "grad_norm": NaN, "learning_rate": 0.0004727272727272727, "loss": 2.4565, "step": 37 }, { "epoch": 0.17272727272727273, "grad_norm": 3.1815218925476074, "learning_rate": 0.000471969696969697, "loss": 2.2242, "step": 38 }, { "epoch": 0.17727272727272728, "grad_norm": 3.4017906188964844, "learning_rate": 0.0004712121212121212, "loss": 2.1975, "step": 39 }, { "epoch": 0.18181818181818182, "grad_norm": 2.7533328533172607, "learning_rate": 0.00047045454545454544, "loss": 2.1, "step": 40 }, { "epoch": 0.18636363636363637, "grad_norm": 2.8896608352661133, "learning_rate": 0.0004696969696969697, "loss": 1.8716, "step": 41 }, { "epoch": 0.19090909090909092, "grad_norm": 2.501896858215332, "learning_rate": 0.00046893939393939394, "loss": 2.3533, "step": 42 }, { "epoch": 0.19545454545454546, "grad_norm": 2.2779133319854736, "learning_rate": 0.0004681818181818182, "loss": 1.8062, "step": 43 }, { "epoch": 0.2, "grad_norm": 2.78344988822937, "learning_rate": 0.00046742424242424244, "loss": 2.4259, "step": 44 }, { "epoch": 0.20454545454545456, "grad_norm": 2.230311393737793, "learning_rate": 0.00046666666666666666, "loss": 1.8689, "step": 45 }, { "epoch": 0.20909090909090908, "grad_norm": 3.4158501625061035, "learning_rate": 0.0004659090909090909, "loss": 2.3214, "step": 46 }, { "epoch": 0.21363636363636362, "grad_norm": 2.355423927307129, "learning_rate": 0.00046515151515151516, "loss": 1.7059, "step": 47 }, { "epoch": 0.21818181818181817, "grad_norm": 2.495224952697754, "learning_rate": 0.00046439393939393944, "loss": 1.3574, "step": 48 }, { "epoch": 0.22272727272727272, "grad_norm": 2.829482078552246, "learning_rate": 0.00046363636363636366, "loss": 2.2364, "step": 49 }, { "epoch": 0.22727272727272727, "grad_norm": 2.390627861022949, "learning_rate": 0.0004628787878787879, "loss": 1.5226, "step": 50 }, { "epoch": 0.2318181818181818, "grad_norm": 2.2006781101226807, "learning_rate": 0.0004621212121212121, "loss": 1.6848, "step": 51 }, { "epoch": 0.23636363636363636, "grad_norm": 2.737412452697754, "learning_rate": 0.0004613636363636364, "loss": 1.9071, "step": 52 }, { "epoch": 0.2409090909090909, "grad_norm": 3.4992029666900635, "learning_rate": 0.00046060606060606066, "loss": 2.0172, "step": 53 }, { "epoch": 0.24545454545454545, "grad_norm": 3.196709394454956, "learning_rate": 0.0004598484848484848, "loss": 2.6357, "step": 54 }, { "epoch": 0.25, "grad_norm": 3.9549436569213867, "learning_rate": 0.0004590909090909091, "loss": 2.9822, "step": 55 }, { "epoch": 0.2545454545454545, "grad_norm": 2.788527488708496, "learning_rate": 0.0004583333333333333, "loss": 2.3458, "step": 56 }, { "epoch": 0.2590909090909091, "grad_norm": 3.2539544105529785, "learning_rate": 0.0004575757575757576, "loss": 1.9336, "step": 57 }, { "epoch": 0.2636363636363636, "grad_norm": 2.859744071960449, "learning_rate": 0.0004568181818181819, "loss": 2.4852, "step": 58 }, { "epoch": 0.2681818181818182, "grad_norm": 2.6832542419433594, "learning_rate": 0.00045606060606060605, "loss": 2.0347, "step": 59 }, { "epoch": 0.2727272727272727, "grad_norm": 4.593046188354492, "learning_rate": 0.0004553030303030303, "loss": 2.2415, "step": 60 }, { "epoch": 0.2772727272727273, "grad_norm": 3.3459599018096924, "learning_rate": 0.00045454545454545455, "loss": 2.2843, "step": 61 }, { "epoch": 0.2818181818181818, "grad_norm": 2.1371498107910156, "learning_rate": 0.0004537878787878788, "loss": 1.9185, "step": 62 }, { "epoch": 0.2863636363636364, "grad_norm": 2.3603177070617676, "learning_rate": 0.000453030303030303, "loss": 1.9426, "step": 63 }, { "epoch": 0.2909090909090909, "grad_norm": 2.5435550212860107, "learning_rate": 0.00045227272727272727, "loss": 2.4551, "step": 64 }, { "epoch": 0.29545454545454547, "grad_norm": 2.5501880645751953, "learning_rate": 0.00045151515151515154, "loss": 1.9113, "step": 65 }, { "epoch": 0.3, "grad_norm": 2.8549928665161133, "learning_rate": 0.00045075757575757577, "loss": 2.2465, "step": 66 }, { "epoch": 0.30454545454545456, "grad_norm": 2.396857976913452, "learning_rate": 0.00045000000000000004, "loss": 2.1836, "step": 67 }, { "epoch": 0.3090909090909091, "grad_norm": 2.7043912410736084, "learning_rate": 0.0004492424242424242, "loss": 2.1715, "step": 68 }, { "epoch": 0.31363636363636366, "grad_norm": 3.1579270362854004, "learning_rate": 0.0004484848484848485, "loss": 2.4971, "step": 69 }, { "epoch": 0.3181818181818182, "grad_norm": 2.3673815727233887, "learning_rate": 0.00044772727272727276, "loss": 1.7927, "step": 70 }, { "epoch": 0.32272727272727275, "grad_norm": 2.776143789291382, "learning_rate": 0.000446969696969697, "loss": 1.9065, "step": 71 }, { "epoch": 0.32727272727272727, "grad_norm": 2.937574625015259, "learning_rate": 0.00044621212121212126, "loss": 1.6579, "step": 72 }, { "epoch": 0.33181818181818185, "grad_norm": 2.937641143798828, "learning_rate": 0.00044545454545454543, "loss": 2.2259, "step": 73 }, { "epoch": 0.33636363636363636, "grad_norm": 2.305844306945801, "learning_rate": 0.0004446969696969697, "loss": 2.0766, "step": 74 }, { "epoch": 0.3409090909090909, "grad_norm": 3.1322784423828125, "learning_rate": 0.000443939393939394, "loss": 2.3208, "step": 75 }, { "epoch": 0.34545454545454546, "grad_norm": 2.7713890075683594, "learning_rate": 0.0004431818181818182, "loss": 2.4553, "step": 76 }, { "epoch": 0.35, "grad_norm": 2.275108575820923, "learning_rate": 0.00044242424242424243, "loss": 1.4249, "step": 77 }, { "epoch": 0.35454545454545455, "grad_norm": 2.502997875213623, "learning_rate": 0.00044166666666666665, "loss": 2.4343, "step": 78 }, { "epoch": 0.35909090909090907, "grad_norm": 2.1204617023468018, "learning_rate": 0.00044090909090909093, "loss": 1.9248, "step": 79 }, { "epoch": 0.36363636363636365, "grad_norm": 2.9564898014068604, "learning_rate": 0.00044015151515151515, "loss": 1.6672, "step": 80 }, { "epoch": 0.36818181818181817, "grad_norm": 3.0879478454589844, "learning_rate": 0.0004393939393939394, "loss": 2.0625, "step": 81 }, { "epoch": 0.37272727272727274, "grad_norm": 3.1532368659973145, "learning_rate": 0.00043863636363636365, "loss": 2.3648, "step": 82 }, { "epoch": 0.37727272727272726, "grad_norm": 2.1850852966308594, "learning_rate": 0.00043787878787878787, "loss": 2.4397, "step": 83 }, { "epoch": 0.38181818181818183, "grad_norm": 2.241144895553589, "learning_rate": 0.00043712121212121215, "loss": 2.5193, "step": 84 }, { "epoch": 0.38636363636363635, "grad_norm": 3.6500165462493896, "learning_rate": 0.00043636363636363637, "loss": 2.1096, "step": 85 }, { "epoch": 0.39090909090909093, "grad_norm": 2.0548837184906006, "learning_rate": 0.0004356060606060606, "loss": 2.359, "step": 86 }, { "epoch": 0.39545454545454545, "grad_norm": 2.4512407779693604, "learning_rate": 0.00043484848484848487, "loss": 1.5353, "step": 87 }, { "epoch": 0.4, "grad_norm": 1.9420820474624634, "learning_rate": 0.0004340909090909091, "loss": 1.688, "step": 88 }, { "epoch": 0.40454545454545454, "grad_norm": 1.845475673675537, "learning_rate": 0.00043333333333333337, "loss": 1.6642, "step": 89 }, { "epoch": 0.4090909090909091, "grad_norm": 2.2735133171081543, "learning_rate": 0.0004325757575757576, "loss": 1.89, "step": 90 }, { "epoch": 0.41363636363636364, "grad_norm": 2.203105926513672, "learning_rate": 0.0004318181818181818, "loss": 1.9556, "step": 91 }, { "epoch": 0.41818181818181815, "grad_norm": 1.8748105764389038, "learning_rate": 0.00043106060606060603, "loss": 1.5452, "step": 92 }, { "epoch": 0.42272727272727273, "grad_norm": 2.8958442211151123, "learning_rate": 0.0004303030303030303, "loss": 1.9343, "step": 93 }, { "epoch": 0.42727272727272725, "grad_norm": 2.7512269020080566, "learning_rate": 0.0004295454545454546, "loss": 2.4008, "step": 94 }, { "epoch": 0.4318181818181818, "grad_norm": 2.748307228088379, "learning_rate": 0.00042878787878787876, "loss": 2.3614, "step": 95 }, { "epoch": 0.43636363636363634, "grad_norm": 3.7091145515441895, "learning_rate": 0.00042803030303030303, "loss": 1.5435, "step": 96 }, { "epoch": 0.4409090909090909, "grad_norm": 2.0227293968200684, "learning_rate": 0.00042727272727272726, "loss": 1.5536, "step": 97 }, { "epoch": 0.44545454545454544, "grad_norm": 1.868477702140808, "learning_rate": 0.00042651515151515153, "loss": 2.0019, "step": 98 }, { "epoch": 0.45, "grad_norm": 2.2410340309143066, "learning_rate": 0.0004257575757575758, "loss": 2.0278, "step": 99 }, { "epoch": 0.45454545454545453, "grad_norm": 2.4206206798553467, "learning_rate": 0.000425, "loss": 2.6757, "step": 100 }, { "epoch": 0.4590909090909091, "grad_norm": 2.6481056213378906, "learning_rate": 0.00042424242424242425, "loss": 2.004, "step": 101 }, { "epoch": 0.4636363636363636, "grad_norm": 2.493495225906372, "learning_rate": 0.0004234848484848485, "loss": 2.2102, "step": 102 }, { "epoch": 0.4681818181818182, "grad_norm": 2.588595390319824, "learning_rate": 0.00042272727272727275, "loss": 2.3133, "step": 103 }, { "epoch": 0.4727272727272727, "grad_norm": 2.185718536376953, "learning_rate": 0.00042196969696969703, "loss": 2.5506, "step": 104 }, { "epoch": 0.4772727272727273, "grad_norm": 2.155470132827759, "learning_rate": 0.0004212121212121212, "loss": 2.2074, "step": 105 }, { "epoch": 0.4818181818181818, "grad_norm": 2.518435001373291, "learning_rate": 0.0004204545454545455, "loss": 1.8589, "step": 106 }, { "epoch": 0.4863636363636364, "grad_norm": 2.5512635707855225, "learning_rate": 0.0004196969696969697, "loss": 1.9953, "step": 107 }, { "epoch": 0.4909090909090909, "grad_norm": 2.238809108734131, "learning_rate": 0.00041893939393939397, "loss": 2.2441, "step": 108 }, { "epoch": 0.4954545454545455, "grad_norm": 1.8442784547805786, "learning_rate": 0.00041818181818181814, "loss": 1.8682, "step": 109 }, { "epoch": 0.5, "grad_norm": 2.4844954013824463, "learning_rate": 0.0004174242424242424, "loss": 1.9522, "step": 110 }, { "epoch": 0.5045454545454545, "grad_norm": 1.9704878330230713, "learning_rate": 0.0004166666666666667, "loss": 1.4167, "step": 111 }, { "epoch": 0.509090909090909, "grad_norm": 2.2447972297668457, "learning_rate": 0.0004159090909090909, "loss": 1.7897, "step": 112 }, { "epoch": 0.5136363636363637, "grad_norm": 2.530410051345825, "learning_rate": 0.0004151515151515152, "loss": 2.5473, "step": 113 }, { "epoch": 0.5181818181818182, "grad_norm": 2.450526714324951, "learning_rate": 0.00041439393939393936, "loss": 2.0436, "step": 114 }, { "epoch": 0.5227272727272727, "grad_norm": 2.4212632179260254, "learning_rate": 0.00041363636363636364, "loss": 2.1118, "step": 115 }, { "epoch": 0.5272727272727272, "grad_norm": 1.9820351600646973, "learning_rate": 0.0004128787878787879, "loss": 1.9614, "step": 116 }, { "epoch": 0.5318181818181819, "grad_norm": 2.467961549758911, "learning_rate": 0.00041212121212121214, "loss": 1.9572, "step": 117 }, { "epoch": 0.5363636363636364, "grad_norm": 2.2693068981170654, "learning_rate": 0.00041136363636363636, "loss": 1.9033, "step": 118 }, { "epoch": 0.5409090909090909, "grad_norm": 2.299119710922241, "learning_rate": 0.0004106060606060606, "loss": 1.9848, "step": 119 }, { "epoch": 0.5454545454545454, "grad_norm": 2.0330560207366943, "learning_rate": 0.00040984848484848486, "loss": 1.3706, "step": 120 }, { "epoch": 0.55, "grad_norm": 2.197603702545166, "learning_rate": 0.00040909090909090913, "loss": 2.1687, "step": 121 }, { "epoch": 0.5545454545454546, "grad_norm": 2.7206549644470215, "learning_rate": 0.00040833333333333336, "loss": 2.3247, "step": 122 }, { "epoch": 0.5590909090909091, "grad_norm": 2.882654905319214, "learning_rate": 0.0004075757575757576, "loss": 1.6946, "step": 123 }, { "epoch": 0.5636363636363636, "grad_norm": 2.3815231323242188, "learning_rate": 0.0004068181818181818, "loss": 1.862, "step": 124 }, { "epoch": 0.5681818181818182, "grad_norm": 2.4142932891845703, "learning_rate": 0.0004060606060606061, "loss": 2.0066, "step": 125 }, { "epoch": 0.5727272727272728, "grad_norm": 2.6641104221343994, "learning_rate": 0.0004053030303030303, "loss": 1.9456, "step": 126 }, { "epoch": 0.5772727272727273, "grad_norm": 2.983633518218994, "learning_rate": 0.0004045454545454546, "loss": 2.0049, "step": 127 }, { "epoch": 0.5818181818181818, "grad_norm": 2.1993696689605713, "learning_rate": 0.0004037878787878788, "loss": 1.6259, "step": 128 }, { "epoch": 0.5863636363636363, "grad_norm": 3.0398480892181396, "learning_rate": 0.000403030303030303, "loss": 1.6535, "step": 129 }, { "epoch": 0.5909090909090909, "grad_norm": 2.298558235168457, "learning_rate": 0.0004022727272727273, "loss": 1.8693, "step": 130 }, { "epoch": 0.5954545454545455, "grad_norm": 2.5126214027404785, "learning_rate": 0.0004015151515151515, "loss": 2.1854, "step": 131 }, { "epoch": 0.6, "grad_norm": 2.0419557094573975, "learning_rate": 0.00040075757575757574, "loss": 1.5857, "step": 132 }, { "epoch": 0.6045454545454545, "grad_norm": 2.4304699897766113, "learning_rate": 0.0004, "loss": 2.347, "step": 133 }, { "epoch": 0.6090909090909091, "grad_norm": 3.098036050796509, "learning_rate": 0.00039924242424242424, "loss": 2.9748, "step": 134 }, { "epoch": 0.6136363636363636, "grad_norm": 2.4403679370880127, "learning_rate": 0.0003984848484848485, "loss": 1.9342, "step": 135 }, { "epoch": 0.6181818181818182, "grad_norm": 2.832394599914551, "learning_rate": 0.00039772727272727274, "loss": 2.132, "step": 136 }, { "epoch": 0.6227272727272727, "grad_norm": 2.601243019104004, "learning_rate": 0.00039696969696969696, "loss": 2.3074, "step": 137 }, { "epoch": 0.6272727272727273, "grad_norm": 2.2306132316589355, "learning_rate": 0.00039621212121212124, "loss": 1.6065, "step": 138 }, { "epoch": 0.6318181818181818, "grad_norm": 2.393157720565796, "learning_rate": 0.00039545454545454546, "loss": 1.7411, "step": 139 }, { "epoch": 0.6363636363636364, "grad_norm": 2.174208164215088, "learning_rate": 0.00039469696969696974, "loss": 1.3876, "step": 140 }, { "epoch": 0.6409090909090909, "grad_norm": 2.3376457691192627, "learning_rate": 0.0003939393939393939, "loss": 2.3752, "step": 141 }, { "epoch": 0.6454545454545455, "grad_norm": 2.141479969024658, "learning_rate": 0.0003931818181818182, "loss": 1.3948, "step": 142 }, { "epoch": 0.65, "grad_norm": 2.4302890300750732, "learning_rate": 0.0003924242424242424, "loss": 2.5493, "step": 143 }, { "epoch": 0.6545454545454545, "grad_norm": 1.9080986976623535, "learning_rate": 0.0003916666666666667, "loss": 1.4652, "step": 144 }, { "epoch": 0.6590909090909091, "grad_norm": 2.438420057296753, "learning_rate": 0.00039090909090909096, "loss": 2.0606, "step": 145 }, { "epoch": 0.6636363636363637, "grad_norm": 2.028294563293457, "learning_rate": 0.0003901515151515151, "loss": 1.8798, "step": 146 }, { "epoch": 0.6681818181818182, "grad_norm": 2.4235315322875977, "learning_rate": 0.0003893939393939394, "loss": 1.8855, "step": 147 }, { "epoch": 0.6727272727272727, "grad_norm": 2.430391788482666, "learning_rate": 0.0003886363636363636, "loss": 1.7753, "step": 148 }, { "epoch": 0.6772727272727272, "grad_norm": 2.053199052810669, "learning_rate": 0.0003878787878787879, "loss": 2.1466, "step": 149 }, { "epoch": 0.6818181818181818, "grad_norm": 2.067093849182129, "learning_rate": 0.0003871212121212122, "loss": 1.7715, "step": 150 }, { "epoch": 0.6863636363636364, "grad_norm": 2.047165632247925, "learning_rate": 0.00038636363636363635, "loss": 1.8703, "step": 151 }, { "epoch": 0.6909090909090909, "grad_norm": 2.567540168762207, "learning_rate": 0.0003856060606060606, "loss": 1.7973, "step": 152 }, { "epoch": 0.6954545454545454, "grad_norm": 2.5282599925994873, "learning_rate": 0.00038484848484848485, "loss": 2.422, "step": 153 }, { "epoch": 0.7, "grad_norm": 2.0428948402404785, "learning_rate": 0.0003840909090909091, "loss": 1.5709, "step": 154 }, { "epoch": 0.7045454545454546, "grad_norm": 2.032672166824341, "learning_rate": 0.00038333333333333334, "loss": 1.8584, "step": 155 }, { "epoch": 0.7090909090909091, "grad_norm": 2.4448535442352295, "learning_rate": 0.00038257575757575757, "loss": 2.069, "step": 156 }, { "epoch": 0.7136363636363636, "grad_norm": 1.6503087282180786, "learning_rate": 0.00038181818181818184, "loss": 1.5194, "step": 157 }, { "epoch": 0.7181818181818181, "grad_norm": 2.5853593349456787, "learning_rate": 0.00038106060606060607, "loss": 2.4603, "step": 158 }, { "epoch": 0.7227272727272728, "grad_norm": 2.353992223739624, "learning_rate": 0.00038030303030303034, "loss": 1.4417, "step": 159 }, { "epoch": 0.7272727272727273, "grad_norm": 2.382633686065674, "learning_rate": 0.0003795454545454545, "loss": 1.9239, "step": 160 }, { "epoch": 0.7318181818181818, "grad_norm": 2.7274303436279297, "learning_rate": 0.0003787878787878788, "loss": 2.1116, "step": 161 }, { "epoch": 0.7363636363636363, "grad_norm": 2.0137476921081543, "learning_rate": 0.00037803030303030306, "loss": 1.5707, "step": 162 }, { "epoch": 0.740909090909091, "grad_norm": 1.977155089378357, "learning_rate": 0.0003772727272727273, "loss": 1.4972, "step": 163 }, { "epoch": 0.7454545454545455, "grad_norm": 2.5506880283355713, "learning_rate": 0.0003765151515151515, "loss": 2.4583, "step": 164 }, { "epoch": 0.75, "grad_norm": 2.2664661407470703, "learning_rate": 0.00037575757575757573, "loss": 2.1239, "step": 165 }, { "epoch": 0.7545454545454545, "grad_norm": 2.393831968307495, "learning_rate": 0.000375, "loss": 2.2104, "step": 166 }, { "epoch": 0.759090909090909, "grad_norm": 2.358670711517334, "learning_rate": 0.0003742424242424243, "loss": 1.7545, "step": 167 }, { "epoch": 0.7636363636363637, "grad_norm": 2.0985164642333984, "learning_rate": 0.0003734848484848485, "loss": 1.7337, "step": 168 }, { "epoch": 0.7681818181818182, "grad_norm": 1.711176872253418, "learning_rate": 0.00037272727272727273, "loss": 1.3195, "step": 169 }, { "epoch": 0.7727272727272727, "grad_norm": 2.20684814453125, "learning_rate": 0.00037196969696969695, "loss": 1.7633, "step": 170 }, { "epoch": 0.7772727272727272, "grad_norm": 2.0569570064544678, "learning_rate": 0.00037121212121212123, "loss": 1.8354, "step": 171 }, { "epoch": 0.7818181818181819, "grad_norm": 2.4895520210266113, "learning_rate": 0.0003704545454545455, "loss": 2.6706, "step": 172 }, { "epoch": 0.7863636363636364, "grad_norm": 2.3134992122650146, "learning_rate": 0.00036969696969696967, "loss": 2.077, "step": 173 }, { "epoch": 0.7909090909090909, "grad_norm": 1.876047968864441, "learning_rate": 0.00036893939393939395, "loss": 1.5816, "step": 174 }, { "epoch": 0.7954545454545454, "grad_norm": 2.301314353942871, "learning_rate": 0.00036818181818181817, "loss": 2.9433, "step": 175 }, { "epoch": 0.8, "grad_norm": 2.4783785343170166, "learning_rate": 0.00036742424242424245, "loss": 2.3913, "step": 176 }, { "epoch": 0.8045454545454546, "grad_norm": 2.3966879844665527, "learning_rate": 0.00036666666666666667, "loss": 2.6103, "step": 177 }, { "epoch": 0.8090909090909091, "grad_norm": 2.1050190925598145, "learning_rate": 0.0003659090909090909, "loss": 1.6801, "step": 178 }, { "epoch": 0.8136363636363636, "grad_norm": 2.3336639404296875, "learning_rate": 0.00036515151515151517, "loss": 1.936, "step": 179 }, { "epoch": 0.8181818181818182, "grad_norm": 2.4867429733276367, "learning_rate": 0.0003643939393939394, "loss": 2.0285, "step": 180 }, { "epoch": 0.8227272727272728, "grad_norm": 1.9529379606246948, "learning_rate": 0.00036363636363636367, "loss": 1.5503, "step": 181 }, { "epoch": 0.8272727272727273, "grad_norm": 2.095381736755371, "learning_rate": 0.00036287878787878784, "loss": 2.5626, "step": 182 }, { "epoch": 0.8318181818181818, "grad_norm": 2.156099319458008, "learning_rate": 0.0003621212121212121, "loss": 1.8788, "step": 183 }, { "epoch": 0.8363636363636363, "grad_norm": 1.8726741075515747, "learning_rate": 0.0003613636363636364, "loss": 1.6386, "step": 184 }, { "epoch": 0.8409090909090909, "grad_norm": 2.6056482791900635, "learning_rate": 0.0003606060606060606, "loss": 1.7965, "step": 185 }, { "epoch": 0.8454545454545455, "grad_norm": 2.65775728225708, "learning_rate": 0.0003598484848484849, "loss": 1.775, "step": 186 }, { "epoch": 0.85, "grad_norm": 1.9309563636779785, "learning_rate": 0.00035909090909090906, "loss": 1.7575, "step": 187 }, { "epoch": 0.8545454545454545, "grad_norm": 1.874107003211975, "learning_rate": 0.00035833333333333333, "loss": 1.4686, "step": 188 }, { "epoch": 0.8590909090909091, "grad_norm": 2.125084400177002, "learning_rate": 0.0003575757575757576, "loss": 1.3713, "step": 189 }, { "epoch": 0.8636363636363636, "grad_norm": 2.016660690307617, "learning_rate": 0.00035681818181818183, "loss": 1.6914, "step": 190 }, { "epoch": 0.8681818181818182, "grad_norm": 2.8699893951416016, "learning_rate": 0.0003560606060606061, "loss": 1.6524, "step": 191 }, { "epoch": 0.8727272727272727, "grad_norm": 2.2474772930145264, "learning_rate": 0.0003553030303030303, "loss": 1.7333, "step": 192 }, { "epoch": 0.8772727272727273, "grad_norm": 2.6996030807495117, "learning_rate": 0.00035454545454545455, "loss": 1.5828, "step": 193 }, { "epoch": 0.8818181818181818, "grad_norm": 2.196274757385254, "learning_rate": 0.0003537878787878788, "loss": 1.6058, "step": 194 }, { "epoch": 0.8863636363636364, "grad_norm": 2.4350290298461914, "learning_rate": 0.00035303030303030305, "loss": 2.0724, "step": 195 }, { "epoch": 0.8909090909090909, "grad_norm": 2.047480821609497, "learning_rate": 0.0003522727272727273, "loss": 1.7112, "step": 196 }, { "epoch": 0.8954545454545455, "grad_norm": 2.4547033309936523, "learning_rate": 0.0003515151515151515, "loss": 1.7747, "step": 197 }, { "epoch": 0.9, "grad_norm": 2.9125847816467285, "learning_rate": 0.0003507575757575758, "loss": 2.0878, "step": 198 }, { "epoch": 0.9045454545454545, "grad_norm": 2.168196678161621, "learning_rate": 0.00035, "loss": 1.5592, "step": 199 }, { "epoch": 0.9090909090909091, "grad_norm": 2.364847183227539, "learning_rate": 0.0003492424242424243, "loss": 2.0301, "step": 200 }, { "epoch": 0.9136363636363637, "grad_norm": 2.743267059326172, "learning_rate": 0.0003484848484848485, "loss": 1.8784, "step": 201 }, { "epoch": 0.9181818181818182, "grad_norm": 2.2784361839294434, "learning_rate": 0.0003477272727272727, "loss": 1.5936, "step": 202 }, { "epoch": 0.9227272727272727, "grad_norm": 2.875943422317505, "learning_rate": 0.000346969696969697, "loss": 1.9961, "step": 203 }, { "epoch": 0.9272727272727272, "grad_norm": 2.0056071281433105, "learning_rate": 0.0003462121212121212, "loss": 1.8795, "step": 204 }, { "epoch": 0.9318181818181818, "grad_norm": 2.3547298908233643, "learning_rate": 0.00034545454545454544, "loss": 2.1429, "step": 205 }, { "epoch": 0.9363636363636364, "grad_norm": 2.7082138061523438, "learning_rate": 0.0003446969696969697, "loss": 1.7504, "step": 206 }, { "epoch": 0.9409090909090909, "grad_norm": 2.335139751434326, "learning_rate": 0.00034393939393939394, "loss": 1.9774, "step": 207 }, { "epoch": 0.9454545454545454, "grad_norm": 2.3550143241882324, "learning_rate": 0.0003431818181818182, "loss": 1.8602, "step": 208 }, { "epoch": 0.95, "grad_norm": 2.622682809829712, "learning_rate": 0.00034242424242424244, "loss": 2.2283, "step": 209 }, { "epoch": 0.9545454545454546, "grad_norm": 1.7282129526138306, "learning_rate": 0.00034166666666666666, "loss": 1.6025, "step": 210 }, { "epoch": 0.9590909090909091, "grad_norm": 2.8252415657043457, "learning_rate": 0.0003409090909090909, "loss": 1.7649, "step": 211 }, { "epoch": 0.9636363636363636, "grad_norm": 2.146219491958618, "learning_rate": 0.00034015151515151516, "loss": 1.9742, "step": 212 }, { "epoch": 0.9681818181818181, "grad_norm": 2.124577045440674, "learning_rate": 0.00033939393939393943, "loss": 1.7412, "step": 213 }, { "epoch": 0.9727272727272728, "grad_norm": 1.7649497985839844, "learning_rate": 0.00033863636363636366, "loss": 1.2667, "step": 214 }, { "epoch": 0.9772727272727273, "grad_norm": 2.375659465789795, "learning_rate": 0.0003378787878787879, "loss": 1.7142, "step": 215 }, { "epoch": 0.9818181818181818, "grad_norm": 1.9995368719100952, "learning_rate": 0.0003371212121212121, "loss": 1.4613, "step": 216 }, { "epoch": 0.9863636363636363, "grad_norm": 2.2640981674194336, "learning_rate": 0.0003363636363636364, "loss": 1.9474, "step": 217 }, { "epoch": 0.990909090909091, "grad_norm": 2.1680893898010254, "learning_rate": 0.00033560606060606066, "loss": 2.5352, "step": 218 }, { "epoch": 0.9954545454545455, "grad_norm": 2.3969366550445557, "learning_rate": 0.0003348484848484848, "loss": 1.9012, "step": 219 }, { "epoch": 1.0, "grad_norm": 3.3696913719177246, "learning_rate": 0.0003340909090909091, "loss": 2.2928, "step": 220 }, { "epoch": 1.0, "eval_f1": 0.891, "eval_gen_len": 41.9182, "eval_loss": 1.8093845844268799, "eval_precision": 0.8891, "eval_recall": 0.8931, "eval_rouge1": 0.466, "eval_rouge2": 0.2146, "eval_rougeL": 0.3912, "eval_rougeLsum": 0.4301, "eval_runtime": 25.1921, "eval_samples_per_second": 4.366, "eval_steps_per_second": 0.556, "step": 220 }, { "epoch": 1.0045454545454546, "grad_norm": 1.7403843402862549, "learning_rate": 0.0003333333333333333, "loss": 1.7294, "step": 221 }, { "epoch": 1.009090909090909, "grad_norm": 1.5273452997207642, "learning_rate": 0.0003325757575757576, "loss": 1.3279, "step": 222 }, { "epoch": 1.0136363636363637, "grad_norm": 1.427538275718689, "learning_rate": 0.0003318181818181819, "loss": 0.9647, "step": 223 }, { "epoch": 1.018181818181818, "grad_norm": 1.5605067014694214, "learning_rate": 0.00033106060606060604, "loss": 1.3178, "step": 224 }, { "epoch": 1.0227272727272727, "grad_norm": 1.6737922430038452, "learning_rate": 0.0003303030303030303, "loss": 1.403, "step": 225 }, { "epoch": 1.0272727272727273, "grad_norm": 2.3249313831329346, "learning_rate": 0.00032954545454545454, "loss": 1.4907, "step": 226 }, { "epoch": 1.0318181818181817, "grad_norm": 1.9939628839492798, "learning_rate": 0.0003287878787878788, "loss": 1.795, "step": 227 }, { "epoch": 1.0363636363636364, "grad_norm": 2.482421398162842, "learning_rate": 0.000328030303030303, "loss": 1.7309, "step": 228 }, { "epoch": 1.040909090909091, "grad_norm": 1.6090010404586792, "learning_rate": 0.00032727272727272726, "loss": 1.0976, "step": 229 }, { "epoch": 1.0454545454545454, "grad_norm": 1.5481090545654297, "learning_rate": 0.00032651515151515154, "loss": 1.1785, "step": 230 }, { "epoch": 1.05, "grad_norm": 2.0420186519622803, "learning_rate": 0.00032575757575757576, "loss": 1.1853, "step": 231 }, { "epoch": 1.0545454545454545, "grad_norm": 2.5020453929901123, "learning_rate": 0.00032500000000000004, "loss": 1.8213, "step": 232 }, { "epoch": 1.059090909090909, "grad_norm": 1.5312837362289429, "learning_rate": 0.0003242424242424242, "loss": 1.016, "step": 233 }, { "epoch": 1.0636363636363637, "grad_norm": 2.062110185623169, "learning_rate": 0.0003234848484848485, "loss": 1.2245, "step": 234 }, { "epoch": 1.0681818181818181, "grad_norm": 1.7521977424621582, "learning_rate": 0.00032272727272727276, "loss": 1.4904, "step": 235 }, { "epoch": 1.0727272727272728, "grad_norm": 1.431998372077942, "learning_rate": 0.000321969696969697, "loss": 1.2364, "step": 236 }, { "epoch": 1.0772727272727272, "grad_norm": 1.8957371711730957, "learning_rate": 0.00032121212121212126, "loss": 1.9241, "step": 237 }, { "epoch": 1.0818181818181818, "grad_norm": 1.6720540523529053, "learning_rate": 0.00032045454545454543, "loss": 1.0261, "step": 238 }, { "epoch": 1.0863636363636364, "grad_norm": 1.8503271341323853, "learning_rate": 0.0003196969696969697, "loss": 1.6694, "step": 239 }, { "epoch": 1.0909090909090908, "grad_norm": 2.055772066116333, "learning_rate": 0.000318939393939394, "loss": 1.6855, "step": 240 }, { "epoch": 1.0954545454545455, "grad_norm": 4.034445285797119, "learning_rate": 0.0003181818181818182, "loss": 1.8592, "step": 241 }, { "epoch": 1.1, "grad_norm": 2.6941239833831787, "learning_rate": 0.0003174242424242424, "loss": 1.1528, "step": 242 }, { "epoch": 1.1045454545454545, "grad_norm": 1.8258893489837646, "learning_rate": 0.00031666666666666665, "loss": 1.1762, "step": 243 }, { "epoch": 1.1090909090909091, "grad_norm": 2.721888303756714, "learning_rate": 0.0003159090909090909, "loss": 1.5523, "step": 244 }, { "epoch": 1.1136363636363635, "grad_norm": 2.35798978805542, "learning_rate": 0.00031515151515151515, "loss": 1.7533, "step": 245 }, { "epoch": 1.1181818181818182, "grad_norm": 2.089695453643799, "learning_rate": 0.0003143939393939394, "loss": 1.4344, "step": 246 }, { "epoch": 1.1227272727272728, "grad_norm": 2.0275492668151855, "learning_rate": 0.00031363636363636365, "loss": 1.5359, "step": 247 }, { "epoch": 1.1272727272727272, "grad_norm": 2.290893077850342, "learning_rate": 0.00031287878787878787, "loss": 1.8292, "step": 248 }, { "epoch": 1.1318181818181818, "grad_norm": 2.3136603832244873, "learning_rate": 0.00031212121212121214, "loss": 1.6828, "step": 249 }, { "epoch": 1.1363636363636362, "grad_norm": 2.1181459426879883, "learning_rate": 0.00031136363636363637, "loss": 1.1531, "step": 250 }, { "epoch": 1.1409090909090909, "grad_norm": 1.9240480661392212, "learning_rate": 0.0003106060606060606, "loss": 1.2515, "step": 251 }, { "epoch": 1.1454545454545455, "grad_norm": 2.522502899169922, "learning_rate": 0.00030984848484848487, "loss": 1.4942, "step": 252 }, { "epoch": 1.15, "grad_norm": 1.5959993600845337, "learning_rate": 0.0003090909090909091, "loss": 1.1412, "step": 253 }, { "epoch": 1.1545454545454545, "grad_norm": 1.740268588066101, "learning_rate": 0.00030833333333333337, "loss": 1.7009, "step": 254 }, { "epoch": 1.1590909090909092, "grad_norm": 2.1791181564331055, "learning_rate": 0.0003075757575757576, "loss": 1.4727, "step": 255 }, { "epoch": 1.1636363636363636, "grad_norm": 2.2325475215911865, "learning_rate": 0.0003068181818181818, "loss": 1.4379, "step": 256 }, { "epoch": 1.1681818181818182, "grad_norm": 2.206281900405884, "learning_rate": 0.00030606060606060603, "loss": 1.5069, "step": 257 }, { "epoch": 1.1727272727272728, "grad_norm": 2.6821632385253906, "learning_rate": 0.0003053030303030303, "loss": 1.7888, "step": 258 }, { "epoch": 1.1772727272727272, "grad_norm": 2.567087173461914, "learning_rate": 0.0003045454545454546, "loss": 1.7501, "step": 259 }, { "epoch": 1.1818181818181819, "grad_norm": 2.131887674331665, "learning_rate": 0.00030378787878787875, "loss": 1.3294, "step": 260 }, { "epoch": 1.1863636363636363, "grad_norm": 1.5638073682785034, "learning_rate": 0.00030303030303030303, "loss": 1.3679, "step": 261 }, { "epoch": 1.190909090909091, "grad_norm": 2.208799362182617, "learning_rate": 0.00030227272727272725, "loss": 1.2585, "step": 262 }, { "epoch": 1.1954545454545455, "grad_norm": 2.379058599472046, "learning_rate": 0.00030151515151515153, "loss": 1.4692, "step": 263 }, { "epoch": 1.2, "grad_norm": 2.2137858867645264, "learning_rate": 0.0003007575757575758, "loss": 1.5648, "step": 264 }, { "epoch": 1.2045454545454546, "grad_norm": 1.94793701171875, "learning_rate": 0.0003, "loss": 1.2642, "step": 265 }, { "epoch": 1.209090909090909, "grad_norm": 2.8321635723114014, "learning_rate": 0.00029924242424242425, "loss": 1.4013, "step": 266 }, { "epoch": 1.2136363636363636, "grad_norm": 2.6073920726776123, "learning_rate": 0.00029848484848484847, "loss": 1.6666, "step": 267 }, { "epoch": 1.2181818181818183, "grad_norm": 1.6753661632537842, "learning_rate": 0.00029772727272727275, "loss": 1.2472, "step": 268 }, { "epoch": 1.2227272727272727, "grad_norm": 2.296635866165161, "learning_rate": 0.000296969696969697, "loss": 1.8099, "step": 269 }, { "epoch": 1.2272727272727273, "grad_norm": 2.8359079360961914, "learning_rate": 0.0002962121212121212, "loss": 1.5522, "step": 270 }, { "epoch": 1.231818181818182, "grad_norm": 2.3741962909698486, "learning_rate": 0.00029545454545454547, "loss": 1.5737, "step": 271 }, { "epoch": 1.2363636363636363, "grad_norm": 1.9859591722488403, "learning_rate": 0.0002946969696969697, "loss": 1.5659, "step": 272 }, { "epoch": 1.240909090909091, "grad_norm": 1.9343222379684448, "learning_rate": 0.00029393939393939397, "loss": 1.1204, "step": 273 }, { "epoch": 1.2454545454545454, "grad_norm": 1.6376460790634155, "learning_rate": 0.00029318181818181814, "loss": 0.8886, "step": 274 }, { "epoch": 1.25, "grad_norm": 1.9865474700927734, "learning_rate": 0.0002924242424242424, "loss": 1.5425, "step": 275 }, { "epoch": 1.2545454545454544, "grad_norm": 2.1017825603485107, "learning_rate": 0.0002916666666666667, "loss": 1.1125, "step": 276 }, { "epoch": 1.259090909090909, "grad_norm": 2.349350690841675, "learning_rate": 0.0002909090909090909, "loss": 1.6496, "step": 277 }, { "epoch": 1.2636363636363637, "grad_norm": 1.8741109371185303, "learning_rate": 0.0002901515151515152, "loss": 1.3404, "step": 278 }, { "epoch": 1.268181818181818, "grad_norm": 2.2605037689208984, "learning_rate": 0.00028939393939393936, "loss": 1.5495, "step": 279 }, { "epoch": 1.2727272727272727, "grad_norm": 2.0851080417633057, "learning_rate": 0.00028863636363636363, "loss": 1.501, "step": 280 }, { "epoch": 1.2772727272727273, "grad_norm": 2.2092325687408447, "learning_rate": 0.0002878787878787879, "loss": 1.5655, "step": 281 }, { "epoch": 1.2818181818181817, "grad_norm": 2.343780755996704, "learning_rate": 0.00028712121212121213, "loss": 1.4229, "step": 282 }, { "epoch": 1.2863636363636364, "grad_norm": 1.684411883354187, "learning_rate": 0.00028636363636363636, "loss": 1.4132, "step": 283 }, { "epoch": 1.290909090909091, "grad_norm": 2.034984827041626, "learning_rate": 0.0002856060606060606, "loss": 1.1224, "step": 284 }, { "epoch": 1.2954545454545454, "grad_norm": 1.9973840713500977, "learning_rate": 0.00028484848484848485, "loss": 1.4387, "step": 285 }, { "epoch": 1.3, "grad_norm": 2.2674074172973633, "learning_rate": 0.00028409090909090913, "loss": 1.6697, "step": 286 }, { "epoch": 1.3045454545454547, "grad_norm": 1.596279501914978, "learning_rate": 0.00028333333333333335, "loss": 1.0433, "step": 287 }, { "epoch": 1.309090909090909, "grad_norm": 1.874055027961731, "learning_rate": 0.0002825757575757576, "loss": 1.291, "step": 288 }, { "epoch": 1.3136363636363637, "grad_norm": 2.2551302909851074, "learning_rate": 0.0002818181818181818, "loss": 1.3771, "step": 289 }, { "epoch": 1.3181818181818181, "grad_norm": 2.502380847930908, "learning_rate": 0.0002810606060606061, "loss": 1.853, "step": 290 }, { "epoch": 1.3227272727272728, "grad_norm": 1.9750282764434814, "learning_rate": 0.0002803030303030303, "loss": 1.4369, "step": 291 }, { "epoch": 1.3272727272727272, "grad_norm": 2.4181363582611084, "learning_rate": 0.0002795454545454546, "loss": 1.3565, "step": 292 }, { "epoch": 1.3318181818181818, "grad_norm": 2.0823867321014404, "learning_rate": 0.0002787878787878788, "loss": 1.4589, "step": 293 }, { "epoch": 1.3363636363636364, "grad_norm": 2.147993326187134, "learning_rate": 0.000278030303030303, "loss": 1.3616, "step": 294 }, { "epoch": 1.3409090909090908, "grad_norm": 3.184967517852783, "learning_rate": 0.0002772727272727273, "loss": 1.8248, "step": 295 }, { "epoch": 1.3454545454545455, "grad_norm": 2.3890221118927, "learning_rate": 0.0002765151515151515, "loss": 1.4599, "step": 296 }, { "epoch": 1.35, "grad_norm": 1.9724668264389038, "learning_rate": 0.00027575757575757574, "loss": 1.2248, "step": 297 }, { "epoch": 1.3545454545454545, "grad_norm": 2.1539180278778076, "learning_rate": 0.000275, "loss": 1.1712, "step": 298 }, { "epoch": 1.3590909090909091, "grad_norm": 2.107490062713623, "learning_rate": 0.00027424242424242424, "loss": 1.3786, "step": 299 }, { "epoch": 1.3636363636363638, "grad_norm": 2.052065372467041, "learning_rate": 0.0002734848484848485, "loss": 1.2121, "step": 300 }, { "epoch": 1.3681818181818182, "grad_norm": 2.310871124267578, "learning_rate": 0.00027272727272727274, "loss": 1.4206, "step": 301 }, { "epoch": 1.3727272727272728, "grad_norm": 2.1283962726593018, "learning_rate": 0.00027196969696969696, "loss": 1.8294, "step": 302 }, { "epoch": 1.3772727272727272, "grad_norm": 1.676561951637268, "learning_rate": 0.00027121212121212124, "loss": 0.9432, "step": 303 }, { "epoch": 1.3818181818181818, "grad_norm": 2.4148755073547363, "learning_rate": 0.00027045454545454546, "loss": 1.8412, "step": 304 }, { "epoch": 1.3863636363636362, "grad_norm": 1.668143391609192, "learning_rate": 0.00026969696969696974, "loss": 0.9952, "step": 305 }, { "epoch": 1.3909090909090909, "grad_norm": 2.411818504333496, "learning_rate": 0.0002689393939393939, "loss": 1.657, "step": 306 }, { "epoch": 1.3954545454545455, "grad_norm": 2.2723898887634277, "learning_rate": 0.0002681818181818182, "loss": 1.6628, "step": 307 }, { "epoch": 1.4, "grad_norm": 1.8225884437561035, "learning_rate": 0.0002674242424242424, "loss": 1.3039, "step": 308 }, { "epoch": 1.4045454545454545, "grad_norm": 1.860181450843811, "learning_rate": 0.0002666666666666667, "loss": 1.4974, "step": 309 }, { "epoch": 1.4090909090909092, "grad_norm": 2.22611927986145, "learning_rate": 0.00026590909090909096, "loss": 1.3242, "step": 310 }, { "epoch": 1.4136363636363636, "grad_norm": 2.4301326274871826, "learning_rate": 0.0002651515151515151, "loss": 1.4631, "step": 311 }, { "epoch": 1.4181818181818182, "grad_norm": 2.2716891765594482, "learning_rate": 0.0002643939393939394, "loss": 1.4076, "step": 312 }, { "epoch": 1.4227272727272728, "grad_norm": 1.8279646635055542, "learning_rate": 0.0002636363636363636, "loss": 1.1232, "step": 313 }, { "epoch": 1.4272727272727272, "grad_norm": 1.9681382179260254, "learning_rate": 0.0002628787878787879, "loss": 1.0339, "step": 314 }, { "epoch": 1.4318181818181819, "grad_norm": 2.138864278793335, "learning_rate": 0.0002621212121212122, "loss": 1.4739, "step": 315 }, { "epoch": 1.4363636363636363, "grad_norm": 1.997995376586914, "learning_rate": 0.00026136363636363634, "loss": 1.6025, "step": 316 }, { "epoch": 1.440909090909091, "grad_norm": 2.493382215499878, "learning_rate": 0.0002606060606060606, "loss": 1.9215, "step": 317 }, { "epoch": 1.4454545454545453, "grad_norm": 2.0182077884674072, "learning_rate": 0.00025984848484848484, "loss": 0.8709, "step": 318 }, { "epoch": 1.45, "grad_norm": 2.0383524894714355, "learning_rate": 0.0002590909090909091, "loss": 1.1791, "step": 319 }, { "epoch": 1.4545454545454546, "grad_norm": 2.159406900405884, "learning_rate": 0.00025833333333333334, "loss": 1.896, "step": 320 }, { "epoch": 1.459090909090909, "grad_norm": 2.154700756072998, "learning_rate": 0.00025757575757575756, "loss": 1.4738, "step": 321 }, { "epoch": 1.4636363636363636, "grad_norm": 2.071272134780884, "learning_rate": 0.00025681818181818184, "loss": 1.4454, "step": 322 }, { "epoch": 1.4681818181818183, "grad_norm": 2.1091556549072266, "learning_rate": 0.00025606060606060606, "loss": 1.2386, "step": 323 }, { "epoch": 1.4727272727272727, "grad_norm": 1.8080791234970093, "learning_rate": 0.00025530303030303034, "loss": 0.9288, "step": 324 }, { "epoch": 1.4772727272727273, "grad_norm": 2.170426607131958, "learning_rate": 0.0002545454545454545, "loss": 1.6025, "step": 325 }, { "epoch": 1.481818181818182, "grad_norm": 2.9302620887756348, "learning_rate": 0.0002537878787878788, "loss": 2.0105, "step": 326 }, { "epoch": 1.4863636363636363, "grad_norm": 2.4640023708343506, "learning_rate": 0.00025303030303030306, "loss": 1.5101, "step": 327 }, { "epoch": 1.490909090909091, "grad_norm": 1.6185390949249268, "learning_rate": 0.0002522727272727273, "loss": 0.9489, "step": 328 }, { "epoch": 1.4954545454545456, "grad_norm": 1.4413659572601318, "learning_rate": 0.0002515151515151515, "loss": 0.8982, "step": 329 }, { "epoch": 1.5, "grad_norm": 1.884208083152771, "learning_rate": 0.00025075757575757573, "loss": 1.2771, "step": 330 }, { "epoch": 1.5045454545454544, "grad_norm": 1.9692877531051636, "learning_rate": 0.00025, "loss": 1.6345, "step": 331 }, { "epoch": 1.509090909090909, "grad_norm": 2.3343496322631836, "learning_rate": 0.00024924242424242423, "loss": 1.6179, "step": 332 }, { "epoch": 1.5136363636363637, "grad_norm": 2.2422614097595215, "learning_rate": 0.0002484848484848485, "loss": 1.3785, "step": 333 }, { "epoch": 1.518181818181818, "grad_norm": 2.130425214767456, "learning_rate": 0.0002477272727272727, "loss": 1.6261, "step": 334 }, { "epoch": 1.5227272727272727, "grad_norm": 2.13519287109375, "learning_rate": 0.000246969696969697, "loss": 1.4136, "step": 335 }, { "epoch": 1.5272727272727273, "grad_norm": 1.268389344215393, "learning_rate": 0.0002462121212121212, "loss": 0.9329, "step": 336 }, { "epoch": 1.5318181818181817, "grad_norm": 2.094594955444336, "learning_rate": 0.00024545454545454545, "loss": 1.3465, "step": 337 }, { "epoch": 1.5363636363636364, "grad_norm": 2.156473159790039, "learning_rate": 0.00024469696969696967, "loss": 1.2741, "step": 338 }, { "epoch": 1.540909090909091, "grad_norm": 1.898336410522461, "learning_rate": 0.00024393939393939392, "loss": 1.059, "step": 339 }, { "epoch": 1.5454545454545454, "grad_norm": 1.8859503269195557, "learning_rate": 0.0002431818181818182, "loss": 1.4959, "step": 340 }, { "epoch": 1.55, "grad_norm": 2.060011863708496, "learning_rate": 0.00024242424242424245, "loss": 1.3644, "step": 341 }, { "epoch": 1.5545454545454547, "grad_norm": 2.3816416263580322, "learning_rate": 0.00024166666666666667, "loss": 1.4375, "step": 342 }, { "epoch": 1.559090909090909, "grad_norm": 3.5078084468841553, "learning_rate": 0.00024090909090909092, "loss": 1.422, "step": 343 }, { "epoch": 1.5636363636363635, "grad_norm": 2.221052885055542, "learning_rate": 0.00024015151515151514, "loss": 1.3024, "step": 344 }, { "epoch": 1.5681818181818183, "grad_norm": 2.4711906909942627, "learning_rate": 0.0002393939393939394, "loss": 1.4838, "step": 345 }, { "epoch": 1.5727272727272728, "grad_norm": 1.9803431034088135, "learning_rate": 0.00023863636363636364, "loss": 1.1055, "step": 346 }, { "epoch": 1.5772727272727272, "grad_norm": 2.3196969032287598, "learning_rate": 0.0002378787878787879, "loss": 1.8282, "step": 347 }, { "epoch": 1.5818181818181818, "grad_norm": 2.094829797744751, "learning_rate": 0.00023712121212121214, "loss": 1.2518, "step": 348 }, { "epoch": 1.5863636363636364, "grad_norm": 2.0144240856170654, "learning_rate": 0.00023636363636363636, "loss": 1.4561, "step": 349 }, { "epoch": 1.5909090909090908, "grad_norm": 1.8540327548980713, "learning_rate": 0.0002356060606060606, "loss": 1.3567, "step": 350 }, { "epoch": 1.5954545454545455, "grad_norm": 2.6601638793945312, "learning_rate": 0.00023484848484848486, "loss": 1.7279, "step": 351 }, { "epoch": 1.6, "grad_norm": 2.5020227432250977, "learning_rate": 0.0002340909090909091, "loss": 1.7535, "step": 352 }, { "epoch": 1.6045454545454545, "grad_norm": 2.0830986499786377, "learning_rate": 0.00023333333333333333, "loss": 1.2985, "step": 353 }, { "epoch": 1.6090909090909091, "grad_norm": 1.8451659679412842, "learning_rate": 0.00023257575757575758, "loss": 1.0175, "step": 354 }, { "epoch": 1.6136363636363638, "grad_norm": 2.271484375, "learning_rate": 0.00023181818181818183, "loss": 1.4269, "step": 355 }, { "epoch": 1.6181818181818182, "grad_norm": 4.305004596710205, "learning_rate": 0.00023106060606060605, "loss": 1.3775, "step": 356 }, { "epoch": 1.6227272727272726, "grad_norm": 1.8406462669372559, "learning_rate": 0.00023030303030303033, "loss": 1.2472, "step": 357 }, { "epoch": 1.6272727272727274, "grad_norm": 2.6075923442840576, "learning_rate": 0.00022954545454545455, "loss": 1.3993, "step": 358 }, { "epoch": 1.6318181818181818, "grad_norm": 2.324907064437866, "learning_rate": 0.0002287878787878788, "loss": 1.3076, "step": 359 }, { "epoch": 1.6363636363636362, "grad_norm": 1.954463005065918, "learning_rate": 0.00022803030303030302, "loss": 1.4135, "step": 360 }, { "epoch": 1.6409090909090909, "grad_norm": 1.8350000381469727, "learning_rate": 0.00022727272727272727, "loss": 1.2618, "step": 361 }, { "epoch": 1.6454545454545455, "grad_norm": 2.2729427814483643, "learning_rate": 0.0002265151515151515, "loss": 1.2483, "step": 362 }, { "epoch": 1.65, "grad_norm": 1.7203510999679565, "learning_rate": 0.00022575757575757577, "loss": 0.9189, "step": 363 }, { "epoch": 1.6545454545454545, "grad_norm": 2.2685649394989014, "learning_rate": 0.00022500000000000002, "loss": 1.4897, "step": 364 }, { "epoch": 1.6590909090909092, "grad_norm": 2.502887010574341, "learning_rate": 0.00022424242424242424, "loss": 1.315, "step": 365 }, { "epoch": 1.6636363636363636, "grad_norm": 1.8945139646530151, "learning_rate": 0.0002234848484848485, "loss": 1.1686, "step": 366 }, { "epoch": 1.6681818181818182, "grad_norm": 2.207409620285034, "learning_rate": 0.00022272727272727272, "loss": 1.9896, "step": 367 }, { "epoch": 1.6727272727272728, "grad_norm": 2.028022527694702, "learning_rate": 0.000221969696969697, "loss": 1.5135, "step": 368 }, { "epoch": 1.6772727272727272, "grad_norm": 2.403134822845459, "learning_rate": 0.00022121212121212121, "loss": 1.4201, "step": 369 }, { "epoch": 1.6818181818181817, "grad_norm": 2.022304058074951, "learning_rate": 0.00022045454545454546, "loss": 1.672, "step": 370 }, { "epoch": 1.6863636363636365, "grad_norm": 2.3457093238830566, "learning_rate": 0.0002196969696969697, "loss": 1.4657, "step": 371 }, { "epoch": 1.690909090909091, "grad_norm": 2.0904908180236816, "learning_rate": 0.00021893939393939394, "loss": 1.5864, "step": 372 }, { "epoch": 1.6954545454545453, "grad_norm": 1.5914294719696045, "learning_rate": 0.00021818181818181818, "loss": 1.2828, "step": 373 }, { "epoch": 1.7, "grad_norm": 1.8519418239593506, "learning_rate": 0.00021742424242424243, "loss": 1.5195, "step": 374 }, { "epoch": 1.7045454545454546, "grad_norm": 2.2076525688171387, "learning_rate": 0.00021666666666666668, "loss": 1.247, "step": 375 }, { "epoch": 1.709090909090909, "grad_norm": 1.8584811687469482, "learning_rate": 0.0002159090909090909, "loss": 1.6614, "step": 376 }, { "epoch": 1.7136363636363636, "grad_norm": 1.9943331480026245, "learning_rate": 0.00021515151515151516, "loss": 1.2409, "step": 377 }, { "epoch": 1.7181818181818183, "grad_norm": 1.5293704271316528, "learning_rate": 0.00021439393939393938, "loss": 1.1563, "step": 378 }, { "epoch": 1.7227272727272727, "grad_norm": 2.835527181625366, "learning_rate": 0.00021363636363636363, "loss": 1.3795, "step": 379 }, { "epoch": 1.7272727272727273, "grad_norm": 2.491914987564087, "learning_rate": 0.0002128787878787879, "loss": 1.3017, "step": 380 }, { "epoch": 1.731818181818182, "grad_norm": 3.2870216369628906, "learning_rate": 0.00021212121212121213, "loss": 1.1947, "step": 381 }, { "epoch": 1.7363636363636363, "grad_norm": 2.1510424613952637, "learning_rate": 0.00021136363636363638, "loss": 1.7048, "step": 382 }, { "epoch": 1.740909090909091, "grad_norm": 2.2372002601623535, "learning_rate": 0.0002106060606060606, "loss": 1.1627, "step": 383 }, { "epoch": 1.7454545454545456, "grad_norm": 2.217090368270874, "learning_rate": 0.00020984848484848485, "loss": 1.8424, "step": 384 }, { "epoch": 1.75, "grad_norm": 1.8897899389266968, "learning_rate": 0.00020909090909090907, "loss": 1.3006, "step": 385 }, { "epoch": 1.7545454545454544, "grad_norm": 1.7469165325164795, "learning_rate": 0.00020833333333333335, "loss": 1.496, "step": 386 }, { "epoch": 1.759090909090909, "grad_norm": 1.8982353210449219, "learning_rate": 0.0002075757575757576, "loss": 1.2682, "step": 387 }, { "epoch": 1.7636363636363637, "grad_norm": 2.0795273780822754, "learning_rate": 0.00020681818181818182, "loss": 2.2314, "step": 388 }, { "epoch": 1.768181818181818, "grad_norm": 1.6244702339172363, "learning_rate": 0.00020606060606060607, "loss": 0.9585, "step": 389 }, { "epoch": 1.7727272727272727, "grad_norm": 2.3025052547454834, "learning_rate": 0.0002053030303030303, "loss": 1.6639, "step": 390 }, { "epoch": 1.7772727272727273, "grad_norm": 2.418119192123413, "learning_rate": 0.00020454545454545457, "loss": 1.8274, "step": 391 }, { "epoch": 1.7818181818181817, "grad_norm": 1.70542573928833, "learning_rate": 0.0002037878787878788, "loss": 1.3115, "step": 392 }, { "epoch": 1.7863636363636364, "grad_norm": 1.9681594371795654, "learning_rate": 0.00020303030303030304, "loss": 0.973, "step": 393 }, { "epoch": 1.790909090909091, "grad_norm": 1.856879711151123, "learning_rate": 0.0002022727272727273, "loss": 1.2958, "step": 394 }, { "epoch": 1.7954545454545454, "grad_norm": 2.1984550952911377, "learning_rate": 0.0002015151515151515, "loss": 1.5643, "step": 395 }, { "epoch": 1.8, "grad_norm": 2.6989824771881104, "learning_rate": 0.00020075757575757576, "loss": 1.4334, "step": 396 }, { "epoch": 1.8045454545454547, "grad_norm": 2.3298850059509277, "learning_rate": 0.0002, "loss": 1.5145, "step": 397 }, { "epoch": 1.809090909090909, "grad_norm": 1.9490818977355957, "learning_rate": 0.00019924242424242426, "loss": 1.5346, "step": 398 }, { "epoch": 1.8136363636363635, "grad_norm": 2.0060818195343018, "learning_rate": 0.00019848484848484848, "loss": 1.2493, "step": 399 }, { "epoch": 1.8181818181818183, "grad_norm": 2.1615042686462402, "learning_rate": 0.00019772727272727273, "loss": 1.8834, "step": 400 }, { "epoch": 1.8227272727272728, "grad_norm": 2.4424972534179688, "learning_rate": 0.00019696969696969695, "loss": 1.6012, "step": 401 }, { "epoch": 1.8272727272727272, "grad_norm": 1.83026921749115, "learning_rate": 0.0001962121212121212, "loss": 1.4308, "step": 402 }, { "epoch": 1.8318181818181818, "grad_norm": 2.6793453693389893, "learning_rate": 0.00019545454545454548, "loss": 1.6356, "step": 403 }, { "epoch": 1.8363636363636364, "grad_norm": 2.1211740970611572, "learning_rate": 0.0001946969696969697, "loss": 1.0859, "step": 404 }, { "epoch": 1.8409090909090908, "grad_norm": 1.6658772230148315, "learning_rate": 0.00019393939393939395, "loss": 1.3467, "step": 405 }, { "epoch": 1.8454545454545455, "grad_norm": 1.7903032302856445, "learning_rate": 0.00019318181818181817, "loss": 1.0118, "step": 406 }, { "epoch": 1.85, "grad_norm": 1.7358275651931763, "learning_rate": 0.00019242424242424242, "loss": 1.4404, "step": 407 }, { "epoch": 1.8545454545454545, "grad_norm": 2.0745861530303955, "learning_rate": 0.00019166666666666667, "loss": 1.3409, "step": 408 }, { "epoch": 1.8590909090909091, "grad_norm": 2.1365160942077637, "learning_rate": 0.00019090909090909092, "loss": 1.4658, "step": 409 }, { "epoch": 1.8636363636363638, "grad_norm": 2.040371894836426, "learning_rate": 0.00019015151515151517, "loss": 1.3165, "step": 410 }, { "epoch": 1.8681818181818182, "grad_norm": 1.9335429668426514, "learning_rate": 0.0001893939393939394, "loss": 1.5063, "step": 411 }, { "epoch": 1.8727272727272726, "grad_norm": 2.0025243759155273, "learning_rate": 0.00018863636363636364, "loss": 1.2584, "step": 412 }, { "epoch": 1.8772727272727274, "grad_norm": 1.7558890581130981, "learning_rate": 0.00018787878787878787, "loss": 1.0937, "step": 413 }, { "epoch": 1.8818181818181818, "grad_norm": 1.886003851890564, "learning_rate": 0.00018712121212121214, "loss": 1.3969, "step": 414 }, { "epoch": 1.8863636363636362, "grad_norm": 2.6999497413635254, "learning_rate": 0.00018636363636363636, "loss": 1.1411, "step": 415 }, { "epoch": 1.8909090909090909, "grad_norm": 1.923196792602539, "learning_rate": 0.00018560606060606061, "loss": 1.6597, "step": 416 }, { "epoch": 1.8954545454545455, "grad_norm": 2.261504650115967, "learning_rate": 0.00018484848484848484, "loss": 1.738, "step": 417 }, { "epoch": 1.9, "grad_norm": 1.9908875226974487, "learning_rate": 0.00018409090909090909, "loss": 1.1275, "step": 418 }, { "epoch": 1.9045454545454545, "grad_norm": 2.348240852355957, "learning_rate": 0.00018333333333333334, "loss": 1.5368, "step": 419 }, { "epoch": 1.9090909090909092, "grad_norm": 3.1023001670837402, "learning_rate": 0.00018257575757575758, "loss": 1.1828, "step": 420 }, { "epoch": 1.9136363636363636, "grad_norm": 2.2243757247924805, "learning_rate": 0.00018181818181818183, "loss": 1.4783, "step": 421 }, { "epoch": 1.9181818181818182, "grad_norm": 1.824317216873169, "learning_rate": 0.00018106060606060606, "loss": 1.178, "step": 422 }, { "epoch": 1.9227272727272728, "grad_norm": 2.1963822841644287, "learning_rate": 0.0001803030303030303, "loss": 1.5811, "step": 423 }, { "epoch": 1.9272727272727272, "grad_norm": 1.8573740720748901, "learning_rate": 0.00017954545454545453, "loss": 1.2361, "step": 424 }, { "epoch": 1.9318181818181817, "grad_norm": 2.223315715789795, "learning_rate": 0.0001787878787878788, "loss": 1.3236, "step": 425 }, { "epoch": 1.9363636363636365, "grad_norm": 2.0890495777130127, "learning_rate": 0.00017803030303030305, "loss": 1.4358, "step": 426 }, { "epoch": 1.940909090909091, "grad_norm": 1.8097957372665405, "learning_rate": 0.00017727272727272728, "loss": 1.0634, "step": 427 }, { "epoch": 1.9454545454545453, "grad_norm": 2.378750801086426, "learning_rate": 0.00017651515151515153, "loss": 1.8565, "step": 428 }, { "epoch": 1.95, "grad_norm": 1.7777960300445557, "learning_rate": 0.00017575757575757575, "loss": 1.5755, "step": 429 }, { "epoch": 1.9545454545454546, "grad_norm": 2.1310245990753174, "learning_rate": 0.000175, "loss": 1.4119, "step": 430 }, { "epoch": 1.959090909090909, "grad_norm": 2.506479024887085, "learning_rate": 0.00017424242424242425, "loss": 1.5025, "step": 431 }, { "epoch": 1.9636363636363636, "grad_norm": 2.091891288757324, "learning_rate": 0.0001734848484848485, "loss": 1.383, "step": 432 }, { "epoch": 1.9681818181818183, "grad_norm": 1.7450625896453857, "learning_rate": 0.00017272727272727272, "loss": 1.4122, "step": 433 }, { "epoch": 1.9727272727272727, "grad_norm": 1.7834117412567139, "learning_rate": 0.00017196969696969697, "loss": 1.0932, "step": 434 }, { "epoch": 1.9772727272727273, "grad_norm": 1.6854950189590454, "learning_rate": 0.00017121212121212122, "loss": 0.9985, "step": 435 }, { "epoch": 1.981818181818182, "grad_norm": 1.4569097757339478, "learning_rate": 0.00017045454545454544, "loss": 1.1354, "step": 436 }, { "epoch": 1.9863636363636363, "grad_norm": 2.3083584308624268, "learning_rate": 0.00016969696969696972, "loss": 1.4856, "step": 437 }, { "epoch": 1.990909090909091, "grad_norm": 2.1282572746276855, "learning_rate": 0.00016893939393939394, "loss": 1.942, "step": 438 }, { "epoch": 1.9954545454545456, "grad_norm": 2.199822187423706, "learning_rate": 0.0001681818181818182, "loss": 1.4891, "step": 439 }, { "epoch": 2.0, "grad_norm": 2.0030460357666016, "learning_rate": 0.0001674242424242424, "loss": 1.2939, "step": 440 }, { "epoch": 2.0, "eval_f1": 0.8928, "eval_gen_len": 42.0, "eval_loss": 1.7928985357284546, "eval_precision": 0.8914, "eval_recall": 0.8944, "eval_rouge1": 0.4605, "eval_rouge2": 0.2125, "eval_rougeL": 0.3897, "eval_rougeLsum": 0.4259, "eval_runtime": 25.1108, "eval_samples_per_second": 4.381, "eval_steps_per_second": 0.558, "step": 440 }, { "epoch": 2.0045454545454544, "grad_norm": 1.4425781965255737, "learning_rate": 0.00016666666666666666, "loss": 0.9057, "step": 441 }, { "epoch": 2.0090909090909093, "grad_norm": 1.579765796661377, "learning_rate": 0.00016590909090909094, "loss": 0.7069, "step": 442 }, { "epoch": 2.0136363636363637, "grad_norm": 1.8639825582504272, "learning_rate": 0.00016515151515151516, "loss": 1.1531, "step": 443 }, { "epoch": 2.018181818181818, "grad_norm": 1.4890676736831665, "learning_rate": 0.0001643939393939394, "loss": 0.8112, "step": 444 }, { "epoch": 2.022727272727273, "grad_norm": 1.7381901741027832, "learning_rate": 0.00016363636363636363, "loss": 1.2108, "step": 445 }, { "epoch": 2.0272727272727273, "grad_norm": 1.6125924587249756, "learning_rate": 0.00016287878787878788, "loss": 1.0529, "step": 446 }, { "epoch": 2.0318181818181817, "grad_norm": 1.8624428510665894, "learning_rate": 0.0001621212121212121, "loss": 1.006, "step": 447 }, { "epoch": 2.036363636363636, "grad_norm": 1.719439148902893, "learning_rate": 0.00016136363636363638, "loss": 1.0881, "step": 448 }, { "epoch": 2.040909090909091, "grad_norm": 2.446216106414795, "learning_rate": 0.00016060606060606063, "loss": 1.2399, "step": 449 }, { "epoch": 2.0454545454545454, "grad_norm": 1.703517198562622, "learning_rate": 0.00015984848484848485, "loss": 0.931, "step": 450 }, { "epoch": 2.05, "grad_norm": 1.780228853225708, "learning_rate": 0.0001590909090909091, "loss": 0.9769, "step": 451 }, { "epoch": 2.0545454545454547, "grad_norm": 2.015679121017456, "learning_rate": 0.00015833333333333332, "loss": 1.6044, "step": 452 }, { "epoch": 2.059090909090909, "grad_norm": 2.084481716156006, "learning_rate": 0.00015757575757575757, "loss": 0.9933, "step": 453 }, { "epoch": 2.0636363636363635, "grad_norm": 2.3098299503326416, "learning_rate": 0.00015681818181818182, "loss": 0.9405, "step": 454 }, { "epoch": 2.0681818181818183, "grad_norm": 1.8041385412216187, "learning_rate": 0.00015606060606060607, "loss": 1.1748, "step": 455 }, { "epoch": 2.0727272727272728, "grad_norm": 1.693158745765686, "learning_rate": 0.0001553030303030303, "loss": 0.9358, "step": 456 }, { "epoch": 2.077272727272727, "grad_norm": 1.5484883785247803, "learning_rate": 0.00015454545454545454, "loss": 1.0664, "step": 457 }, { "epoch": 2.081818181818182, "grad_norm": 1.4313092231750488, "learning_rate": 0.0001537878787878788, "loss": 0.6624, "step": 458 }, { "epoch": 2.0863636363636364, "grad_norm": 2.218092679977417, "learning_rate": 0.00015303030303030302, "loss": 0.9856, "step": 459 }, { "epoch": 2.090909090909091, "grad_norm": 2.030869960784912, "learning_rate": 0.0001522727272727273, "loss": 1.143, "step": 460 }, { "epoch": 2.0954545454545457, "grad_norm": 2.190603017807007, "learning_rate": 0.00015151515151515152, "loss": 1.077, "step": 461 }, { "epoch": 2.1, "grad_norm": 1.3030821084976196, "learning_rate": 0.00015075757575757576, "loss": 0.6711, "step": 462 }, { "epoch": 2.1045454545454545, "grad_norm": 1.8678494691848755, "learning_rate": 0.00015, "loss": 1.0674, "step": 463 }, { "epoch": 2.109090909090909, "grad_norm": 1.407085657119751, "learning_rate": 0.00014924242424242424, "loss": 0.7024, "step": 464 }, { "epoch": 2.1136363636363638, "grad_norm": 2.004911422729492, "learning_rate": 0.0001484848484848485, "loss": 0.795, "step": 465 }, { "epoch": 2.118181818181818, "grad_norm": 2.25128173828125, "learning_rate": 0.00014772727272727274, "loss": 1.2232, "step": 466 }, { "epoch": 2.1227272727272726, "grad_norm": 1.960771918296814, "learning_rate": 0.00014696969696969698, "loss": 1.0019, "step": 467 }, { "epoch": 2.1272727272727274, "grad_norm": 1.9563887119293213, "learning_rate": 0.0001462121212121212, "loss": 0.9798, "step": 468 }, { "epoch": 2.131818181818182, "grad_norm": 1.687361240386963, "learning_rate": 0.00014545454545454546, "loss": 0.755, "step": 469 }, { "epoch": 2.1363636363636362, "grad_norm": 2.191286325454712, "learning_rate": 0.00014469696969696968, "loss": 1.0018, "step": 470 }, { "epoch": 2.140909090909091, "grad_norm": 2.046880006790161, "learning_rate": 0.00014393939393939396, "loss": 1.2281, "step": 471 }, { "epoch": 2.1454545454545455, "grad_norm": 2.4996211528778076, "learning_rate": 0.00014318181818181818, "loss": 1.0795, "step": 472 }, { "epoch": 2.15, "grad_norm": 1.8937994241714478, "learning_rate": 0.00014242424242424243, "loss": 1.0556, "step": 473 }, { "epoch": 2.1545454545454543, "grad_norm": 2.250491142272949, "learning_rate": 0.00014166666666666668, "loss": 0.8816, "step": 474 }, { "epoch": 2.159090909090909, "grad_norm": NaN, "learning_rate": 0.00014166666666666668, "loss": 0.0, "step": 475 }, { "epoch": 2.1636363636363636, "grad_norm": 2.231706142425537, "learning_rate": 0.0001409090909090909, "loss": 1.2344, "step": 476 }, { "epoch": 2.168181818181818, "grad_norm": 2.2170498371124268, "learning_rate": 0.00014015151515151515, "loss": 1.4409, "step": 477 }, { "epoch": 2.172727272727273, "grad_norm": 2.3106095790863037, "learning_rate": 0.0001393939393939394, "loss": 0.9081, "step": 478 }, { "epoch": 2.1772727272727272, "grad_norm": 1.9665738344192505, "learning_rate": 0.00013863636363636365, "loss": 1.3029, "step": 479 }, { "epoch": 2.1818181818181817, "grad_norm": 2.321331739425659, "learning_rate": 0.00013787878787878787, "loss": 1.4714, "step": 480 }, { "epoch": 2.1863636363636365, "grad_norm": 2.0038533210754395, "learning_rate": 0.00013712121212121212, "loss": 1.0879, "step": 481 }, { "epoch": 2.190909090909091, "grad_norm": 1.6077767610549927, "learning_rate": 0.00013636363636363637, "loss": 0.6456, "step": 482 }, { "epoch": 2.1954545454545453, "grad_norm": 1.5018125772476196, "learning_rate": 0.00013560606060606062, "loss": 0.6937, "step": 483 }, { "epoch": 2.2, "grad_norm": 1.5473967790603638, "learning_rate": 0.00013484848484848487, "loss": 0.6191, "step": 484 }, { "epoch": 2.2045454545454546, "grad_norm": 2.5554354190826416, "learning_rate": 0.0001340909090909091, "loss": 1.4345, "step": 485 }, { "epoch": 2.209090909090909, "grad_norm": 2.1666858196258545, "learning_rate": 0.00013333333333333334, "loss": 1.4182, "step": 486 }, { "epoch": 2.213636363636364, "grad_norm": 2.2915759086608887, "learning_rate": 0.00013257575757575756, "loss": 0.8752, "step": 487 }, { "epoch": 2.2181818181818183, "grad_norm": 2.24314546585083, "learning_rate": 0.0001318181818181818, "loss": 1.3214, "step": 488 }, { "epoch": 2.2227272727272727, "grad_norm": 2.269216537475586, "learning_rate": 0.0001310606060606061, "loss": 0.9968, "step": 489 }, { "epoch": 2.227272727272727, "grad_norm": 2.3108322620391846, "learning_rate": 0.0001303030303030303, "loss": 0.9695, "step": 490 }, { "epoch": 2.231818181818182, "grad_norm": 2.3146250247955322, "learning_rate": 0.00012954545454545456, "loss": 1.4007, "step": 491 }, { "epoch": 2.2363636363636363, "grad_norm": 1.9747002124786377, "learning_rate": 0.00012878787878787878, "loss": 0.8876, "step": 492 }, { "epoch": 2.2409090909090907, "grad_norm": 2.0410826206207275, "learning_rate": 0.00012803030303030303, "loss": 0.9588, "step": 493 }, { "epoch": 2.2454545454545456, "grad_norm": 2.2743778228759766, "learning_rate": 0.00012727272727272725, "loss": 1.2062, "step": 494 }, { "epoch": 2.25, "grad_norm": 2.272749662399292, "learning_rate": 0.00012651515151515153, "loss": 0.975, "step": 495 }, { "epoch": 2.2545454545454544, "grad_norm": 2.297175884246826, "learning_rate": 0.00012575757575757575, "loss": 1.0806, "step": 496 }, { "epoch": 2.2590909090909093, "grad_norm": 2.2274718284606934, "learning_rate": 0.000125, "loss": 0.9391, "step": 497 }, { "epoch": 2.2636363636363637, "grad_norm": 2.4175453186035156, "learning_rate": 0.00012424242424242425, "loss": 1.2736, "step": 498 }, { "epoch": 2.268181818181818, "grad_norm": 1.7530089616775513, "learning_rate": 0.0001234848484848485, "loss": 1.1917, "step": 499 }, { "epoch": 2.2727272727272725, "grad_norm": 2.598747730255127, "learning_rate": 0.00012272727272727272, "loss": 1.5901, "step": 500 }, { "epoch": 2.2772727272727273, "grad_norm": 2.0590171813964844, "learning_rate": 0.00012196969696969696, "loss": 1.0049, "step": 501 }, { "epoch": 2.2818181818181817, "grad_norm": 1.6530457735061646, "learning_rate": 0.00012121212121212122, "loss": 0.6991, "step": 502 }, { "epoch": 2.286363636363636, "grad_norm": 1.4000625610351562, "learning_rate": 0.00012045454545454546, "loss": 0.7258, "step": 503 }, { "epoch": 2.290909090909091, "grad_norm": 3.4282798767089844, "learning_rate": 0.0001196969696969697, "loss": 0.7331, "step": 504 }, { "epoch": 2.2954545454545454, "grad_norm": 2.0328640937805176, "learning_rate": 0.00011893939393939394, "loss": 1.0245, "step": 505 }, { "epoch": 2.3, "grad_norm": 2.639125347137451, "learning_rate": 0.00011818181818181818, "loss": 0.939, "step": 506 }, { "epoch": 2.3045454545454547, "grad_norm": 2.069645643234253, "learning_rate": 0.00011742424242424243, "loss": 1.186, "step": 507 }, { "epoch": 2.309090909090909, "grad_norm": 2.103675603866577, "learning_rate": 0.00011666666666666667, "loss": 1.0986, "step": 508 }, { "epoch": 2.3136363636363635, "grad_norm": 2.022813320159912, "learning_rate": 0.00011590909090909091, "loss": 1.1106, "step": 509 }, { "epoch": 2.3181818181818183, "grad_norm": 2.1240341663360596, "learning_rate": 0.00011515151515151516, "loss": 0.9754, "step": 510 }, { "epoch": 2.3227272727272728, "grad_norm": 2.11362361907959, "learning_rate": 0.0001143939393939394, "loss": 1.243, "step": 511 }, { "epoch": 2.327272727272727, "grad_norm": 1.9033676385879517, "learning_rate": 0.00011363636363636364, "loss": 0.7314, "step": 512 }, { "epoch": 2.331818181818182, "grad_norm": 2.7902817726135254, "learning_rate": 0.00011287878787878789, "loss": 1.2161, "step": 513 }, { "epoch": 2.3363636363636364, "grad_norm": 2.1139214038848877, "learning_rate": 0.00011212121212121212, "loss": 1.4216, "step": 514 }, { "epoch": 2.340909090909091, "grad_norm": 2.2380800247192383, "learning_rate": 0.00011136363636363636, "loss": 1.0319, "step": 515 }, { "epoch": 2.3454545454545457, "grad_norm": 1.9591755867004395, "learning_rate": 0.00011060606060606061, "loss": 0.7923, "step": 516 }, { "epoch": 2.35, "grad_norm": 2.6767358779907227, "learning_rate": 0.00010984848484848486, "loss": 0.9721, "step": 517 }, { "epoch": 2.3545454545454545, "grad_norm": 2.350008487701416, "learning_rate": 0.00010909090909090909, "loss": 1.1793, "step": 518 }, { "epoch": 2.359090909090909, "grad_norm": 2.0240652561187744, "learning_rate": 0.00010833333333333334, "loss": 1.1184, "step": 519 }, { "epoch": 2.3636363636363638, "grad_norm": 2.058748960494995, "learning_rate": 0.00010757575757575758, "loss": 1.1886, "step": 520 }, { "epoch": 2.368181818181818, "grad_norm": 1.7921950817108154, "learning_rate": 0.00010681818181818181, "loss": 0.8511, "step": 521 }, { "epoch": 2.3727272727272726, "grad_norm": 2.027445077896118, "learning_rate": 0.00010606060606060606, "loss": 0.8641, "step": 522 }, { "epoch": 2.3772727272727274, "grad_norm": 1.8156445026397705, "learning_rate": 0.0001053030303030303, "loss": 0.8234, "step": 523 }, { "epoch": 2.381818181818182, "grad_norm": 2.3511455059051514, "learning_rate": 0.00010454545454545454, "loss": 1.048, "step": 524 }, { "epoch": 2.3863636363636362, "grad_norm": 1.489744782447815, "learning_rate": 0.0001037878787878788, "loss": 0.4886, "step": 525 }, { "epoch": 2.390909090909091, "grad_norm": 2.0359721183776855, "learning_rate": 0.00010303030303030303, "loss": 1.0011, "step": 526 }, { "epoch": 2.3954545454545455, "grad_norm": 2.8290212154388428, "learning_rate": 0.00010227272727272728, "loss": 1.4443, "step": 527 }, { "epoch": 2.4, "grad_norm": 1.991904616355896, "learning_rate": 0.00010151515151515152, "loss": 0.9877, "step": 528 }, { "epoch": 2.4045454545454543, "grad_norm": 1.8174313306808472, "learning_rate": 0.00010075757575757576, "loss": 0.9048, "step": 529 }, { "epoch": 2.409090909090909, "grad_norm": 1.66022527217865, "learning_rate": 0.0001, "loss": 0.9039, "step": 530 }, { "epoch": 2.4136363636363636, "grad_norm": 1.6025142669677734, "learning_rate": 9.924242424242424e-05, "loss": 0.8169, "step": 531 }, { "epoch": 2.418181818181818, "grad_norm": 1.871733546257019, "learning_rate": 9.848484848484848e-05, "loss": 1.039, "step": 532 }, { "epoch": 2.422727272727273, "grad_norm": 2.35320782661438, "learning_rate": 9.772727272727274e-05, "loss": 1.3449, "step": 533 }, { "epoch": 2.4272727272727272, "grad_norm": 1.9311728477478027, "learning_rate": 9.696969696969698e-05, "loss": 1.0332, "step": 534 }, { "epoch": 2.4318181818181817, "grad_norm": 1.6838319301605225, "learning_rate": 9.621212121212121e-05, "loss": 0.6631, "step": 535 }, { "epoch": 2.4363636363636365, "grad_norm": 1.9957849979400635, "learning_rate": 9.545454545454546e-05, "loss": 1.0397, "step": 536 }, { "epoch": 2.440909090909091, "grad_norm": 2.338730573654175, "learning_rate": 9.46969696969697e-05, "loss": 1.1155, "step": 537 }, { "epoch": 2.4454545454545453, "grad_norm": 2.0578792095184326, "learning_rate": 9.393939393939393e-05, "loss": 1.0634, "step": 538 }, { "epoch": 2.45, "grad_norm": 2.0512609481811523, "learning_rate": 9.318181818181818e-05, "loss": 0.9052, "step": 539 }, { "epoch": 2.4545454545454546, "grad_norm": 2.2808845043182373, "learning_rate": 9.242424242424242e-05, "loss": 1.2479, "step": 540 }, { "epoch": 2.459090909090909, "grad_norm": 1.7963327169418335, "learning_rate": 9.166666666666667e-05, "loss": 0.8655, "step": 541 }, { "epoch": 2.463636363636364, "grad_norm": 2.378777265548706, "learning_rate": 9.090909090909092e-05, "loss": 1.1019, "step": 542 }, { "epoch": 2.4681818181818183, "grad_norm": 1.7346596717834473, "learning_rate": 9.015151515151515e-05, "loss": 0.6478, "step": 543 }, { "epoch": 2.4727272727272727, "grad_norm": 1.8121107816696167, "learning_rate": 8.93939393939394e-05, "loss": 0.9549, "step": 544 }, { "epoch": 2.4772727272727275, "grad_norm": 1.9102083444595337, "learning_rate": 8.863636363636364e-05, "loss": 0.9103, "step": 545 }, { "epoch": 2.481818181818182, "grad_norm": 2.3148677349090576, "learning_rate": 8.787878787878787e-05, "loss": 1.1075, "step": 546 }, { "epoch": 2.4863636363636363, "grad_norm": 2.3098530769348145, "learning_rate": 8.712121212121212e-05, "loss": 1.0885, "step": 547 }, { "epoch": 2.4909090909090907, "grad_norm": 2.061582565307617, "learning_rate": 8.636363636363636e-05, "loss": 0.7894, "step": 548 }, { "epoch": 2.4954545454545456, "grad_norm": 2.3829803466796875, "learning_rate": 8.560606060606061e-05, "loss": 1.2397, "step": 549 }, { "epoch": 2.5, "grad_norm": 2.111055612564087, "learning_rate": 8.484848484848486e-05, "loss": 1.0463, "step": 550 }, { "epoch": 2.5045454545454544, "grad_norm": 1.883468508720398, "learning_rate": 8.40909090909091e-05, "loss": 0.9837, "step": 551 }, { "epoch": 2.509090909090909, "grad_norm": 1.8480087518692017, "learning_rate": 8.333333333333333e-05, "loss": 0.897, "step": 552 }, { "epoch": 2.5136363636363637, "grad_norm": 1.9513871669769287, "learning_rate": 8.257575757575758e-05, "loss": 0.9668, "step": 553 }, { "epoch": 2.518181818181818, "grad_norm": 1.5687415599822998, "learning_rate": 8.181818181818182e-05, "loss": 0.8729, "step": 554 }, { "epoch": 2.5227272727272725, "grad_norm": 1.959887981414795, "learning_rate": 8.106060606060605e-05, "loss": 0.9612, "step": 555 }, { "epoch": 2.5272727272727273, "grad_norm": 2.1609091758728027, "learning_rate": 8.030303030303031e-05, "loss": 1.133, "step": 556 }, { "epoch": 2.5318181818181817, "grad_norm": 2.534611225128174, "learning_rate": 7.954545454545455e-05, "loss": 1.3566, "step": 557 }, { "epoch": 2.536363636363636, "grad_norm": 2.731877088546753, "learning_rate": 7.878787878787879e-05, "loss": 1.1991, "step": 558 }, { "epoch": 2.540909090909091, "grad_norm": 2.1953938007354736, "learning_rate": 7.803030303030304e-05, "loss": 0.932, "step": 559 }, { "epoch": 2.5454545454545454, "grad_norm": 2.260007381439209, "learning_rate": 7.727272727272727e-05, "loss": 1.0682, "step": 560 }, { "epoch": 2.55, "grad_norm": 2.9932045936584473, "learning_rate": 7.651515151515151e-05, "loss": 1.2489, "step": 561 }, { "epoch": 2.5545454545454547, "grad_norm": 2.4135005474090576, "learning_rate": 7.575757575757576e-05, "loss": 0.7289, "step": 562 }, { "epoch": 2.559090909090909, "grad_norm": 2.2235300540924072, "learning_rate": 7.5e-05, "loss": 0.7027, "step": 563 }, { "epoch": 2.5636363636363635, "grad_norm": 2.6621127128601074, "learning_rate": 7.424242424242426e-05, "loss": 1.2601, "step": 564 }, { "epoch": 2.5681818181818183, "grad_norm": 2.574686050415039, "learning_rate": 7.348484848484849e-05, "loss": 1.1076, "step": 565 }, { "epoch": 2.5727272727272728, "grad_norm": 2.416339874267578, "learning_rate": 7.272727272727273e-05, "loss": 0.9473, "step": 566 }, { "epoch": 2.577272727272727, "grad_norm": 1.7082793712615967, "learning_rate": 7.196969696969698e-05, "loss": 0.7671, "step": 567 }, { "epoch": 2.581818181818182, "grad_norm": 2.220196008682251, "learning_rate": 7.121212121212121e-05, "loss": 1.1754, "step": 568 }, { "epoch": 2.5863636363636364, "grad_norm": 2.26267409324646, "learning_rate": 7.045454545454545e-05, "loss": 1.4229, "step": 569 }, { "epoch": 2.590909090909091, "grad_norm": 1.7881556749343872, "learning_rate": 6.96969696969697e-05, "loss": 0.8333, "step": 570 }, { "epoch": 2.5954545454545457, "grad_norm": 2.156179904937744, "learning_rate": 6.893939393939393e-05, "loss": 1.1788, "step": 571 }, { "epoch": 2.6, "grad_norm": 1.6093627214431763, "learning_rate": 6.818181818181818e-05, "loss": 0.6442, "step": 572 }, { "epoch": 2.6045454545454545, "grad_norm": 1.984737753868103, "learning_rate": 6.742424242424243e-05, "loss": 0.9969, "step": 573 }, { "epoch": 2.6090909090909093, "grad_norm": 1.958917498588562, "learning_rate": 6.666666666666667e-05, "loss": 0.8534, "step": 574 }, { "epoch": 2.6136363636363638, "grad_norm": 3.060192346572876, "learning_rate": 6.59090909090909e-05, "loss": 1.4748, "step": 575 }, { "epoch": 2.618181818181818, "grad_norm": 1.9157240390777588, "learning_rate": 6.515151515151516e-05, "loss": 0.8512, "step": 576 }, { "epoch": 2.6227272727272726, "grad_norm": 1.6468448638916016, "learning_rate": 6.439393939393939e-05, "loss": 0.7536, "step": 577 }, { "epoch": 2.6272727272727274, "grad_norm": 1.9406344890594482, "learning_rate": 6.363636363636363e-05, "loss": 0.9798, "step": 578 }, { "epoch": 2.631818181818182, "grad_norm": 1.8992547988891602, "learning_rate": 6.287878787878788e-05, "loss": 1.1394, "step": 579 }, { "epoch": 2.6363636363636362, "grad_norm": 1.8168598413467407, "learning_rate": 6.212121212121213e-05, "loss": 0.8848, "step": 580 }, { "epoch": 2.6409090909090907, "grad_norm": 2.8009986877441406, "learning_rate": 6.136363636363636e-05, "loss": 1.1817, "step": 581 }, { "epoch": 2.6454545454545455, "grad_norm": 1.8650470972061157, "learning_rate": 6.060606060606061e-05, "loss": 0.9148, "step": 582 }, { "epoch": 2.65, "grad_norm": 2.132161855697632, "learning_rate": 5.984848484848485e-05, "loss": 1.0103, "step": 583 }, { "epoch": 2.6545454545454543, "grad_norm": 2.488576650619507, "learning_rate": 5.909090909090909e-05, "loss": 1.5804, "step": 584 }, { "epoch": 2.659090909090909, "grad_norm": 1.7953377962112427, "learning_rate": 5.833333333333333e-05, "loss": 0.8419, "step": 585 }, { "epoch": 2.6636363636363636, "grad_norm": 2.563900947570801, "learning_rate": 5.757575757575758e-05, "loss": 1.1122, "step": 586 }, { "epoch": 2.668181818181818, "grad_norm": 2.112504243850708, "learning_rate": 5.681818181818182e-05, "loss": 0.8345, "step": 587 }, { "epoch": 2.672727272727273, "grad_norm": 2.874629020690918, "learning_rate": 5.606060606060606e-05, "loss": 1.257, "step": 588 }, { "epoch": 2.6772727272727272, "grad_norm": 2.3965139389038086, "learning_rate": 5.5303030303030304e-05, "loss": 1.4174, "step": 589 }, { "epoch": 2.6818181818181817, "grad_norm": 2.149787425994873, "learning_rate": 5.4545454545454546e-05, "loss": 1.0162, "step": 590 }, { "epoch": 2.6863636363636365, "grad_norm": 3.67689847946167, "learning_rate": 5.378787878787879e-05, "loss": 0.9925, "step": 591 }, { "epoch": 2.690909090909091, "grad_norm": 2.144545316696167, "learning_rate": 5.303030303030303e-05, "loss": 1.2257, "step": 592 }, { "epoch": 2.6954545454545453, "grad_norm": 1.9149094820022583, "learning_rate": 5.227272727272727e-05, "loss": 0.7236, "step": 593 }, { "epoch": 2.7, "grad_norm": 2.803966999053955, "learning_rate": 5.151515151515152e-05, "loss": 1.1317, "step": 594 }, { "epoch": 2.7045454545454546, "grad_norm": 2.1107089519500732, "learning_rate": 5.075757575757576e-05, "loss": 0.9265, "step": 595 }, { "epoch": 2.709090909090909, "grad_norm": 2.037118911743164, "learning_rate": 5e-05, "loss": 0.6859, "step": 596 }, { "epoch": 2.713636363636364, "grad_norm": 2.310952663421631, "learning_rate": 4.924242424242424e-05, "loss": 0.98, "step": 597 }, { "epoch": 2.7181818181818183, "grad_norm": 1.9925788640975952, "learning_rate": 4.848484848484849e-05, "loss": 0.8919, "step": 598 }, { "epoch": 2.7227272727272727, "grad_norm": 2.466705083847046, "learning_rate": 4.772727272727273e-05, "loss": 1.3115, "step": 599 }, { "epoch": 2.7272727272727275, "grad_norm": 2.8948001861572266, "learning_rate": 4.6969696969696966e-05, "loss": 1.4843, "step": 600 }, { "epoch": 2.731818181818182, "grad_norm": 1.8009178638458252, "learning_rate": 4.621212121212121e-05, "loss": 0.8387, "step": 601 }, { "epoch": 2.7363636363636363, "grad_norm": 1.7695908546447754, "learning_rate": 4.545454545454546e-05, "loss": 0.6376, "step": 602 }, { "epoch": 2.740909090909091, "grad_norm": 2.255938768386841, "learning_rate": 4.46969696969697e-05, "loss": 0.9277, "step": 603 }, { "epoch": 2.7454545454545456, "grad_norm": 2.6216013431549072, "learning_rate": 4.393939393939394e-05, "loss": 0.8539, "step": 604 }, { "epoch": 2.75, "grad_norm": 2.33111310005188, "learning_rate": 4.318181818181818e-05, "loss": 0.766, "step": 605 }, { "epoch": 2.7545454545454544, "grad_norm": 2.256770610809326, "learning_rate": 4.242424242424243e-05, "loss": 0.9658, "step": 606 }, { "epoch": 2.759090909090909, "grad_norm": 2.4762847423553467, "learning_rate": 4.1666666666666665e-05, "loss": 0.8902, "step": 607 }, { "epoch": 2.7636363636363637, "grad_norm": 1.8913813829421997, "learning_rate": 4.090909090909091e-05, "loss": 0.6938, "step": 608 }, { "epoch": 2.768181818181818, "grad_norm": 1.928743839263916, "learning_rate": 4.015151515151516e-05, "loss": 0.637, "step": 609 }, { "epoch": 2.7727272727272725, "grad_norm": 3.3095438480377197, "learning_rate": 3.939393939393939e-05, "loss": 0.9913, "step": 610 }, { "epoch": 2.7772727272727273, "grad_norm": 2.611701011657715, "learning_rate": 3.8636363636363636e-05, "loss": 1.1573, "step": 611 }, { "epoch": 2.7818181818181817, "grad_norm": 2.022073745727539, "learning_rate": 3.787878787878788e-05, "loss": 0.9537, "step": 612 }, { "epoch": 2.786363636363636, "grad_norm": 2.9264447689056396, "learning_rate": 3.712121212121213e-05, "loss": 1.4692, "step": 613 }, { "epoch": 2.790909090909091, "grad_norm": 2.22469425201416, "learning_rate": 3.6363636363636364e-05, "loss": 1.0733, "step": 614 }, { "epoch": 2.7954545454545454, "grad_norm": 2.8329367637634277, "learning_rate": 3.560606060606061e-05, "loss": 1.5153, "step": 615 }, { "epoch": 2.8, "grad_norm": 1.8949047327041626, "learning_rate": 3.484848484848485e-05, "loss": 0.6316, "step": 616 }, { "epoch": 2.8045454545454547, "grad_norm": 2.597440242767334, "learning_rate": 3.409090909090909e-05, "loss": 1.6049, "step": 617 }, { "epoch": 2.809090909090909, "grad_norm": 2.0482330322265625, "learning_rate": 3.3333333333333335e-05, "loss": 1.0083, "step": 618 }, { "epoch": 2.8136363636363635, "grad_norm": 1.7359944581985474, "learning_rate": 3.257575757575758e-05, "loss": 0.9097, "step": 619 }, { "epoch": 2.8181818181818183, "grad_norm": 1.9825539588928223, "learning_rate": 3.1818181818181814e-05, "loss": 0.7107, "step": 620 }, { "epoch": 2.8227272727272728, "grad_norm": 1.7564197778701782, "learning_rate": 3.106060606060606e-05, "loss": 0.9089, "step": 621 }, { "epoch": 2.827272727272727, "grad_norm": 2.735137939453125, "learning_rate": 3.0303030303030306e-05, "loss": 1.4978, "step": 622 }, { "epoch": 2.831818181818182, "grad_norm": 2.702873706817627, "learning_rate": 2.9545454545454545e-05, "loss": 1.2556, "step": 623 }, { "epoch": 2.8363636363636364, "grad_norm": 1.9755101203918457, "learning_rate": 2.878787878787879e-05, "loss": 0.8022, "step": 624 }, { "epoch": 2.840909090909091, "grad_norm": 2.0104050636291504, "learning_rate": 2.803030303030303e-05, "loss": 0.8993, "step": 625 }, { "epoch": 2.8454545454545457, "grad_norm": 2.915588855743408, "learning_rate": 2.7272727272727273e-05, "loss": 1.7374, "step": 626 }, { "epoch": 2.85, "grad_norm": 2.5364370346069336, "learning_rate": 2.6515151515151516e-05, "loss": 0.7967, "step": 627 }, { "epoch": 2.8545454545454545, "grad_norm": 2.731673002243042, "learning_rate": 2.575757575757576e-05, "loss": 1.5847, "step": 628 }, { "epoch": 2.8590909090909093, "grad_norm": 2.1468403339385986, "learning_rate": 2.5e-05, "loss": 1.0101, "step": 629 }, { "epoch": 2.8636363636363638, "grad_norm": 1.8162040710449219, "learning_rate": 2.4242424242424244e-05, "loss": 0.6688, "step": 630 }, { "epoch": 2.868181818181818, "grad_norm": 2.285930871963501, "learning_rate": 2.3484848484848483e-05, "loss": 1.1906, "step": 631 }, { "epoch": 2.8727272727272726, "grad_norm": 2.0448861122131348, "learning_rate": 2.272727272727273e-05, "loss": 0.8868, "step": 632 }, { "epoch": 2.8772727272727274, "grad_norm": 2.98807692527771, "learning_rate": 2.196969696969697e-05, "loss": 1.1679, "step": 633 }, { "epoch": 2.881818181818182, "grad_norm": 1.9618700742721558, "learning_rate": 2.1212121212121215e-05, "loss": 1.0659, "step": 634 }, { "epoch": 2.8863636363636362, "grad_norm": 2.200741767883301, "learning_rate": 2.0454545454545454e-05, "loss": 1.0582, "step": 635 }, { "epoch": 2.8909090909090907, "grad_norm": 2.1953506469726562, "learning_rate": 1.9696969696969697e-05, "loss": 1.4356, "step": 636 }, { "epoch": 2.8954545454545455, "grad_norm": 2.1912357807159424, "learning_rate": 1.893939393939394e-05, "loss": 1.0532, "step": 637 }, { "epoch": 2.9, "grad_norm": 2.125601053237915, "learning_rate": 1.8181818181818182e-05, "loss": 1.458, "step": 638 }, { "epoch": 2.9045454545454543, "grad_norm": 1.9430787563323975, "learning_rate": 1.7424242424242425e-05, "loss": 0.8313, "step": 639 }, { "epoch": 2.909090909090909, "grad_norm": 1.7255802154541016, "learning_rate": 1.6666666666666667e-05, "loss": 0.7894, "step": 640 }, { "epoch": 2.9136363636363636, "grad_norm": 2.259798526763916, "learning_rate": 1.5909090909090907e-05, "loss": 1.0942, "step": 641 }, { "epoch": 2.918181818181818, "grad_norm": 2.4443533420562744, "learning_rate": 1.5151515151515153e-05, "loss": 1.5392, "step": 642 }, { "epoch": 2.922727272727273, "grad_norm": 2.437310218811035, "learning_rate": 1.4393939393939396e-05, "loss": 0.9475, "step": 643 }, { "epoch": 2.9272727272727272, "grad_norm": 2.1248443126678467, "learning_rate": 1.3636363636363637e-05, "loss": 0.6917, "step": 644 }, { "epoch": 2.9318181818181817, "grad_norm": 2.0161659717559814, "learning_rate": 1.287878787878788e-05, "loss": 1.0465, "step": 645 }, { "epoch": 2.9363636363636365, "grad_norm": 1.825695514678955, "learning_rate": 1.2121212121212122e-05, "loss": 0.8001, "step": 646 }, { "epoch": 2.940909090909091, "grad_norm": 2.58219575881958, "learning_rate": 1.1363636363636365e-05, "loss": 0.9649, "step": 647 }, { "epoch": 2.9454545454545453, "grad_norm": 1.9554407596588135, "learning_rate": 1.0606060606060607e-05, "loss": 1.1447, "step": 648 }, { "epoch": 2.95, "grad_norm": 2.89900541305542, "learning_rate": 9.848484848484848e-06, "loss": 0.9461, "step": 649 }, { "epoch": 2.9545454545454546, "grad_norm": 1.8475868701934814, "learning_rate": 9.090909090909091e-06, "loss": 1.1863, "step": 650 }, { "epoch": 2.959090909090909, "grad_norm": 2.264302968978882, "learning_rate": 8.333333333333334e-06, "loss": 0.9459, "step": 651 }, { "epoch": 2.963636363636364, "grad_norm": 2.157198190689087, "learning_rate": 7.5757575757575764e-06, "loss": 1.4461, "step": 652 }, { "epoch": 2.9681818181818183, "grad_norm": 2.3027210235595703, "learning_rate": 6.818181818181818e-06, "loss": 0.8407, "step": 653 }, { "epoch": 2.9727272727272727, "grad_norm": 1.786800503730774, "learning_rate": 6.060606060606061e-06, "loss": 0.7051, "step": 654 }, { "epoch": 2.9772727272727275, "grad_norm": 2.4173872470855713, "learning_rate": 5.303030303030304e-06, "loss": 1.0276, "step": 655 }, { "epoch": 2.981818181818182, "grad_norm": 3.755701780319214, "learning_rate": 4.5454545454545455e-06, "loss": 0.9572, "step": 656 }, { "epoch": 2.9863636363636363, "grad_norm": 2.0097804069519043, "learning_rate": 3.7878787878787882e-06, "loss": 0.9736, "step": 657 }, { "epoch": 2.990909090909091, "grad_norm": 1.773881196975708, "learning_rate": 3.0303030303030305e-06, "loss": 0.8404, "step": 658 }, { "epoch": 2.9954545454545456, "grad_norm": 2.139065980911255, "learning_rate": 2.2727272727272728e-06, "loss": 1.1301, "step": 659 }, { "epoch": 3.0, "grad_norm": 1.8846311569213867, "learning_rate": 1.5151515151515152e-06, "loss": 0.7227, "step": 660 }, { "epoch": 3.0, "eval_f1": 0.8952, "eval_gen_len": 41.8455, "eval_loss": 1.8697103261947632, "eval_precision": 0.8934, "eval_recall": 0.8971, "eval_rouge1": 0.4709, "eval_rouge2": 0.2223, "eval_rougeL": 0.3999, "eval_rougeLsum": 0.4391, "eval_runtime": 25.3276, "eval_samples_per_second": 4.343, "eval_steps_per_second": 0.553, "step": 660 }, { "epoch": 3.0, "step": 660, "total_flos": 2484005840363520.0, "train_loss": 1.4827006761774872, "train_runtime": 507.3071, "train_samples_per_second": 5.198, "train_steps_per_second": 1.301 } ], "logging_steps": 1, "max_steps": 660, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2484005840363520.0, "train_batch_size": 4, "trial_name": null, "trial_params": null }