|
{ |
|
"best_metric": 0.4325978755950928, |
|
"best_model_checkpoint": "miner_id_24/checkpoint-200", |
|
"epoch": 0.007839909057054938, |
|
"eval_steps": 50, |
|
"global_step": 200, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 3.919954528527469e-05, |
|
"grad_norm": 0.16042552888393402, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4326, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 3.919954528527469e-05, |
|
"eval_loss": 0.5350391268730164, |
|
"eval_runtime": 2393.8126, |
|
"eval_samples_per_second": 17.948, |
|
"eval_steps_per_second": 8.974, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 7.839909057054938e-05, |
|
"grad_norm": 0.15639245510101318, |
|
"learning_rate": 1e-05, |
|
"loss": 0.5717, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.00011759863585582407, |
|
"grad_norm": 0.14500007033348083, |
|
"learning_rate": 1.5e-05, |
|
"loss": 0.5217, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.00015679818114109877, |
|
"grad_norm": 0.13646380603313446, |
|
"learning_rate": 2e-05, |
|
"loss": 0.4166, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.00019599772642637345, |
|
"grad_norm": 0.14864464104175568, |
|
"learning_rate": 2.5e-05, |
|
"loss": 0.4906, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.00023519727171164814, |
|
"grad_norm": 0.15161864459514618, |
|
"learning_rate": 3e-05, |
|
"loss": 0.568, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.00027439681699692285, |
|
"grad_norm": 0.14943252503871918, |
|
"learning_rate": 3.5e-05, |
|
"loss": 0.597, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.00031359636228219754, |
|
"grad_norm": 0.14887230098247528, |
|
"learning_rate": 4e-05, |
|
"loss": 0.4819, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.0003527959075674722, |
|
"grad_norm": 0.18022413551807404, |
|
"learning_rate": 4.5e-05, |
|
"loss": 0.5437, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.0003919954528527469, |
|
"grad_norm": 0.18863749504089355, |
|
"learning_rate": 5e-05, |
|
"loss": 0.5148, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.0004311949981380216, |
|
"grad_norm": 0.19053728878498077, |
|
"learning_rate": 5.500000000000001e-05, |
|
"loss": 0.6177, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.0004703945434232963, |
|
"grad_norm": 0.16379624605178833, |
|
"learning_rate": 6e-05, |
|
"loss": 0.6024, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.000509594088708571, |
|
"grad_norm": 0.17053860425949097, |
|
"learning_rate": 6.500000000000001e-05, |
|
"loss": 0.5195, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.0005487936339938457, |
|
"grad_norm": 0.1678755134344101, |
|
"learning_rate": 7e-05, |
|
"loss": 0.5825, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.0005879931792791204, |
|
"grad_norm": 0.20500046014785767, |
|
"learning_rate": 7.500000000000001e-05, |
|
"loss": 0.6271, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.0006271927245643951, |
|
"grad_norm": 0.17075373232364655, |
|
"learning_rate": 8e-05, |
|
"loss": 0.5008, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.0006663922698496698, |
|
"grad_norm": 0.15935742855072021, |
|
"learning_rate": 8.5e-05, |
|
"loss": 0.4639, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.0007055918151349444, |
|
"grad_norm": 0.182022824883461, |
|
"learning_rate": 9e-05, |
|
"loss": 0.4297, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.0007447913604202191, |
|
"grad_norm": 0.16236171126365662, |
|
"learning_rate": 9.5e-05, |
|
"loss": 0.4011, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.0007839909057054938, |
|
"grad_norm": 0.18881011009216309, |
|
"learning_rate": 0.0001, |
|
"loss": 0.5249, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.0008231904509907685, |
|
"grad_norm": 0.18972285091876984, |
|
"learning_rate": 9.999238475781957e-05, |
|
"loss": 0.4455, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.0008623899962760432, |
|
"grad_norm": 0.18663617968559265, |
|
"learning_rate": 9.99695413509548e-05, |
|
"loss": 0.4337, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.0009015895415613179, |
|
"grad_norm": 0.1874244064092636, |
|
"learning_rate": 9.99314767377287e-05, |
|
"loss": 0.4353, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.0009407890868465926, |
|
"grad_norm": 0.1862955540418625, |
|
"learning_rate": 9.987820251299122e-05, |
|
"loss": 0.5082, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.0009799886321318672, |
|
"grad_norm": 0.17161045968532562, |
|
"learning_rate": 9.980973490458728e-05, |
|
"loss": 0.4783, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.001019188177417142, |
|
"grad_norm": 0.17733538150787354, |
|
"learning_rate": 9.972609476841367e-05, |
|
"loss": 0.4794, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.0010583877227024166, |
|
"grad_norm": 0.17916053533554077, |
|
"learning_rate": 9.962730758206611e-05, |
|
"loss": 0.4799, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.0010975872679876914, |
|
"grad_norm": 0.17826522886753082, |
|
"learning_rate": 9.951340343707852e-05, |
|
"loss": 0.4, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.001136786813272966, |
|
"grad_norm": 0.19639909267425537, |
|
"learning_rate": 9.938441702975689e-05, |
|
"loss": 0.372, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.0011759863585582408, |
|
"grad_norm": 0.19923311471939087, |
|
"learning_rate": 9.924038765061042e-05, |
|
"loss": 0.423, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.0012151859038435154, |
|
"grad_norm": 0.20051322877407074, |
|
"learning_rate": 9.908135917238321e-05, |
|
"loss": 0.4233, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.0012543854491287902, |
|
"grad_norm": 0.20958496630191803, |
|
"learning_rate": 9.890738003669029e-05, |
|
"loss": 0.4288, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.0012935849944140647, |
|
"grad_norm": 0.20028801262378693, |
|
"learning_rate": 9.871850323926177e-05, |
|
"loss": 0.3726, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.0013327845396993395, |
|
"grad_norm": 0.20952534675598145, |
|
"learning_rate": 9.851478631379982e-05, |
|
"loss": 0.5309, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.001371984084984614, |
|
"grad_norm": 0.21508319675922394, |
|
"learning_rate": 9.829629131445342e-05, |
|
"loss": 0.4658, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.001411183630269889, |
|
"grad_norm": 0.2109171450138092, |
|
"learning_rate": 9.806308479691595e-05, |
|
"loss": 0.4026, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.0014503831755551635, |
|
"grad_norm": 0.2147839516401291, |
|
"learning_rate": 9.781523779815179e-05, |
|
"loss": 0.3535, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.0014895827208404383, |
|
"grad_norm": 0.20986714959144592, |
|
"learning_rate": 9.755282581475769e-05, |
|
"loss": 0.3854, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.0015287822661257128, |
|
"grad_norm": 0.22886401414871216, |
|
"learning_rate": 9.727592877996585e-05, |
|
"loss": 0.3524, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.0015679818114109876, |
|
"grad_norm": 0.2274533212184906, |
|
"learning_rate": 9.698463103929542e-05, |
|
"loss": 0.3901, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.0016071813566962624, |
|
"grad_norm": 0.2675640285015106, |
|
"learning_rate": 9.667902132486009e-05, |
|
"loss": 0.3687, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.001646380901981537, |
|
"grad_norm": 0.23128102719783783, |
|
"learning_rate": 9.635919272833938e-05, |
|
"loss": 0.4233, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.0016855804472668118, |
|
"grad_norm": 0.22661566734313965, |
|
"learning_rate": 9.602524267262203e-05, |
|
"loss": 0.4152, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.0017247799925520864, |
|
"grad_norm": 0.2415810078382492, |
|
"learning_rate": 9.567727288213005e-05, |
|
"loss": 0.3817, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.0017639795378373612, |
|
"grad_norm": 0.2653232216835022, |
|
"learning_rate": 9.53153893518325e-05, |
|
"loss": 0.4235, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.0018031790831226358, |
|
"grad_norm": 0.2872054874897003, |
|
"learning_rate": 9.493970231495835e-05, |
|
"loss": 0.4208, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.0018423786284079105, |
|
"grad_norm": 0.2797979712486267, |
|
"learning_rate": 9.45503262094184e-05, |
|
"loss": 0.4168, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.0018815781736931851, |
|
"grad_norm": 0.2640951871871948, |
|
"learning_rate": 9.414737964294636e-05, |
|
"loss": 0.3265, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.00192077771897846, |
|
"grad_norm": 0.33999350666999817, |
|
"learning_rate": 9.373098535696979e-05, |
|
"loss": 0.3443, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.0019599772642637345, |
|
"grad_norm": 0.49079790711402893, |
|
"learning_rate": 9.330127018922194e-05, |
|
"loss": 0.3703, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.0019599772642637345, |
|
"eval_loss": 0.44784459471702576, |
|
"eval_runtime": 2398.5451, |
|
"eval_samples_per_second": 17.913, |
|
"eval_steps_per_second": 8.957, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.001999176809549009, |
|
"grad_norm": 0.1652698665857315, |
|
"learning_rate": 9.285836503510562e-05, |
|
"loss": 0.4374, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.002038376354834284, |
|
"grad_norm": 0.15927405655384064, |
|
"learning_rate": 9.24024048078213e-05, |
|
"loss": 0.5079, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.0020775759001195587, |
|
"grad_norm": 0.13452063500881195, |
|
"learning_rate": 9.193352839727121e-05, |
|
"loss": 0.4709, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.0021167754454048332, |
|
"grad_norm": 0.12576071918010712, |
|
"learning_rate": 9.145187862775209e-05, |
|
"loss": 0.5341, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.002155974990690108, |
|
"grad_norm": 0.11464632302522659, |
|
"learning_rate": 9.09576022144496e-05, |
|
"loss": 0.4928, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.002195174535975383, |
|
"grad_norm": 0.12085549533367157, |
|
"learning_rate": 9.045084971874738e-05, |
|
"loss": 0.5811, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.0022343740812606574, |
|
"grad_norm": 0.12558674812316895, |
|
"learning_rate": 8.993177550236464e-05, |
|
"loss": 0.596, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.002273573626545932, |
|
"grad_norm": 0.12457916140556335, |
|
"learning_rate": 8.940053768033609e-05, |
|
"loss": 0.5087, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.0023127731718312066, |
|
"grad_norm": 0.12604419887065887, |
|
"learning_rate": 8.885729807284856e-05, |
|
"loss": 0.5619, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.0023519727171164816, |
|
"grad_norm": 0.1314886063337326, |
|
"learning_rate": 8.83022221559489e-05, |
|
"loss": 0.5879, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.002391172262401756, |
|
"grad_norm": 0.12995365262031555, |
|
"learning_rate": 8.773547901113862e-05, |
|
"loss": 0.4307, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.0024303718076870307, |
|
"grad_norm": 0.13908788561820984, |
|
"learning_rate": 8.715724127386972e-05, |
|
"loss": 0.538, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.0024695713529723057, |
|
"grad_norm": 0.13697293400764465, |
|
"learning_rate": 8.656768508095853e-05, |
|
"loss": 0.4955, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.0025087708982575803, |
|
"grad_norm": 0.12941715121269226, |
|
"learning_rate": 8.596699001693255e-05, |
|
"loss": 0.5089, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.002547970443542855, |
|
"grad_norm": 0.1320359855890274, |
|
"learning_rate": 8.535533905932738e-05, |
|
"loss": 0.4805, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.0025871699888281295, |
|
"grad_norm": 0.13320119678974152, |
|
"learning_rate": 8.473291852294987e-05, |
|
"loss": 0.4316, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.0026263695341134045, |
|
"grad_norm": 0.13988560438156128, |
|
"learning_rate": 8.409991800312493e-05, |
|
"loss": 0.5421, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.002665569079398679, |
|
"grad_norm": 0.13911867141723633, |
|
"learning_rate": 8.345653031794292e-05, |
|
"loss": 0.4276, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.0027047686246839536, |
|
"grad_norm": 0.14803774654865265, |
|
"learning_rate": 8.280295144952536e-05, |
|
"loss": 0.3929, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.002743968169969228, |
|
"grad_norm": 0.14860422909259796, |
|
"learning_rate": 8.213938048432697e-05, |
|
"loss": 0.4329, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.002783167715254503, |
|
"grad_norm": 0.1481442004442215, |
|
"learning_rate": 8.146601955249188e-05, |
|
"loss": 0.4106, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.002822367260539778, |
|
"grad_norm": 0.1563064306974411, |
|
"learning_rate": 8.07830737662829e-05, |
|
"loss": 0.482, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.0028615668058250524, |
|
"grad_norm": 0.146785169839859, |
|
"learning_rate": 8.009075115760243e-05, |
|
"loss": 0.4506, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.002900766351110327, |
|
"grad_norm": 0.15569761395454407, |
|
"learning_rate": 7.938926261462366e-05, |
|
"loss": 0.4138, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.002939965896395602, |
|
"grad_norm": 0.16449235379695892, |
|
"learning_rate": 7.86788218175523e-05, |
|
"loss": 0.5195, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.0029791654416808765, |
|
"grad_norm": 0.15461541712284088, |
|
"learning_rate": 7.795964517353735e-05, |
|
"loss": 0.4168, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.003018364986966151, |
|
"grad_norm": 0.16411833465099335, |
|
"learning_rate": 7.723195175075136e-05, |
|
"loss": 0.4094, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.0030575645322514257, |
|
"grad_norm": 0.1622595489025116, |
|
"learning_rate": 7.649596321166024e-05, |
|
"loss": 0.4647, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.0030967640775367007, |
|
"grad_norm": 0.16740792989730835, |
|
"learning_rate": 7.575190374550272e-05, |
|
"loss": 0.4236, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.0031359636228219753, |
|
"grad_norm": 0.17070157825946808, |
|
"learning_rate": 7.500000000000001e-05, |
|
"loss": 0.4204, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.00317516316810725, |
|
"grad_norm": 0.1682911217212677, |
|
"learning_rate": 7.424048101231686e-05, |
|
"loss": 0.3919, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.003214362713392525, |
|
"grad_norm": 0.17516714334487915, |
|
"learning_rate": 7.347357813929454e-05, |
|
"loss": 0.4194, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.0032535622586777994, |
|
"grad_norm": 0.17585694789886475, |
|
"learning_rate": 7.269952498697734e-05, |
|
"loss": 0.3488, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.003292761803963074, |
|
"grad_norm": 0.16620786488056183, |
|
"learning_rate": 7.191855733945387e-05, |
|
"loss": 0.3378, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.0033319613492483486, |
|
"grad_norm": 0.17664961516857147, |
|
"learning_rate": 7.113091308703498e-05, |
|
"loss": 0.3861, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.0033711608945336236, |
|
"grad_norm": 0.1843758374452591, |
|
"learning_rate": 7.033683215379002e-05, |
|
"loss": 0.3457, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.003410360439818898, |
|
"grad_norm": 0.19849351048469543, |
|
"learning_rate": 6.953655642446368e-05, |
|
"loss": 0.3744, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.0034495599851041728, |
|
"grad_norm": 0.1878618746995926, |
|
"learning_rate": 6.873032967079561e-05, |
|
"loss": 0.3847, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.0034887595303894473, |
|
"grad_norm": 0.19447298347949982, |
|
"learning_rate": 6.7918397477265e-05, |
|
"loss": 0.425, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.0035279590756747223, |
|
"grad_norm": 0.1870015263557434, |
|
"learning_rate": 6.710100716628344e-05, |
|
"loss": 0.3633, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.003567158620959997, |
|
"grad_norm": 0.19948124885559082, |
|
"learning_rate": 6.627840772285784e-05, |
|
"loss": 0.3673, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.0036063581662452715, |
|
"grad_norm": 0.23523126542568207, |
|
"learning_rate": 6.545084971874738e-05, |
|
"loss": 0.4524, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.003645557711530546, |
|
"grad_norm": 0.21212515234947205, |
|
"learning_rate": 6.461858523613684e-05, |
|
"loss": 0.3936, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.003684757256815821, |
|
"grad_norm": 0.2363279014825821, |
|
"learning_rate": 6.378186779084995e-05, |
|
"loss": 0.4356, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.0037239568021010957, |
|
"grad_norm": 0.21925963461399078, |
|
"learning_rate": 6.294095225512603e-05, |
|
"loss": 0.3833, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.0037631563473863702, |
|
"grad_norm": 0.20751087367534637, |
|
"learning_rate": 6.209609477998338e-05, |
|
"loss": 0.336, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.003802355892671645, |
|
"grad_norm": 0.2536580264568329, |
|
"learning_rate": 6.124755271719325e-05, |
|
"loss": 0.4993, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.00384155543795692, |
|
"grad_norm": 0.25841864943504333, |
|
"learning_rate": 6.0395584540887963e-05, |
|
"loss": 0.3797, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.0038807549832421944, |
|
"grad_norm": 0.2994214594364166, |
|
"learning_rate": 5.9540449768827246e-05, |
|
"loss": 0.332, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.003919954528527469, |
|
"grad_norm": 0.3704180121421814, |
|
"learning_rate": 5.868240888334653e-05, |
|
"loss": 0.4043, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.003919954528527469, |
|
"eval_loss": 0.43801045417785645, |
|
"eval_runtime": 2400.0319, |
|
"eval_samples_per_second": 17.902, |
|
"eval_steps_per_second": 8.951, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.0039591540738127436, |
|
"grad_norm": 0.10259481519460678, |
|
"learning_rate": 5.782172325201155e-05, |
|
"loss": 0.4114, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.003998353619098018, |
|
"grad_norm": 0.1059873104095459, |
|
"learning_rate": 5.695865504800327e-05, |
|
"loss": 0.4391, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.004037553164383294, |
|
"grad_norm": 0.10306891053915024, |
|
"learning_rate": 5.6093467170257374e-05, |
|
"loss": 0.4589, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.004076752709668568, |
|
"grad_norm": 0.09604241698980331, |
|
"learning_rate": 5.522642316338268e-05, |
|
"loss": 0.4462, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.004115952254953843, |
|
"grad_norm": 0.1092570349574089, |
|
"learning_rate": 5.435778713738292e-05, |
|
"loss": 0.4617, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.004155151800239117, |
|
"grad_norm": 0.10691061615943909, |
|
"learning_rate": 5.348782368720626e-05, |
|
"loss": 0.5161, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.004194351345524392, |
|
"grad_norm": 0.10758673399686813, |
|
"learning_rate": 5.26167978121472e-05, |
|
"loss": 0.6515, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.0042335508908096665, |
|
"grad_norm": 0.10809138417243958, |
|
"learning_rate": 5.174497483512506e-05, |
|
"loss": 0.4775, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.004272750436094941, |
|
"grad_norm": 0.10625655204057693, |
|
"learning_rate": 5.0872620321864185e-05, |
|
"loss": 0.534, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.004311949981380216, |
|
"grad_norm": 0.11111942678689957, |
|
"learning_rate": 5e-05, |
|
"loss": 0.5024, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.004351149526665491, |
|
"grad_norm": 0.10645933449268341, |
|
"learning_rate": 4.912737967813583e-05, |
|
"loss": 0.459, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.004390349071950766, |
|
"grad_norm": 0.11491873860359192, |
|
"learning_rate": 4.825502516487497e-05, |
|
"loss": 0.6275, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.00442954861723604, |
|
"grad_norm": 0.11639122664928436, |
|
"learning_rate": 4.738320218785281e-05, |
|
"loss": 0.5297, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.004468748162521315, |
|
"grad_norm": 0.12297159433364868, |
|
"learning_rate": 4.6512176312793736e-05, |
|
"loss": 0.4937, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.004507947707806589, |
|
"grad_norm": 0.11242052167654037, |
|
"learning_rate": 4.564221286261709e-05, |
|
"loss": 0.4692, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.004547147253091864, |
|
"grad_norm": 0.1251300573348999, |
|
"learning_rate": 4.477357683661734e-05, |
|
"loss": 0.5091, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.0045863467983771385, |
|
"grad_norm": 0.12641409039497375, |
|
"learning_rate": 4.390653282974264e-05, |
|
"loss": 0.539, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.004625546343662413, |
|
"grad_norm": 0.1277226060628891, |
|
"learning_rate": 4.3041344951996746e-05, |
|
"loss": 0.5769, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.0046647458889476886, |
|
"grad_norm": 0.12659917771816254, |
|
"learning_rate": 4.2178276747988446e-05, |
|
"loss": 0.3997, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.004703945434232963, |
|
"grad_norm": 0.1406412273645401, |
|
"learning_rate": 4.131759111665349e-05, |
|
"loss": 0.4552, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.004743144979518238, |
|
"grad_norm": 0.1345863938331604, |
|
"learning_rate": 4.045955023117276e-05, |
|
"loss": 0.4727, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.004782344524803512, |
|
"grad_norm": 0.13614259660243988, |
|
"learning_rate": 3.960441545911204e-05, |
|
"loss": 0.4239, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.004821544070088787, |
|
"grad_norm": 0.13695776462554932, |
|
"learning_rate": 3.875244728280676e-05, |
|
"loss": 0.3773, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.0048607436153740614, |
|
"grad_norm": 0.14491674304008484, |
|
"learning_rate": 3.790390522001662e-05, |
|
"loss": 0.39, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.004899943160659336, |
|
"grad_norm": 0.14301298558712006, |
|
"learning_rate": 3.705904774487396e-05, |
|
"loss": 0.3741, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.0049391427059446115, |
|
"grad_norm": 0.14596934616565704, |
|
"learning_rate": 3.6218132209150045e-05, |
|
"loss": 0.4123, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.004978342251229886, |
|
"grad_norm": 0.12781678140163422, |
|
"learning_rate": 3.5381414763863166e-05, |
|
"loss": 0.3334, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.005017541796515161, |
|
"grad_norm": 0.1531793177127838, |
|
"learning_rate": 3.4549150281252636e-05, |
|
"loss": 0.3903, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.005056741341800435, |
|
"grad_norm": 0.1685943901538849, |
|
"learning_rate": 3.372159227714218e-05, |
|
"loss": 0.514, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.00509594088708571, |
|
"grad_norm": 0.15798459947109222, |
|
"learning_rate": 3.289899283371657e-05, |
|
"loss": 0.4733, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.005135140432370984, |
|
"grad_norm": 0.16000813245773315, |
|
"learning_rate": 3.2081602522734986e-05, |
|
"loss": 0.4515, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.005174339977656259, |
|
"grad_norm": 0.15483549237251282, |
|
"learning_rate": 3.12696703292044e-05, |
|
"loss": 0.3787, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.0052135395229415335, |
|
"grad_norm": 0.18001963198184967, |
|
"learning_rate": 3.046344357553632e-05, |
|
"loss": 0.4393, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.005252739068226809, |
|
"grad_norm": 0.1790371835231781, |
|
"learning_rate": 2.9663167846209998e-05, |
|
"loss": 0.3586, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.0052919386135120835, |
|
"grad_norm": 0.1700044572353363, |
|
"learning_rate": 2.886908691296504e-05, |
|
"loss": 0.4281, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.005331138158797358, |
|
"grad_norm": 0.1701839715242386, |
|
"learning_rate": 2.8081442660546125e-05, |
|
"loss": 0.2861, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.005370337704082633, |
|
"grad_norm": 0.17230786383152008, |
|
"learning_rate": 2.7300475013022663e-05, |
|
"loss": 0.3672, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.005409537249367907, |
|
"grad_norm": 0.19381862878799438, |
|
"learning_rate": 2.6526421860705473e-05, |
|
"loss": 0.4293, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.005448736794653182, |
|
"grad_norm": 0.17442116141319275, |
|
"learning_rate": 2.575951898768315e-05, |
|
"loss": 0.356, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.005487936339938456, |
|
"grad_norm": 0.18775367736816406, |
|
"learning_rate": 2.500000000000001e-05, |
|
"loss": 0.3817, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.005527135885223731, |
|
"grad_norm": 0.20612426102161407, |
|
"learning_rate": 2.4248096254497288e-05, |
|
"loss": 0.433, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.005566335430509006, |
|
"grad_norm": 0.19330035150051117, |
|
"learning_rate": 2.350403678833976e-05, |
|
"loss": 0.4553, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.005605534975794281, |
|
"grad_norm": 0.19368550181388855, |
|
"learning_rate": 2.2768048249248648e-05, |
|
"loss": 0.4106, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.005644734521079556, |
|
"grad_norm": 0.21053840219974518, |
|
"learning_rate": 2.2040354826462668e-05, |
|
"loss": 0.3231, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.00568393406636483, |
|
"grad_norm": 0.2086849808692932, |
|
"learning_rate": 2.132117818244771e-05, |
|
"loss": 0.368, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.005723133611650105, |
|
"grad_norm": 0.20636354386806488, |
|
"learning_rate": 2.061073738537635e-05, |
|
"loss": 0.3489, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.005762333156935379, |
|
"grad_norm": 0.23584240674972534, |
|
"learning_rate": 1.9909248842397584e-05, |
|
"loss": 0.4045, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.005801532702220654, |
|
"grad_norm": 0.2612581253051758, |
|
"learning_rate": 1.9216926233717085e-05, |
|
"loss": 0.39, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.005840732247505929, |
|
"grad_norm": 0.2916366159915924, |
|
"learning_rate": 1.8533980447508137e-05, |
|
"loss": 0.4611, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.005879931792791204, |
|
"grad_norm": 0.429953396320343, |
|
"learning_rate": 1.7860619515673033e-05, |
|
"loss": 0.3297, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.005879931792791204, |
|
"eval_loss": 0.43333882093429565, |
|
"eval_runtime": 2400.7314, |
|
"eval_samples_per_second": 17.897, |
|
"eval_steps_per_second": 8.949, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.0059191313380764785, |
|
"grad_norm": 0.08383464813232422, |
|
"learning_rate": 1.7197048550474643e-05, |
|
"loss": 0.4218, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.005958330883361753, |
|
"grad_norm": 0.08397100120782852, |
|
"learning_rate": 1.6543469682057106e-05, |
|
"loss": 0.4163, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.005997530428647028, |
|
"grad_norm": 0.08535988628864288, |
|
"learning_rate": 1.5900081996875083e-05, |
|
"loss": 0.4975, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.006036729973932302, |
|
"grad_norm": 0.08967166393995285, |
|
"learning_rate": 1.526708147705013e-05, |
|
"loss": 0.4722, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.006075929519217577, |
|
"grad_norm": 0.09084314107894897, |
|
"learning_rate": 1.4644660940672627e-05, |
|
"loss": 0.428, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.006115129064502851, |
|
"grad_norm": 0.10152707993984222, |
|
"learning_rate": 1.4033009983067452e-05, |
|
"loss": 0.5041, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.006154328609788127, |
|
"grad_norm": 0.10302560031414032, |
|
"learning_rate": 1.3432314919041478e-05, |
|
"loss": 0.4389, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.006193528155073401, |
|
"grad_norm": 0.10672393441200256, |
|
"learning_rate": 1.2842758726130283e-05, |
|
"loss": 0.5093, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.006232727700358676, |
|
"grad_norm": 0.11030876636505127, |
|
"learning_rate": 1.22645209888614e-05, |
|
"loss": 0.5054, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.0062719272456439506, |
|
"grad_norm": 0.12015864998102188, |
|
"learning_rate": 1.1697777844051105e-05, |
|
"loss": 0.5693, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.006311126790929225, |
|
"grad_norm": 0.1129220575094223, |
|
"learning_rate": 1.1142701927151456e-05, |
|
"loss": 0.5293, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.0063503263362145, |
|
"grad_norm": 0.11352799832820892, |
|
"learning_rate": 1.0599462319663905e-05, |
|
"loss": 0.4903, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.006389525881499774, |
|
"grad_norm": 0.13408681750297546, |
|
"learning_rate": 1.006822449763537e-05, |
|
"loss": 0.5941, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.00642872542678505, |
|
"grad_norm": 0.12811416387557983, |
|
"learning_rate": 9.549150281252633e-06, |
|
"loss": 0.5844, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.006467924972070324, |
|
"grad_norm": 0.12660935521125793, |
|
"learning_rate": 9.042397785550405e-06, |
|
"loss": 0.4857, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.006507124517355599, |
|
"grad_norm": 0.12374024838209152, |
|
"learning_rate": 8.548121372247918e-06, |
|
"loss": 0.5069, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.0065463240626408735, |
|
"grad_norm": 0.1355339139699936, |
|
"learning_rate": 8.066471602728803e-06, |
|
"loss": 0.616, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.006585523607926148, |
|
"grad_norm": 0.1416226476430893, |
|
"learning_rate": 7.597595192178702e-06, |
|
"loss": 0.5935, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.006624723153211423, |
|
"grad_norm": 0.12920598685741425, |
|
"learning_rate": 7.1416349648943894e-06, |
|
"loss": 0.5472, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.006663922698496697, |
|
"grad_norm": 0.1332990676164627, |
|
"learning_rate": 6.698729810778065e-06, |
|
"loss": 0.3977, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.006703122243781972, |
|
"grad_norm": 0.1304275095462799, |
|
"learning_rate": 6.269014643030213e-06, |
|
"loss": 0.4423, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.006742321789067247, |
|
"grad_norm": 0.13337342441082, |
|
"learning_rate": 5.852620357053651e-06, |
|
"loss": 0.3449, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.006781521334352522, |
|
"grad_norm": 0.14355811476707458, |
|
"learning_rate": 5.449673790581611e-06, |
|
"loss": 0.3999, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.006820720879637796, |
|
"grad_norm": 0.1402300000190735, |
|
"learning_rate": 5.060297685041659e-06, |
|
"loss": 0.4032, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.006859920424923071, |
|
"grad_norm": 0.14491398632526398, |
|
"learning_rate": 4.684610648167503e-06, |
|
"loss": 0.3728, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.0068991199702083455, |
|
"grad_norm": 0.1472439020872116, |
|
"learning_rate": 4.322727117869951e-06, |
|
"loss": 0.3707, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.00693831951549362, |
|
"grad_norm": 0.14669029414653778, |
|
"learning_rate": 3.974757327377981e-06, |
|
"loss": 0.4789, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.006977519060778895, |
|
"grad_norm": 0.15367992222309113, |
|
"learning_rate": 3.6408072716606346e-06, |
|
"loss": 0.3934, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.007016718606064169, |
|
"grad_norm": 0.14292077720165253, |
|
"learning_rate": 3.3209786751399187e-06, |
|
"loss": 0.3769, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.007055918151349445, |
|
"grad_norm": 0.14761218428611755, |
|
"learning_rate": 3.0153689607045845e-06, |
|
"loss": 0.3591, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.007095117696634719, |
|
"grad_norm": 0.16149510443210602, |
|
"learning_rate": 2.724071220034158e-06, |
|
"loss": 0.4567, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.007134317241919994, |
|
"grad_norm": 0.1627466082572937, |
|
"learning_rate": 2.4471741852423237e-06, |
|
"loss": 0.3955, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.007173516787205268, |
|
"grad_norm": 0.16270063817501068, |
|
"learning_rate": 2.1847622018482283e-06, |
|
"loss": 0.2732, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.007212716332490543, |
|
"grad_norm": 0.17279773950576782, |
|
"learning_rate": 1.9369152030840556e-06, |
|
"loss": 0.3798, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.007251915877775818, |
|
"grad_norm": 0.17936405539512634, |
|
"learning_rate": 1.70370868554659e-06, |
|
"loss": 0.409, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.007291115423061092, |
|
"grad_norm": 0.1672297567129135, |
|
"learning_rate": 1.4852136862001764e-06, |
|
"loss": 0.3581, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.007330314968346368, |
|
"grad_norm": 0.1915307492017746, |
|
"learning_rate": 1.2814967607382432e-06, |
|
"loss": 0.4487, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.007369514513631642, |
|
"grad_norm": 0.17655663192272186, |
|
"learning_rate": 1.0926199633097157e-06, |
|
"loss": 0.3467, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.007408714058916917, |
|
"grad_norm": 0.17965379357337952, |
|
"learning_rate": 9.186408276168013e-07, |
|
"loss": 0.3978, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.007447913604202191, |
|
"grad_norm": 0.1944323629140854, |
|
"learning_rate": 7.596123493895991e-07, |
|
"loss": 0.3575, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.007487113149487466, |
|
"grad_norm": 0.2105947881937027, |
|
"learning_rate": 6.15582970243117e-07, |
|
"loss": 0.3813, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.0075263126947727405, |
|
"grad_norm": 0.19838747382164001, |
|
"learning_rate": 4.865965629214819e-07, |
|
"loss": 0.4062, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.007565512240058015, |
|
"grad_norm": 0.18956390023231506, |
|
"learning_rate": 3.7269241793390085e-07, |
|
"loss": 0.3249, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.00760471178534329, |
|
"grad_norm": 0.20137059688568115, |
|
"learning_rate": 2.7390523158633554e-07, |
|
"loss": 0.3665, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.007643911330628565, |
|
"grad_norm": 0.22965259850025177, |
|
"learning_rate": 1.9026509541272275e-07, |
|
"loss": 0.3771, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.00768311087591384, |
|
"grad_norm": 0.23276779055595398, |
|
"learning_rate": 1.2179748700879012e-07, |
|
"loss": 0.3388, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.007722310421199114, |
|
"grad_norm": 0.23765169084072113, |
|
"learning_rate": 6.852326227130834e-08, |
|
"loss": 0.3759, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.007761509966484389, |
|
"grad_norm": 0.3065294027328491, |
|
"learning_rate": 3.04586490452119e-08, |
|
"loss": 0.3336, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.007800709511769663, |
|
"grad_norm": 0.2954770028591156, |
|
"learning_rate": 7.615242180436522e-09, |
|
"loss": 0.3031, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.007839909057054938, |
|
"grad_norm": 0.38077908754348755, |
|
"learning_rate": 0.0, |
|
"loss": 0.3565, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.007839909057054938, |
|
"eval_loss": 0.4325978755950928, |
|
"eval_runtime": 2401.2634, |
|
"eval_samples_per_second": 17.893, |
|
"eval_steps_per_second": 8.947, |
|
"step": 200 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 200, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 50, |
|
"stateful_callbacks": { |
|
"EarlyStoppingCallback": { |
|
"args": { |
|
"early_stopping_patience": 5, |
|
"early_stopping_threshold": 0.0 |
|
}, |
|
"attributes": { |
|
"early_stopping_patience_counter": 0 |
|
} |
|
}, |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1.2178385512326758e+17, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|