{ "best_metric": null, "best_model_checkpoint": null, "epoch": 12.421052631578947, "eval_steps": 500, "global_step": 56, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.21052631578947367, "grad_norm": 19.218000411987305, "learning_rate": 2e-05, "loss": 3.886, "step": 1 }, { "epoch": 0.42105263157894735, "grad_norm": 19.777868270874023, "learning_rate": 4e-05, "loss": 4.4615, "step": 2 }, { "epoch": 0.631578947368421, "grad_norm": 17.180343627929688, "learning_rate": 6e-05, "loss": 3.7595, "step": 3 }, { "epoch": 0.8421052631578947, "grad_norm": 17.781330108642578, "learning_rate": 8e-05, "loss": 3.3302, "step": 4 }, { "epoch": 1.1052631578947367, "grad_norm": 20.543649673461914, "learning_rate": 0.0001, "loss": 4.6154, "step": 5 }, { "epoch": 1.3157894736842106, "grad_norm": 11.29702091217041, "learning_rate": 0.00012, "loss": 2.5755, "step": 6 }, { "epoch": 1.526315789473684, "grad_norm": 8.194554328918457, "learning_rate": 0.00014, "loss": 1.6726, "step": 7 }, { "epoch": 1.736842105263158, "grad_norm": 6.040785312652588, "learning_rate": 0.00016, "loss": 1.2606, "step": 8 }, { "epoch": 1.9473684210526314, "grad_norm": 12.296614646911621, "learning_rate": 0.00018, "loss": 1.9246, "step": 9 }, { "epoch": 2.2105263157894735, "grad_norm": 6.707492828369141, "learning_rate": 0.0002, "loss": 1.0017, "step": 10 }, { "epoch": 2.4210526315789473, "grad_norm": 4.79848575592041, "learning_rate": 0.00019976687691905393, "loss": 0.6978, "step": 11 }, { "epoch": 2.6315789473684212, "grad_norm": 3.3441014289855957, "learning_rate": 0.00019906859460363307, "loss": 0.5996, "step": 12 }, { "epoch": 2.8421052631578947, "grad_norm": 3.113868474960327, "learning_rate": 0.00019790840876823232, "loss": 0.5782, "step": 13 }, { "epoch": 3.1052631578947367, "grad_norm": 7.5327229499816895, "learning_rate": 0.00019629172873477995, "loss": 0.8014, "step": 14 }, { "epoch": 3.3157894736842106, "grad_norm": 2.6903626918792725, "learning_rate": 0.00019422609221188207, "loss": 0.3285, "step": 15 }, { "epoch": 3.526315789473684, "grad_norm": 3.672346591949463, "learning_rate": 0.00019172113015054532, "loss": 0.3814, "step": 16 }, { "epoch": 3.736842105263158, "grad_norm": 1.4505695104599, "learning_rate": 0.0001887885218402375, "loss": 0.2401, "step": 17 }, { "epoch": 3.9473684210526314, "grad_norm": 2.6290504932403564, "learning_rate": 0.00018544194045464886, "loss": 0.3272, "step": 18 }, { "epoch": 4.2105263157894735, "grad_norm": 1.3445160388946533, "learning_rate": 0.0001816969893010442, "loss": 0.1871, "step": 19 }, { "epoch": 4.421052631578947, "grad_norm": 1.851083755493164, "learning_rate": 0.000177571129070442, "loss": 0.1781, "step": 20 }, { "epoch": 4.631578947368421, "grad_norm": 1.0952624082565308, "learning_rate": 0.00017308359642781242, "loss": 0.1966, "step": 21 }, { "epoch": 4.842105263157895, "grad_norm": 2.4012107849121094, "learning_rate": 0.00016825531432186543, "loss": 0.1738, "step": 22 }, { "epoch": 5.105263157894737, "grad_norm": 1.6361955404281616, "learning_rate": 0.00016310879443260528, "loss": 0.2131, "step": 23 }, { "epoch": 5.315789473684211, "grad_norm": 2.245840072631836, "learning_rate": 0.00015766803221148673, "loss": 0.1646, "step": 24 }, { "epoch": 5.526315789473684, "grad_norm": 1.684137225151062, "learning_rate": 0.00015195839500354335, "loss": 0.184, "step": 25 }, { "epoch": 5.7368421052631575, "grad_norm": 0.9733522534370422, "learning_rate": 0.00014600650377311522, "loss": 0.1465, "step": 26 }, { "epoch": 5.947368421052632, "grad_norm": 2.4283876419067383, "learning_rate": 0.00013984010898462416, "loss": 0.2177, "step": 27 }, { "epoch": 6.2105263157894735, "grad_norm": 1.4879282712936401, "learning_rate": 0.00013348796121709862, "loss": 0.1299, "step": 28 }, { "epoch": 6.421052631578947, "grad_norm": 1.0862969160079956, "learning_rate": 0.00012697967711570242, "loss": 0.1115, "step": 29 }, { "epoch": 6.631578947368421, "grad_norm": 0.7216919660568237, "learning_rate": 0.0001203456013052634, "loss": 0.0971, "step": 30 }, { "epoch": 6.842105263157895, "grad_norm": 1.4882885217666626, "learning_rate": 0.00011361666490962468, "loss": 0.1453, "step": 31 }, { "epoch": 7.105263157894737, "grad_norm": 1.220837950706482, "learning_rate": 0.0001068242413364671, "loss": 0.1634, "step": 32 }, { "epoch": 7.315789473684211, "grad_norm": 0.6117944121360779, "learning_rate": 0.0001, "loss": 0.1026, "step": 33 }, { "epoch": 7.526315789473684, "grad_norm": 0.9110010266304016, "learning_rate": 9.317575866353292e-05, "loss": 0.103, "step": 34 }, { "epoch": 7.7368421052631575, "grad_norm": 0.9700250625610352, "learning_rate": 8.638333509037536e-05, "loss": 0.1205, "step": 35 }, { "epoch": 7.947368421052632, "grad_norm": 0.9124184250831604, "learning_rate": 7.965439869473664e-05, "loss": 0.1095, "step": 36 }, { "epoch": 8.210526315789474, "grad_norm": 0.49950140714645386, "learning_rate": 7.302032288429756e-05, "loss": 0.0796, "step": 37 }, { "epoch": 8.421052631578947, "grad_norm": 0.8700776100158691, "learning_rate": 6.651203878290139e-05, "loss": 0.0774, "step": 38 }, { "epoch": 8.631578947368421, "grad_norm": 0.7889358401298523, "learning_rate": 6.015989101537586e-05, "loss": 0.0835, "step": 39 }, { "epoch": 8.842105263157894, "grad_norm": 0.5236338973045349, "learning_rate": 5.399349622688479e-05, "loss": 0.0581, "step": 40 }, { "epoch": 9.105263157894736, "grad_norm": 0.9376251697540283, "learning_rate": 4.804160499645667e-05, "loss": 0.0936, "step": 41 }, { "epoch": 9.31578947368421, "grad_norm": 0.7259882092475891, "learning_rate": 4.2331967788513295e-05, "loss": 0.0712, "step": 42 }, { "epoch": 9.526315789473685, "grad_norm": 0.5035570859909058, "learning_rate": 3.689120556739475e-05, "loss": 0.0776, "step": 43 }, { "epoch": 9.736842105263158, "grad_norm": 0.4928078055381775, "learning_rate": 3.174468567813461e-05, "loss": 0.0483, "step": 44 }, { "epoch": 9.947368421052632, "grad_norm": 1.1208339929580688, "learning_rate": 2.691640357218759e-05, "loss": 0.1032, "step": 45 }, { "epoch": 10.210526315789474, "grad_norm": 0.5036032795906067, "learning_rate": 2.242887092955801e-05, "loss": 0.0745, "step": 46 }, { "epoch": 10.421052631578947, "grad_norm": 0.4454878866672516, "learning_rate": 1.8303010698955804e-05, "loss": 0.0526, "step": 47 }, { "epoch": 10.631578947368421, "grad_norm": 0.48385077714920044, "learning_rate": 1.4558059545351143e-05, "loss": 0.0609, "step": 48 }, { "epoch": 10.842105263157894, "grad_norm": 0.8407765030860901, "learning_rate": 1.1211478159762478e-05, "loss": 0.0638, "step": 49 }, { "epoch": 11.105263157894736, "grad_norm": 0.4083283841609955, "learning_rate": 8.278869849454718e-06, "loss": 0.065, "step": 50 }, { "epoch": 11.31578947368421, "grad_norm": 0.4340791702270508, "learning_rate": 5.77390778811796e-06, "loss": 0.0545, "step": 51 }, { "epoch": 11.526315789473685, "grad_norm": 0.5537849068641663, "learning_rate": 3.7082712652200867e-06, "loss": 0.0511, "step": 52 }, { "epoch": 11.736842105263158, "grad_norm": 0.36356380581855774, "learning_rate": 2.091591231767709e-06, "loss": 0.0553, "step": 53 }, { "epoch": 11.947368421052632, "grad_norm": 1.3945286273956299, "learning_rate": 9.314053963669245e-07, "loss": 0.1255, "step": 54 }, { "epoch": 12.210526315789474, "grad_norm": 0.5195576548576355, "learning_rate": 2.3312308094607382e-07, "loss": 0.0559, "step": 55 }, { "epoch": 12.421052631578947, "grad_norm": 0.5838670134544373, "learning_rate": 0.0, "loss": 0.0631, "step": 56 } ], "logging_steps": 1, "max_steps": 56, "num_input_tokens_seen": 0, "num_train_epochs": 14, "save_steps": 4, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 50350673362944.0, "train_batch_size": 1, "trial_name": null, "trial_params": null }