{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.8806026624471123, "eval_steps": 200, "global_step": 3200, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0005503766640294451, "grad_norm": 2.2015435695648193, "learning_rate": 1.4775011317868612e-06, "loss": 0.7802, "step": 2 }, { "epoch": 0.0011007533280588903, "grad_norm": 2.0623114109039307, "learning_rate": 2.9550022635737224e-06, "loss": 0.6659, "step": 4 }, { "epoch": 0.0016511299920883354, "grad_norm": 0.8444932699203491, "learning_rate": 3.819285020442103e-06, "loss": 0.6275, "step": 6 }, { "epoch": 0.0022015066561177805, "grad_norm": 0.7291238307952881, "learning_rate": 4.432503395360583e-06, "loss": 0.5955, "step": 8 }, { "epoch": 0.0027518833201472257, "grad_norm": 0.591098427772522, "learning_rate": 4.90815251991065e-06, "loss": 0.5906, "step": 10 }, { "epoch": 0.003302259984176671, "grad_norm": 0.6075527667999268, "learning_rate": 5.2967861522289644e-06, "loss": 0.5451, "step": 12 }, { "epoch": 0.003852636648206116, "grad_norm": 0.5598031878471375, "learning_rate": 5.625371206454386e-06, "loss": 0.5539, "step": 14 }, { "epoch": 0.004403013312235561, "grad_norm": 0.5352339148521423, "learning_rate": 5.910004527147445e-06, "loss": 0.5452, "step": 16 }, { "epoch": 0.004953389976265006, "grad_norm": 0.524741530418396, "learning_rate": 6.161068909097345e-06, "loss": 0.5536, "step": 18 }, { "epoch": 0.005503766640294451, "grad_norm": 0.4852159321308136, "learning_rate": 6.38565365169751e-06, "loss": 0.5439, "step": 20 }, { "epoch": 0.0060541433043238965, "grad_norm": 0.4764852225780487, "learning_rate": 6.5888152636627215e-06, "loss": 0.5468, "step": 22 }, { "epoch": 0.006604519968353342, "grad_norm": 0.463278591632843, "learning_rate": 6.774287284015826e-06, "loss": 0.541, "step": 24 }, { "epoch": 0.007154896632382787, "grad_norm": 0.4566305875778198, "learning_rate": 6.944905003449378e-06, "loss": 0.5258, "step": 26 }, { "epoch": 0.007705273296412232, "grad_norm": 0.4572094678878784, "learning_rate": 7.102872338241248e-06, "loss": 0.5385, "step": 28 }, { "epoch": 0.008255649960441678, "grad_norm": 0.4581094980239868, "learning_rate": 7.2499364085658915e-06, "loss": 0.5258, "step": 30 }, { "epoch": 0.008806026624471122, "grad_norm": 0.4602491557598114, "learning_rate": 7.387505658934305e-06, "loss": 0.5239, "step": 32 }, { "epoch": 0.009356403288500568, "grad_norm": 0.4633028507232666, "learning_rate": 7.516732105870977e-06, "loss": 0.5237, "step": 34 }, { "epoch": 0.009906779952530012, "grad_norm": 0.4267115592956543, "learning_rate": 7.638570040884206e-06, "loss": 0.5467, "step": 36 }, { "epoch": 0.010457156616559458, "grad_norm": 0.4587521255016327, "learning_rate": 7.753818840648305e-06, "loss": 0.5282, "step": 38 }, { "epoch": 0.011007533280588903, "grad_norm": 0.44529175758361816, "learning_rate": 7.863154783484372e-06, "loss": 0.536, "step": 40 }, { "epoch": 0.011557909944618349, "grad_norm": 0.436199814081192, "learning_rate": 7.967155095109629e-06, "loss": 0.5259, "step": 42 }, { "epoch": 0.012108286608647793, "grad_norm": 0.43157511949539185, "learning_rate": 8.066316395449581e-06, "loss": 0.5173, "step": 44 }, { "epoch": 0.012658663272677239, "grad_norm": 0.4393196105957031, "learning_rate": 8.161069041569085e-06, "loss": 0.5037, "step": 46 }, { "epoch": 0.013209039936706683, "grad_norm": 0.5085024237632751, "learning_rate": 8.251788415802687e-06, "loss": 0.5015, "step": 48 }, { "epoch": 0.01375941660073613, "grad_norm": 0.40056705474853516, "learning_rate": 8.338803908034438e-06, "loss": 0.5194, "step": 50 }, { "epoch": 0.014309793264765574, "grad_norm": 0.43617382645606995, "learning_rate": 8.422406135236239e-06, "loss": 0.5327, "step": 52 }, { "epoch": 0.01486016992879502, "grad_norm": 0.42528873682022095, "learning_rate": 8.502852797752587e-06, "loss": 0.5245, "step": 54 }, { "epoch": 0.015410546592824464, "grad_norm": 0.40340831875801086, "learning_rate": 8.58037347002811e-06, "loss": 0.5059, "step": 56 }, { "epoch": 0.015960923256853908, "grad_norm": 0.40859195590019226, "learning_rate": 8.65517355028691e-06, "loss": 0.5136, "step": 58 }, { "epoch": 0.016511299920883356, "grad_norm": 0.42262887954711914, "learning_rate": 8.727437540352753e-06, "loss": 0.5011, "step": 60 }, { "epoch": 0.0170616765849128, "grad_norm": 0.4255228638648987, "learning_rate": 8.79733178747776e-06, "loss": 0.5218, "step": 62 }, { "epoch": 0.017612053248942244, "grad_norm": 0.4350854754447937, "learning_rate": 8.865006790721166e-06, "loss": 0.5334, "step": 64 }, { "epoch": 0.01816242991297169, "grad_norm": 0.41395291686058044, "learning_rate": 8.930599152317962e-06, "loss": 0.5233, "step": 66 }, { "epoch": 0.018712806577001136, "grad_norm": 0.4127484858036041, "learning_rate": 8.99423323765784e-06, "loss": 0.5143, "step": 68 }, { "epoch": 0.01926318324103058, "grad_norm": 0.42464280128479004, "learning_rate": 9.056022594578175e-06, "loss": 0.5164, "step": 70 }, { "epoch": 0.019813559905060025, "grad_norm": 0.4011682868003845, "learning_rate": 9.116071172671068e-06, "loss": 0.5036, "step": 72 }, { "epoch": 0.02036393656908947, "grad_norm": 0.39912551641464233, "learning_rate": 9.174474375494509e-06, "loss": 0.5038, "step": 74 }, { "epoch": 0.020914313233118917, "grad_norm": 0.40526625514030457, "learning_rate": 9.231319972435167e-06, "loss": 0.518, "step": 76 }, { "epoch": 0.02146468989714836, "grad_norm": 0.4195484220981598, "learning_rate": 9.28668889210462e-06, "loss": 0.5065, "step": 78 }, { "epoch": 0.022015066561177805, "grad_norm": 0.4483351409435272, "learning_rate": 9.340655915271231e-06, "loss": 0.5128, "step": 80 }, { "epoch": 0.02256544322520725, "grad_norm": 0.42789894342422485, "learning_rate": 9.393290282217048e-06, "loss": 0.525, "step": 82 }, { "epoch": 0.023115819889236697, "grad_norm": 0.434644490480423, "learning_rate": 9.444656226896488e-06, "loss": 0.5248, "step": 84 }, { "epoch": 0.02366619655326614, "grad_norm": 0.4532856345176697, "learning_rate": 9.494813448234365e-06, "loss": 0.5226, "step": 86 }, { "epoch": 0.024216573217295586, "grad_norm": 0.4103749692440033, "learning_rate": 9.543817527236444e-06, "loss": 0.5034, "step": 88 }, { "epoch": 0.02476694988132503, "grad_norm": 0.4208613336086273, "learning_rate": 9.591720297221133e-06, "loss": 0.5214, "step": 90 }, { "epoch": 0.025317326545354478, "grad_norm": 0.4020327627658844, "learning_rate": 9.638570173355947e-06, "loss": 0.5047, "step": 92 }, { "epoch": 0.025867703209383922, "grad_norm": 0.4074559509754181, "learning_rate": 9.684412446751251e-06, "loss": 0.4999, "step": 94 }, { "epoch": 0.026418079873413366, "grad_norm": 0.43330731987953186, "learning_rate": 9.729289547589548e-06, "loss": 0.5089, "step": 96 }, { "epoch": 0.02696845653744281, "grad_norm": 0.42775431275367737, "learning_rate": 9.773241281121913e-06, "loss": 0.5169, "step": 98 }, { "epoch": 0.02751883320147226, "grad_norm": 0.421403706073761, "learning_rate": 9.8163050398213e-06, "loss": 0.5123, "step": 100 }, { "epoch": 0.028069209865501703, "grad_norm": 0.42337778210639954, "learning_rate": 9.858515994526218e-06, "loss": 0.5116, "step": 102 }, { "epoch": 0.028619586529531147, "grad_norm": 0.4156826138496399, "learning_rate": 9.8999072670231e-06, "loss": 0.5077, "step": 104 }, { "epoch": 0.02916996319356059, "grad_norm": 0.4544354975223541, "learning_rate": 9.9405100861891e-06, "loss": 0.5099, "step": 106 }, { "epoch": 0.02972033985759004, "grad_norm": 0.4015970528125763, "learning_rate": 9.980353929539448e-06, "loss": 0.5049, "step": 108 }, { "epoch": 0.030270716521619483, "grad_norm": 0.3907098174095154, "learning_rate": 1e-05, "loss": 0.5202, "step": 110 }, { "epoch": 0.030821093185648928, "grad_norm": 0.4184499979019165, "learning_rate": 1e-05, "loss": 0.5085, "step": 112 }, { "epoch": 0.031371469849678375, "grad_norm": 0.47195565700531006, "learning_rate": 1e-05, "loss": 0.5161, "step": 114 }, { "epoch": 0.031921846513707816, "grad_norm": 0.43992695212364197, "learning_rate": 1e-05, "loss": 0.4978, "step": 116 }, { "epoch": 0.032472223177737264, "grad_norm": 0.43099331855773926, "learning_rate": 1e-05, "loss": 0.5035, "step": 118 }, { "epoch": 0.03302259984176671, "grad_norm": 0.44256317615509033, "learning_rate": 1e-05, "loss": 0.4991, "step": 120 }, { "epoch": 0.03357297650579615, "grad_norm": 0.42082124948501587, "learning_rate": 1e-05, "loss": 0.5028, "step": 122 }, { "epoch": 0.0341233531698256, "grad_norm": 0.38576358556747437, "learning_rate": 1e-05, "loss": 0.5081, "step": 124 }, { "epoch": 0.03467372983385505, "grad_norm": 0.3880733251571655, "learning_rate": 1e-05, "loss": 0.5001, "step": 126 }, { "epoch": 0.03522410649788449, "grad_norm": 0.41802075505256653, "learning_rate": 1e-05, "loss": 0.5056, "step": 128 }, { "epoch": 0.035774483161913936, "grad_norm": 0.3949527144432068, "learning_rate": 1e-05, "loss": 0.5155, "step": 130 }, { "epoch": 0.03632485982594338, "grad_norm": 0.4038969576358795, "learning_rate": 1e-05, "loss": 0.5056, "step": 132 }, { "epoch": 0.036875236489972825, "grad_norm": 0.40195325016975403, "learning_rate": 1e-05, "loss": 0.4968, "step": 134 }, { "epoch": 0.03742561315400227, "grad_norm": 0.3946043848991394, "learning_rate": 1e-05, "loss": 0.4981, "step": 136 }, { "epoch": 0.037975989818031713, "grad_norm": 0.3914756774902344, "learning_rate": 1e-05, "loss": 0.5, "step": 138 }, { "epoch": 0.03852636648206116, "grad_norm": 0.4295148551464081, "learning_rate": 1e-05, "loss": 0.5147, "step": 140 }, { "epoch": 0.03907674314609061, "grad_norm": 0.40092742443084717, "learning_rate": 1e-05, "loss": 0.5196, "step": 142 }, { "epoch": 0.03962711981012005, "grad_norm": 0.41200628876686096, "learning_rate": 1e-05, "loss": 0.5031, "step": 144 }, { "epoch": 0.0401774964741495, "grad_norm": 0.43834391236305237, "learning_rate": 1e-05, "loss": 0.5047, "step": 146 }, { "epoch": 0.04072787313817894, "grad_norm": 0.3940436542034149, "learning_rate": 1e-05, "loss": 0.4912, "step": 148 }, { "epoch": 0.041278249802208386, "grad_norm": 0.3873765170574188, "learning_rate": 1e-05, "loss": 0.482, "step": 150 }, { "epoch": 0.041828626466237834, "grad_norm": 0.4272858798503876, "learning_rate": 1e-05, "loss": 0.4923, "step": 152 }, { "epoch": 0.042379003130267275, "grad_norm": 0.40542730689048767, "learning_rate": 1e-05, "loss": 0.4892, "step": 154 }, { "epoch": 0.04292937979429672, "grad_norm": 0.38277357816696167, "learning_rate": 1e-05, "loss": 0.517, "step": 156 }, { "epoch": 0.04347975645832617, "grad_norm": 0.39421385526657104, "learning_rate": 1e-05, "loss": 0.503, "step": 158 }, { "epoch": 0.04403013312235561, "grad_norm": 0.3984109163284302, "learning_rate": 1e-05, "loss": 0.5074, "step": 160 }, { "epoch": 0.04458050978638506, "grad_norm": 0.40513876080513, "learning_rate": 1e-05, "loss": 0.5092, "step": 162 }, { "epoch": 0.0451308864504145, "grad_norm": 0.45850449800491333, "learning_rate": 1e-05, "loss": 0.5086, "step": 164 }, { "epoch": 0.04568126311444395, "grad_norm": 0.4050631821155548, "learning_rate": 1e-05, "loss": 0.5073, "step": 166 }, { "epoch": 0.046231639778473395, "grad_norm": 0.41050952672958374, "learning_rate": 1e-05, "loss": 0.5007, "step": 168 }, { "epoch": 0.046782016442502836, "grad_norm": 0.39902788400650024, "learning_rate": 1e-05, "loss": 0.4941, "step": 170 }, { "epoch": 0.04733239310653228, "grad_norm": 0.4421572983264923, "learning_rate": 1e-05, "loss": 0.4988, "step": 172 }, { "epoch": 0.04788276977056173, "grad_norm": 0.4092646837234497, "learning_rate": 1e-05, "loss": 0.5001, "step": 174 }, { "epoch": 0.04843314643459117, "grad_norm": 0.4195966124534607, "learning_rate": 1e-05, "loss": 0.4964, "step": 176 }, { "epoch": 0.04898352309862062, "grad_norm": 0.3937481641769409, "learning_rate": 1e-05, "loss": 0.4977, "step": 178 }, { "epoch": 0.04953389976265006, "grad_norm": 0.434950590133667, "learning_rate": 1e-05, "loss": 0.5054, "step": 180 }, { "epoch": 0.05008427642667951, "grad_norm": 0.40112894773483276, "learning_rate": 1e-05, "loss": 0.494, "step": 182 }, { "epoch": 0.050634653090708956, "grad_norm": 0.42001938819885254, "learning_rate": 1e-05, "loss": 0.4744, "step": 184 }, { "epoch": 0.0511850297547384, "grad_norm": 0.4066455364227295, "learning_rate": 1e-05, "loss": 0.4838, "step": 186 }, { "epoch": 0.051735406418767844, "grad_norm": 0.3934157192707062, "learning_rate": 1e-05, "loss": 0.5017, "step": 188 }, { "epoch": 0.05228578308279729, "grad_norm": 0.38877320289611816, "learning_rate": 1e-05, "loss": 0.5018, "step": 190 }, { "epoch": 0.05283615974682673, "grad_norm": 0.39771756529808044, "learning_rate": 1e-05, "loss": 0.485, "step": 192 }, { "epoch": 0.05338653641085618, "grad_norm": 0.3938674330711365, "learning_rate": 1e-05, "loss": 0.5034, "step": 194 }, { "epoch": 0.05393691307488562, "grad_norm": 0.40473559498786926, "learning_rate": 1e-05, "loss": 0.5082, "step": 196 }, { "epoch": 0.05448728973891507, "grad_norm": 0.3977149426937103, "learning_rate": 1e-05, "loss": 0.4997, "step": 198 }, { "epoch": 0.05503766640294452, "grad_norm": 0.39340054988861084, "learning_rate": 1e-05, "loss": 0.4859, "step": 200 }, { "epoch": 0.05503766640294452, "eval_merge_loss": 0.4414624571800232, "eval_merge_runtime": 600.1539, "eval_merge_samples_per_second": 56.239, "eval_merge_steps_per_second": 2.344, "step": 200 }, { "epoch": 0.05503766640294452, "eval_new_aug_datas_filtered.json_loss": 0.5691156983375549, "eval_new_aug_datas_filtered.json_runtime": 10.6767, "eval_new_aug_datas_filtered.json_samples_per_second": 71.839, "eval_new_aug_datas_filtered.json_steps_per_second": 2.997, "step": 200 }, { "epoch": 0.05503766640294452, "eval_sharegpt_gpt4.json_loss": 0.8223738670349121, "eval_sharegpt_gpt4.json_runtime": 31.6183, "eval_sharegpt_gpt4.json_samples_per_second": 58.858, "eval_sharegpt_gpt4.json_steps_per_second": 2.467, "step": 200 }, { "epoch": 0.05503766640294452, "eval_Table_GPT.json_loss": 0.09253557026386261, "eval_Table_GPT.json_runtime": 24.9748, "eval_Table_GPT.json_samples_per_second": 83.804, "eval_Table_GPT.json_steps_per_second": 3.524, "step": 200 }, { "epoch": 0.05503766640294452, "eval_gpt_4o_200k.json_loss": 0.849287211894989, "eval_gpt_4o_200k.json_runtime": 48.5339, "eval_gpt_4o_200k.json_samples_per_second": 129.415, "eval_gpt_4o_200k.json_steps_per_second": 5.398, "step": 200 }, { "epoch": 0.05503766640294452, "eval_multi_turn_datas.json_loss": 0.3907540738582611, "eval_multi_turn_datas.json_runtime": 75.6133, "eval_multi_turn_datas.json_samples_per_second": 52.927, "eval_multi_turn_datas.json_steps_per_second": 2.209, "step": 200 }, { "epoch": 0.05503766640294452, "eval_table_python_code_datas.json_loss": 0.33119720220565796, "eval_table_python_code_datas.json_runtime": 43.1313, "eval_table_python_code_datas.json_samples_per_second": 50.056, "eval_table_python_code_datas.json_steps_per_second": 2.087, "step": 200 }, { "epoch": 0.05503766640294452, "eval_tabular_llm_data.json_loss": 0.14601922035217285, "eval_tabular_llm_data.json_runtime": 8.7785, "eval_tabular_llm_data.json_samples_per_second": 28.023, "eval_tabular_llm_data.json_steps_per_second": 1.253, "step": 200 }, { "epoch": 0.05503766640294452, "eval_python_code_critic_21k.json_loss": 0.625038743019104, "eval_python_code_critic_21k.json_runtime": 3.237, "eval_python_code_critic_21k.json_samples_per_second": 184.43, "eval_python_code_critic_21k.json_steps_per_second": 7.723, "step": 200 }, { "epoch": 0.05503766640294452, "eval_all_merge_table_dataset.json_loss": 0.09772461652755737, "eval_all_merge_table_dataset.json_runtime": 24.3077, "eval_all_merge_table_dataset.json_samples_per_second": 29.291, "eval_all_merge_table_dataset.json_steps_per_second": 1.234, "step": 200 }, { "epoch": 0.05503766640294452, "eval_code_feedback_multi_turn.json_loss": 0.6093290448188782, "eval_code_feedback_multi_turn.json_runtime": 32.4589, "eval_code_feedback_multi_turn.json_samples_per_second": 67.809, "eval_code_feedback_multi_turn.json_steps_per_second": 2.834, "step": 200 }, { "epoch": 0.05503766640294452, "eval_ultrainteract_sft.json_loss": 0.4469935894012451, "eval_ultrainteract_sft.json_runtime": 8.6702, "eval_ultrainteract_sft.json_samples_per_second": 167.931, "eval_ultrainteract_sft.json_steps_per_second": 7.036, "step": 200 }, { "epoch": 0.05503766640294452, "eval_synthetic_text_to_sql.json_loss": 0.11159003525972366, "eval_synthetic_text_to_sql.json_runtime": 0.1306, "eval_synthetic_text_to_sql.json_samples_per_second": 260.355, "eval_synthetic_text_to_sql.json_steps_per_second": 15.315, "step": 200 }, { "epoch": 0.05503766640294452, "eval_sft_react_sql_datas.json_loss": 0.6847189664840698, "eval_sft_react_sql_datas.json_runtime": 7.8434, "eval_sft_react_sql_datas.json_samples_per_second": 40.034, "eval_sft_react_sql_datas.json_steps_per_second": 1.785, "step": 200 }, { "epoch": 0.05503766640294452, "eval_all_merge_code.json_loss": 0.32269543409347534, "eval_all_merge_code.json_runtime": 0.3287, "eval_all_merge_code.json_samples_per_second": 191.649, "eval_all_merge_code.json_steps_per_second": 9.126, "step": 200 }, { "epoch": 0.05503766640294452, "eval_magpie_datas.json_loss": 0.4600640833377838, "eval_magpie_datas.json_runtime": 2.2095, "eval_magpie_datas.json_samples_per_second": 77.844, "eval_magpie_datas.json_steps_per_second": 3.621, "step": 200 }, { "epoch": 0.05503766640294452, "eval_train_data_for_qwen.json_loss": 0.017207294702529907, "eval_train_data_for_qwen.json_runtime": 0.2494, "eval_train_data_for_qwen.json_samples_per_second": 40.095, "eval_train_data_for_qwen.json_steps_per_second": 4.01, "step": 200 }, { "epoch": 0.05503766640294452, "eval_alpaca_cleaned.json_loss": 0.9374485015869141, "eval_alpaca_cleaned.json_runtime": 0.1149, "eval_alpaca_cleaned.json_samples_per_second": 234.896, "eval_alpaca_cleaned.json_steps_per_second": 17.4, "step": 200 }, { "epoch": 0.05503766640294452, "eval_agent_instruct.json_loss": 0.23996739089488983, "eval_agent_instruct.json_runtime": 0.5126, "eval_agent_instruct.json_samples_per_second": 93.639, "eval_agent_instruct.json_steps_per_second": 3.902, "step": 200 }, { "epoch": 0.05503766640294452, "eval_MathInstruct.json_loss": 0.2269323617219925, "eval_MathInstruct.json_runtime": 0.3472, "eval_MathInstruct.json_samples_per_second": 164.184, "eval_MathInstruct.json_steps_per_second": 8.641, "step": 200 }, { "epoch": 0.05503766640294452, "eval_tested_143k_python_alpaca.json_loss": 0.4513254165649414, "eval_tested_143k_python_alpaca.json_runtime": 0.3017, "eval_tested_143k_python_alpaca.json_samples_per_second": 112.684, "eval_tested_143k_python_alpaca.json_steps_per_second": 6.628, "step": 200 }, { "epoch": 0.05503766640294452, "eval_xlam_function_calling_60k.json_loss": 0.011208846233785152, "eval_xlam_function_calling_60k.json_runtime": 0.1011, "eval_xlam_function_calling_60k.json_samples_per_second": 227.556, "eval_xlam_function_calling_60k.json_steps_per_second": 9.894, "step": 200 }, { "epoch": 0.05503766640294452, "eval_alpaca_data_gpt4_chinese.json_loss": 1.6813441514968872, "eval_alpaca_data_gpt4_chinese.json_runtime": 0.0523, "eval_alpaca_data_gpt4_chinese.json_samples_per_second": 306.208, "eval_alpaca_data_gpt4_chinese.json_steps_per_second": 19.138, "step": 200 }, { "epoch": 0.05503766640294452, "eval_alpaca_gpt4_zh.json_loss": 1.0053786039352417, "eval_alpaca_gpt4_zh.json_runtime": 0.0504, "eval_alpaca_gpt4_zh.json_samples_per_second": 218.451, "eval_alpaca_gpt4_zh.json_steps_per_second": 19.859, "step": 200 }, { "epoch": 0.05503766640294452, "eval_codefeedback_filtered_instruction.json_loss": 0.5965134501457214, "eval_codefeedback_filtered_instruction.json_runtime": 0.4841, "eval_codefeedback_filtered_instruction.json_samples_per_second": 41.316, "eval_codefeedback_filtered_instruction.json_steps_per_second": 2.066, "step": 200 }, { "epoch": 0.05558804306697396, "grad_norm": 0.38687607645988464, "learning_rate": 1e-05, "loss": 0.49, "step": 202 }, { "epoch": 0.056138419731003406, "grad_norm": 0.39803430438041687, "learning_rate": 1e-05, "loss": 0.5047, "step": 204 }, { "epoch": 0.05668879639503285, "grad_norm": 0.41770851612091064, "learning_rate": 1e-05, "loss": 0.4874, "step": 206 }, { "epoch": 0.057239173059062294, "grad_norm": 0.3909968435764313, "learning_rate": 1e-05, "loss": 0.4992, "step": 208 }, { "epoch": 0.05778954972309174, "grad_norm": 0.3818782866001129, "learning_rate": 1e-05, "loss": 0.5006, "step": 210 }, { "epoch": 0.05833992638712118, "grad_norm": 0.4179542362689972, "learning_rate": 1e-05, "loss": 0.4945, "step": 212 }, { "epoch": 0.05889030305115063, "grad_norm": 0.3872973322868347, "learning_rate": 1e-05, "loss": 0.4918, "step": 214 }, { "epoch": 0.05944067971518008, "grad_norm": 0.4249219298362732, "learning_rate": 1e-05, "loss": 0.5039, "step": 216 }, { "epoch": 0.05999105637920952, "grad_norm": 0.43381986021995544, "learning_rate": 1e-05, "loss": 0.4873, "step": 218 }, { "epoch": 0.06054143304323897, "grad_norm": 0.40741005539894104, "learning_rate": 1e-05, "loss": 0.4771, "step": 220 }, { "epoch": 0.061091809707268414, "grad_norm": 0.37800464034080505, "learning_rate": 1e-05, "loss": 0.5015, "step": 222 }, { "epoch": 0.061642186371297855, "grad_norm": 0.42365899682044983, "learning_rate": 1e-05, "loss": 0.4906, "step": 224 }, { "epoch": 0.0621925630353273, "grad_norm": 0.39279666543006897, "learning_rate": 1e-05, "loss": 0.51, "step": 226 }, { "epoch": 0.06274293969935675, "grad_norm": 0.4037010073661804, "learning_rate": 1e-05, "loss": 0.5162, "step": 228 }, { "epoch": 0.0632933163633862, "grad_norm": 0.37650179862976074, "learning_rate": 1e-05, "loss": 0.4984, "step": 230 }, { "epoch": 0.06384369302741563, "grad_norm": 0.42879757285118103, "learning_rate": 1e-05, "loss": 0.492, "step": 232 }, { "epoch": 0.06439406969144508, "grad_norm": 0.42225000262260437, "learning_rate": 1e-05, "loss": 0.5215, "step": 234 }, { "epoch": 0.06494444635547453, "grad_norm": 0.3948579430580139, "learning_rate": 1e-05, "loss": 0.5045, "step": 236 }, { "epoch": 0.06549482301950398, "grad_norm": 0.40142592787742615, "learning_rate": 1e-05, "loss": 0.5083, "step": 238 }, { "epoch": 0.06604519968353342, "grad_norm": 0.41938111186027527, "learning_rate": 1e-05, "loss": 0.5094, "step": 240 }, { "epoch": 0.06659557634756286, "grad_norm": 0.4345923066139221, "learning_rate": 1e-05, "loss": 0.5076, "step": 242 }, { "epoch": 0.0671459530115923, "grad_norm": 0.3985568881034851, "learning_rate": 1e-05, "loss": 0.5007, "step": 244 }, { "epoch": 0.06769632967562175, "grad_norm": 0.37891215085983276, "learning_rate": 1e-05, "loss": 0.513, "step": 246 }, { "epoch": 0.0682467063396512, "grad_norm": 0.413566917181015, "learning_rate": 1e-05, "loss": 0.493, "step": 248 }, { "epoch": 0.06879708300368065, "grad_norm": 0.3980996608734131, "learning_rate": 1e-05, "loss": 0.5161, "step": 250 }, { "epoch": 0.0693474596677101, "grad_norm": 0.4525178372859955, "learning_rate": 1e-05, "loss": 0.5077, "step": 252 }, { "epoch": 0.06989783633173953, "grad_norm": 0.3720250427722931, "learning_rate": 1e-05, "loss": 0.4809, "step": 254 }, { "epoch": 0.07044821299576898, "grad_norm": 0.37366852164268494, "learning_rate": 1e-05, "loss": 0.4724, "step": 256 }, { "epoch": 0.07099858965979843, "grad_norm": 0.38189247250556946, "learning_rate": 1e-05, "loss": 0.5062, "step": 258 }, { "epoch": 0.07154896632382787, "grad_norm": 0.39108410477638245, "learning_rate": 1e-05, "loss": 0.4894, "step": 260 }, { "epoch": 0.07209934298785732, "grad_norm": 0.4071044921875, "learning_rate": 1e-05, "loss": 0.4916, "step": 262 }, { "epoch": 0.07264971965188675, "grad_norm": 0.38570597767829895, "learning_rate": 1e-05, "loss": 0.4925, "step": 264 }, { "epoch": 0.0732000963159162, "grad_norm": 0.409600168466568, "learning_rate": 1e-05, "loss": 0.4987, "step": 266 }, { "epoch": 0.07375047297994565, "grad_norm": 0.3844049274921417, "learning_rate": 1e-05, "loss": 0.5011, "step": 268 }, { "epoch": 0.0743008496439751, "grad_norm": 0.41260388493537903, "learning_rate": 1e-05, "loss": 0.5014, "step": 270 }, { "epoch": 0.07485122630800455, "grad_norm": 0.402567982673645, "learning_rate": 1e-05, "loss": 0.4926, "step": 272 }, { "epoch": 0.07540160297203398, "grad_norm": 0.4058002233505249, "learning_rate": 1e-05, "loss": 0.4879, "step": 274 }, { "epoch": 0.07595197963606343, "grad_norm": 0.42676812410354614, "learning_rate": 1e-05, "loss": 0.5073, "step": 276 }, { "epoch": 0.07650235630009287, "grad_norm": 0.3878956735134125, "learning_rate": 1e-05, "loss": 0.4831, "step": 278 }, { "epoch": 0.07705273296412232, "grad_norm": 0.37560945749282837, "learning_rate": 1e-05, "loss": 0.4705, "step": 280 }, { "epoch": 0.07760310962815177, "grad_norm": 0.4071865379810333, "learning_rate": 1e-05, "loss": 0.489, "step": 282 }, { "epoch": 0.07815348629218122, "grad_norm": 0.3832094073295593, "learning_rate": 1e-05, "loss": 0.4843, "step": 284 }, { "epoch": 0.07870386295621065, "grad_norm": 0.3808830976486206, "learning_rate": 1e-05, "loss": 0.5019, "step": 286 }, { "epoch": 0.0792542396202401, "grad_norm": 0.40182846784591675, "learning_rate": 1e-05, "loss": 0.4921, "step": 288 }, { "epoch": 0.07980461628426955, "grad_norm": 0.4483119249343872, "learning_rate": 1e-05, "loss": 0.5042, "step": 290 }, { "epoch": 0.080354992948299, "grad_norm": 0.3664950132369995, "learning_rate": 1e-05, "loss": 0.4758, "step": 292 }, { "epoch": 0.08090536961232844, "grad_norm": 0.39573603868484497, "learning_rate": 1e-05, "loss": 0.4945, "step": 294 }, { "epoch": 0.08145574627635788, "grad_norm": 0.44645532965660095, "learning_rate": 1e-05, "loss": 0.4964, "step": 296 }, { "epoch": 0.08200612294038732, "grad_norm": 0.39092323184013367, "learning_rate": 1e-05, "loss": 0.4947, "step": 298 }, { "epoch": 0.08255649960441677, "grad_norm": 0.41762229800224304, "learning_rate": 1e-05, "loss": 0.4949, "step": 300 }, { "epoch": 0.08310687626844622, "grad_norm": 0.39803358912467957, "learning_rate": 1e-05, "loss": 0.4822, "step": 302 }, { "epoch": 0.08365725293247567, "grad_norm": 0.39895498752593994, "learning_rate": 1e-05, "loss": 0.4893, "step": 304 }, { "epoch": 0.0842076295965051, "grad_norm": 0.3883228600025177, "learning_rate": 1e-05, "loss": 0.5062, "step": 306 }, { "epoch": 0.08475800626053455, "grad_norm": 0.4112294018268585, "learning_rate": 1e-05, "loss": 0.4979, "step": 308 }, { "epoch": 0.085308382924564, "grad_norm": 0.3851683437824249, "learning_rate": 1e-05, "loss": 0.4934, "step": 310 }, { "epoch": 0.08585875958859344, "grad_norm": 0.39728567004203796, "learning_rate": 1e-05, "loss": 0.4746, "step": 312 }, { "epoch": 0.08640913625262289, "grad_norm": 0.3943733274936676, "learning_rate": 1e-05, "loss": 0.4904, "step": 314 }, { "epoch": 0.08695951291665234, "grad_norm": 0.3954530656337738, "learning_rate": 1e-05, "loss": 0.4796, "step": 316 }, { "epoch": 0.08750988958068177, "grad_norm": 0.41237205266952515, "learning_rate": 1e-05, "loss": 0.4908, "step": 318 }, { "epoch": 0.08806026624471122, "grad_norm": 0.3923771381378174, "learning_rate": 1e-05, "loss": 0.4988, "step": 320 }, { "epoch": 0.08861064290874067, "grad_norm": 0.38542094826698303, "learning_rate": 1e-05, "loss": 0.5027, "step": 322 }, { "epoch": 0.08916101957277012, "grad_norm": 0.41598251461982727, "learning_rate": 1e-05, "loss": 0.4976, "step": 324 }, { "epoch": 0.08971139623679956, "grad_norm": 0.40826794505119324, "learning_rate": 1e-05, "loss": 0.4929, "step": 326 }, { "epoch": 0.090261772900829, "grad_norm": 0.39970022439956665, "learning_rate": 1e-05, "loss": 0.4946, "step": 328 }, { "epoch": 0.09081214956485845, "grad_norm": 0.3739086985588074, "learning_rate": 1e-05, "loss": 0.4678, "step": 330 }, { "epoch": 0.0913625262288879, "grad_norm": 0.3746420741081238, "learning_rate": 1e-05, "loss": 0.4757, "step": 332 }, { "epoch": 0.09191290289291734, "grad_norm": 0.3976924419403076, "learning_rate": 1e-05, "loss": 0.487, "step": 334 }, { "epoch": 0.09246327955694679, "grad_norm": 0.398971289396286, "learning_rate": 1e-05, "loss": 0.5077, "step": 336 }, { "epoch": 0.09301365622097624, "grad_norm": 0.3937431871891022, "learning_rate": 1e-05, "loss": 0.4885, "step": 338 }, { "epoch": 0.09356403288500567, "grad_norm": 0.395084410905838, "learning_rate": 1e-05, "loss": 0.4871, "step": 340 }, { "epoch": 0.09411440954903512, "grad_norm": 0.3677273690700531, "learning_rate": 1e-05, "loss": 0.4813, "step": 342 }, { "epoch": 0.09466478621306457, "grad_norm": 0.39645129442214966, "learning_rate": 1e-05, "loss": 0.4842, "step": 344 }, { "epoch": 0.09521516287709401, "grad_norm": 0.3642916679382324, "learning_rate": 1e-05, "loss": 0.504, "step": 346 }, { "epoch": 0.09576553954112346, "grad_norm": 0.40385907888412476, "learning_rate": 1e-05, "loss": 0.4933, "step": 348 }, { "epoch": 0.0963159162051529, "grad_norm": 0.39063799381256104, "learning_rate": 1e-05, "loss": 0.4856, "step": 350 }, { "epoch": 0.09686629286918234, "grad_norm": 0.38000059127807617, "learning_rate": 1e-05, "loss": 0.5001, "step": 352 }, { "epoch": 0.09741666953321179, "grad_norm": 0.39380577206611633, "learning_rate": 1e-05, "loss": 0.4961, "step": 354 }, { "epoch": 0.09796704619724124, "grad_norm": 0.39326363801956177, "learning_rate": 1e-05, "loss": 0.498, "step": 356 }, { "epoch": 0.09851742286127069, "grad_norm": 0.3775707185268402, "learning_rate": 1e-05, "loss": 0.4792, "step": 358 }, { "epoch": 0.09906779952530012, "grad_norm": 0.3770863115787506, "learning_rate": 1e-05, "loss": 0.4837, "step": 360 }, { "epoch": 0.09961817618932957, "grad_norm": 0.41484272480010986, "learning_rate": 1e-05, "loss": 0.4739, "step": 362 }, { "epoch": 0.10016855285335902, "grad_norm": 0.39758750796318054, "learning_rate": 1e-05, "loss": 0.4957, "step": 364 }, { "epoch": 0.10071892951738846, "grad_norm": 0.43485164642333984, "learning_rate": 1e-05, "loss": 0.492, "step": 366 }, { "epoch": 0.10126930618141791, "grad_norm": 0.40296798944473267, "learning_rate": 1e-05, "loss": 0.4977, "step": 368 }, { "epoch": 0.10181968284544736, "grad_norm": 0.3818409740924835, "learning_rate": 1e-05, "loss": 0.481, "step": 370 }, { "epoch": 0.1023700595094768, "grad_norm": 0.3949006199836731, "learning_rate": 1e-05, "loss": 0.5021, "step": 372 }, { "epoch": 0.10292043617350624, "grad_norm": 0.4327391982078552, "learning_rate": 1e-05, "loss": 0.5036, "step": 374 }, { "epoch": 0.10347081283753569, "grad_norm": 0.4008086025714874, "learning_rate": 1e-05, "loss": 0.4854, "step": 376 }, { "epoch": 0.10402118950156514, "grad_norm": 0.4146427810192108, "learning_rate": 1e-05, "loss": 0.4933, "step": 378 }, { "epoch": 0.10457156616559458, "grad_norm": 0.4073733389377594, "learning_rate": 1e-05, "loss": 0.4923, "step": 380 }, { "epoch": 0.10512194282962402, "grad_norm": 0.40570083260536194, "learning_rate": 1e-05, "loss": 0.4806, "step": 382 }, { "epoch": 0.10567231949365347, "grad_norm": 0.39516401290893555, "learning_rate": 1e-05, "loss": 0.5038, "step": 384 }, { "epoch": 0.10622269615768291, "grad_norm": 0.3886268138885498, "learning_rate": 1e-05, "loss": 0.4737, "step": 386 }, { "epoch": 0.10677307282171236, "grad_norm": 0.3846561014652252, "learning_rate": 1e-05, "loss": 0.4852, "step": 388 }, { "epoch": 0.10732344948574181, "grad_norm": 0.3952987492084503, "learning_rate": 1e-05, "loss": 0.496, "step": 390 }, { "epoch": 0.10787382614977124, "grad_norm": 0.3840448558330536, "learning_rate": 1e-05, "loss": 0.4976, "step": 392 }, { "epoch": 0.10842420281380069, "grad_norm": 0.38074344396591187, "learning_rate": 1e-05, "loss": 0.508, "step": 394 }, { "epoch": 0.10897457947783014, "grad_norm": 0.4216584861278534, "learning_rate": 1e-05, "loss": 0.4841, "step": 396 }, { "epoch": 0.10952495614185959, "grad_norm": 0.39932167530059814, "learning_rate": 1e-05, "loss": 0.4783, "step": 398 }, { "epoch": 0.11007533280588903, "grad_norm": 0.3687106966972351, "learning_rate": 1e-05, "loss": 0.4747, "step": 400 }, { "epoch": 0.11007533280588903, "eval_merge_loss": 0.42975950241088867, "eval_merge_runtime": 600.4283, "eval_merge_samples_per_second": 56.213, "eval_merge_steps_per_second": 2.343, "step": 400 }, { "epoch": 0.11007533280588903, "eval_new_aug_datas_filtered.json_loss": 0.558424174785614, "eval_new_aug_datas_filtered.json_runtime": 10.4015, "eval_new_aug_datas_filtered.json_samples_per_second": 73.74, "eval_new_aug_datas_filtered.json_steps_per_second": 3.076, "step": 400 }, { "epoch": 0.11007533280588903, "eval_sharegpt_gpt4.json_loss": 0.8120941519737244, "eval_sharegpt_gpt4.json_runtime": 31.6378, "eval_sharegpt_gpt4.json_samples_per_second": 58.822, "eval_sharegpt_gpt4.json_steps_per_second": 2.465, "step": 400 }, { "epoch": 0.11007533280588903, "eval_Table_GPT.json_loss": 0.08201506733894348, "eval_Table_GPT.json_runtime": 24.9859, "eval_Table_GPT.json_samples_per_second": 83.767, "eval_Table_GPT.json_steps_per_second": 3.522, "step": 400 }, { "epoch": 0.11007533280588903, "eval_gpt_4o_200k.json_loss": 0.8391836881637573, "eval_gpt_4o_200k.json_runtime": 48.456, "eval_gpt_4o_200k.json_samples_per_second": 129.623, "eval_gpt_4o_200k.json_steps_per_second": 5.407, "step": 400 }, { "epoch": 0.11007533280588903, "eval_multi_turn_datas.json_loss": 0.37471804022789, "eval_multi_turn_datas.json_runtime": 75.4526, "eval_multi_turn_datas.json_samples_per_second": 53.04, "eval_multi_turn_datas.json_steps_per_second": 2.213, "step": 400 }, { "epoch": 0.11007533280588903, "eval_table_python_code_datas.json_loss": 0.3102189898490906, "eval_table_python_code_datas.json_runtime": 42.9961, "eval_table_python_code_datas.json_samples_per_second": 50.214, "eval_table_python_code_datas.json_steps_per_second": 2.093, "step": 400 }, { "epoch": 0.11007533280588903, "eval_tabular_llm_data.json_loss": 0.16565443575382233, "eval_tabular_llm_data.json_runtime": 8.5095, "eval_tabular_llm_data.json_samples_per_second": 28.909, "eval_tabular_llm_data.json_steps_per_second": 1.293, "step": 400 }, { "epoch": 0.11007533280588903, "eval_python_code_critic_21k.json_loss": 0.6095640063285828, "eval_python_code_critic_21k.json_runtime": 3.2106, "eval_python_code_critic_21k.json_samples_per_second": 185.945, "eval_python_code_critic_21k.json_steps_per_second": 7.787, "step": 400 }, { "epoch": 0.11007533280588903, "eval_all_merge_table_dataset.json_loss": 0.09411227703094482, "eval_all_merge_table_dataset.json_runtime": 23.2459, "eval_all_merge_table_dataset.json_samples_per_second": 30.629, "eval_all_merge_table_dataset.json_steps_per_second": 1.291, "step": 400 }, { "epoch": 0.11007533280588903, "eval_code_feedback_multi_turn.json_loss": 0.6033111810684204, "eval_code_feedback_multi_turn.json_runtime": 32.3176, "eval_code_feedback_multi_turn.json_samples_per_second": 68.105, "eval_code_feedback_multi_turn.json_steps_per_second": 2.847, "step": 400 }, { "epoch": 0.11007533280588903, "eval_ultrainteract_sft.json_loss": 0.4417967200279236, "eval_ultrainteract_sft.json_runtime": 8.6225, "eval_ultrainteract_sft.json_samples_per_second": 168.86, "eval_ultrainteract_sft.json_steps_per_second": 7.075, "step": 400 }, { "epoch": 0.11007533280588903, "eval_synthetic_text_to_sql.json_loss": 0.10689640045166016, "eval_synthetic_text_to_sql.json_runtime": 0.1258, "eval_synthetic_text_to_sql.json_samples_per_second": 270.238, "eval_synthetic_text_to_sql.json_steps_per_second": 15.896, "step": 400 }, { "epoch": 0.11007533280588903, "eval_sft_react_sql_datas.json_loss": 0.6745051145553589, "eval_sft_react_sql_datas.json_runtime": 7.8354, "eval_sft_react_sql_datas.json_samples_per_second": 40.074, "eval_sft_react_sql_datas.json_steps_per_second": 1.787, "step": 400 }, { "epoch": 0.11007533280588903, "eval_all_merge_code.json_loss": 0.3035649061203003, "eval_all_merge_code.json_runtime": 0.3282, "eval_all_merge_code.json_samples_per_second": 191.936, "eval_all_merge_code.json_steps_per_second": 9.14, "step": 400 }, { "epoch": 0.11007533280588903, "eval_magpie_datas.json_loss": 0.4511661231517792, "eval_magpie_datas.json_runtime": 2.2095, "eval_magpie_datas.json_samples_per_second": 77.847, "eval_magpie_datas.json_steps_per_second": 3.621, "step": 400 }, { "epoch": 0.11007533280588903, "eval_train_data_for_qwen.json_loss": 0.012529651634395123, "eval_train_data_for_qwen.json_runtime": 0.2431, "eval_train_data_for_qwen.json_samples_per_second": 41.135, "eval_train_data_for_qwen.json_steps_per_second": 4.113, "step": 400 }, { "epoch": 0.11007533280588903, "eval_alpaca_cleaned.json_loss": 0.9377387166023254, "eval_alpaca_cleaned.json_runtime": 0.115, "eval_alpaca_cleaned.json_samples_per_second": 234.777, "eval_alpaca_cleaned.json_steps_per_second": 17.391, "step": 400 }, { "epoch": 0.11007533280588903, "eval_agent_instruct.json_loss": 0.2363067865371704, "eval_agent_instruct.json_runtime": 0.5121, "eval_agent_instruct.json_samples_per_second": 93.728, "eval_agent_instruct.json_steps_per_second": 3.905, "step": 400 }, { "epoch": 0.11007533280588903, "eval_MathInstruct.json_loss": 0.21367128193378448, "eval_MathInstruct.json_runtime": 0.3647, "eval_MathInstruct.json_samples_per_second": 156.291, "eval_MathInstruct.json_steps_per_second": 8.226, "step": 400 }, { "epoch": 0.11007533280588903, "eval_tested_143k_python_alpaca.json_loss": 0.4474259316921234, "eval_tested_143k_python_alpaca.json_runtime": 0.3019, "eval_tested_143k_python_alpaca.json_samples_per_second": 112.631, "eval_tested_143k_python_alpaca.json_steps_per_second": 6.625, "step": 400 }, { "epoch": 0.11007533280588903, "eval_xlam_function_calling_60k.json_loss": 0.011296543292701244, "eval_xlam_function_calling_60k.json_runtime": 0.1, "eval_xlam_function_calling_60k.json_samples_per_second": 230.102, "eval_xlam_function_calling_60k.json_steps_per_second": 10.004, "step": 400 }, { "epoch": 0.11007533280588903, "eval_alpaca_data_gpt4_chinese.json_loss": 1.645748496055603, "eval_alpaca_data_gpt4_chinese.json_runtime": 0.0515, "eval_alpaca_data_gpt4_chinese.json_samples_per_second": 310.905, "eval_alpaca_data_gpt4_chinese.json_steps_per_second": 19.432, "step": 400 }, { "epoch": 0.11007533280588903, "eval_alpaca_gpt4_zh.json_loss": 0.9886136651039124, "eval_alpaca_gpt4_zh.json_runtime": 0.0503, "eval_alpaca_gpt4_zh.json_samples_per_second": 218.827, "eval_alpaca_gpt4_zh.json_steps_per_second": 19.893, "step": 400 }, { "epoch": 0.11007533280588903, "eval_codefeedback_filtered_instruction.json_loss": 0.5969922542572021, "eval_codefeedback_filtered_instruction.json_runtime": 0.4851, "eval_codefeedback_filtered_instruction.json_samples_per_second": 41.226, "eval_codefeedback_filtered_instruction.json_steps_per_second": 2.061, "step": 400 }, { "epoch": 0.11062570946991848, "grad_norm": 0.40487441420555115, "learning_rate": 1e-05, "loss": 0.4811, "step": 402 }, { "epoch": 0.11117608613394792, "grad_norm": 0.39143064618110657, "learning_rate": 1e-05, "loss": 0.4704, "step": 404 }, { "epoch": 0.11172646279797736, "grad_norm": 0.46816787123680115, "learning_rate": 1e-05, "loss": 0.4941, "step": 406 }, { "epoch": 0.11227683946200681, "grad_norm": 0.37707188725471497, "learning_rate": 1e-05, "loss": 0.4839, "step": 408 }, { "epoch": 0.11282721612603626, "grad_norm": 0.3780951201915741, "learning_rate": 1e-05, "loss": 0.4889, "step": 410 }, { "epoch": 0.1133775927900657, "grad_norm": 0.36941519379615784, "learning_rate": 1e-05, "loss": 0.4747, "step": 412 }, { "epoch": 0.11392796945409514, "grad_norm": 0.39626002311706543, "learning_rate": 1e-05, "loss": 0.4872, "step": 414 }, { "epoch": 0.11447834611812459, "grad_norm": 0.38315075635910034, "learning_rate": 1e-05, "loss": 0.471, "step": 416 }, { "epoch": 0.11502872278215404, "grad_norm": 0.37200862169265747, "learning_rate": 1e-05, "loss": 0.4891, "step": 418 }, { "epoch": 0.11557909944618348, "grad_norm": 0.39199399948120117, "learning_rate": 1e-05, "loss": 0.4807, "step": 420 }, { "epoch": 0.11612947611021293, "grad_norm": 0.37726107239723206, "learning_rate": 1e-05, "loss": 0.4834, "step": 422 }, { "epoch": 0.11667985277424237, "grad_norm": 0.38188016414642334, "learning_rate": 1e-05, "loss": 0.4853, "step": 424 }, { "epoch": 0.11723022943827181, "grad_norm": 0.39772850275039673, "learning_rate": 1e-05, "loss": 0.4895, "step": 426 }, { "epoch": 0.11778060610230126, "grad_norm": 0.3797503411769867, "learning_rate": 1e-05, "loss": 0.4818, "step": 428 }, { "epoch": 0.11833098276633071, "grad_norm": 0.39962416887283325, "learning_rate": 1e-05, "loss": 0.4802, "step": 430 }, { "epoch": 0.11888135943036016, "grad_norm": 0.37405237555503845, "learning_rate": 1e-05, "loss": 0.4879, "step": 432 }, { "epoch": 0.1194317360943896, "grad_norm": 0.39297720789909363, "learning_rate": 1e-05, "loss": 0.4853, "step": 434 }, { "epoch": 0.11998211275841904, "grad_norm": 0.3871022164821625, "learning_rate": 1e-05, "loss": 0.4845, "step": 436 }, { "epoch": 0.12053248942244849, "grad_norm": 0.43845734000205994, "learning_rate": 1e-05, "loss": 0.4865, "step": 438 }, { "epoch": 0.12108286608647793, "grad_norm": 0.3888757526874542, "learning_rate": 1e-05, "loss": 0.4862, "step": 440 }, { "epoch": 0.12163324275050738, "grad_norm": 0.3801029920578003, "learning_rate": 1e-05, "loss": 0.4751, "step": 442 }, { "epoch": 0.12218361941453683, "grad_norm": 0.3861992657184601, "learning_rate": 1e-05, "loss": 0.5026, "step": 444 }, { "epoch": 0.12273399607856626, "grad_norm": 0.40307343006134033, "learning_rate": 1e-05, "loss": 0.4901, "step": 446 }, { "epoch": 0.12328437274259571, "grad_norm": 0.36803606152534485, "learning_rate": 1e-05, "loss": 0.4927, "step": 448 }, { "epoch": 0.12383474940662516, "grad_norm": 0.40266790986061096, "learning_rate": 1e-05, "loss": 0.4663, "step": 450 }, { "epoch": 0.1243851260706546, "grad_norm": 0.3870522975921631, "learning_rate": 1e-05, "loss": 0.475, "step": 452 }, { "epoch": 0.12493550273468405, "grad_norm": 0.3978688716888428, "learning_rate": 1e-05, "loss": 0.4979, "step": 454 }, { "epoch": 0.1254858793987135, "grad_norm": 0.3799881935119629, "learning_rate": 1e-05, "loss": 0.4802, "step": 456 }, { "epoch": 0.12603625606274294, "grad_norm": 0.3795452415943146, "learning_rate": 1e-05, "loss": 0.4878, "step": 458 }, { "epoch": 0.1265866327267724, "grad_norm": 0.3865358233451843, "learning_rate": 1e-05, "loss": 0.4825, "step": 460 }, { "epoch": 0.12713700939080183, "grad_norm": 0.3646644353866577, "learning_rate": 1e-05, "loss": 0.4725, "step": 462 }, { "epoch": 0.12768738605483126, "grad_norm": 0.3851023018360138, "learning_rate": 1e-05, "loss": 0.4849, "step": 464 }, { "epoch": 0.12823776271886073, "grad_norm": 0.37587490677833557, "learning_rate": 1e-05, "loss": 0.4729, "step": 466 }, { "epoch": 0.12878813938289016, "grad_norm": 0.3559257686138153, "learning_rate": 1e-05, "loss": 0.4826, "step": 468 }, { "epoch": 0.12933851604691962, "grad_norm": 0.3967975974082947, "learning_rate": 1e-05, "loss": 0.4917, "step": 470 }, { "epoch": 0.12988889271094906, "grad_norm": 0.4064919650554657, "learning_rate": 1e-05, "loss": 0.5018, "step": 472 }, { "epoch": 0.1304392693749785, "grad_norm": 0.3609434962272644, "learning_rate": 1e-05, "loss": 0.4805, "step": 474 }, { "epoch": 0.13098964603900795, "grad_norm": 0.4229820668697357, "learning_rate": 1e-05, "loss": 0.4756, "step": 476 }, { "epoch": 0.13154002270303738, "grad_norm": 0.3882080018520355, "learning_rate": 1e-05, "loss": 0.4946, "step": 478 }, { "epoch": 0.13209039936706685, "grad_norm": 0.37811529636383057, "learning_rate": 1e-05, "loss": 0.495, "step": 480 }, { "epoch": 0.13264077603109628, "grad_norm": 0.4139231741428375, "learning_rate": 1e-05, "loss": 0.4722, "step": 482 }, { "epoch": 0.13319115269512571, "grad_norm": 0.3836536705493927, "learning_rate": 1e-05, "loss": 0.4795, "step": 484 }, { "epoch": 0.13374152935915518, "grad_norm": 0.39434006810188293, "learning_rate": 1e-05, "loss": 0.4783, "step": 486 }, { "epoch": 0.1342919060231846, "grad_norm": 0.3847144544124603, "learning_rate": 1e-05, "loss": 0.4751, "step": 488 }, { "epoch": 0.13484228268721407, "grad_norm": 0.4081107974052429, "learning_rate": 1e-05, "loss": 0.4947, "step": 490 }, { "epoch": 0.1353926593512435, "grad_norm": 0.3780671954154968, "learning_rate": 1e-05, "loss": 0.4932, "step": 492 }, { "epoch": 0.13594303601527294, "grad_norm": 0.39522022008895874, "learning_rate": 1e-05, "loss": 0.4868, "step": 494 }, { "epoch": 0.1364934126793024, "grad_norm": 0.3978594243526459, "learning_rate": 1e-05, "loss": 0.4895, "step": 496 }, { "epoch": 0.13704378934333183, "grad_norm": 0.40067028999328613, "learning_rate": 1e-05, "loss": 0.4841, "step": 498 }, { "epoch": 0.1375941660073613, "grad_norm": 0.38525891304016113, "learning_rate": 1e-05, "loss": 0.4769, "step": 500 }, { "epoch": 0.13814454267139073, "grad_norm": 0.3708615303039551, "learning_rate": 1e-05, "loss": 0.4787, "step": 502 }, { "epoch": 0.1386949193354202, "grad_norm": 0.3583269417285919, "learning_rate": 1e-05, "loss": 0.4905, "step": 504 }, { "epoch": 0.13924529599944963, "grad_norm": 0.4004143178462982, "learning_rate": 1e-05, "loss": 0.4797, "step": 506 }, { "epoch": 0.13979567266347906, "grad_norm": 0.3877711594104767, "learning_rate": 1e-05, "loss": 0.4968, "step": 508 }, { "epoch": 0.14034604932750852, "grad_norm": 0.394502729177475, "learning_rate": 1e-05, "loss": 0.4743, "step": 510 }, { "epoch": 0.14089642599153795, "grad_norm": 0.3829086720943451, "learning_rate": 1e-05, "loss": 0.4769, "step": 512 }, { "epoch": 0.14144680265556742, "grad_norm": 0.3849917948246002, "learning_rate": 1e-05, "loss": 0.4763, "step": 514 }, { "epoch": 0.14199717931959685, "grad_norm": 0.40810078382492065, "learning_rate": 1e-05, "loss": 0.4904, "step": 516 }, { "epoch": 0.14254755598362628, "grad_norm": 0.3982490599155426, "learning_rate": 1e-05, "loss": 0.4762, "step": 518 }, { "epoch": 0.14309793264765575, "grad_norm": 0.36841145157814026, "learning_rate": 1e-05, "loss": 0.4745, "step": 520 }, { "epoch": 0.14364830931168518, "grad_norm": 0.3805830180644989, "learning_rate": 1e-05, "loss": 0.4811, "step": 522 }, { "epoch": 0.14419868597571464, "grad_norm": 0.40074169635772705, "learning_rate": 1e-05, "loss": 0.4923, "step": 524 }, { "epoch": 0.14474906263974407, "grad_norm": 0.42140403389930725, "learning_rate": 1e-05, "loss": 0.4972, "step": 526 }, { "epoch": 0.1452994393037735, "grad_norm": 0.38489535450935364, "learning_rate": 1e-05, "loss": 0.4921, "step": 528 }, { "epoch": 0.14584981596780297, "grad_norm": 0.38449668884277344, "learning_rate": 1e-05, "loss": 0.4883, "step": 530 }, { "epoch": 0.1464001926318324, "grad_norm": 0.38009950518608093, "learning_rate": 1e-05, "loss": 0.4808, "step": 532 }, { "epoch": 0.14695056929586187, "grad_norm": 0.3916541337966919, "learning_rate": 1e-05, "loss": 0.4708, "step": 534 }, { "epoch": 0.1475009459598913, "grad_norm": 0.39856135845184326, "learning_rate": 1e-05, "loss": 0.4933, "step": 536 }, { "epoch": 0.14805132262392073, "grad_norm": 0.3804597556591034, "learning_rate": 1e-05, "loss": 0.4772, "step": 538 }, { "epoch": 0.1486016992879502, "grad_norm": 0.39584964513778687, "learning_rate": 1e-05, "loss": 0.4746, "step": 540 }, { "epoch": 0.14915207595197963, "grad_norm": 0.36922863125801086, "learning_rate": 1e-05, "loss": 0.4911, "step": 542 }, { "epoch": 0.1497024526160091, "grad_norm": 0.38762298226356506, "learning_rate": 1e-05, "loss": 0.4744, "step": 544 }, { "epoch": 0.15025282928003852, "grad_norm": 0.38803887367248535, "learning_rate": 1e-05, "loss": 0.4776, "step": 546 }, { "epoch": 0.15080320594406796, "grad_norm": 0.39409226179122925, "learning_rate": 1e-05, "loss": 0.4789, "step": 548 }, { "epoch": 0.15135358260809742, "grad_norm": 0.4141768217086792, "learning_rate": 1e-05, "loss": 0.4752, "step": 550 }, { "epoch": 0.15190395927212685, "grad_norm": 0.3770216703414917, "learning_rate": 1e-05, "loss": 0.4689, "step": 552 }, { "epoch": 0.15245433593615632, "grad_norm": 0.3929697573184967, "learning_rate": 1e-05, "loss": 0.4861, "step": 554 }, { "epoch": 0.15300471260018575, "grad_norm": 0.3859105706214905, "learning_rate": 1e-05, "loss": 0.4799, "step": 556 }, { "epoch": 0.1535550892642152, "grad_norm": 0.41044744849205017, "learning_rate": 1e-05, "loss": 0.4911, "step": 558 }, { "epoch": 0.15410546592824464, "grad_norm": 0.36859771609306335, "learning_rate": 1e-05, "loss": 0.4653, "step": 560 }, { "epoch": 0.15465584259227408, "grad_norm": 0.39258813858032227, "learning_rate": 1e-05, "loss": 0.4769, "step": 562 }, { "epoch": 0.15520621925630354, "grad_norm": 0.38241100311279297, "learning_rate": 1e-05, "loss": 0.4821, "step": 564 }, { "epoch": 0.15575659592033297, "grad_norm": 0.4107513427734375, "learning_rate": 1e-05, "loss": 0.4746, "step": 566 }, { "epoch": 0.15630697258436244, "grad_norm": 0.3872488737106323, "learning_rate": 1e-05, "loss": 0.4817, "step": 568 }, { "epoch": 0.15685734924839187, "grad_norm": 0.3712390065193176, "learning_rate": 1e-05, "loss": 0.4944, "step": 570 }, { "epoch": 0.1574077259124213, "grad_norm": 0.413503497838974, "learning_rate": 1e-05, "loss": 0.4774, "step": 572 }, { "epoch": 0.15795810257645077, "grad_norm": 0.35706543922424316, "learning_rate": 1e-05, "loss": 0.4743, "step": 574 }, { "epoch": 0.1585084792404802, "grad_norm": 0.39815768599510193, "learning_rate": 1e-05, "loss": 0.4846, "step": 576 }, { "epoch": 0.15905885590450966, "grad_norm": 0.38346678018569946, "learning_rate": 1e-05, "loss": 0.4633, "step": 578 }, { "epoch": 0.1596092325685391, "grad_norm": 0.3905611038208008, "learning_rate": 1e-05, "loss": 0.4776, "step": 580 }, { "epoch": 0.16015960923256853, "grad_norm": 0.3790382742881775, "learning_rate": 1e-05, "loss": 0.4892, "step": 582 }, { "epoch": 0.160709985896598, "grad_norm": 0.37033775448799133, "learning_rate": 1e-05, "loss": 0.4848, "step": 584 }, { "epoch": 0.16126036256062742, "grad_norm": 0.3686079680919647, "learning_rate": 1e-05, "loss": 0.4514, "step": 586 }, { "epoch": 0.16181073922465689, "grad_norm": 0.3836509883403778, "learning_rate": 1e-05, "loss": 0.4859, "step": 588 }, { "epoch": 0.16236111588868632, "grad_norm": 0.40387076139450073, "learning_rate": 1e-05, "loss": 0.485, "step": 590 }, { "epoch": 0.16291149255271575, "grad_norm": 0.3850373327732086, "learning_rate": 1e-05, "loss": 0.4843, "step": 592 }, { "epoch": 0.16346186921674521, "grad_norm": 0.3814505934715271, "learning_rate": 1e-05, "loss": 0.4749, "step": 594 }, { "epoch": 0.16401224588077465, "grad_norm": 0.35501739382743835, "learning_rate": 1e-05, "loss": 0.4645, "step": 596 }, { "epoch": 0.1645626225448041, "grad_norm": 0.34997090697288513, "learning_rate": 1e-05, "loss": 0.4687, "step": 598 }, { "epoch": 0.16511299920883354, "grad_norm": 0.365212619304657, "learning_rate": 1e-05, "loss": 0.4956, "step": 600 }, { "epoch": 0.16511299920883354, "eval_merge_loss": 0.4216049909591675, "eval_merge_runtime": 600.0688, "eval_merge_samples_per_second": 56.247, "eval_merge_steps_per_second": 2.345, "step": 600 }, { "epoch": 0.16511299920883354, "eval_new_aug_datas_filtered.json_loss": 0.5500591397285461, "eval_new_aug_datas_filtered.json_runtime": 10.5295, "eval_new_aug_datas_filtered.json_samples_per_second": 72.843, "eval_new_aug_datas_filtered.json_steps_per_second": 3.039, "step": 600 }, { "epoch": 0.16511299920883354, "eval_sharegpt_gpt4.json_loss": 0.8046284914016724, "eval_sharegpt_gpt4.json_runtime": 31.7366, "eval_sharegpt_gpt4.json_samples_per_second": 58.639, "eval_sharegpt_gpt4.json_steps_per_second": 2.458, "step": 600 }, { "epoch": 0.16511299920883354, "eval_Table_GPT.json_loss": 0.07981107383966446, "eval_Table_GPT.json_runtime": 25.0085, "eval_Table_GPT.json_samples_per_second": 83.691, "eval_Table_GPT.json_steps_per_second": 3.519, "step": 600 }, { "epoch": 0.16511299920883354, "eval_gpt_4o_200k.json_loss": 0.8323716521263123, "eval_gpt_4o_200k.json_runtime": 48.5988, "eval_gpt_4o_200k.json_samples_per_second": 129.242, "eval_gpt_4o_200k.json_steps_per_second": 5.391, "step": 600 }, { "epoch": 0.16511299920883354, "eval_multi_turn_datas.json_loss": 0.36492469906806946, "eval_multi_turn_datas.json_runtime": 75.8696, "eval_multi_turn_datas.json_samples_per_second": 52.748, "eval_multi_turn_datas.json_steps_per_second": 2.201, "step": 600 }, { "epoch": 0.16511299920883354, "eval_table_python_code_datas.json_loss": 0.29984766244888306, "eval_table_python_code_datas.json_runtime": 43.1945, "eval_table_python_code_datas.json_samples_per_second": 49.983, "eval_table_python_code_datas.json_steps_per_second": 2.084, "step": 600 }, { "epoch": 0.16511299920883354, "eval_tabular_llm_data.json_loss": 0.13250145316123962, "eval_tabular_llm_data.json_runtime": 8.5476, "eval_tabular_llm_data.json_samples_per_second": 28.78, "eval_tabular_llm_data.json_steps_per_second": 1.287, "step": 600 }, { "epoch": 0.16511299920883354, "eval_python_code_critic_21k.json_loss": 0.599878191947937, "eval_python_code_critic_21k.json_runtime": 3.2358, "eval_python_code_critic_21k.json_samples_per_second": 184.496, "eval_python_code_critic_21k.json_steps_per_second": 7.726, "step": 600 }, { "epoch": 0.16511299920883354, "eval_all_merge_table_dataset.json_loss": 0.08627181500196457, "eval_all_merge_table_dataset.json_runtime": 23.3808, "eval_all_merge_table_dataset.json_samples_per_second": 30.452, "eval_all_merge_table_dataset.json_steps_per_second": 1.283, "step": 600 }, { "epoch": 0.16511299920883354, "eval_code_feedback_multi_turn.json_loss": 0.5982062220573425, "eval_code_feedback_multi_turn.json_runtime": 32.4617, "eval_code_feedback_multi_turn.json_samples_per_second": 67.803, "eval_code_feedback_multi_turn.json_steps_per_second": 2.834, "step": 600 }, { "epoch": 0.16511299920883354, "eval_ultrainteract_sft.json_loss": 0.4367137849330902, "eval_ultrainteract_sft.json_runtime": 8.672, "eval_ultrainteract_sft.json_samples_per_second": 167.896, "eval_ultrainteract_sft.json_steps_per_second": 7.034, "step": 600 }, { "epoch": 0.16511299920883354, "eval_synthetic_text_to_sql.json_loss": 0.1079096570611, "eval_synthetic_text_to_sql.json_runtime": 0.1265, "eval_synthetic_text_to_sql.json_samples_per_second": 268.769, "eval_synthetic_text_to_sql.json_steps_per_second": 15.81, "step": 600 }, { "epoch": 0.16511299920883354, "eval_sft_react_sql_datas.json_loss": 0.6610473990440369, "eval_sft_react_sql_datas.json_runtime": 7.8536, "eval_sft_react_sql_datas.json_samples_per_second": 39.982, "eval_sft_react_sql_datas.json_steps_per_second": 1.783, "step": 600 }, { "epoch": 0.16511299920883354, "eval_all_merge_code.json_loss": 0.3021065890789032, "eval_all_merge_code.json_runtime": 0.3377, "eval_all_merge_code.json_samples_per_second": 186.571, "eval_all_merge_code.json_steps_per_second": 8.884, "step": 600 }, { "epoch": 0.16511299920883354, "eval_magpie_datas.json_loss": 0.4455747604370117, "eval_magpie_datas.json_runtime": 2.2122, "eval_magpie_datas.json_samples_per_second": 77.751, "eval_magpie_datas.json_steps_per_second": 3.616, "step": 600 }, { "epoch": 0.16511299920883354, "eval_train_data_for_qwen.json_loss": 0.009937227703630924, "eval_train_data_for_qwen.json_runtime": 0.2454, "eval_train_data_for_qwen.json_samples_per_second": 40.745, "eval_train_data_for_qwen.json_steps_per_second": 4.075, "step": 600 }, { "epoch": 0.16511299920883354, "eval_alpaca_cleaned.json_loss": 0.9349167943000793, "eval_alpaca_cleaned.json_runtime": 0.1148, "eval_alpaca_cleaned.json_samples_per_second": 235.251, "eval_alpaca_cleaned.json_steps_per_second": 17.426, "step": 600 }, { "epoch": 0.16511299920883354, "eval_agent_instruct.json_loss": 0.2310038059949875, "eval_agent_instruct.json_runtime": 0.5119, "eval_agent_instruct.json_samples_per_second": 93.766, "eval_agent_instruct.json_steps_per_second": 3.907, "step": 600 }, { "epoch": 0.16511299920883354, "eval_MathInstruct.json_loss": 0.21358835697174072, "eval_MathInstruct.json_runtime": 0.3581, "eval_MathInstruct.json_samples_per_second": 159.182, "eval_MathInstruct.json_steps_per_second": 8.378, "step": 600 }, { "epoch": 0.16511299920883354, "eval_tested_143k_python_alpaca.json_loss": 0.4455429017543793, "eval_tested_143k_python_alpaca.json_runtime": 0.3013, "eval_tested_143k_python_alpaca.json_samples_per_second": 112.849, "eval_tested_143k_python_alpaca.json_steps_per_second": 6.638, "step": 600 }, { "epoch": 0.16511299920883354, "eval_xlam_function_calling_60k.json_loss": 0.00893339328467846, "eval_xlam_function_calling_60k.json_runtime": 0.1004, "eval_xlam_function_calling_60k.json_samples_per_second": 228.974, "eval_xlam_function_calling_60k.json_steps_per_second": 9.955, "step": 600 }, { "epoch": 0.16511299920883354, "eval_alpaca_data_gpt4_chinese.json_loss": 1.6295539140701294, "eval_alpaca_data_gpt4_chinese.json_runtime": 0.0512, "eval_alpaca_data_gpt4_chinese.json_samples_per_second": 312.613, "eval_alpaca_data_gpt4_chinese.json_steps_per_second": 19.538, "step": 600 }, { "epoch": 0.16511299920883354, "eval_alpaca_gpt4_zh.json_loss": 0.9761592745780945, "eval_alpaca_gpt4_zh.json_runtime": 0.0499, "eval_alpaca_gpt4_zh.json_samples_per_second": 220.289, "eval_alpaca_gpt4_zh.json_steps_per_second": 20.026, "step": 600 }, { "epoch": 0.16511299920883354, "eval_codefeedback_filtered_instruction.json_loss": 0.5956905484199524, "eval_codefeedback_filtered_instruction.json_runtime": 0.4851, "eval_codefeedback_filtered_instruction.json_samples_per_second": 41.232, "eval_codefeedback_filtered_instruction.json_steps_per_second": 2.062, "step": 600 }, { "epoch": 0.16566337587286298, "grad_norm": 0.38436150550842285, "learning_rate": 1e-05, "loss": 0.4609, "step": 602 }, { "epoch": 0.16621375253689244, "grad_norm": 0.3946292996406555, "learning_rate": 1e-05, "loss": 0.4699, "step": 604 }, { "epoch": 0.16676412920092187, "grad_norm": 0.4069615304470062, "learning_rate": 1e-05, "loss": 0.4722, "step": 606 }, { "epoch": 0.16731450586495134, "grad_norm": 0.371660977602005, "learning_rate": 1e-05, "loss": 0.4856, "step": 608 }, { "epoch": 0.16786488252898077, "grad_norm": 0.394911527633667, "learning_rate": 1e-05, "loss": 0.4804, "step": 610 }, { "epoch": 0.1684152591930102, "grad_norm": 0.4873884916305542, "learning_rate": 1e-05, "loss": 0.4686, "step": 612 }, { "epoch": 0.16896563585703966, "grad_norm": 0.3943842649459839, "learning_rate": 1e-05, "loss": 0.4887, "step": 614 }, { "epoch": 0.1695160125210691, "grad_norm": 0.3716658055782318, "learning_rate": 1e-05, "loss": 0.4898, "step": 616 }, { "epoch": 0.17006638918509856, "grad_norm": 0.36271047592163086, "learning_rate": 1e-05, "loss": 0.4861, "step": 618 }, { "epoch": 0.170616765849128, "grad_norm": 0.3833015263080597, "learning_rate": 1e-05, "loss": 0.4814, "step": 620 }, { "epoch": 0.17116714251315746, "grad_norm": 0.3661365211009979, "learning_rate": 1e-05, "loss": 0.4873, "step": 622 }, { "epoch": 0.1717175191771869, "grad_norm": 0.3613869845867157, "learning_rate": 1e-05, "loss": 0.4537, "step": 624 }, { "epoch": 0.17226789584121632, "grad_norm": 0.34498724341392517, "learning_rate": 1e-05, "loss": 0.483, "step": 626 }, { "epoch": 0.17281827250524578, "grad_norm": 0.41466256976127625, "learning_rate": 1e-05, "loss": 0.4765, "step": 628 }, { "epoch": 0.17336864916927522, "grad_norm": 0.36220455169677734, "learning_rate": 1e-05, "loss": 0.4842, "step": 630 }, { "epoch": 0.17391902583330468, "grad_norm": 0.38009753823280334, "learning_rate": 1e-05, "loss": 0.482, "step": 632 }, { "epoch": 0.17446940249733411, "grad_norm": 0.3589475452899933, "learning_rate": 1e-05, "loss": 0.4714, "step": 634 }, { "epoch": 0.17501977916136355, "grad_norm": 0.37625178694725037, "learning_rate": 1e-05, "loss": 0.4487, "step": 636 }, { "epoch": 0.175570155825393, "grad_norm": 0.3818652331829071, "learning_rate": 1e-05, "loss": 0.4757, "step": 638 }, { "epoch": 0.17612053248942244, "grad_norm": 0.39498913288116455, "learning_rate": 1e-05, "loss": 0.4879, "step": 640 }, { "epoch": 0.1766709091534519, "grad_norm": 0.3864663243293762, "learning_rate": 1e-05, "loss": 0.4815, "step": 642 }, { "epoch": 0.17722128581748134, "grad_norm": 0.37452608346939087, "learning_rate": 1e-05, "loss": 0.4773, "step": 644 }, { "epoch": 0.17777166248151077, "grad_norm": 0.3754761219024658, "learning_rate": 1e-05, "loss": 0.4916, "step": 646 }, { "epoch": 0.17832203914554023, "grad_norm": 0.3797055780887604, "learning_rate": 1e-05, "loss": 0.4663, "step": 648 }, { "epoch": 0.17887241580956967, "grad_norm": 0.3640367090702057, "learning_rate": 1e-05, "loss": 0.4737, "step": 650 }, { "epoch": 0.17942279247359913, "grad_norm": 0.35961100459098816, "learning_rate": 1e-05, "loss": 0.4757, "step": 652 }, { "epoch": 0.17997316913762856, "grad_norm": 0.40443646907806396, "learning_rate": 1e-05, "loss": 0.4789, "step": 654 }, { "epoch": 0.180523545801658, "grad_norm": 0.35993334650993347, "learning_rate": 1e-05, "loss": 0.4902, "step": 656 }, { "epoch": 0.18107392246568746, "grad_norm": 0.3933318853378296, "learning_rate": 1e-05, "loss": 0.4726, "step": 658 }, { "epoch": 0.1816242991297169, "grad_norm": 0.3923085033893585, "learning_rate": 1e-05, "loss": 0.4714, "step": 660 }, { "epoch": 0.18217467579374635, "grad_norm": 0.37387627363204956, "learning_rate": 1e-05, "loss": 0.478, "step": 662 }, { "epoch": 0.1827250524577758, "grad_norm": 0.3787866532802582, "learning_rate": 1e-05, "loss": 0.4849, "step": 664 }, { "epoch": 0.18327542912180522, "grad_norm": 0.39361730217933655, "learning_rate": 1e-05, "loss": 0.4836, "step": 666 }, { "epoch": 0.18382580578583468, "grad_norm": 0.37430262565612793, "learning_rate": 1e-05, "loss": 0.4876, "step": 668 }, { "epoch": 0.18437618244986412, "grad_norm": 0.3914833068847656, "learning_rate": 1e-05, "loss": 0.48, "step": 670 }, { "epoch": 0.18492655911389358, "grad_norm": 0.36528506875038147, "learning_rate": 1e-05, "loss": 0.4583, "step": 672 }, { "epoch": 0.185476935777923, "grad_norm": 0.3779620826244354, "learning_rate": 1e-05, "loss": 0.483, "step": 674 }, { "epoch": 0.18602731244195247, "grad_norm": 0.3712228834629059, "learning_rate": 1e-05, "loss": 0.4833, "step": 676 }, { "epoch": 0.1865776891059819, "grad_norm": 0.3959150016307831, "learning_rate": 1e-05, "loss": 0.4678, "step": 678 }, { "epoch": 0.18712806577001134, "grad_norm": 0.38113903999328613, "learning_rate": 1e-05, "loss": 0.4794, "step": 680 }, { "epoch": 0.1876784424340408, "grad_norm": 0.3872113525867462, "learning_rate": 1e-05, "loss": 0.4627, "step": 682 }, { "epoch": 0.18822881909807024, "grad_norm": 0.35678407549858093, "learning_rate": 1e-05, "loss": 0.4666, "step": 684 }, { "epoch": 0.1887791957620997, "grad_norm": 0.37833312153816223, "learning_rate": 1e-05, "loss": 0.4734, "step": 686 }, { "epoch": 0.18932957242612913, "grad_norm": 0.3900817930698395, "learning_rate": 1e-05, "loss": 0.4834, "step": 688 }, { "epoch": 0.18987994909015857, "grad_norm": 0.37114864587783813, "learning_rate": 1e-05, "loss": 0.4682, "step": 690 }, { "epoch": 0.19043032575418803, "grad_norm": 0.37264662981033325, "learning_rate": 1e-05, "loss": 0.4815, "step": 692 }, { "epoch": 0.19098070241821746, "grad_norm": 0.3758707344532013, "learning_rate": 1e-05, "loss": 0.4847, "step": 694 }, { "epoch": 0.19153107908224692, "grad_norm": 0.38832512497901917, "learning_rate": 1e-05, "loss": 0.486, "step": 696 }, { "epoch": 0.19208145574627636, "grad_norm": 0.382926344871521, "learning_rate": 1e-05, "loss": 0.4844, "step": 698 }, { "epoch": 0.1926318324103058, "grad_norm": 0.3953557312488556, "learning_rate": 1e-05, "loss": 0.472, "step": 700 }, { "epoch": 0.19318220907433525, "grad_norm": 0.36295419931411743, "learning_rate": 1e-05, "loss": 0.4792, "step": 702 }, { "epoch": 0.1937325857383647, "grad_norm": 0.35859328508377075, "learning_rate": 1e-05, "loss": 0.4665, "step": 704 }, { "epoch": 0.19428296240239415, "grad_norm": 0.3658142685890198, "learning_rate": 1e-05, "loss": 0.4724, "step": 706 }, { "epoch": 0.19483333906642358, "grad_norm": 0.3860156834125519, "learning_rate": 1e-05, "loss": 0.4803, "step": 708 }, { "epoch": 0.19538371573045302, "grad_norm": 0.38030922412872314, "learning_rate": 1e-05, "loss": 0.4692, "step": 710 }, { "epoch": 0.19593409239448248, "grad_norm": 0.417516827583313, "learning_rate": 1e-05, "loss": 0.4833, "step": 712 }, { "epoch": 0.1964844690585119, "grad_norm": 0.39626750349998474, "learning_rate": 1e-05, "loss": 0.4808, "step": 714 }, { "epoch": 0.19703484572254137, "grad_norm": 0.3886042535305023, "learning_rate": 1e-05, "loss": 0.4716, "step": 716 }, { "epoch": 0.1975852223865708, "grad_norm": 0.3816077411174774, "learning_rate": 1e-05, "loss": 0.468, "step": 718 }, { "epoch": 0.19813559905060024, "grad_norm": 0.39385372400283813, "learning_rate": 1e-05, "loss": 0.4671, "step": 720 }, { "epoch": 0.1986859757146297, "grad_norm": 0.35457953810691833, "learning_rate": 1e-05, "loss": 0.4667, "step": 722 }, { "epoch": 0.19923635237865914, "grad_norm": 0.39437657594680786, "learning_rate": 1e-05, "loss": 0.4637, "step": 724 }, { "epoch": 0.1997867290426886, "grad_norm": 0.41132184863090515, "learning_rate": 1e-05, "loss": 0.4723, "step": 726 }, { "epoch": 0.20033710570671803, "grad_norm": 0.3640534281730652, "learning_rate": 1e-05, "loss": 0.4623, "step": 728 }, { "epoch": 0.20088748237074747, "grad_norm": 0.39893659949302673, "learning_rate": 1e-05, "loss": 0.4881, "step": 730 }, { "epoch": 0.20143785903477693, "grad_norm": 0.3677632212638855, "learning_rate": 1e-05, "loss": 0.4572, "step": 732 }, { "epoch": 0.20198823569880636, "grad_norm": 0.40594953298568726, "learning_rate": 1e-05, "loss": 0.4726, "step": 734 }, { "epoch": 0.20253861236283582, "grad_norm": 0.39571645855903625, "learning_rate": 1e-05, "loss": 0.4751, "step": 736 }, { "epoch": 0.20308898902686526, "grad_norm": 0.3569906949996948, "learning_rate": 1e-05, "loss": 0.4855, "step": 738 }, { "epoch": 0.20363936569089472, "grad_norm": 0.39166778326034546, "learning_rate": 1e-05, "loss": 0.4864, "step": 740 }, { "epoch": 0.20418974235492415, "grad_norm": 0.36861687898635864, "learning_rate": 1e-05, "loss": 0.4659, "step": 742 }, { "epoch": 0.2047401190189536, "grad_norm": 0.3691236078739166, "learning_rate": 1e-05, "loss": 0.4688, "step": 744 }, { "epoch": 0.20529049568298305, "grad_norm": 0.41912853717803955, "learning_rate": 1e-05, "loss": 0.4787, "step": 746 }, { "epoch": 0.20584087234701248, "grad_norm": 0.4022221565246582, "learning_rate": 1e-05, "loss": 0.4758, "step": 748 }, { "epoch": 0.20639124901104194, "grad_norm": 0.402567595243454, "learning_rate": 1e-05, "loss": 0.4766, "step": 750 }, { "epoch": 0.20694162567507138, "grad_norm": 0.3741600811481476, "learning_rate": 1e-05, "loss": 0.4833, "step": 752 }, { "epoch": 0.2074920023391008, "grad_norm": 0.3958164155483246, "learning_rate": 1e-05, "loss": 0.4786, "step": 754 }, { "epoch": 0.20804237900313027, "grad_norm": 0.37908801436424255, "learning_rate": 1e-05, "loss": 0.4715, "step": 756 }, { "epoch": 0.2085927556671597, "grad_norm": 0.38426473736763, "learning_rate": 1e-05, "loss": 0.4874, "step": 758 }, { "epoch": 0.20914313233118917, "grad_norm": 0.3873310983181, "learning_rate": 1e-05, "loss": 0.4677, "step": 760 }, { "epoch": 0.2096935089952186, "grad_norm": 0.4033788740634918, "learning_rate": 1e-05, "loss": 0.4815, "step": 762 }, { "epoch": 0.21024388565924804, "grad_norm": 0.40875962376594543, "learning_rate": 1e-05, "loss": 0.4879, "step": 764 }, { "epoch": 0.2107942623232775, "grad_norm": 0.38724496960639954, "learning_rate": 1e-05, "loss": 0.4646, "step": 766 }, { "epoch": 0.21134463898730693, "grad_norm": 0.39307013154029846, "learning_rate": 1e-05, "loss": 0.4762, "step": 768 }, { "epoch": 0.2118950156513364, "grad_norm": 0.37346333265304565, "learning_rate": 1e-05, "loss": 0.4775, "step": 770 }, { "epoch": 0.21244539231536583, "grad_norm": 0.3753449618816376, "learning_rate": 1e-05, "loss": 0.4707, "step": 772 }, { "epoch": 0.21299576897939526, "grad_norm": 0.3829357922077179, "learning_rate": 1e-05, "loss": 0.4825, "step": 774 }, { "epoch": 0.21354614564342472, "grad_norm": 0.35514822602272034, "learning_rate": 1e-05, "loss": 0.4779, "step": 776 }, { "epoch": 0.21409652230745416, "grad_norm": 0.3714098036289215, "learning_rate": 1e-05, "loss": 0.4598, "step": 778 }, { "epoch": 0.21464689897148362, "grad_norm": 0.3754241168498993, "learning_rate": 1e-05, "loss": 0.4736, "step": 780 }, { "epoch": 0.21519727563551305, "grad_norm": 0.36637604236602783, "learning_rate": 1e-05, "loss": 0.4652, "step": 782 }, { "epoch": 0.2157476522995425, "grad_norm": 0.367357075214386, "learning_rate": 1e-05, "loss": 0.466, "step": 784 }, { "epoch": 0.21629802896357195, "grad_norm": 0.3747154176235199, "learning_rate": 1e-05, "loss": 0.4668, "step": 786 }, { "epoch": 0.21684840562760138, "grad_norm": 0.3824009895324707, "learning_rate": 1e-05, "loss": 0.4748, "step": 788 }, { "epoch": 0.21739878229163084, "grad_norm": 0.385030061006546, "learning_rate": 1e-05, "loss": 0.4882, "step": 790 }, { "epoch": 0.21794915895566028, "grad_norm": 0.35460343956947327, "learning_rate": 1e-05, "loss": 0.4664, "step": 792 }, { "epoch": 0.21849953561968974, "grad_norm": 0.3792308270931244, "learning_rate": 1e-05, "loss": 0.4874, "step": 794 }, { "epoch": 0.21904991228371917, "grad_norm": 0.37190011143684387, "learning_rate": 1e-05, "loss": 0.4838, "step": 796 }, { "epoch": 0.2196002889477486, "grad_norm": 0.3757864832878113, "learning_rate": 1e-05, "loss": 0.4538, "step": 798 }, { "epoch": 0.22015066561177807, "grad_norm": 0.3677947223186493, "learning_rate": 1e-05, "loss": 0.4605, "step": 800 }, { "epoch": 0.22015066561177807, "eval_merge_loss": 0.4150216579437256, "eval_merge_runtime": 600.1194, "eval_merge_samples_per_second": 56.242, "eval_merge_steps_per_second": 2.345, "step": 800 }, { "epoch": 0.22015066561177807, "eval_new_aug_datas_filtered.json_loss": 0.5434484481811523, "eval_new_aug_datas_filtered.json_runtime": 10.4424, "eval_new_aug_datas_filtered.json_samples_per_second": 73.45, "eval_new_aug_datas_filtered.json_steps_per_second": 3.064, "step": 800 }, { "epoch": 0.22015066561177807, "eval_sharegpt_gpt4.json_loss": 0.7981637716293335, "eval_sharegpt_gpt4.json_runtime": 31.6015, "eval_sharegpt_gpt4.json_samples_per_second": 58.89, "eval_sharegpt_gpt4.json_steps_per_second": 2.468, "step": 800 }, { "epoch": 0.22015066561177807, "eval_Table_GPT.json_loss": 0.0783885195851326, "eval_Table_GPT.json_runtime": 24.9448, "eval_Table_GPT.json_samples_per_second": 83.905, "eval_Table_GPT.json_steps_per_second": 3.528, "step": 800 }, { "epoch": 0.22015066561177807, "eval_gpt_4o_200k.json_loss": 0.8245088458061218, "eval_gpt_4o_200k.json_runtime": 48.4135, "eval_gpt_4o_200k.json_samples_per_second": 129.737, "eval_gpt_4o_200k.json_steps_per_second": 5.412, "step": 800 }, { "epoch": 0.22015066561177807, "eval_multi_turn_datas.json_loss": 0.35650402307510376, "eval_multi_turn_datas.json_runtime": 75.5012, "eval_multi_turn_datas.json_samples_per_second": 53.006, "eval_multi_turn_datas.json_steps_per_second": 2.212, "step": 800 }, { "epoch": 0.22015066561177807, "eval_table_python_code_datas.json_loss": 0.2912423610687256, "eval_table_python_code_datas.json_runtime": 43.0138, "eval_table_python_code_datas.json_samples_per_second": 50.193, "eval_table_python_code_datas.json_steps_per_second": 2.092, "step": 800 }, { "epoch": 0.22015066561177807, "eval_tabular_llm_data.json_loss": 0.11931464821100235, "eval_tabular_llm_data.json_runtime": 8.524, "eval_tabular_llm_data.json_samples_per_second": 28.86, "eval_tabular_llm_data.json_steps_per_second": 1.29, "step": 800 }, { "epoch": 0.22015066561177807, "eval_python_code_critic_21k.json_loss": 0.5899094343185425, "eval_python_code_critic_21k.json_runtime": 3.2108, "eval_python_code_critic_21k.json_samples_per_second": 185.935, "eval_python_code_critic_21k.json_steps_per_second": 7.786, "step": 800 }, { "epoch": 0.22015066561177807, "eval_all_merge_table_dataset.json_loss": 0.08210163563489914, "eval_all_merge_table_dataset.json_runtime": 23.2334, "eval_all_merge_table_dataset.json_samples_per_second": 30.646, "eval_all_merge_table_dataset.json_steps_per_second": 1.291, "step": 800 }, { "epoch": 0.22015066561177807, "eval_code_feedback_multi_turn.json_loss": 0.5942392349243164, "eval_code_feedback_multi_turn.json_runtime": 32.3672, "eval_code_feedback_multi_turn.json_samples_per_second": 68.001, "eval_code_feedback_multi_turn.json_steps_per_second": 2.842, "step": 800 }, { "epoch": 0.22015066561177807, "eval_ultrainteract_sft.json_loss": 0.43230774998664856, "eval_ultrainteract_sft.json_runtime": 8.6469, "eval_ultrainteract_sft.json_samples_per_second": 168.384, "eval_ultrainteract_sft.json_steps_per_second": 7.055, "step": 800 }, { "epoch": 0.22015066561177807, "eval_synthetic_text_to_sql.json_loss": 0.10562511533498764, "eval_synthetic_text_to_sql.json_runtime": 0.1256, "eval_synthetic_text_to_sql.json_samples_per_second": 270.776, "eval_synthetic_text_to_sql.json_steps_per_second": 15.928, "step": 800 }, { "epoch": 0.22015066561177807, "eval_sft_react_sql_datas.json_loss": 0.6536443829536438, "eval_sft_react_sql_datas.json_runtime": 7.8424, "eval_sft_react_sql_datas.json_samples_per_second": 40.039, "eval_sft_react_sql_datas.json_steps_per_second": 1.785, "step": 800 }, { "epoch": 0.22015066561177807, "eval_all_merge_code.json_loss": 0.2989647090435028, "eval_all_merge_code.json_runtime": 0.3335, "eval_all_merge_code.json_samples_per_second": 188.9, "eval_all_merge_code.json_steps_per_second": 8.995, "step": 800 }, { "epoch": 0.22015066561177807, "eval_magpie_datas.json_loss": 0.4389919340610504, "eval_magpie_datas.json_runtime": 2.209, "eval_magpie_datas.json_samples_per_second": 77.862, "eval_magpie_datas.json_steps_per_second": 3.621, "step": 800 }, { "epoch": 0.22015066561177807, "eval_train_data_for_qwen.json_loss": 0.0057810284197330475, "eval_train_data_for_qwen.json_runtime": 0.2434, "eval_train_data_for_qwen.json_samples_per_second": 41.087, "eval_train_data_for_qwen.json_steps_per_second": 4.109, "step": 800 }, { "epoch": 0.22015066561177807, "eval_alpaca_cleaned.json_loss": 0.9368440508842468, "eval_alpaca_cleaned.json_runtime": 0.1149, "eval_alpaca_cleaned.json_samples_per_second": 234.893, "eval_alpaca_cleaned.json_steps_per_second": 17.399, "step": 800 }, { "epoch": 0.22015066561177807, "eval_agent_instruct.json_loss": 0.2261410802602768, "eval_agent_instruct.json_runtime": 0.5137, "eval_agent_instruct.json_samples_per_second": 93.432, "eval_agent_instruct.json_steps_per_second": 3.893, "step": 800 }, { "epoch": 0.22015066561177807, "eval_MathInstruct.json_loss": 0.208473339676857, "eval_MathInstruct.json_runtime": 0.3639, "eval_MathInstruct.json_samples_per_second": 156.645, "eval_MathInstruct.json_steps_per_second": 8.244, "step": 800 }, { "epoch": 0.22015066561177807, "eval_tested_143k_python_alpaca.json_loss": 0.44293999671936035, "eval_tested_143k_python_alpaca.json_runtime": 0.2997, "eval_tested_143k_python_alpaca.json_samples_per_second": 113.456, "eval_tested_143k_python_alpaca.json_steps_per_second": 6.674, "step": 800 }, { "epoch": 0.22015066561177807, "eval_xlam_function_calling_60k.json_loss": 0.010015022940933704, "eval_xlam_function_calling_60k.json_runtime": 0.1001, "eval_xlam_function_calling_60k.json_samples_per_second": 229.814, "eval_xlam_function_calling_60k.json_steps_per_second": 9.992, "step": 800 }, { "epoch": 0.22015066561177807, "eval_alpaca_data_gpt4_chinese.json_loss": 1.591582179069519, "eval_alpaca_data_gpt4_chinese.json_runtime": 0.0515, "eval_alpaca_data_gpt4_chinese.json_samples_per_second": 310.705, "eval_alpaca_data_gpt4_chinese.json_steps_per_second": 19.419, "step": 800 }, { "epoch": 0.22015066561177807, "eval_alpaca_gpt4_zh.json_loss": 0.9911380410194397, "eval_alpaca_gpt4_zh.json_runtime": 0.0498, "eval_alpaca_gpt4_zh.json_samples_per_second": 221.019, "eval_alpaca_gpt4_zh.json_steps_per_second": 20.093, "step": 800 }, { "epoch": 0.22015066561177807, "eval_codefeedback_filtered_instruction.json_loss": 0.5947377681732178, "eval_codefeedback_filtered_instruction.json_runtime": 0.485, "eval_codefeedback_filtered_instruction.json_samples_per_second": 41.237, "eval_codefeedback_filtered_instruction.json_steps_per_second": 2.062, "step": 800 }, { "epoch": 0.2207010422758075, "grad_norm": 0.36098968982696533, "learning_rate": 1e-05, "loss": 0.4646, "step": 802 }, { "epoch": 0.22125141893983696, "grad_norm": 0.3653786778450012, "learning_rate": 1e-05, "loss": 0.4501, "step": 804 }, { "epoch": 0.2218017956038664, "grad_norm": 0.36442849040031433, "learning_rate": 1e-05, "loss": 0.4686, "step": 806 }, { "epoch": 0.22235217226789583, "grad_norm": 0.3782612383365631, "learning_rate": 1e-05, "loss": 0.4598, "step": 808 }, { "epoch": 0.2229025489319253, "grad_norm": 0.39521896839141846, "learning_rate": 1e-05, "loss": 0.4679, "step": 810 }, { "epoch": 0.22345292559595473, "grad_norm": 0.3727470636367798, "learning_rate": 1e-05, "loss": 0.4803, "step": 812 }, { "epoch": 0.2240033022599842, "grad_norm": 0.3883068263530731, "learning_rate": 1e-05, "loss": 0.4773, "step": 814 }, { "epoch": 0.22455367892401362, "grad_norm": 0.37147605419158936, "learning_rate": 1e-05, "loss": 0.4825, "step": 816 }, { "epoch": 0.22510405558804306, "grad_norm": 0.3924333155155182, "learning_rate": 1e-05, "loss": 0.4698, "step": 818 }, { "epoch": 0.22565443225207252, "grad_norm": 0.38133057951927185, "learning_rate": 1e-05, "loss": 0.4842, "step": 820 }, { "epoch": 0.22620480891610195, "grad_norm": 0.36132821440696716, "learning_rate": 1e-05, "loss": 0.4594, "step": 822 }, { "epoch": 0.2267551855801314, "grad_norm": 0.39988580346107483, "learning_rate": 1e-05, "loss": 0.4795, "step": 824 }, { "epoch": 0.22730556224416085, "grad_norm": 0.38140830397605896, "learning_rate": 1e-05, "loss": 0.4649, "step": 826 }, { "epoch": 0.22785593890819028, "grad_norm": 0.3726978898048401, "learning_rate": 1e-05, "loss": 0.4603, "step": 828 }, { "epoch": 0.22840631557221974, "grad_norm": 0.3880995512008667, "learning_rate": 1e-05, "loss": 0.4739, "step": 830 }, { "epoch": 0.22895669223624918, "grad_norm": 0.4118787944316864, "learning_rate": 1e-05, "loss": 0.4733, "step": 832 }, { "epoch": 0.22950706890027864, "grad_norm": 0.37878745794296265, "learning_rate": 1e-05, "loss": 0.4922, "step": 834 }, { "epoch": 0.23005744556430807, "grad_norm": 0.3838474154472351, "learning_rate": 1e-05, "loss": 0.4646, "step": 836 }, { "epoch": 0.2306078222283375, "grad_norm": 0.37345945835113525, "learning_rate": 1e-05, "loss": 0.4798, "step": 838 }, { "epoch": 0.23115819889236697, "grad_norm": 0.36341801285743713, "learning_rate": 1e-05, "loss": 0.4773, "step": 840 }, { "epoch": 0.2317085755563964, "grad_norm": 0.38800522685050964, "learning_rate": 1e-05, "loss": 0.4806, "step": 842 }, { "epoch": 0.23225895222042586, "grad_norm": 0.38882526755332947, "learning_rate": 1e-05, "loss": 0.4765, "step": 844 }, { "epoch": 0.2328093288844553, "grad_norm": 0.37744489312171936, "learning_rate": 1e-05, "loss": 0.4842, "step": 846 }, { "epoch": 0.23335970554848473, "grad_norm": 0.39916718006134033, "learning_rate": 1e-05, "loss": 0.467, "step": 848 }, { "epoch": 0.2339100822125142, "grad_norm": 0.36556801199913025, "learning_rate": 1e-05, "loss": 0.4711, "step": 850 }, { "epoch": 0.23446045887654363, "grad_norm": 0.3993853032588959, "learning_rate": 1e-05, "loss": 0.4656, "step": 852 }, { "epoch": 0.2350108355405731, "grad_norm": 0.39630356431007385, "learning_rate": 1e-05, "loss": 0.4734, "step": 854 }, { "epoch": 0.23556121220460252, "grad_norm": 0.3797578513622284, "learning_rate": 1e-05, "loss": 0.4718, "step": 856 }, { "epoch": 0.23611158886863198, "grad_norm": 0.38648873567581177, "learning_rate": 1e-05, "loss": 0.4751, "step": 858 }, { "epoch": 0.23666196553266142, "grad_norm": 0.3934420347213745, "learning_rate": 1e-05, "loss": 0.4653, "step": 860 }, { "epoch": 0.23721234219669085, "grad_norm": 0.3899431824684143, "learning_rate": 1e-05, "loss": 0.4644, "step": 862 }, { "epoch": 0.2377627188607203, "grad_norm": 0.3696826696395874, "learning_rate": 1e-05, "loss": 0.482, "step": 864 }, { "epoch": 0.23831309552474975, "grad_norm": 0.352923184633255, "learning_rate": 1e-05, "loss": 0.4707, "step": 866 }, { "epoch": 0.2388634721887792, "grad_norm": 0.36678972840309143, "learning_rate": 1e-05, "loss": 0.4687, "step": 868 }, { "epoch": 0.23941384885280864, "grad_norm": 0.38986021280288696, "learning_rate": 1e-05, "loss": 0.4613, "step": 870 }, { "epoch": 0.23996422551683808, "grad_norm": 0.3684535622596741, "learning_rate": 1e-05, "loss": 0.4734, "step": 872 }, { "epoch": 0.24051460218086754, "grad_norm": 0.36672261357307434, "learning_rate": 1e-05, "loss": 0.4796, "step": 874 }, { "epoch": 0.24106497884489697, "grad_norm": 0.39910420775413513, "learning_rate": 1e-05, "loss": 0.4681, "step": 876 }, { "epoch": 0.24161535550892643, "grad_norm": 0.38694077730178833, "learning_rate": 1e-05, "loss": 0.4821, "step": 878 }, { "epoch": 0.24216573217295587, "grad_norm": 0.4555080831050873, "learning_rate": 1e-05, "loss": 0.4882, "step": 880 }, { "epoch": 0.2427161088369853, "grad_norm": 0.3934450149536133, "learning_rate": 1e-05, "loss": 0.4778, "step": 882 }, { "epoch": 0.24326648550101476, "grad_norm": 0.35743412375450134, "learning_rate": 1e-05, "loss": 0.4793, "step": 884 }, { "epoch": 0.2438168621650442, "grad_norm": 0.3518178462982178, "learning_rate": 1e-05, "loss": 0.472, "step": 886 }, { "epoch": 0.24436723882907366, "grad_norm": 0.35367751121520996, "learning_rate": 1e-05, "loss": 0.4747, "step": 888 }, { "epoch": 0.2449176154931031, "grad_norm": 0.3810805678367615, "learning_rate": 1e-05, "loss": 0.4834, "step": 890 }, { "epoch": 0.24546799215713253, "grad_norm": 0.38103243708610535, "learning_rate": 1e-05, "loss": 0.4763, "step": 892 }, { "epoch": 0.246018368821162, "grad_norm": 0.3839399218559265, "learning_rate": 1e-05, "loss": 0.4696, "step": 894 }, { "epoch": 0.24656874548519142, "grad_norm": 0.41292649507522583, "learning_rate": 1e-05, "loss": 0.4777, "step": 896 }, { "epoch": 0.24711912214922088, "grad_norm": 0.36179229617118835, "learning_rate": 1e-05, "loss": 0.4668, "step": 898 }, { "epoch": 0.24766949881325032, "grad_norm": 0.3638279139995575, "learning_rate": 1e-05, "loss": 0.4645, "step": 900 }, { "epoch": 0.24821987547727975, "grad_norm": 0.3458470106124878, "learning_rate": 1e-05, "loss": 0.4746, "step": 902 }, { "epoch": 0.2487702521413092, "grad_norm": 0.3822806775569916, "learning_rate": 1e-05, "loss": 0.4715, "step": 904 }, { "epoch": 0.24932062880533865, "grad_norm": 0.3655596077442169, "learning_rate": 1e-05, "loss": 0.4659, "step": 906 }, { "epoch": 0.2498710054693681, "grad_norm": 0.3868783116340637, "learning_rate": 1e-05, "loss": 0.4743, "step": 908 }, { "epoch": 0.25042138213339754, "grad_norm": 0.3778232932090759, "learning_rate": 1e-05, "loss": 0.4652, "step": 910 }, { "epoch": 0.250971758797427, "grad_norm": 0.36664894223213196, "learning_rate": 1e-05, "loss": 0.4554, "step": 912 }, { "epoch": 0.2515221354614564, "grad_norm": 0.3995139002799988, "learning_rate": 1e-05, "loss": 0.4683, "step": 914 }, { "epoch": 0.25207251212548587, "grad_norm": 0.40083470940589905, "learning_rate": 1e-05, "loss": 0.4673, "step": 916 }, { "epoch": 0.25262288878951533, "grad_norm": 0.37919968366622925, "learning_rate": 1e-05, "loss": 0.4776, "step": 918 }, { "epoch": 0.2531732654535448, "grad_norm": 0.3586704432964325, "learning_rate": 1e-05, "loss": 0.4792, "step": 920 }, { "epoch": 0.2537236421175742, "grad_norm": 0.3744722902774811, "learning_rate": 1e-05, "loss": 0.463, "step": 922 }, { "epoch": 0.25427401878160366, "grad_norm": 0.37209680676460266, "learning_rate": 1e-05, "loss": 0.4805, "step": 924 }, { "epoch": 0.2548243954456331, "grad_norm": 0.40809133648872375, "learning_rate": 1e-05, "loss": 0.4781, "step": 926 }, { "epoch": 0.25537477210966253, "grad_norm": 0.37261903285980225, "learning_rate": 1e-05, "loss": 0.4617, "step": 928 }, { "epoch": 0.255925148773692, "grad_norm": 0.37391313910484314, "learning_rate": 1e-05, "loss": 0.4617, "step": 930 }, { "epoch": 0.25647552543772145, "grad_norm": 0.36610838770866394, "learning_rate": 1e-05, "loss": 0.4642, "step": 932 }, { "epoch": 0.25702590210175086, "grad_norm": 0.3854142129421234, "learning_rate": 1e-05, "loss": 0.4652, "step": 934 }, { "epoch": 0.2575762787657803, "grad_norm": 0.365159809589386, "learning_rate": 1e-05, "loss": 0.4714, "step": 936 }, { "epoch": 0.2581266554298098, "grad_norm": 0.41678836941719055, "learning_rate": 1e-05, "loss": 0.4854, "step": 938 }, { "epoch": 0.25867703209383924, "grad_norm": 0.380215585231781, "learning_rate": 1e-05, "loss": 0.4785, "step": 940 }, { "epoch": 0.25922740875786865, "grad_norm": 0.3704361617565155, "learning_rate": 1e-05, "loss": 0.4433, "step": 942 }, { "epoch": 0.2597777854218981, "grad_norm": 0.34440556168556213, "learning_rate": 1e-05, "loss": 0.4642, "step": 944 }, { "epoch": 0.2603281620859276, "grad_norm": 0.36701446771621704, "learning_rate": 1e-05, "loss": 0.4533, "step": 946 }, { "epoch": 0.260878538749957, "grad_norm": 0.3694971799850464, "learning_rate": 1e-05, "loss": 0.4942, "step": 948 }, { "epoch": 0.26142891541398644, "grad_norm": 0.3697713017463684, "learning_rate": 1e-05, "loss": 0.4586, "step": 950 }, { "epoch": 0.2619792920780159, "grad_norm": 0.36559173464775085, "learning_rate": 1e-05, "loss": 0.4679, "step": 952 }, { "epoch": 0.26252966874204536, "grad_norm": 0.3704969584941864, "learning_rate": 1e-05, "loss": 0.4624, "step": 954 }, { "epoch": 0.26308004540607477, "grad_norm": 0.3804495334625244, "learning_rate": 1e-05, "loss": 0.4603, "step": 956 }, { "epoch": 0.26363042207010423, "grad_norm": 0.34987303614616394, "learning_rate": 1e-05, "loss": 0.4679, "step": 958 }, { "epoch": 0.2641807987341337, "grad_norm": 0.3723856508731842, "learning_rate": 1e-05, "loss": 0.4631, "step": 960 }, { "epoch": 0.2647311753981631, "grad_norm": 0.35623612999916077, "learning_rate": 1e-05, "loss": 0.4627, "step": 962 }, { "epoch": 0.26528155206219256, "grad_norm": 0.37969711422920227, "learning_rate": 1e-05, "loss": 0.4815, "step": 964 }, { "epoch": 0.265831928726222, "grad_norm": 0.3889734447002411, "learning_rate": 1e-05, "loss": 0.471, "step": 966 }, { "epoch": 0.26638230539025143, "grad_norm": 0.39106228947639465, "learning_rate": 1e-05, "loss": 0.4542, "step": 968 }, { "epoch": 0.2669326820542809, "grad_norm": 0.38163650035858154, "learning_rate": 1e-05, "loss": 0.4604, "step": 970 }, { "epoch": 0.26748305871831035, "grad_norm": 0.3733852505683899, "learning_rate": 1e-05, "loss": 0.4768, "step": 972 }, { "epoch": 0.2680334353823398, "grad_norm": 0.3894038796424866, "learning_rate": 1e-05, "loss": 0.4816, "step": 974 }, { "epoch": 0.2685838120463692, "grad_norm": 0.3697439432144165, "learning_rate": 1e-05, "loss": 0.4731, "step": 976 }, { "epoch": 0.2691341887103987, "grad_norm": 0.39549171924591064, "learning_rate": 1e-05, "loss": 0.473, "step": 978 }, { "epoch": 0.26968456537442814, "grad_norm": 0.38712403178215027, "learning_rate": 1e-05, "loss": 0.4717, "step": 980 }, { "epoch": 0.27023494203845755, "grad_norm": 0.3775619864463806, "learning_rate": 1e-05, "loss": 0.4638, "step": 982 }, { "epoch": 0.270785318702487, "grad_norm": 0.38664135336875916, "learning_rate": 1e-05, "loss": 0.4655, "step": 984 }, { "epoch": 0.27133569536651647, "grad_norm": 0.3730804920196533, "learning_rate": 1e-05, "loss": 0.4672, "step": 986 }, { "epoch": 0.2718860720305459, "grad_norm": 0.36626750230789185, "learning_rate": 1e-05, "loss": 0.4562, "step": 988 }, { "epoch": 0.27243644869457534, "grad_norm": 0.38708406686782837, "learning_rate": 1e-05, "loss": 0.4583, "step": 990 }, { "epoch": 0.2729868253586048, "grad_norm": 0.37348565459251404, "learning_rate": 1e-05, "loss": 0.4709, "step": 992 }, { "epoch": 0.27353720202263426, "grad_norm": 0.39145755767822266, "learning_rate": 1e-05, "loss": 0.4667, "step": 994 }, { "epoch": 0.27408757868666367, "grad_norm": 0.3615020215511322, "learning_rate": 1e-05, "loss": 0.4585, "step": 996 }, { "epoch": 0.27463795535069313, "grad_norm": 0.38545548915863037, "learning_rate": 1e-05, "loss": 0.472, "step": 998 }, { "epoch": 0.2751883320147226, "grad_norm": 0.3605005741119385, "learning_rate": 1e-05, "loss": 0.4575, "step": 1000 }, { "epoch": 0.2751883320147226, "eval_merge_loss": 0.4092504382133484, "eval_merge_runtime": 599.649, "eval_merge_samples_per_second": 56.286, "eval_merge_steps_per_second": 2.346, "step": 1000 }, { "epoch": 0.2751883320147226, "eval_new_aug_datas_filtered.json_loss": 0.534787118434906, "eval_new_aug_datas_filtered.json_runtime": 10.3465, "eval_new_aug_datas_filtered.json_samples_per_second": 74.131, "eval_new_aug_datas_filtered.json_steps_per_second": 3.093, "step": 1000 }, { "epoch": 0.2751883320147226, "eval_sharegpt_gpt4.json_loss": 0.7911589741706848, "eval_sharegpt_gpt4.json_runtime": 31.721, "eval_sharegpt_gpt4.json_samples_per_second": 58.668, "eval_sharegpt_gpt4.json_steps_per_second": 2.459, "step": 1000 }, { "epoch": 0.2751883320147226, "eval_Table_GPT.json_loss": 0.07000603526830673, "eval_Table_GPT.json_runtime": 24.9973, "eval_Table_GPT.json_samples_per_second": 83.729, "eval_Table_GPT.json_steps_per_second": 3.52, "step": 1000 }, { "epoch": 0.2751883320147226, "eval_gpt_4o_200k.json_loss": 0.8180866837501526, "eval_gpt_4o_200k.json_runtime": 48.5388, "eval_gpt_4o_200k.json_samples_per_second": 129.402, "eval_gpt_4o_200k.json_steps_per_second": 5.398, "step": 1000 }, { "epoch": 0.2751883320147226, "eval_multi_turn_datas.json_loss": 0.34955134987831116, "eval_multi_turn_datas.json_runtime": 75.86, "eval_multi_turn_datas.json_samples_per_second": 52.755, "eval_multi_turn_datas.json_steps_per_second": 2.201, "step": 1000 }, { "epoch": 0.2751883320147226, "eval_table_python_code_datas.json_loss": 0.285086989402771, "eval_table_python_code_datas.json_runtime": 43.1585, "eval_table_python_code_datas.json_samples_per_second": 50.025, "eval_table_python_code_datas.json_steps_per_second": 2.085, "step": 1000 }, { "epoch": 0.2751883320147226, "eval_tabular_llm_data.json_loss": 0.12198314070701599, "eval_tabular_llm_data.json_runtime": 8.5654, "eval_tabular_llm_data.json_samples_per_second": 28.72, "eval_tabular_llm_data.json_steps_per_second": 1.284, "step": 1000 }, { "epoch": 0.2751883320147226, "eval_python_code_critic_21k.json_loss": 0.5841899514198303, "eval_python_code_critic_21k.json_runtime": 3.2248, "eval_python_code_critic_21k.json_samples_per_second": 185.125, "eval_python_code_critic_21k.json_steps_per_second": 7.752, "step": 1000 }, { "epoch": 0.2751883320147226, "eval_all_merge_table_dataset.json_loss": 0.08110550791025162, "eval_all_merge_table_dataset.json_runtime": 23.4122, "eval_all_merge_table_dataset.json_samples_per_second": 30.411, "eval_all_merge_table_dataset.json_steps_per_second": 1.281, "step": 1000 }, { "epoch": 0.2751883320147226, "eval_code_feedback_multi_turn.json_loss": 0.5908513069152832, "eval_code_feedback_multi_turn.json_runtime": 32.4627, "eval_code_feedback_multi_turn.json_samples_per_second": 67.801, "eval_code_feedback_multi_turn.json_steps_per_second": 2.834, "step": 1000 }, { "epoch": 0.2751883320147226, "eval_ultrainteract_sft.json_loss": 0.42869675159454346, "eval_ultrainteract_sft.json_runtime": 8.6816, "eval_ultrainteract_sft.json_samples_per_second": 167.711, "eval_ultrainteract_sft.json_steps_per_second": 7.026, "step": 1000 }, { "epoch": 0.2751883320147226, "eval_synthetic_text_to_sql.json_loss": 0.10359195619821548, "eval_synthetic_text_to_sql.json_runtime": 0.1301, "eval_synthetic_text_to_sql.json_samples_per_second": 261.368, "eval_synthetic_text_to_sql.json_steps_per_second": 15.375, "step": 1000 }, { "epoch": 0.2751883320147226, "eval_sft_react_sql_datas.json_loss": 0.6493918895721436, "eval_sft_react_sql_datas.json_runtime": 7.8489, "eval_sft_react_sql_datas.json_samples_per_second": 40.006, "eval_sft_react_sql_datas.json_steps_per_second": 1.784, "step": 1000 }, { "epoch": 0.2751883320147226, "eval_all_merge_code.json_loss": 0.29959577322006226, "eval_all_merge_code.json_runtime": 0.3379, "eval_all_merge_code.json_samples_per_second": 186.458, "eval_all_merge_code.json_steps_per_second": 8.879, "step": 1000 }, { "epoch": 0.2751883320147226, "eval_magpie_datas.json_loss": 0.4377444088459015, "eval_magpie_datas.json_runtime": 2.2091, "eval_magpie_datas.json_samples_per_second": 77.86, "eval_magpie_datas.json_steps_per_second": 3.621, "step": 1000 }, { "epoch": 0.2751883320147226, "eval_train_data_for_qwen.json_loss": 0.003975613508373499, "eval_train_data_for_qwen.json_runtime": 0.2434, "eval_train_data_for_qwen.json_samples_per_second": 41.087, "eval_train_data_for_qwen.json_steps_per_second": 4.109, "step": 1000 }, { "epoch": 0.2751883320147226, "eval_alpaca_cleaned.json_loss": 0.9270830750465393, "eval_alpaca_cleaned.json_runtime": 0.1147, "eval_alpaca_cleaned.json_samples_per_second": 235.404, "eval_alpaca_cleaned.json_steps_per_second": 17.437, "step": 1000 }, { "epoch": 0.2751883320147226, "eval_agent_instruct.json_loss": 0.2235051840543747, "eval_agent_instruct.json_runtime": 0.5147, "eval_agent_instruct.json_samples_per_second": 93.255, "eval_agent_instruct.json_steps_per_second": 3.886, "step": 1000 }, { "epoch": 0.2751883320147226, "eval_MathInstruct.json_loss": 0.20924758911132812, "eval_MathInstruct.json_runtime": 0.3588, "eval_MathInstruct.json_samples_per_second": 158.853, "eval_MathInstruct.json_steps_per_second": 8.361, "step": 1000 }, { "epoch": 0.2751883320147226, "eval_tested_143k_python_alpaca.json_loss": 0.44443246722221375, "eval_tested_143k_python_alpaca.json_runtime": 0.3017, "eval_tested_143k_python_alpaca.json_samples_per_second": 112.684, "eval_tested_143k_python_alpaca.json_steps_per_second": 6.628, "step": 1000 }, { "epoch": 0.2751883320147226, "eval_xlam_function_calling_60k.json_loss": 0.008116651326417923, "eval_xlam_function_calling_60k.json_runtime": 0.1004, "eval_xlam_function_calling_60k.json_samples_per_second": 229.157, "eval_xlam_function_calling_60k.json_steps_per_second": 9.963, "step": 1000 }, { "epoch": 0.2751883320147226, "eval_alpaca_data_gpt4_chinese.json_loss": 1.588812232017517, "eval_alpaca_data_gpt4_chinese.json_runtime": 0.0516, "eval_alpaca_data_gpt4_chinese.json_samples_per_second": 310.032, "eval_alpaca_data_gpt4_chinese.json_steps_per_second": 19.377, "step": 1000 }, { "epoch": 0.2751883320147226, "eval_alpaca_gpt4_zh.json_loss": 0.9696416258811951, "eval_alpaca_gpt4_zh.json_runtime": 0.0501, "eval_alpaca_gpt4_zh.json_samples_per_second": 219.488, "eval_alpaca_gpt4_zh.json_steps_per_second": 19.953, "step": 1000 }, { "epoch": 0.2751883320147226, "eval_codefeedback_filtered_instruction.json_loss": 0.5965829491615295, "eval_codefeedback_filtered_instruction.json_runtime": 0.4872, "eval_codefeedback_filtered_instruction.json_samples_per_second": 41.049, "eval_codefeedback_filtered_instruction.json_steps_per_second": 2.052, "step": 1000 }, { "epoch": 0.275738708678752, "grad_norm": 0.3598334789276123, "learning_rate": 1e-05, "loss": 0.4624, "step": 1002 }, { "epoch": 0.27628908534278146, "grad_norm": 0.3716166615486145, "learning_rate": 1e-05, "loss": 0.4593, "step": 1004 }, { "epoch": 0.2768394620068109, "grad_norm": 0.3814164996147156, "learning_rate": 1e-05, "loss": 0.4581, "step": 1006 }, { "epoch": 0.2773898386708404, "grad_norm": 0.3595026731491089, "learning_rate": 1e-05, "loss": 0.4826, "step": 1008 }, { "epoch": 0.2779402153348698, "grad_norm": 0.3587126135826111, "learning_rate": 1e-05, "loss": 0.4563, "step": 1010 }, { "epoch": 0.27849059199889925, "grad_norm": 0.36048388481140137, "learning_rate": 1e-05, "loss": 0.4646, "step": 1012 }, { "epoch": 0.2790409686629287, "grad_norm": 0.37650784850120544, "learning_rate": 1e-05, "loss": 0.4658, "step": 1014 }, { "epoch": 0.2795913453269581, "grad_norm": 0.34934109449386597, "learning_rate": 1e-05, "loss": 0.4535, "step": 1016 }, { "epoch": 0.2801417219909876, "grad_norm": 0.375130295753479, "learning_rate": 1e-05, "loss": 0.4802, "step": 1018 }, { "epoch": 0.28069209865501704, "grad_norm": 0.3595198094844818, "learning_rate": 1e-05, "loss": 0.4736, "step": 1020 }, { "epoch": 0.28124247531904645, "grad_norm": 0.37816157937049866, "learning_rate": 1e-05, "loss": 0.4639, "step": 1022 }, { "epoch": 0.2817928519830759, "grad_norm": 0.39598193764686584, "learning_rate": 1e-05, "loss": 0.4544, "step": 1024 }, { "epoch": 0.28234322864710537, "grad_norm": 0.35407206416130066, "learning_rate": 1e-05, "loss": 0.4342, "step": 1026 }, { "epoch": 0.28289360531113483, "grad_norm": 0.3630298972129822, "learning_rate": 1e-05, "loss": 0.48, "step": 1028 }, { "epoch": 0.28344398197516424, "grad_norm": 0.35917675495147705, "learning_rate": 1e-05, "loss": 0.4647, "step": 1030 }, { "epoch": 0.2839943586391937, "grad_norm": 0.36868980526924133, "learning_rate": 1e-05, "loss": 0.4633, "step": 1032 }, { "epoch": 0.28454473530322316, "grad_norm": 0.38559168577194214, "learning_rate": 1e-05, "loss": 0.4786, "step": 1034 }, { "epoch": 0.28509511196725257, "grad_norm": 0.3563440442085266, "learning_rate": 1e-05, "loss": 0.4703, "step": 1036 }, { "epoch": 0.28564548863128203, "grad_norm": 0.3761630654335022, "learning_rate": 1e-05, "loss": 0.4712, "step": 1038 }, { "epoch": 0.2861958652953115, "grad_norm": 0.3870238661766052, "learning_rate": 1e-05, "loss": 0.4622, "step": 1040 }, { "epoch": 0.2867462419593409, "grad_norm": 0.36192306876182556, "learning_rate": 1e-05, "loss": 0.4619, "step": 1042 }, { "epoch": 0.28729661862337036, "grad_norm": 0.3688748776912689, "learning_rate": 1e-05, "loss": 0.457, "step": 1044 }, { "epoch": 0.2878469952873998, "grad_norm": 0.38211309909820557, "learning_rate": 1e-05, "loss": 0.4661, "step": 1046 }, { "epoch": 0.2883973719514293, "grad_norm": 0.36421847343444824, "learning_rate": 1e-05, "loss": 0.4647, "step": 1048 }, { "epoch": 0.2889477486154587, "grad_norm": 0.38917919993400574, "learning_rate": 1e-05, "loss": 0.4573, "step": 1050 }, { "epoch": 0.28949812527948815, "grad_norm": 0.3668692111968994, "learning_rate": 1e-05, "loss": 0.4545, "step": 1052 }, { "epoch": 0.2900485019435176, "grad_norm": 0.3869079649448395, "learning_rate": 1e-05, "loss": 0.4569, "step": 1054 }, { "epoch": 0.290598878607547, "grad_norm": 0.3763209283351898, "learning_rate": 1e-05, "loss": 0.451, "step": 1056 }, { "epoch": 0.2911492552715765, "grad_norm": 0.37899014353752136, "learning_rate": 1e-05, "loss": 0.4658, "step": 1058 }, { "epoch": 0.29169963193560594, "grad_norm": 0.38784778118133545, "learning_rate": 1e-05, "loss": 0.4589, "step": 1060 }, { "epoch": 0.2922500085996354, "grad_norm": 0.38340142369270325, "learning_rate": 1e-05, "loss": 0.4644, "step": 1062 }, { "epoch": 0.2928003852636648, "grad_norm": 0.3758372962474823, "learning_rate": 1e-05, "loss": 0.4597, "step": 1064 }, { "epoch": 0.29335076192769427, "grad_norm": 0.36990198493003845, "learning_rate": 1e-05, "loss": 0.4577, "step": 1066 }, { "epoch": 0.29390113859172373, "grad_norm": 0.35997095704078674, "learning_rate": 1e-05, "loss": 0.452, "step": 1068 }, { "epoch": 0.29445151525575314, "grad_norm": 0.3728466331958771, "learning_rate": 1e-05, "loss": 0.4567, "step": 1070 }, { "epoch": 0.2950018919197826, "grad_norm": 0.3471437990665436, "learning_rate": 1e-05, "loss": 0.4661, "step": 1072 }, { "epoch": 0.29555226858381206, "grad_norm": 0.39197105169296265, "learning_rate": 1e-05, "loss": 0.4738, "step": 1074 }, { "epoch": 0.29610264524784147, "grad_norm": 0.366745263338089, "learning_rate": 1e-05, "loss": 0.4555, "step": 1076 }, { "epoch": 0.29665302191187093, "grad_norm": 0.3721451759338379, "learning_rate": 1e-05, "loss": 0.4784, "step": 1078 }, { "epoch": 0.2972033985759004, "grad_norm": 0.3505246341228485, "learning_rate": 1e-05, "loss": 0.4486, "step": 1080 }, { "epoch": 0.29775377523992985, "grad_norm": 0.37022680044174194, "learning_rate": 1e-05, "loss": 0.4631, "step": 1082 }, { "epoch": 0.29830415190395926, "grad_norm": 0.3808286190032959, "learning_rate": 1e-05, "loss": 0.472, "step": 1084 }, { "epoch": 0.2988545285679887, "grad_norm": 0.3860435485839844, "learning_rate": 1e-05, "loss": 0.4541, "step": 1086 }, { "epoch": 0.2994049052320182, "grad_norm": 0.35552406311035156, "learning_rate": 1e-05, "loss": 0.4565, "step": 1088 }, { "epoch": 0.2999552818960476, "grad_norm": 0.3758242428302765, "learning_rate": 1e-05, "loss": 0.4803, "step": 1090 }, { "epoch": 0.30050565856007705, "grad_norm": 0.3900710940361023, "learning_rate": 1e-05, "loss": 0.4658, "step": 1092 }, { "epoch": 0.3010560352241065, "grad_norm": 0.38439512252807617, "learning_rate": 1e-05, "loss": 0.4677, "step": 1094 }, { "epoch": 0.3016064118881359, "grad_norm": 0.3970472812652588, "learning_rate": 1e-05, "loss": 0.4751, "step": 1096 }, { "epoch": 0.3021567885521654, "grad_norm": 0.36555778980255127, "learning_rate": 1e-05, "loss": 0.4556, "step": 1098 }, { "epoch": 0.30270716521619484, "grad_norm": 0.3682638108730316, "learning_rate": 1e-05, "loss": 0.4504, "step": 1100 }, { "epoch": 0.3032575418802243, "grad_norm": 0.4228995442390442, "learning_rate": 1e-05, "loss": 0.4736, "step": 1102 }, { "epoch": 0.3038079185442537, "grad_norm": 0.35070449113845825, "learning_rate": 1e-05, "loss": 0.4589, "step": 1104 }, { "epoch": 0.30435829520828317, "grad_norm": 0.40524446964263916, "learning_rate": 1e-05, "loss": 0.4616, "step": 1106 }, { "epoch": 0.30490867187231263, "grad_norm": 0.3461023271083832, "learning_rate": 1e-05, "loss": 0.4679, "step": 1108 }, { "epoch": 0.30545904853634204, "grad_norm": 0.3741723299026489, "learning_rate": 1e-05, "loss": 0.4618, "step": 1110 }, { "epoch": 0.3060094252003715, "grad_norm": 0.37440451979637146, "learning_rate": 1e-05, "loss": 0.4638, "step": 1112 }, { "epoch": 0.30655980186440096, "grad_norm": 0.34469377994537354, "learning_rate": 1e-05, "loss": 0.4426, "step": 1114 }, { "epoch": 0.3071101785284304, "grad_norm": 0.35499683022499084, "learning_rate": 1e-05, "loss": 0.4548, "step": 1116 }, { "epoch": 0.30766055519245983, "grad_norm": 0.3623688220977783, "learning_rate": 1e-05, "loss": 0.4574, "step": 1118 }, { "epoch": 0.3082109318564893, "grad_norm": 0.3487359583377838, "learning_rate": 1e-05, "loss": 0.4632, "step": 1120 }, { "epoch": 0.30876130852051875, "grad_norm": 0.36232292652130127, "learning_rate": 1e-05, "loss": 0.462, "step": 1122 }, { "epoch": 0.30931168518454816, "grad_norm": 0.38301897048950195, "learning_rate": 1e-05, "loss": 0.4545, "step": 1124 }, { "epoch": 0.3098620618485776, "grad_norm": 0.3788921535015106, "learning_rate": 1e-05, "loss": 0.4614, "step": 1126 }, { "epoch": 0.3104124385126071, "grad_norm": 0.3723096251487732, "learning_rate": 1e-05, "loss": 0.4658, "step": 1128 }, { "epoch": 0.3109628151766365, "grad_norm": 0.3926720917224884, "learning_rate": 1e-05, "loss": 0.4602, "step": 1130 }, { "epoch": 0.31151319184066595, "grad_norm": 0.3565811514854431, "learning_rate": 1e-05, "loss": 0.4692, "step": 1132 }, { "epoch": 0.3120635685046954, "grad_norm": 0.38179391622543335, "learning_rate": 1e-05, "loss": 0.4581, "step": 1134 }, { "epoch": 0.31261394516872487, "grad_norm": 0.3732840418815613, "learning_rate": 1e-05, "loss": 0.4628, "step": 1136 }, { "epoch": 0.3131643218327543, "grad_norm": 0.3934018313884735, "learning_rate": 1e-05, "loss": 0.4634, "step": 1138 }, { "epoch": 0.31371469849678374, "grad_norm": 0.3575834035873413, "learning_rate": 1e-05, "loss": 0.4507, "step": 1140 }, { "epoch": 0.3142650751608132, "grad_norm": 0.3623636066913605, "learning_rate": 1e-05, "loss": 0.4547, "step": 1142 }, { "epoch": 0.3148154518248426, "grad_norm": 0.3794458508491516, "learning_rate": 1e-05, "loss": 0.4661, "step": 1144 }, { "epoch": 0.31536582848887207, "grad_norm": 0.3896718919277191, "learning_rate": 1e-05, "loss": 0.4646, "step": 1146 }, { "epoch": 0.31591620515290153, "grad_norm": 0.3608621060848236, "learning_rate": 1e-05, "loss": 0.4522, "step": 1148 }, { "epoch": 0.31646658181693094, "grad_norm": 0.37019404768943787, "learning_rate": 1e-05, "loss": 0.4548, "step": 1150 }, { "epoch": 0.3170169584809604, "grad_norm": 0.37957248091697693, "learning_rate": 1e-05, "loss": 0.4554, "step": 1152 }, { "epoch": 0.31756733514498986, "grad_norm": 0.3605276048183441, "learning_rate": 1e-05, "loss": 0.4679, "step": 1154 }, { "epoch": 0.3181177118090193, "grad_norm": 0.37218716740608215, "learning_rate": 1e-05, "loss": 0.4686, "step": 1156 }, { "epoch": 0.3186680884730487, "grad_norm": 0.37037035822868347, "learning_rate": 1e-05, "loss": 0.4898, "step": 1158 }, { "epoch": 0.3192184651370782, "grad_norm": 0.3569047749042511, "learning_rate": 1e-05, "loss": 0.4619, "step": 1160 }, { "epoch": 0.31976884180110765, "grad_norm": 0.3728378117084503, "learning_rate": 1e-05, "loss": 0.4544, "step": 1162 }, { "epoch": 0.32031921846513706, "grad_norm": 0.35970696806907654, "learning_rate": 1e-05, "loss": 0.4704, "step": 1164 }, { "epoch": 0.3208695951291665, "grad_norm": 0.36476969718933105, "learning_rate": 1e-05, "loss": 0.4605, "step": 1166 }, { "epoch": 0.321419971793196, "grad_norm": 0.35015928745269775, "learning_rate": 1e-05, "loss": 0.4653, "step": 1168 }, { "epoch": 0.3219703484572254, "grad_norm": 0.3600417375564575, "learning_rate": 1e-05, "loss": 0.4557, "step": 1170 }, { "epoch": 0.32252072512125485, "grad_norm": 0.36994755268096924, "learning_rate": 1e-05, "loss": 0.4601, "step": 1172 }, { "epoch": 0.3230711017852843, "grad_norm": 0.39908286929130554, "learning_rate": 1e-05, "loss": 0.472, "step": 1174 }, { "epoch": 0.32362147844931377, "grad_norm": 0.3717789947986603, "learning_rate": 1e-05, "loss": 0.4646, "step": 1176 }, { "epoch": 0.3241718551133432, "grad_norm": 0.3617453873157501, "learning_rate": 1e-05, "loss": 0.4606, "step": 1178 }, { "epoch": 0.32472223177737264, "grad_norm": 0.35809728503227234, "learning_rate": 1e-05, "loss": 0.4548, "step": 1180 }, { "epoch": 0.3252726084414021, "grad_norm": 0.3767383396625519, "learning_rate": 1e-05, "loss": 0.4785, "step": 1182 }, { "epoch": 0.3258229851054315, "grad_norm": 0.3819461166858673, "learning_rate": 1e-05, "loss": 0.4695, "step": 1184 }, { "epoch": 0.32637336176946097, "grad_norm": 0.3590524196624756, "learning_rate": 1e-05, "loss": 0.468, "step": 1186 }, { "epoch": 0.32692373843349043, "grad_norm": 0.37356823682785034, "learning_rate": 1e-05, "loss": 0.4628, "step": 1188 }, { "epoch": 0.3274741150975199, "grad_norm": 0.39389410614967346, "learning_rate": 1e-05, "loss": 0.4686, "step": 1190 }, { "epoch": 0.3280244917615493, "grad_norm": 0.36901354789733887, "learning_rate": 1e-05, "loss": 0.4623, "step": 1192 }, { "epoch": 0.32857486842557876, "grad_norm": 0.35733821988105774, "learning_rate": 1e-05, "loss": 0.457, "step": 1194 }, { "epoch": 0.3291252450896082, "grad_norm": 0.3803520202636719, "learning_rate": 1e-05, "loss": 0.4661, "step": 1196 }, { "epoch": 0.3296756217536376, "grad_norm": 0.36812326312065125, "learning_rate": 1e-05, "loss": 0.453, "step": 1198 }, { "epoch": 0.3302259984176671, "grad_norm": 0.37463024258613586, "learning_rate": 1e-05, "loss": 0.4611, "step": 1200 }, { "epoch": 0.3302259984176671, "eval_merge_loss": 0.4038620591163635, "eval_merge_runtime": 600.528, "eval_merge_samples_per_second": 56.204, "eval_merge_steps_per_second": 2.343, "step": 1200 }, { "epoch": 0.3302259984176671, "eval_new_aug_datas_filtered.json_loss": 0.5311903953552246, "eval_new_aug_datas_filtered.json_runtime": 10.3899, "eval_new_aug_datas_filtered.json_samples_per_second": 73.822, "eval_new_aug_datas_filtered.json_steps_per_second": 3.08, "step": 1200 }, { "epoch": 0.3302259984176671, "eval_sharegpt_gpt4.json_loss": 0.7848892211914062, "eval_sharegpt_gpt4.json_runtime": 31.7548, "eval_sharegpt_gpt4.json_samples_per_second": 58.605, "eval_sharegpt_gpt4.json_steps_per_second": 2.456, "step": 1200 }, { "epoch": 0.3302259984176671, "eval_Table_GPT.json_loss": 0.07294219732284546, "eval_Table_GPT.json_runtime": 25.0251, "eval_Table_GPT.json_samples_per_second": 83.636, "eval_Table_GPT.json_steps_per_second": 3.516, "step": 1200 }, { "epoch": 0.3302259984176671, "eval_gpt_4o_200k.json_loss": 0.8128483295440674, "eval_gpt_4o_200k.json_runtime": 48.5727, "eval_gpt_4o_200k.json_samples_per_second": 129.311, "eval_gpt_4o_200k.json_steps_per_second": 5.394, "step": 1200 }, { "epoch": 0.3302259984176671, "eval_multi_turn_datas.json_loss": 0.343874990940094, "eval_multi_turn_datas.json_runtime": 75.779, "eval_multi_turn_datas.json_samples_per_second": 52.811, "eval_multi_turn_datas.json_steps_per_second": 2.204, "step": 1200 }, { "epoch": 0.3302259984176671, "eval_table_python_code_datas.json_loss": 0.2791996896266937, "eval_table_python_code_datas.json_runtime": 43.1703, "eval_table_python_code_datas.json_samples_per_second": 50.011, "eval_table_python_code_datas.json_steps_per_second": 2.085, "step": 1200 }, { "epoch": 0.3302259984176671, "eval_tabular_llm_data.json_loss": 0.11510641872882843, "eval_tabular_llm_data.json_runtime": 8.5754, "eval_tabular_llm_data.json_samples_per_second": 28.687, "eval_tabular_llm_data.json_steps_per_second": 1.283, "step": 1200 }, { "epoch": 0.3302259984176671, "eval_python_code_critic_21k.json_loss": 0.5806341171264648, "eval_python_code_critic_21k.json_runtime": 3.2355, "eval_python_code_critic_21k.json_samples_per_second": 184.517, "eval_python_code_critic_21k.json_steps_per_second": 7.727, "step": 1200 }, { "epoch": 0.3302259984176671, "eval_all_merge_table_dataset.json_loss": 0.0781954750418663, "eval_all_merge_table_dataset.json_runtime": 23.3576, "eval_all_merge_table_dataset.json_samples_per_second": 30.483, "eval_all_merge_table_dataset.json_steps_per_second": 1.284, "step": 1200 }, { "epoch": 0.3302259984176671, "eval_code_feedback_multi_turn.json_loss": 0.5880293846130371, "eval_code_feedback_multi_turn.json_runtime": 32.5337, "eval_code_feedback_multi_turn.json_samples_per_second": 67.653, "eval_code_feedback_multi_turn.json_steps_per_second": 2.828, "step": 1200 }, { "epoch": 0.3302259984176671, "eval_ultrainteract_sft.json_loss": 0.42568570375442505, "eval_ultrainteract_sft.json_runtime": 8.665, "eval_ultrainteract_sft.json_samples_per_second": 168.033, "eval_ultrainteract_sft.json_steps_per_second": 7.04, "step": 1200 }, { "epoch": 0.3302259984176671, "eval_synthetic_text_to_sql.json_loss": 0.10025755316019058, "eval_synthetic_text_to_sql.json_runtime": 0.127, "eval_synthetic_text_to_sql.json_samples_per_second": 267.683, "eval_synthetic_text_to_sql.json_steps_per_second": 15.746, "step": 1200 }, { "epoch": 0.3302259984176671, "eval_sft_react_sql_datas.json_loss": 0.6435717344284058, "eval_sft_react_sql_datas.json_runtime": 7.8854, "eval_sft_react_sql_datas.json_samples_per_second": 39.82, "eval_sft_react_sql_datas.json_steps_per_second": 1.775, "step": 1200 }, { "epoch": 0.3302259984176671, "eval_all_merge_code.json_loss": 0.29655295610427856, "eval_all_merge_code.json_runtime": 0.3333, "eval_all_merge_code.json_samples_per_second": 189.039, "eval_all_merge_code.json_steps_per_second": 9.002, "step": 1200 }, { "epoch": 0.3302259984176671, "eval_magpie_datas.json_loss": 0.4353857934474945, "eval_magpie_datas.json_runtime": 2.22, "eval_magpie_datas.json_samples_per_second": 77.478, "eval_magpie_datas.json_steps_per_second": 3.604, "step": 1200 }, { "epoch": 0.3302259984176671, "eval_train_data_for_qwen.json_loss": 0.0036680654156953096, "eval_train_data_for_qwen.json_runtime": 0.2448, "eval_train_data_for_qwen.json_samples_per_second": 40.856, "eval_train_data_for_qwen.json_steps_per_second": 4.086, "step": 1200 }, { "epoch": 0.3302259984176671, "eval_alpaca_cleaned.json_loss": 0.9278478622436523, "eval_alpaca_cleaned.json_runtime": 0.1139, "eval_alpaca_cleaned.json_samples_per_second": 237.139, "eval_alpaca_cleaned.json_steps_per_second": 17.566, "step": 1200 }, { "epoch": 0.3302259984176671, "eval_agent_instruct.json_loss": 0.22283704578876495, "eval_agent_instruct.json_runtime": 0.5129, "eval_agent_instruct.json_samples_per_second": 93.582, "eval_agent_instruct.json_steps_per_second": 3.899, "step": 1200 }, { "epoch": 0.3302259984176671, "eval_MathInstruct.json_loss": 0.20810073614120483, "eval_MathInstruct.json_runtime": 0.3587, "eval_MathInstruct.json_samples_per_second": 158.905, "eval_MathInstruct.json_steps_per_second": 8.363, "step": 1200 }, { "epoch": 0.3302259984176671, "eval_tested_143k_python_alpaca.json_loss": 0.44691047072410583, "eval_tested_143k_python_alpaca.json_runtime": 0.3024, "eval_tested_143k_python_alpaca.json_samples_per_second": 112.419, "eval_tested_143k_python_alpaca.json_steps_per_second": 6.613, "step": 1200 }, { "epoch": 0.3302259984176671, "eval_xlam_function_calling_60k.json_loss": 0.009029570966959, "eval_xlam_function_calling_60k.json_runtime": 0.1005, "eval_xlam_function_calling_60k.json_samples_per_second": 228.948, "eval_xlam_function_calling_60k.json_steps_per_second": 9.954, "step": 1200 }, { "epoch": 0.3302259984176671, "eval_alpaca_data_gpt4_chinese.json_loss": 1.5715256929397583, "eval_alpaca_data_gpt4_chinese.json_runtime": 0.0514, "eval_alpaca_data_gpt4_chinese.json_samples_per_second": 311.088, "eval_alpaca_data_gpt4_chinese.json_steps_per_second": 19.443, "step": 1200 }, { "epoch": 0.3302259984176671, "eval_alpaca_gpt4_zh.json_loss": 0.9568694233894348, "eval_alpaca_gpt4_zh.json_runtime": 0.0501, "eval_alpaca_gpt4_zh.json_samples_per_second": 219.517, "eval_alpaca_gpt4_zh.json_steps_per_second": 19.956, "step": 1200 }, { "epoch": 0.3302259984176671, "eval_codefeedback_filtered_instruction.json_loss": 0.5982481837272644, "eval_codefeedback_filtered_instruction.json_runtime": 0.487, "eval_codefeedback_filtered_instruction.json_samples_per_second": 41.068, "eval_codefeedback_filtered_instruction.json_steps_per_second": 2.053, "step": 1200 }, { "epoch": 0.33077637508169655, "grad_norm": 0.3862474262714386, "learning_rate": 1e-05, "loss": 0.467, "step": 1202 }, { "epoch": 0.33132675174572596, "grad_norm": 0.3586987555027008, "learning_rate": 1e-05, "loss": 0.4586, "step": 1204 }, { "epoch": 0.3318771284097554, "grad_norm": 0.36768838763237, "learning_rate": 1e-05, "loss": 0.4658, "step": 1206 }, { "epoch": 0.3324275050737849, "grad_norm": 0.36789608001708984, "learning_rate": 1e-05, "loss": 0.4479, "step": 1208 }, { "epoch": 0.33297788173781434, "grad_norm": 0.3875747323036194, "learning_rate": 1e-05, "loss": 0.4651, "step": 1210 }, { "epoch": 0.33352825840184375, "grad_norm": 0.37122058868408203, "learning_rate": 1e-05, "loss": 0.4474, "step": 1212 }, { "epoch": 0.3340786350658732, "grad_norm": 0.3785482347011566, "learning_rate": 1e-05, "loss": 0.4573, "step": 1214 }, { "epoch": 0.33462901172990267, "grad_norm": 0.3795594871044159, "learning_rate": 1e-05, "loss": 0.4633, "step": 1216 }, { "epoch": 0.3351793883939321, "grad_norm": 0.35303714871406555, "learning_rate": 1e-05, "loss": 0.4701, "step": 1218 }, { "epoch": 0.33572976505796154, "grad_norm": 0.3473946154117584, "learning_rate": 1e-05, "loss": 0.4565, "step": 1220 }, { "epoch": 0.336280141721991, "grad_norm": 0.36495375633239746, "learning_rate": 1e-05, "loss": 0.4528, "step": 1222 }, { "epoch": 0.3368305183860204, "grad_norm": 0.3617894649505615, "learning_rate": 1e-05, "loss": 0.4756, "step": 1224 }, { "epoch": 0.33738089505004987, "grad_norm": 0.36371487379074097, "learning_rate": 1e-05, "loss": 0.4606, "step": 1226 }, { "epoch": 0.33793127171407933, "grad_norm": 0.39192309975624084, "learning_rate": 1e-05, "loss": 0.4435, "step": 1228 }, { "epoch": 0.3384816483781088, "grad_norm": 0.3902663588523865, "learning_rate": 1e-05, "loss": 0.4699, "step": 1230 }, { "epoch": 0.3390320250421382, "grad_norm": 0.3662269115447998, "learning_rate": 1e-05, "loss": 0.4627, "step": 1232 }, { "epoch": 0.33958240170616766, "grad_norm": 0.3659150004386902, "learning_rate": 1e-05, "loss": 0.4663, "step": 1234 }, { "epoch": 0.3401327783701971, "grad_norm": 0.3632274568080902, "learning_rate": 1e-05, "loss": 0.4499, "step": 1236 }, { "epoch": 0.3406831550342265, "grad_norm": 0.38413625955581665, "learning_rate": 1e-05, "loss": 0.4516, "step": 1238 }, { "epoch": 0.341233531698256, "grad_norm": 0.35747644305229187, "learning_rate": 1e-05, "loss": 0.4718, "step": 1240 }, { "epoch": 0.34178390836228545, "grad_norm": 0.36938604712486267, "learning_rate": 1e-05, "loss": 0.4568, "step": 1242 }, { "epoch": 0.3423342850263149, "grad_norm": 0.38448217511177063, "learning_rate": 1e-05, "loss": 0.474, "step": 1244 }, { "epoch": 0.3428846616903443, "grad_norm": 0.3694998323917389, "learning_rate": 1e-05, "loss": 0.4516, "step": 1246 }, { "epoch": 0.3434350383543738, "grad_norm": 0.41237321496009827, "learning_rate": 1e-05, "loss": 0.4569, "step": 1248 }, { "epoch": 0.34398541501840324, "grad_norm": 0.4058983325958252, "learning_rate": 1e-05, "loss": 0.4657, "step": 1250 }, { "epoch": 0.34453579168243265, "grad_norm": 0.3610474467277527, "learning_rate": 1e-05, "loss": 0.4587, "step": 1252 }, { "epoch": 0.3450861683464621, "grad_norm": 0.3664454221725464, "learning_rate": 1e-05, "loss": 0.4656, "step": 1254 }, { "epoch": 0.34563654501049157, "grad_norm": 0.35148540139198303, "learning_rate": 1e-05, "loss": 0.4471, "step": 1256 }, { "epoch": 0.346186921674521, "grad_norm": 0.35331565141677856, "learning_rate": 1e-05, "loss": 0.4674, "step": 1258 }, { "epoch": 0.34673729833855044, "grad_norm": 0.35367992520332336, "learning_rate": 1e-05, "loss": 0.4572, "step": 1260 }, { "epoch": 0.3472876750025799, "grad_norm": 0.36106035113334656, "learning_rate": 1e-05, "loss": 0.466, "step": 1262 }, { "epoch": 0.34783805166660936, "grad_norm": 0.36034414172172546, "learning_rate": 1e-05, "loss": 0.4412, "step": 1264 }, { "epoch": 0.34838842833063877, "grad_norm": 0.3532898426055908, "learning_rate": 1e-05, "loss": 0.4573, "step": 1266 }, { "epoch": 0.34893880499466823, "grad_norm": 0.35383620858192444, "learning_rate": 1e-05, "loss": 0.4644, "step": 1268 }, { "epoch": 0.3494891816586977, "grad_norm": 0.3757399022579193, "learning_rate": 1e-05, "loss": 0.4548, "step": 1270 }, { "epoch": 0.3500395583227271, "grad_norm": 0.35997340083122253, "learning_rate": 1e-05, "loss": 0.4664, "step": 1272 }, { "epoch": 0.35058993498675656, "grad_norm": 0.3761090636253357, "learning_rate": 1e-05, "loss": 0.4601, "step": 1274 }, { "epoch": 0.351140311650786, "grad_norm": 0.33666959404945374, "learning_rate": 1e-05, "loss": 0.4596, "step": 1276 }, { "epoch": 0.3516906883148154, "grad_norm": 0.36252304911613464, "learning_rate": 1e-05, "loss": 0.4688, "step": 1278 }, { "epoch": 0.3522410649788449, "grad_norm": 0.3987884819507599, "learning_rate": 1e-05, "loss": 0.4444, "step": 1280 }, { "epoch": 0.35279144164287435, "grad_norm": 0.35914021730422974, "learning_rate": 1e-05, "loss": 0.4508, "step": 1282 }, { "epoch": 0.3533418183069038, "grad_norm": 0.36508429050445557, "learning_rate": 1e-05, "loss": 0.4597, "step": 1284 }, { "epoch": 0.3538921949709332, "grad_norm": 0.3923473060131073, "learning_rate": 1e-05, "loss": 0.4594, "step": 1286 }, { "epoch": 0.3544425716349627, "grad_norm": 0.38775792717933655, "learning_rate": 1e-05, "loss": 0.4573, "step": 1288 }, { "epoch": 0.35499294829899214, "grad_norm": 0.4628289043903351, "learning_rate": 1e-05, "loss": 0.4732, "step": 1290 }, { "epoch": 0.35554332496302155, "grad_norm": 0.35442307591438293, "learning_rate": 1e-05, "loss": 0.4621, "step": 1292 }, { "epoch": 0.356093701627051, "grad_norm": 0.3809347152709961, "learning_rate": 1e-05, "loss": 0.4696, "step": 1294 }, { "epoch": 0.35664407829108047, "grad_norm": 0.3683224618434906, "learning_rate": 1e-05, "loss": 0.4649, "step": 1296 }, { "epoch": 0.35719445495510993, "grad_norm": 0.3792459964752197, "learning_rate": 1e-05, "loss": 0.465, "step": 1298 }, { "epoch": 0.35774483161913934, "grad_norm": 0.3704141080379486, "learning_rate": 1e-05, "loss": 0.4572, "step": 1300 }, { "epoch": 0.3582952082831688, "grad_norm": 0.3618161678314209, "learning_rate": 1e-05, "loss": 0.4497, "step": 1302 }, { "epoch": 0.35884558494719826, "grad_norm": 0.36538904905319214, "learning_rate": 1e-05, "loss": 0.4525, "step": 1304 }, { "epoch": 0.35939596161122767, "grad_norm": 0.36815035343170166, "learning_rate": 1e-05, "loss": 0.4767, "step": 1306 }, { "epoch": 0.3599463382752571, "grad_norm": 0.39006996154785156, "learning_rate": 1e-05, "loss": 0.4809, "step": 1308 }, { "epoch": 0.3604967149392866, "grad_norm": 0.3829619288444519, "learning_rate": 1e-05, "loss": 0.4714, "step": 1310 }, { "epoch": 0.361047091603316, "grad_norm": 0.37935730814933777, "learning_rate": 1e-05, "loss": 0.4518, "step": 1312 }, { "epoch": 0.36159746826734546, "grad_norm": 0.371320903301239, "learning_rate": 1e-05, "loss": 0.4437, "step": 1314 }, { "epoch": 0.3621478449313749, "grad_norm": 0.35784757137298584, "learning_rate": 1e-05, "loss": 0.4579, "step": 1316 }, { "epoch": 0.3626982215954044, "grad_norm": 0.36308974027633667, "learning_rate": 1e-05, "loss": 0.4561, "step": 1318 }, { "epoch": 0.3632485982594338, "grad_norm": 0.3538898825645447, "learning_rate": 1e-05, "loss": 0.4574, "step": 1320 }, { "epoch": 0.36379897492346325, "grad_norm": 0.3715920150279999, "learning_rate": 1e-05, "loss": 0.4649, "step": 1322 }, { "epoch": 0.3643493515874927, "grad_norm": 0.3698347806930542, "learning_rate": 1e-05, "loss": 0.4648, "step": 1324 }, { "epoch": 0.3648997282515221, "grad_norm": 0.3725499212741852, "learning_rate": 1e-05, "loss": 0.4669, "step": 1326 }, { "epoch": 0.3654501049155516, "grad_norm": 0.37399542331695557, "learning_rate": 1e-05, "loss": 0.4615, "step": 1328 }, { "epoch": 0.36600048157958104, "grad_norm": 0.35364219546318054, "learning_rate": 1e-05, "loss": 0.4573, "step": 1330 }, { "epoch": 0.36655085824361044, "grad_norm": 0.3651660084724426, "learning_rate": 1e-05, "loss": 0.4485, "step": 1332 }, { "epoch": 0.3671012349076399, "grad_norm": 0.3659324645996094, "learning_rate": 1e-05, "loss": 0.4492, "step": 1334 }, { "epoch": 0.36765161157166937, "grad_norm": 0.35941600799560547, "learning_rate": 1e-05, "loss": 0.4727, "step": 1336 }, { "epoch": 0.36820198823569883, "grad_norm": 0.35083696246147156, "learning_rate": 1e-05, "loss": 0.4453, "step": 1338 }, { "epoch": 0.36875236489972824, "grad_norm": 0.3690749406814575, "learning_rate": 1e-05, "loss": 0.4582, "step": 1340 }, { "epoch": 0.3693027415637577, "grad_norm": 0.3743647038936615, "learning_rate": 1e-05, "loss": 0.4383, "step": 1342 }, { "epoch": 0.36985311822778716, "grad_norm": 0.37354332208633423, "learning_rate": 1e-05, "loss": 0.466, "step": 1344 }, { "epoch": 0.37040349489181656, "grad_norm": 0.3735334575176239, "learning_rate": 1e-05, "loss": 0.4535, "step": 1346 }, { "epoch": 0.370953871555846, "grad_norm": 0.37339311838150024, "learning_rate": 1e-05, "loss": 0.454, "step": 1348 }, { "epoch": 0.3715042482198755, "grad_norm": 0.35196128487586975, "learning_rate": 1e-05, "loss": 0.4685, "step": 1350 }, { "epoch": 0.37205462488390495, "grad_norm": 0.4031345546245575, "learning_rate": 1e-05, "loss": 0.4689, "step": 1352 }, { "epoch": 0.37260500154793436, "grad_norm": 0.363320529460907, "learning_rate": 1e-05, "loss": 0.459, "step": 1354 }, { "epoch": 0.3731553782119638, "grad_norm": 0.36146363615989685, "learning_rate": 1e-05, "loss": 0.4446, "step": 1356 }, { "epoch": 0.3737057548759933, "grad_norm": 0.36425283551216125, "learning_rate": 1e-05, "loss": 0.468, "step": 1358 }, { "epoch": 0.3742561315400227, "grad_norm": 0.3795093894004822, "learning_rate": 1e-05, "loss": 0.4513, "step": 1360 }, { "epoch": 0.37480650820405215, "grad_norm": 0.37901571393013, "learning_rate": 1e-05, "loss": 0.464, "step": 1362 }, { "epoch": 0.3753568848680816, "grad_norm": 0.3682788014411926, "learning_rate": 1e-05, "loss": 0.4535, "step": 1364 }, { "epoch": 0.375907261532111, "grad_norm": 0.38756048679351807, "learning_rate": 1e-05, "loss": 0.4421, "step": 1366 }, { "epoch": 0.3764576381961405, "grad_norm": 0.3859202563762665, "learning_rate": 1e-05, "loss": 0.4601, "step": 1368 }, { "epoch": 0.37700801486016994, "grad_norm": 0.3959304392337799, "learning_rate": 1e-05, "loss": 0.4427, "step": 1370 }, { "epoch": 0.3775583915241994, "grad_norm": 0.3768652379512787, "learning_rate": 1e-05, "loss": 0.4483, "step": 1372 }, { "epoch": 0.3781087681882288, "grad_norm": 0.37339305877685547, "learning_rate": 1e-05, "loss": 0.4605, "step": 1374 }, { "epoch": 0.37865914485225827, "grad_norm": 0.4036271572113037, "learning_rate": 1e-05, "loss": 0.4546, "step": 1376 }, { "epoch": 0.37920952151628773, "grad_norm": 0.35173818469047546, "learning_rate": 1e-05, "loss": 0.4675, "step": 1378 }, { "epoch": 0.37975989818031713, "grad_norm": 0.3682287335395813, "learning_rate": 1e-05, "loss": 0.4676, "step": 1380 }, { "epoch": 0.3803102748443466, "grad_norm": 0.37660422921180725, "learning_rate": 1e-05, "loss": 0.4483, "step": 1382 }, { "epoch": 0.38086065150837606, "grad_norm": 0.37428486347198486, "learning_rate": 1e-05, "loss": 0.4537, "step": 1384 }, { "epoch": 0.38141102817240546, "grad_norm": 0.36140507459640503, "learning_rate": 1e-05, "loss": 0.4542, "step": 1386 }, { "epoch": 0.3819614048364349, "grad_norm": 0.3818880021572113, "learning_rate": 1e-05, "loss": 0.4546, "step": 1388 }, { "epoch": 0.3825117815004644, "grad_norm": 0.3840683698654175, "learning_rate": 1e-05, "loss": 0.4419, "step": 1390 }, { "epoch": 0.38306215816449385, "grad_norm": 0.36933979392051697, "learning_rate": 1e-05, "loss": 0.4632, "step": 1392 }, { "epoch": 0.38361253482852326, "grad_norm": 0.3724002540111542, "learning_rate": 1e-05, "loss": 0.455, "step": 1394 }, { "epoch": 0.3841629114925527, "grad_norm": 0.35783514380455017, "learning_rate": 1e-05, "loss": 0.4652, "step": 1396 }, { "epoch": 0.3847132881565822, "grad_norm": 0.36758366227149963, "learning_rate": 1e-05, "loss": 0.4647, "step": 1398 }, { "epoch": 0.3852636648206116, "grad_norm": 0.3690735995769501, "learning_rate": 1e-05, "loss": 0.4572, "step": 1400 }, { "epoch": 0.3852636648206116, "eval_merge_loss": 0.39922505617141724, "eval_merge_runtime": 600.7214, "eval_merge_samples_per_second": 56.186, "eval_merge_steps_per_second": 2.342, "step": 1400 }, { "epoch": 0.3852636648206116, "eval_new_aug_datas_filtered.json_loss": 0.5239847898483276, "eval_new_aug_datas_filtered.json_runtime": 10.3569, "eval_new_aug_datas_filtered.json_samples_per_second": 74.057, "eval_new_aug_datas_filtered.json_steps_per_second": 3.09, "step": 1400 }, { "epoch": 0.3852636648206116, "eval_sharegpt_gpt4.json_loss": 0.7791606187820435, "eval_sharegpt_gpt4.json_runtime": 31.6746, "eval_sharegpt_gpt4.json_samples_per_second": 58.754, "eval_sharegpt_gpt4.json_steps_per_second": 2.463, "step": 1400 }, { "epoch": 0.3852636648206116, "eval_Table_GPT.json_loss": 0.0626993179321289, "eval_Table_GPT.json_runtime": 24.9542, "eval_Table_GPT.json_samples_per_second": 83.874, "eval_Table_GPT.json_steps_per_second": 3.526, "step": 1400 }, { "epoch": 0.3852636648206116, "eval_gpt_4o_200k.json_loss": 0.8082922101020813, "eval_gpt_4o_200k.json_runtime": 48.466, "eval_gpt_4o_200k.json_samples_per_second": 129.596, "eval_gpt_4o_200k.json_steps_per_second": 5.406, "step": 1400 }, { "epoch": 0.3852636648206116, "eval_multi_turn_datas.json_loss": 0.3381649851799011, "eval_multi_turn_datas.json_runtime": 75.5711, "eval_multi_turn_datas.json_samples_per_second": 52.957, "eval_multi_turn_datas.json_steps_per_second": 2.21, "step": 1400 }, { "epoch": 0.3852636648206116, "eval_table_python_code_datas.json_loss": 0.2752579152584076, "eval_table_python_code_datas.json_runtime": 43.0439, "eval_table_python_code_datas.json_samples_per_second": 50.158, "eval_table_python_code_datas.json_steps_per_second": 2.091, "step": 1400 }, { "epoch": 0.3852636648206116, "eval_tabular_llm_data.json_loss": 0.11023548245429993, "eval_tabular_llm_data.json_runtime": 8.5291, "eval_tabular_llm_data.json_samples_per_second": 28.843, "eval_tabular_llm_data.json_steps_per_second": 1.29, "step": 1400 }, { "epoch": 0.3852636648206116, "eval_python_code_critic_21k.json_loss": 0.5756029486656189, "eval_python_code_critic_21k.json_runtime": 3.2275, "eval_python_code_critic_21k.json_samples_per_second": 184.973, "eval_python_code_critic_21k.json_steps_per_second": 7.746, "step": 1400 }, { "epoch": 0.3852636648206116, "eval_all_merge_table_dataset.json_loss": 0.08007320761680603, "eval_all_merge_table_dataset.json_runtime": 23.3, "eval_all_merge_table_dataset.json_samples_per_second": 30.558, "eval_all_merge_table_dataset.json_steps_per_second": 1.288, "step": 1400 }, { "epoch": 0.3852636648206116, "eval_code_feedback_multi_turn.json_loss": 0.5849318504333496, "eval_code_feedback_multi_turn.json_runtime": 32.4131, "eval_code_feedback_multi_turn.json_samples_per_second": 67.905, "eval_code_feedback_multi_turn.json_steps_per_second": 2.838, "step": 1400 }, { "epoch": 0.3852636648206116, "eval_ultrainteract_sft.json_loss": 0.4235917031764984, "eval_ultrainteract_sft.json_runtime": 8.6815, "eval_ultrainteract_sft.json_samples_per_second": 167.713, "eval_ultrainteract_sft.json_steps_per_second": 7.026, "step": 1400 }, { "epoch": 0.3852636648206116, "eval_synthetic_text_to_sql.json_loss": 0.10058007389307022, "eval_synthetic_text_to_sql.json_runtime": 0.1256, "eval_synthetic_text_to_sql.json_samples_per_second": 270.794, "eval_synthetic_text_to_sql.json_steps_per_second": 15.929, "step": 1400 }, { "epoch": 0.3852636648206116, "eval_sft_react_sql_datas.json_loss": 0.63919597864151, "eval_sft_react_sql_datas.json_runtime": 7.8177, "eval_sft_react_sql_datas.json_samples_per_second": 40.165, "eval_sft_react_sql_datas.json_steps_per_second": 1.791, "step": 1400 }, { "epoch": 0.3852636648206116, "eval_all_merge_code.json_loss": 0.293491929769516, "eval_all_merge_code.json_runtime": 0.3331, "eval_all_merge_code.json_samples_per_second": 189.11, "eval_all_merge_code.json_steps_per_second": 9.005, "step": 1400 }, { "epoch": 0.3852636648206116, "eval_magpie_datas.json_loss": 0.43307721614837646, "eval_magpie_datas.json_runtime": 2.214, "eval_magpie_datas.json_samples_per_second": 77.687, "eval_magpie_datas.json_steps_per_second": 3.613, "step": 1400 }, { "epoch": 0.3852636648206116, "eval_train_data_for_qwen.json_loss": 0.004504092503339052, "eval_train_data_for_qwen.json_runtime": 0.2448, "eval_train_data_for_qwen.json_samples_per_second": 40.845, "eval_train_data_for_qwen.json_steps_per_second": 4.084, "step": 1400 }, { "epoch": 0.3852636648206116, "eval_alpaca_cleaned.json_loss": 0.9073267579078674, "eval_alpaca_cleaned.json_runtime": 0.1148, "eval_alpaca_cleaned.json_samples_per_second": 235.221, "eval_alpaca_cleaned.json_steps_per_second": 17.424, "step": 1400 }, { "epoch": 0.3852636648206116, "eval_agent_instruct.json_loss": 0.22197985649108887, "eval_agent_instruct.json_runtime": 0.5129, "eval_agent_instruct.json_samples_per_second": 93.586, "eval_agent_instruct.json_steps_per_second": 3.899, "step": 1400 }, { "epoch": 0.3852636648206116, "eval_MathInstruct.json_loss": 0.2014550119638443, "eval_MathInstruct.json_runtime": 0.3655, "eval_MathInstruct.json_samples_per_second": 155.953, "eval_MathInstruct.json_steps_per_second": 8.208, "step": 1400 }, { "epoch": 0.3852636648206116, "eval_tested_143k_python_alpaca.json_loss": 0.44645121693611145, "eval_tested_143k_python_alpaca.json_runtime": 0.3016, "eval_tested_143k_python_alpaca.json_samples_per_second": 112.74, "eval_tested_143k_python_alpaca.json_steps_per_second": 6.632, "step": 1400 }, { "epoch": 0.3852636648206116, "eval_xlam_function_calling_60k.json_loss": 0.009633864276111126, "eval_xlam_function_calling_60k.json_runtime": 0.0999, "eval_xlam_function_calling_60k.json_samples_per_second": 230.172, "eval_xlam_function_calling_60k.json_steps_per_second": 10.007, "step": 1400 }, { "epoch": 0.3852636648206116, "eval_alpaca_data_gpt4_chinese.json_loss": 1.5636402368545532, "eval_alpaca_data_gpt4_chinese.json_runtime": 0.0503, "eval_alpaca_data_gpt4_chinese.json_samples_per_second": 318.002, "eval_alpaca_data_gpt4_chinese.json_steps_per_second": 19.875, "step": 1400 }, { "epoch": 0.3852636648206116, "eval_alpaca_gpt4_zh.json_loss": 0.9602435231208801, "eval_alpaca_gpt4_zh.json_runtime": 0.0502, "eval_alpaca_gpt4_zh.json_samples_per_second": 219.067, "eval_alpaca_gpt4_zh.json_steps_per_second": 19.915, "step": 1400 }, { "epoch": 0.3852636648206116, "eval_codefeedback_filtered_instruction.json_loss": 0.5993592143058777, "eval_codefeedback_filtered_instruction.json_runtime": 0.4852, "eval_codefeedback_filtered_instruction.json_samples_per_second": 41.223, "eval_codefeedback_filtered_instruction.json_steps_per_second": 2.061, "step": 1400 }, { "epoch": 0.38581404148464105, "grad_norm": 0.36705121397972107, "learning_rate": 1e-05, "loss": 0.4617, "step": 1402 }, { "epoch": 0.3863644181486705, "grad_norm": 0.3653152883052826, "learning_rate": 1e-05, "loss": 0.4528, "step": 1404 }, { "epoch": 0.38691479481269997, "grad_norm": 0.34426313638687134, "learning_rate": 1e-05, "loss": 0.4464, "step": 1406 }, { "epoch": 0.3874651714767294, "grad_norm": 0.3493911623954773, "learning_rate": 1e-05, "loss": 0.4638, "step": 1408 }, { "epoch": 0.38801554814075884, "grad_norm": 0.3841487765312195, "learning_rate": 1e-05, "loss": 0.4471, "step": 1410 }, { "epoch": 0.3885659248047883, "grad_norm": 0.3770912289619446, "learning_rate": 1e-05, "loss": 0.4623, "step": 1412 }, { "epoch": 0.3891163014688177, "grad_norm": 0.38141822814941406, "learning_rate": 1e-05, "loss": 0.4583, "step": 1414 }, { "epoch": 0.38966667813284717, "grad_norm": 0.3774464726448059, "learning_rate": 1e-05, "loss": 0.4574, "step": 1416 }, { "epoch": 0.39021705479687663, "grad_norm": 0.35681846737861633, "learning_rate": 1e-05, "loss": 0.4443, "step": 1418 }, { "epoch": 0.39076743146090603, "grad_norm": 0.3700469732284546, "learning_rate": 1e-05, "loss": 0.4468, "step": 1420 }, { "epoch": 0.3913178081249355, "grad_norm": 0.35229384899139404, "learning_rate": 1e-05, "loss": 0.456, "step": 1422 }, { "epoch": 0.39186818478896496, "grad_norm": 0.3469116687774658, "learning_rate": 1e-05, "loss": 0.451, "step": 1424 }, { "epoch": 0.3924185614529944, "grad_norm": 0.36313918232917786, "learning_rate": 1e-05, "loss": 0.4679, "step": 1426 }, { "epoch": 0.3929689381170238, "grad_norm": 0.3543436527252197, "learning_rate": 1e-05, "loss": 0.464, "step": 1428 }, { "epoch": 0.3935193147810533, "grad_norm": 0.3992765545845032, "learning_rate": 1e-05, "loss": 0.486, "step": 1430 }, { "epoch": 0.39406969144508275, "grad_norm": 0.36149340867996216, "learning_rate": 1e-05, "loss": 0.4426, "step": 1432 }, { "epoch": 0.39462006810911215, "grad_norm": 0.37118762731552124, "learning_rate": 1e-05, "loss": 0.4531, "step": 1434 }, { "epoch": 0.3951704447731416, "grad_norm": 0.3618330955505371, "learning_rate": 1e-05, "loss": 0.4621, "step": 1436 }, { "epoch": 0.3957208214371711, "grad_norm": 0.37272128462791443, "learning_rate": 1e-05, "loss": 0.4616, "step": 1438 }, { "epoch": 0.3962711981012005, "grad_norm": 0.3678719997406006, "learning_rate": 1e-05, "loss": 0.4477, "step": 1440 }, { "epoch": 0.39682157476522995, "grad_norm": 0.342907190322876, "learning_rate": 1e-05, "loss": 0.4484, "step": 1442 }, { "epoch": 0.3973719514292594, "grad_norm": 0.3722037374973297, "learning_rate": 1e-05, "loss": 0.4576, "step": 1444 }, { "epoch": 0.39792232809328887, "grad_norm": 0.3829335868358612, "learning_rate": 1e-05, "loss": 0.4568, "step": 1446 }, { "epoch": 0.3984727047573183, "grad_norm": 0.36857596039772034, "learning_rate": 1e-05, "loss": 0.4509, "step": 1448 }, { "epoch": 0.39902308142134774, "grad_norm": 0.36784934997558594, "learning_rate": 1e-05, "loss": 0.46, "step": 1450 }, { "epoch": 0.3995734580853772, "grad_norm": 0.36996331810951233, "learning_rate": 1e-05, "loss": 0.4435, "step": 1452 }, { "epoch": 0.4001238347494066, "grad_norm": 0.3608056604862213, "learning_rate": 1e-05, "loss": 0.4467, "step": 1454 }, { "epoch": 0.40067421141343607, "grad_norm": 0.3827229142189026, "learning_rate": 1e-05, "loss": 0.4576, "step": 1456 }, { "epoch": 0.4012245880774655, "grad_norm": 0.38073116540908813, "learning_rate": 1e-05, "loss": 0.4433, "step": 1458 }, { "epoch": 0.40177496474149493, "grad_norm": 0.3861468434333801, "learning_rate": 1e-05, "loss": 0.4466, "step": 1460 }, { "epoch": 0.4023253414055244, "grad_norm": 0.36093631386756897, "learning_rate": 1e-05, "loss": 0.4409, "step": 1462 }, { "epoch": 0.40287571806955386, "grad_norm": 0.34549927711486816, "learning_rate": 1e-05, "loss": 0.4507, "step": 1464 }, { "epoch": 0.4034260947335833, "grad_norm": 0.3782083988189697, "learning_rate": 1e-05, "loss": 0.4648, "step": 1466 }, { "epoch": 0.4039764713976127, "grad_norm": 0.366914302110672, "learning_rate": 1e-05, "loss": 0.462, "step": 1468 }, { "epoch": 0.4045268480616422, "grad_norm": 0.3604414761066437, "learning_rate": 1e-05, "loss": 0.4639, "step": 1470 }, { "epoch": 0.40507722472567165, "grad_norm": 0.3806079924106598, "learning_rate": 1e-05, "loss": 0.452, "step": 1472 }, { "epoch": 0.40562760138970105, "grad_norm": 0.36079150438308716, "learning_rate": 1e-05, "loss": 0.4534, "step": 1474 }, { "epoch": 0.4061779780537305, "grad_norm": 0.3526926040649414, "learning_rate": 1e-05, "loss": 0.4483, "step": 1476 }, { "epoch": 0.40672835471776, "grad_norm": 0.36440181732177734, "learning_rate": 1e-05, "loss": 0.4445, "step": 1478 }, { "epoch": 0.40727873138178944, "grad_norm": 0.3452344238758087, "learning_rate": 1e-05, "loss": 0.4531, "step": 1480 }, { "epoch": 0.40782910804581884, "grad_norm": 0.3774935007095337, "learning_rate": 1e-05, "loss": 0.4644, "step": 1482 }, { "epoch": 0.4083794847098483, "grad_norm": 0.3485760986804962, "learning_rate": 1e-05, "loss": 0.4489, "step": 1484 }, { "epoch": 0.40892986137387777, "grad_norm": 0.3787960708141327, "learning_rate": 1e-05, "loss": 0.4682, "step": 1486 }, { "epoch": 0.4094802380379072, "grad_norm": 0.38031846284866333, "learning_rate": 1e-05, "loss": 0.462, "step": 1488 }, { "epoch": 0.41003061470193664, "grad_norm": 0.3756881654262543, "learning_rate": 1e-05, "loss": 0.4514, "step": 1490 }, { "epoch": 0.4105809913659661, "grad_norm": 0.3663581311702728, "learning_rate": 1e-05, "loss": 0.4482, "step": 1492 }, { "epoch": 0.4111313680299955, "grad_norm": 0.35938966274261475, "learning_rate": 1e-05, "loss": 0.4471, "step": 1494 }, { "epoch": 0.41168174469402496, "grad_norm": 0.3561854064464569, "learning_rate": 1e-05, "loss": 0.4514, "step": 1496 }, { "epoch": 0.4122321213580544, "grad_norm": 0.36052775382995605, "learning_rate": 1e-05, "loss": 0.4564, "step": 1498 }, { "epoch": 0.4127824980220839, "grad_norm": 0.3753555119037628, "learning_rate": 1e-05, "loss": 0.4543, "step": 1500 }, { "epoch": 0.4133328746861133, "grad_norm": 0.3747691810131073, "learning_rate": 1e-05, "loss": 0.4588, "step": 1502 }, { "epoch": 0.41388325135014276, "grad_norm": 0.3654341399669647, "learning_rate": 1e-05, "loss": 0.451, "step": 1504 }, { "epoch": 0.4144336280141722, "grad_norm": 0.3624642491340637, "learning_rate": 1e-05, "loss": 0.4528, "step": 1506 }, { "epoch": 0.4149840046782016, "grad_norm": 0.3465966284275055, "learning_rate": 1e-05, "loss": 0.45, "step": 1508 }, { "epoch": 0.4155343813422311, "grad_norm": 0.38202422857284546, "learning_rate": 1e-05, "loss": 0.4459, "step": 1510 }, { "epoch": 0.41608475800626055, "grad_norm": 0.3562781512737274, "learning_rate": 1e-05, "loss": 0.4375, "step": 1512 }, { "epoch": 0.41663513467028995, "grad_norm": 0.36660805344581604, "learning_rate": 1e-05, "loss": 0.4511, "step": 1514 }, { "epoch": 0.4171855113343194, "grad_norm": 0.36541464924812317, "learning_rate": 1e-05, "loss": 0.4618, "step": 1516 }, { "epoch": 0.4177358879983489, "grad_norm": 0.3570851981639862, "learning_rate": 1e-05, "loss": 0.4568, "step": 1518 }, { "epoch": 0.41828626466237834, "grad_norm": 0.3508870005607605, "learning_rate": 1e-05, "loss": 0.4492, "step": 1520 }, { "epoch": 0.41883664132640774, "grad_norm": 0.35050973296165466, "learning_rate": 1e-05, "loss": 0.4481, "step": 1522 }, { "epoch": 0.4193870179904372, "grad_norm": 0.3564668297767639, "learning_rate": 1e-05, "loss": 0.4461, "step": 1524 }, { "epoch": 0.41993739465446667, "grad_norm": 0.3646043539047241, "learning_rate": 1e-05, "loss": 0.4554, "step": 1526 }, { "epoch": 0.4204877713184961, "grad_norm": 0.3904356360435486, "learning_rate": 1e-05, "loss": 0.4731, "step": 1528 }, { "epoch": 0.42103814798252553, "grad_norm": 0.37373483180999756, "learning_rate": 1e-05, "loss": 0.4679, "step": 1530 }, { "epoch": 0.421588524646555, "grad_norm": 0.3704439699649811, "learning_rate": 1e-05, "loss": 0.4706, "step": 1532 }, { "epoch": 0.42213890131058446, "grad_norm": 0.37894484400749207, "learning_rate": 1e-05, "loss": 0.4515, "step": 1534 }, { "epoch": 0.42268927797461386, "grad_norm": 0.3871210217475891, "learning_rate": 1e-05, "loss": 0.4477, "step": 1536 }, { "epoch": 0.4232396546386433, "grad_norm": 0.3755747079849243, "learning_rate": 1e-05, "loss": 0.4633, "step": 1538 }, { "epoch": 0.4237900313026728, "grad_norm": 0.359764039516449, "learning_rate": 1e-05, "loss": 0.4798, "step": 1540 }, { "epoch": 0.4243404079667022, "grad_norm": 0.37172380089759827, "learning_rate": 1e-05, "loss": 0.4383, "step": 1542 }, { "epoch": 0.42489078463073165, "grad_norm": 0.3501332700252533, "learning_rate": 1e-05, "loss": 0.442, "step": 1544 }, { "epoch": 0.4254411612947611, "grad_norm": 0.3552211821079254, "learning_rate": 1e-05, "loss": 0.4539, "step": 1546 }, { "epoch": 0.4259915379587905, "grad_norm": 0.35052230954170227, "learning_rate": 1e-05, "loss": 0.428, "step": 1548 }, { "epoch": 0.42654191462282, "grad_norm": 0.3710823357105255, "learning_rate": 1e-05, "loss": 0.4297, "step": 1550 }, { "epoch": 0.42709229128684945, "grad_norm": 0.37135034799575806, "learning_rate": 1e-05, "loss": 0.4587, "step": 1552 }, { "epoch": 0.4276426679508789, "grad_norm": 0.3729698061943054, "learning_rate": 1e-05, "loss": 0.4585, "step": 1554 }, { "epoch": 0.4281930446149083, "grad_norm": 0.3525015711784363, "learning_rate": 1e-05, "loss": 0.459, "step": 1556 }, { "epoch": 0.4287434212789378, "grad_norm": 0.38500455021858215, "learning_rate": 1e-05, "loss": 0.4469, "step": 1558 }, { "epoch": 0.42929379794296724, "grad_norm": 0.3852159380912781, "learning_rate": 1e-05, "loss": 0.4421, "step": 1560 }, { "epoch": 0.42984417460699664, "grad_norm": 0.3567640781402588, "learning_rate": 1e-05, "loss": 0.4538, "step": 1562 }, { "epoch": 0.4303945512710261, "grad_norm": 0.36795344948768616, "learning_rate": 1e-05, "loss": 0.4432, "step": 1564 }, { "epoch": 0.43094492793505557, "grad_norm": 0.37614256143569946, "learning_rate": 1e-05, "loss": 0.4631, "step": 1566 }, { "epoch": 0.431495304599085, "grad_norm": 0.356991171836853, "learning_rate": 1e-05, "loss": 0.4389, "step": 1568 }, { "epoch": 0.43204568126311443, "grad_norm": 0.3793700933456421, "learning_rate": 1e-05, "loss": 0.4609, "step": 1570 }, { "epoch": 0.4325960579271439, "grad_norm": 0.36675581336021423, "learning_rate": 1e-05, "loss": 0.4484, "step": 1572 }, { "epoch": 0.43314643459117336, "grad_norm": 0.36404114961624146, "learning_rate": 1e-05, "loss": 0.45, "step": 1574 }, { "epoch": 0.43369681125520276, "grad_norm": 0.3868160843849182, "learning_rate": 1e-05, "loss": 0.4652, "step": 1576 }, { "epoch": 0.4342471879192322, "grad_norm": 0.3898649215698242, "learning_rate": 1e-05, "loss": 0.4612, "step": 1578 }, { "epoch": 0.4347975645832617, "grad_norm": 0.36762335896492004, "learning_rate": 1e-05, "loss": 0.4543, "step": 1580 }, { "epoch": 0.4353479412472911, "grad_norm": 0.3434213101863861, "learning_rate": 1e-05, "loss": 0.4423, "step": 1582 }, { "epoch": 0.43589831791132055, "grad_norm": 0.3741122782230377, "learning_rate": 1e-05, "loss": 0.4638, "step": 1584 }, { "epoch": 0.43644869457535, "grad_norm": 0.38991764187812805, "learning_rate": 1e-05, "loss": 0.438, "step": 1586 }, { "epoch": 0.4369990712393795, "grad_norm": 0.35284510254859924, "learning_rate": 1e-05, "loss": 0.4559, "step": 1588 }, { "epoch": 0.4375494479034089, "grad_norm": 0.36775341629981995, "learning_rate": 1e-05, "loss": 0.4594, "step": 1590 }, { "epoch": 0.43809982456743835, "grad_norm": 0.3677217364311218, "learning_rate": 1e-05, "loss": 0.451, "step": 1592 }, { "epoch": 0.4386502012314678, "grad_norm": 0.35295674204826355, "learning_rate": 1e-05, "loss": 0.4506, "step": 1594 }, { "epoch": 0.4392005778954972, "grad_norm": 0.3770224452018738, "learning_rate": 1e-05, "loss": 0.4506, "step": 1596 }, { "epoch": 0.4397509545595267, "grad_norm": 0.3824670612812042, "learning_rate": 1e-05, "loss": 0.4633, "step": 1598 }, { "epoch": 0.44030133122355614, "grad_norm": 0.38165828585624695, "learning_rate": 1e-05, "loss": 0.4458, "step": 1600 }, { "epoch": 0.44030133122355614, "eval_merge_loss": 0.39449170231819153, "eval_merge_runtime": 599.3899, "eval_merge_samples_per_second": 56.311, "eval_merge_steps_per_second": 2.347, "step": 1600 }, { "epoch": 0.44030133122355614, "eval_new_aug_datas_filtered.json_loss": 0.5198476314544678, "eval_new_aug_datas_filtered.json_runtime": 10.3548, "eval_new_aug_datas_filtered.json_samples_per_second": 74.072, "eval_new_aug_datas_filtered.json_steps_per_second": 3.09, "step": 1600 }, { "epoch": 0.44030133122355614, "eval_sharegpt_gpt4.json_loss": 0.7743993997573853, "eval_sharegpt_gpt4.json_runtime": 31.7173, "eval_sharegpt_gpt4.json_samples_per_second": 58.675, "eval_sharegpt_gpt4.json_steps_per_second": 2.459, "step": 1600 }, { "epoch": 0.44030133122355614, "eval_Table_GPT.json_loss": 0.05817935988306999, "eval_Table_GPT.json_runtime": 25.0301, "eval_Table_GPT.json_samples_per_second": 83.619, "eval_Table_GPT.json_steps_per_second": 3.516, "step": 1600 }, { "epoch": 0.44030133122355614, "eval_gpt_4o_200k.json_loss": 0.8023759126663208, "eval_gpt_4o_200k.json_runtime": 48.5498, "eval_gpt_4o_200k.json_samples_per_second": 129.372, "eval_gpt_4o_200k.json_steps_per_second": 5.397, "step": 1600 }, { "epoch": 0.44030133122355614, "eval_multi_turn_datas.json_loss": 0.3328835964202881, "eval_multi_turn_datas.json_runtime": 75.669, "eval_multi_turn_datas.json_samples_per_second": 52.888, "eval_multi_turn_datas.json_steps_per_second": 2.207, "step": 1600 }, { "epoch": 0.44030133122355614, "eval_table_python_code_datas.json_loss": 0.2713072597980499, "eval_table_python_code_datas.json_runtime": 43.1148, "eval_table_python_code_datas.json_samples_per_second": 50.076, "eval_table_python_code_datas.json_steps_per_second": 2.087, "step": 1600 }, { "epoch": 0.44030133122355614, "eval_tabular_llm_data.json_loss": 0.10233539342880249, "eval_tabular_llm_data.json_runtime": 8.5788, "eval_tabular_llm_data.json_samples_per_second": 28.675, "eval_tabular_llm_data.json_steps_per_second": 1.282, "step": 1600 }, { "epoch": 0.44030133122355614, "eval_python_code_critic_21k.json_loss": 0.5702229142189026, "eval_python_code_critic_21k.json_runtime": 3.2319, "eval_python_code_critic_21k.json_samples_per_second": 184.719, "eval_python_code_critic_21k.json_steps_per_second": 7.735, "step": 1600 }, { "epoch": 0.44030133122355614, "eval_all_merge_table_dataset.json_loss": 0.07606548815965652, "eval_all_merge_table_dataset.json_runtime": 23.3911, "eval_all_merge_table_dataset.json_samples_per_second": 30.439, "eval_all_merge_table_dataset.json_steps_per_second": 1.283, "step": 1600 }, { "epoch": 0.44030133122355614, "eval_code_feedback_multi_turn.json_loss": 0.5824379324913025, "eval_code_feedback_multi_turn.json_runtime": 32.5207, "eval_code_feedback_multi_turn.json_samples_per_second": 67.68, "eval_code_feedback_multi_turn.json_steps_per_second": 2.829, "step": 1600 }, { "epoch": 0.44030133122355614, "eval_ultrainteract_sft.json_loss": 0.42119815945625305, "eval_ultrainteract_sft.json_runtime": 8.677, "eval_ultrainteract_sft.json_samples_per_second": 167.801, "eval_ultrainteract_sft.json_steps_per_second": 7.03, "step": 1600 }, { "epoch": 0.44030133122355614, "eval_synthetic_text_to_sql.json_loss": 0.09474331140518188, "eval_synthetic_text_to_sql.json_runtime": 0.1262, "eval_synthetic_text_to_sql.json_samples_per_second": 269.361, "eval_synthetic_text_to_sql.json_steps_per_second": 15.845, "step": 1600 }, { "epoch": 0.44030133122355614, "eval_sft_react_sql_datas.json_loss": 0.6350359320640564, "eval_sft_react_sql_datas.json_runtime": 7.869, "eval_sft_react_sql_datas.json_samples_per_second": 39.903, "eval_sft_react_sql_datas.json_steps_per_second": 1.779, "step": 1600 }, { "epoch": 0.44030133122355614, "eval_all_merge_code.json_loss": 0.2929154634475708, "eval_all_merge_code.json_runtime": 0.3373, "eval_all_merge_code.json_samples_per_second": 186.752, "eval_all_merge_code.json_steps_per_second": 8.893, "step": 1600 }, { "epoch": 0.44030133122355614, "eval_magpie_datas.json_loss": 0.4318141043186188, "eval_magpie_datas.json_runtime": 2.2195, "eval_magpie_datas.json_samples_per_second": 77.496, "eval_magpie_datas.json_steps_per_second": 3.604, "step": 1600 }, { "epoch": 0.44030133122355614, "eval_train_data_for_qwen.json_loss": 0.00419951044023037, "eval_train_data_for_qwen.json_runtime": 0.2455, "eval_train_data_for_qwen.json_samples_per_second": 40.733, "eval_train_data_for_qwen.json_steps_per_second": 4.073, "step": 1600 }, { "epoch": 0.44030133122355614, "eval_alpaca_cleaned.json_loss": 0.910367488861084, "eval_alpaca_cleaned.json_runtime": 0.1147, "eval_alpaca_cleaned.json_samples_per_second": 235.312, "eval_alpaca_cleaned.json_steps_per_second": 17.431, "step": 1600 }, { "epoch": 0.44030133122355614, "eval_agent_instruct.json_loss": 0.21950356662273407, "eval_agent_instruct.json_runtime": 0.5156, "eval_agent_instruct.json_samples_per_second": 93.094, "eval_agent_instruct.json_steps_per_second": 3.879, "step": 1600 }, { "epoch": 0.44030133122355614, "eval_MathInstruct.json_loss": 0.19855839014053345, "eval_MathInstruct.json_runtime": 0.3654, "eval_MathInstruct.json_samples_per_second": 155.99, "eval_MathInstruct.json_steps_per_second": 8.21, "step": 1600 }, { "epoch": 0.44030133122355614, "eval_tested_143k_python_alpaca.json_loss": 0.4433169662952423, "eval_tested_143k_python_alpaca.json_runtime": 0.3031, "eval_tested_143k_python_alpaca.json_samples_per_second": 112.164, "eval_tested_143k_python_alpaca.json_steps_per_second": 6.598, "step": 1600 }, { "epoch": 0.44030133122355614, "eval_xlam_function_calling_60k.json_loss": 0.008965943939983845, "eval_xlam_function_calling_60k.json_runtime": 0.1008, "eval_xlam_function_calling_60k.json_samples_per_second": 228.26, "eval_xlam_function_calling_60k.json_steps_per_second": 9.924, "step": 1600 }, { "epoch": 0.44030133122355614, "eval_alpaca_data_gpt4_chinese.json_loss": 1.560943603515625, "eval_alpaca_data_gpt4_chinese.json_runtime": 0.0505, "eval_alpaca_data_gpt4_chinese.json_samples_per_second": 316.662, "eval_alpaca_data_gpt4_chinese.json_steps_per_second": 19.791, "step": 1600 }, { "epoch": 0.44030133122355614, "eval_alpaca_gpt4_zh.json_loss": 0.9813264012336731, "eval_alpaca_gpt4_zh.json_runtime": 0.05, "eval_alpaca_gpt4_zh.json_samples_per_second": 219.867, "eval_alpaca_gpt4_zh.json_steps_per_second": 19.988, "step": 1600 }, { "epoch": 0.44030133122355614, "eval_codefeedback_filtered_instruction.json_loss": 0.5885769128799438, "eval_codefeedback_filtered_instruction.json_runtime": 0.4829, "eval_codefeedback_filtered_instruction.json_samples_per_second": 41.42, "eval_codefeedback_filtered_instruction.json_steps_per_second": 2.071, "step": 1600 }, { "epoch": 0.44085170788758554, "grad_norm": 0.36969345808029175, "learning_rate": 1e-05, "loss": 0.4474, "step": 1602 }, { "epoch": 0.441402084551615, "grad_norm": 0.3673281967639923, "learning_rate": 1e-05, "loss": 0.4566, "step": 1604 }, { "epoch": 0.44195246121564447, "grad_norm": 0.3695686459541321, "learning_rate": 1e-05, "loss": 0.4602, "step": 1606 }, { "epoch": 0.4425028378796739, "grad_norm": 0.3653704822063446, "learning_rate": 1e-05, "loss": 0.4489, "step": 1608 }, { "epoch": 0.44305321454370333, "grad_norm": 0.37890321016311646, "learning_rate": 1e-05, "loss": 0.4588, "step": 1610 }, { "epoch": 0.4436035912077328, "grad_norm": 0.34637650847435, "learning_rate": 1e-05, "loss": 0.4554, "step": 1612 }, { "epoch": 0.44415396787176226, "grad_norm": 0.3733616769313812, "learning_rate": 1e-05, "loss": 0.4477, "step": 1614 }, { "epoch": 0.44470434453579166, "grad_norm": 0.3740238547325134, "learning_rate": 1e-05, "loss": 0.4528, "step": 1616 }, { "epoch": 0.4452547211998211, "grad_norm": 0.35610541701316833, "learning_rate": 1e-05, "loss": 0.4487, "step": 1618 }, { "epoch": 0.4458050978638506, "grad_norm": 0.362763911485672, "learning_rate": 1e-05, "loss": 0.4619, "step": 1620 }, { "epoch": 0.44635547452788, "grad_norm": 0.3781318962574005, "learning_rate": 1e-05, "loss": 0.4481, "step": 1622 }, { "epoch": 0.44690585119190945, "grad_norm": 0.40836694836616516, "learning_rate": 1e-05, "loss": 0.4597, "step": 1624 }, { "epoch": 0.4474562278559389, "grad_norm": 0.3662070035934448, "learning_rate": 1e-05, "loss": 0.4466, "step": 1626 }, { "epoch": 0.4480066045199684, "grad_norm": 0.37797635793685913, "learning_rate": 1e-05, "loss": 0.4589, "step": 1628 }, { "epoch": 0.4485569811839978, "grad_norm": 0.3544275462627411, "learning_rate": 1e-05, "loss": 0.4549, "step": 1630 }, { "epoch": 0.44910735784802724, "grad_norm": 0.36321336030960083, "learning_rate": 1e-05, "loss": 0.443, "step": 1632 }, { "epoch": 0.4496577345120567, "grad_norm": 0.45478886365890503, "learning_rate": 1e-05, "loss": 0.4343, "step": 1634 }, { "epoch": 0.4502081111760861, "grad_norm": 0.3670060336589813, "learning_rate": 1e-05, "loss": 0.4463, "step": 1636 }, { "epoch": 0.4507584878401156, "grad_norm": 0.381145715713501, "learning_rate": 1e-05, "loss": 0.4512, "step": 1638 }, { "epoch": 0.45130886450414504, "grad_norm": 0.3729204833507538, "learning_rate": 1e-05, "loss": 0.451, "step": 1640 }, { "epoch": 0.4518592411681745, "grad_norm": 0.36986637115478516, "learning_rate": 1e-05, "loss": 0.4622, "step": 1642 }, { "epoch": 0.4524096178322039, "grad_norm": 0.37230783700942993, "learning_rate": 1e-05, "loss": 0.4377, "step": 1644 }, { "epoch": 0.45295999449623336, "grad_norm": 0.3671816885471344, "learning_rate": 1e-05, "loss": 0.4433, "step": 1646 }, { "epoch": 0.4535103711602628, "grad_norm": 0.359372615814209, "learning_rate": 1e-05, "loss": 0.4512, "step": 1648 }, { "epoch": 0.45406074782429223, "grad_norm": 0.3682217001914978, "learning_rate": 1e-05, "loss": 0.4478, "step": 1650 }, { "epoch": 0.4546111244883217, "grad_norm": 0.3779531419277191, "learning_rate": 1e-05, "loss": 0.4446, "step": 1652 }, { "epoch": 0.45516150115235116, "grad_norm": 0.3579237759113312, "learning_rate": 1e-05, "loss": 0.4432, "step": 1654 }, { "epoch": 0.45571187781638056, "grad_norm": 0.35086673498153687, "learning_rate": 1e-05, "loss": 0.4511, "step": 1656 }, { "epoch": 0.45626225448041, "grad_norm": 0.36263635754585266, "learning_rate": 1e-05, "loss": 0.4552, "step": 1658 }, { "epoch": 0.4568126311444395, "grad_norm": 0.3715769648551941, "learning_rate": 1e-05, "loss": 0.4549, "step": 1660 }, { "epoch": 0.45736300780846895, "grad_norm": 0.36989322304725647, "learning_rate": 1e-05, "loss": 0.4468, "step": 1662 }, { "epoch": 0.45791338447249835, "grad_norm": 0.35716795921325684, "learning_rate": 1e-05, "loss": 0.4506, "step": 1664 }, { "epoch": 0.4584637611365278, "grad_norm": 0.36870133876800537, "learning_rate": 1e-05, "loss": 0.4581, "step": 1666 }, { "epoch": 0.4590141378005573, "grad_norm": 0.36808547377586365, "learning_rate": 1e-05, "loss": 0.4518, "step": 1668 }, { "epoch": 0.4595645144645867, "grad_norm": 0.3777028024196625, "learning_rate": 1e-05, "loss": 0.4526, "step": 1670 }, { "epoch": 0.46011489112861614, "grad_norm": 0.3849789798259735, "learning_rate": 1e-05, "loss": 0.452, "step": 1672 }, { "epoch": 0.4606652677926456, "grad_norm": 0.38168811798095703, "learning_rate": 1e-05, "loss": 0.4408, "step": 1674 }, { "epoch": 0.461215644456675, "grad_norm": 0.3601077198982239, "learning_rate": 1e-05, "loss": 0.4415, "step": 1676 }, { "epoch": 0.4617660211207045, "grad_norm": 0.3658849596977234, "learning_rate": 1e-05, "loss": 0.4461, "step": 1678 }, { "epoch": 0.46231639778473393, "grad_norm": 0.3822179138660431, "learning_rate": 1e-05, "loss": 0.4585, "step": 1680 }, { "epoch": 0.4628667744487634, "grad_norm": 0.38321495056152344, "learning_rate": 1e-05, "loss": 0.4469, "step": 1682 }, { "epoch": 0.4634171511127928, "grad_norm": 0.3911297917366028, "learning_rate": 1e-05, "loss": 0.4522, "step": 1684 }, { "epoch": 0.46396752777682226, "grad_norm": 0.38053110241889954, "learning_rate": 1e-05, "loss": 0.4487, "step": 1686 }, { "epoch": 0.4645179044408517, "grad_norm": 0.3704802691936493, "learning_rate": 1e-05, "loss": 0.4436, "step": 1688 }, { "epoch": 0.46506828110488113, "grad_norm": 0.3804566562175751, "learning_rate": 1e-05, "loss": 0.4419, "step": 1690 }, { "epoch": 0.4656186577689106, "grad_norm": 0.3807014524936676, "learning_rate": 1e-05, "loss": 0.4526, "step": 1692 }, { "epoch": 0.46616903443294005, "grad_norm": 0.3678591549396515, "learning_rate": 1e-05, "loss": 0.4579, "step": 1694 }, { "epoch": 0.46671941109696946, "grad_norm": 0.37586984038352966, "learning_rate": 1e-05, "loss": 0.4404, "step": 1696 }, { "epoch": 0.4672697877609989, "grad_norm": 0.36084264516830444, "learning_rate": 1e-05, "loss": 0.4398, "step": 1698 }, { "epoch": 0.4678201644250284, "grad_norm": 0.36694666743278503, "learning_rate": 1e-05, "loss": 0.4369, "step": 1700 }, { "epoch": 0.46837054108905785, "grad_norm": 0.4061066210269928, "learning_rate": 1e-05, "loss": 0.4495, "step": 1702 }, { "epoch": 0.46892091775308725, "grad_norm": 0.37329551577568054, "learning_rate": 1e-05, "loss": 0.4482, "step": 1704 }, { "epoch": 0.4694712944171167, "grad_norm": 0.39072346687316895, "learning_rate": 1e-05, "loss": 0.4506, "step": 1706 }, { "epoch": 0.4700216710811462, "grad_norm": 0.3565053343772888, "learning_rate": 1e-05, "loss": 0.447, "step": 1708 }, { "epoch": 0.4705720477451756, "grad_norm": 0.39754360914230347, "learning_rate": 1e-05, "loss": 0.4468, "step": 1710 }, { "epoch": 0.47112242440920504, "grad_norm": 0.34416159987449646, "learning_rate": 1e-05, "loss": 0.4509, "step": 1712 }, { "epoch": 0.4716728010732345, "grad_norm": 0.3646188974380493, "learning_rate": 1e-05, "loss": 0.4436, "step": 1714 }, { "epoch": 0.47222317773726397, "grad_norm": 0.372549831867218, "learning_rate": 1e-05, "loss": 0.4622, "step": 1716 }, { "epoch": 0.47277355440129337, "grad_norm": 0.34616753458976746, "learning_rate": 1e-05, "loss": 0.4513, "step": 1718 }, { "epoch": 0.47332393106532283, "grad_norm": 0.39396756887435913, "learning_rate": 1e-05, "loss": 0.4464, "step": 1720 }, { "epoch": 0.4738743077293523, "grad_norm": 0.3681057095527649, "learning_rate": 1e-05, "loss": 0.4514, "step": 1722 }, { "epoch": 0.4744246843933817, "grad_norm": 0.38942328095436096, "learning_rate": 1e-05, "loss": 0.4603, "step": 1724 }, { "epoch": 0.47497506105741116, "grad_norm": 0.380278617143631, "learning_rate": 1e-05, "loss": 0.4463, "step": 1726 }, { "epoch": 0.4755254377214406, "grad_norm": 0.37930282950401306, "learning_rate": 1e-05, "loss": 0.4377, "step": 1728 }, { "epoch": 0.47607581438547003, "grad_norm": 0.36719146370887756, "learning_rate": 1e-05, "loss": 0.4285, "step": 1730 }, { "epoch": 0.4766261910494995, "grad_norm": 0.3802686035633087, "learning_rate": 1e-05, "loss": 0.4346, "step": 1732 }, { "epoch": 0.47717656771352895, "grad_norm": 0.3655955493450165, "learning_rate": 1e-05, "loss": 0.4504, "step": 1734 }, { "epoch": 0.4777269443775584, "grad_norm": 0.34403982758522034, "learning_rate": 1e-05, "loss": 0.4502, "step": 1736 }, { "epoch": 0.4782773210415878, "grad_norm": 0.35954922437667847, "learning_rate": 1e-05, "loss": 0.4313, "step": 1738 }, { "epoch": 0.4788276977056173, "grad_norm": 0.3489810824394226, "learning_rate": 1e-05, "loss": 0.4479, "step": 1740 }, { "epoch": 0.47937807436964675, "grad_norm": 0.3789598047733307, "learning_rate": 1e-05, "loss": 0.4488, "step": 1742 }, { "epoch": 0.47992845103367615, "grad_norm": 0.38226747512817383, "learning_rate": 1e-05, "loss": 0.4612, "step": 1744 }, { "epoch": 0.4804788276977056, "grad_norm": 0.36648547649383545, "learning_rate": 1e-05, "loss": 0.4521, "step": 1746 }, { "epoch": 0.4810292043617351, "grad_norm": 0.36434775590896606, "learning_rate": 1e-05, "loss": 0.4579, "step": 1748 }, { "epoch": 0.4815795810257645, "grad_norm": 0.3805695176124573, "learning_rate": 1e-05, "loss": 0.437, "step": 1750 }, { "epoch": 0.48212995768979394, "grad_norm": 0.34234747290611267, "learning_rate": 1e-05, "loss": 0.4411, "step": 1752 }, { "epoch": 0.4826803343538234, "grad_norm": 0.356953501701355, "learning_rate": 1e-05, "loss": 0.4563, "step": 1754 }, { "epoch": 0.48323071101785287, "grad_norm": 0.35372647643089294, "learning_rate": 1e-05, "loss": 0.4506, "step": 1756 }, { "epoch": 0.48378108768188227, "grad_norm": 0.3776678442955017, "learning_rate": 1e-05, "loss": 0.4517, "step": 1758 }, { "epoch": 0.48433146434591173, "grad_norm": 0.336029052734375, "learning_rate": 1e-05, "loss": 0.4387, "step": 1760 }, { "epoch": 0.4848818410099412, "grad_norm": 0.35482755303382874, "learning_rate": 1e-05, "loss": 0.4456, "step": 1762 }, { "epoch": 0.4854322176739706, "grad_norm": 0.3713533580303192, "learning_rate": 1e-05, "loss": 0.4616, "step": 1764 }, { "epoch": 0.48598259433800006, "grad_norm": 0.348069965839386, "learning_rate": 1e-05, "loss": 0.4504, "step": 1766 }, { "epoch": 0.4865329710020295, "grad_norm": 0.36832061409950256, "learning_rate": 1e-05, "loss": 0.45, "step": 1768 }, { "epoch": 0.487083347666059, "grad_norm": 0.3665439486503601, "learning_rate": 1e-05, "loss": 0.4525, "step": 1770 }, { "epoch": 0.4876337243300884, "grad_norm": 0.39572247862815857, "learning_rate": 1e-05, "loss": 0.4521, "step": 1772 }, { "epoch": 0.48818410099411785, "grad_norm": 0.36583212018013, "learning_rate": 1e-05, "loss": 0.4298, "step": 1774 }, { "epoch": 0.4887344776581473, "grad_norm": 0.35969898104667664, "learning_rate": 1e-05, "loss": 0.4497, "step": 1776 }, { "epoch": 0.4892848543221767, "grad_norm": 0.3651510775089264, "learning_rate": 1e-05, "loss": 0.4355, "step": 1778 }, { "epoch": 0.4898352309862062, "grad_norm": 0.3885847330093384, "learning_rate": 1e-05, "loss": 0.4633, "step": 1780 }, { "epoch": 0.49038560765023564, "grad_norm": 0.357166588306427, "learning_rate": 1e-05, "loss": 0.4512, "step": 1782 }, { "epoch": 0.49093598431426505, "grad_norm": 0.34748879075050354, "learning_rate": 1e-05, "loss": 0.437, "step": 1784 }, { "epoch": 0.4914863609782945, "grad_norm": 0.371999055147171, "learning_rate": 1e-05, "loss": 0.4493, "step": 1786 }, { "epoch": 0.492036737642324, "grad_norm": 0.3602544665336609, "learning_rate": 1e-05, "loss": 0.4413, "step": 1788 }, { "epoch": 0.49258711430635344, "grad_norm": 0.38811835646629333, "learning_rate": 1e-05, "loss": 0.4406, "step": 1790 }, { "epoch": 0.49313749097038284, "grad_norm": 0.366616427898407, "learning_rate": 1e-05, "loss": 0.4587, "step": 1792 }, { "epoch": 0.4936878676344123, "grad_norm": 0.39588844776153564, "learning_rate": 1e-05, "loss": 0.4525, "step": 1794 }, { "epoch": 0.49423824429844176, "grad_norm": 0.3641244173049927, "learning_rate": 1e-05, "loss": 0.4533, "step": 1796 }, { "epoch": 0.49478862096247117, "grad_norm": 0.35738009214401245, "learning_rate": 1e-05, "loss": 0.4542, "step": 1798 }, { "epoch": 0.49533899762650063, "grad_norm": 0.36343181133270264, "learning_rate": 1e-05, "loss": 0.4527, "step": 1800 }, { "epoch": 0.49533899762650063, "eval_merge_loss": 0.3900485932826996, "eval_merge_runtime": 600.246, "eval_merge_samples_per_second": 56.23, "eval_merge_steps_per_second": 2.344, "step": 1800 }, { "epoch": 0.49533899762650063, "eval_new_aug_datas_filtered.json_loss": 0.5161438584327698, "eval_new_aug_datas_filtered.json_runtime": 10.4655, "eval_new_aug_datas_filtered.json_samples_per_second": 73.288, "eval_new_aug_datas_filtered.json_steps_per_second": 3.058, "step": 1800 }, { "epoch": 0.49533899762650063, "eval_sharegpt_gpt4.json_loss": 0.7699668407440186, "eval_sharegpt_gpt4.json_runtime": 31.6447, "eval_sharegpt_gpt4.json_samples_per_second": 58.809, "eval_sharegpt_gpt4.json_steps_per_second": 2.465, "step": 1800 }, { "epoch": 0.49533899762650063, "eval_Table_GPT.json_loss": 0.057397227734327316, "eval_Table_GPT.json_runtime": 24.974, "eval_Table_GPT.json_samples_per_second": 83.807, "eval_Table_GPT.json_steps_per_second": 3.524, "step": 1800 }, { "epoch": 0.49533899762650063, "eval_gpt_4o_200k.json_loss": 0.7959992289543152, "eval_gpt_4o_200k.json_runtime": 48.4474, "eval_gpt_4o_200k.json_samples_per_second": 129.646, "eval_gpt_4o_200k.json_steps_per_second": 5.408, "step": 1800 }, { "epoch": 0.49533899762650063, "eval_multi_turn_datas.json_loss": 0.326607346534729, "eval_multi_turn_datas.json_runtime": 75.6077, "eval_multi_turn_datas.json_samples_per_second": 52.931, "eval_multi_turn_datas.json_steps_per_second": 2.209, "step": 1800 }, { "epoch": 0.49533899762650063, "eval_table_python_code_datas.json_loss": 0.26808008551597595, "eval_table_python_code_datas.json_runtime": 43.0557, "eval_table_python_code_datas.json_samples_per_second": 50.144, "eval_table_python_code_datas.json_steps_per_second": 2.09, "step": 1800 }, { "epoch": 0.49533899762650063, "eval_tabular_llm_data.json_loss": 0.1004142090678215, "eval_tabular_llm_data.json_runtime": 8.5429, "eval_tabular_llm_data.json_samples_per_second": 28.796, "eval_tabular_llm_data.json_steps_per_second": 1.288, "step": 1800 }, { "epoch": 0.49533899762650063, "eval_python_code_critic_21k.json_loss": 0.5654606223106384, "eval_python_code_critic_21k.json_runtime": 3.2351, "eval_python_code_critic_21k.json_samples_per_second": 184.538, "eval_python_code_critic_21k.json_steps_per_second": 7.728, "step": 1800 }, { "epoch": 0.49533899762650063, "eval_all_merge_table_dataset.json_loss": 0.07576768845319748, "eval_all_merge_table_dataset.json_runtime": 23.2598, "eval_all_merge_table_dataset.json_samples_per_second": 30.611, "eval_all_merge_table_dataset.json_steps_per_second": 1.29, "step": 1800 }, { "epoch": 0.49533899762650063, "eval_code_feedback_multi_turn.json_loss": 0.579846203327179, "eval_code_feedback_multi_turn.json_runtime": 32.4188, "eval_code_feedback_multi_turn.json_samples_per_second": 67.893, "eval_code_feedback_multi_turn.json_steps_per_second": 2.838, "step": 1800 }, { "epoch": 0.49533899762650063, "eval_ultrainteract_sft.json_loss": 0.4181068241596222, "eval_ultrainteract_sft.json_runtime": 8.6461, "eval_ultrainteract_sft.json_samples_per_second": 168.4, "eval_ultrainteract_sft.json_steps_per_second": 7.055, "step": 1800 }, { "epoch": 0.49533899762650063, "eval_synthetic_text_to_sql.json_loss": 0.09818249940872192, "eval_synthetic_text_to_sql.json_runtime": 0.1264, "eval_synthetic_text_to_sql.json_samples_per_second": 269.092, "eval_synthetic_text_to_sql.json_steps_per_second": 15.829, "step": 1800 }, { "epoch": 0.49533899762650063, "eval_sft_react_sql_datas.json_loss": 0.6291559338569641, "eval_sft_react_sql_datas.json_runtime": 7.8451, "eval_sft_react_sql_datas.json_samples_per_second": 40.025, "eval_sft_react_sql_datas.json_steps_per_second": 1.785, "step": 1800 }, { "epoch": 0.49533899762650063, "eval_all_merge_code.json_loss": 0.29108163714408875, "eval_all_merge_code.json_runtime": 0.3447, "eval_all_merge_code.json_samples_per_second": 182.771, "eval_all_merge_code.json_steps_per_second": 8.703, "step": 1800 }, { "epoch": 0.49533899762650063, "eval_magpie_datas.json_loss": 0.43020525574684143, "eval_magpie_datas.json_runtime": 2.2179, "eval_magpie_datas.json_samples_per_second": 77.551, "eval_magpie_datas.json_steps_per_second": 3.607, "step": 1800 }, { "epoch": 0.49533899762650063, "eval_train_data_for_qwen.json_loss": 0.0027856978122144938, "eval_train_data_for_qwen.json_runtime": 0.2444, "eval_train_data_for_qwen.json_samples_per_second": 40.919, "eval_train_data_for_qwen.json_steps_per_second": 4.092, "step": 1800 }, { "epoch": 0.49533899762650063, "eval_alpaca_cleaned.json_loss": 0.9129724502563477, "eval_alpaca_cleaned.json_runtime": 0.1153, "eval_alpaca_cleaned.json_samples_per_second": 234.093, "eval_alpaca_cleaned.json_steps_per_second": 17.34, "step": 1800 }, { "epoch": 0.49533899762650063, "eval_agent_instruct.json_loss": 0.22024483978748322, "eval_agent_instruct.json_runtime": 0.5149, "eval_agent_instruct.json_samples_per_second": 93.222, "eval_agent_instruct.json_steps_per_second": 3.884, "step": 1800 }, { "epoch": 0.49533899762650063, "eval_MathInstruct.json_loss": 0.20060402154922485, "eval_MathInstruct.json_runtime": 0.3648, "eval_MathInstruct.json_samples_per_second": 156.23, "eval_MathInstruct.json_steps_per_second": 8.223, "step": 1800 }, { "epoch": 0.49533899762650063, "eval_tested_143k_python_alpaca.json_loss": 0.44536128640174866, "eval_tested_143k_python_alpaca.json_runtime": 0.3002, "eval_tested_143k_python_alpaca.json_samples_per_second": 113.24, "eval_tested_143k_python_alpaca.json_steps_per_second": 6.661, "step": 1800 }, { "epoch": 0.49533899762650063, "eval_xlam_function_calling_60k.json_loss": 0.00967579148709774, "eval_xlam_function_calling_60k.json_runtime": 0.1002, "eval_xlam_function_calling_60k.json_samples_per_second": 229.452, "eval_xlam_function_calling_60k.json_steps_per_second": 9.976, "step": 1800 }, { "epoch": 0.49533899762650063, "eval_alpaca_data_gpt4_chinese.json_loss": 1.5544477701187134, "eval_alpaca_data_gpt4_chinese.json_runtime": 0.0511, "eval_alpaca_data_gpt4_chinese.json_samples_per_second": 313.214, "eval_alpaca_data_gpt4_chinese.json_steps_per_second": 19.576, "step": 1800 }, { "epoch": 0.49533899762650063, "eval_alpaca_gpt4_zh.json_loss": 0.977000892162323, "eval_alpaca_gpt4_zh.json_runtime": 0.0508, "eval_alpaca_gpt4_zh.json_samples_per_second": 216.666, "eval_alpaca_gpt4_zh.json_steps_per_second": 19.697, "step": 1800 }, { "epoch": 0.49533899762650063, "eval_codefeedback_filtered_instruction.json_loss": 0.5895399451255798, "eval_codefeedback_filtered_instruction.json_runtime": 0.4883, "eval_codefeedback_filtered_instruction.json_samples_per_second": 40.957, "eval_codefeedback_filtered_instruction.json_steps_per_second": 2.048, "step": 1800 }, { "epoch": 0.4958893742905301, "grad_norm": 0.36430442333221436, "learning_rate": 1e-05, "loss": 0.4418, "step": 1802 }, { "epoch": 0.4964397509545595, "grad_norm": 0.35012543201446533, "learning_rate": 1e-05, "loss": 0.4437, "step": 1804 }, { "epoch": 0.49699012761858896, "grad_norm": 0.3726542294025421, "learning_rate": 1e-05, "loss": 0.4332, "step": 1806 }, { "epoch": 0.4975405042826184, "grad_norm": 0.3564360439777374, "learning_rate": 1e-05, "loss": 0.4378, "step": 1808 }, { "epoch": 0.4980908809466479, "grad_norm": 0.3730456233024597, "learning_rate": 1e-05, "loss": 0.443, "step": 1810 }, { "epoch": 0.4986412576106773, "grad_norm": 0.3588622212409973, "learning_rate": 1e-05, "loss": 0.4387, "step": 1812 }, { "epoch": 0.49919163427470675, "grad_norm": 0.36861783266067505, "learning_rate": 1e-05, "loss": 0.4392, "step": 1814 }, { "epoch": 0.4997420109387362, "grad_norm": 0.3537515699863434, "learning_rate": 1e-05, "loss": 0.4331, "step": 1816 }, { "epoch": 0.5002923876027656, "grad_norm": 0.3723071813583374, "learning_rate": 1e-05, "loss": 0.4429, "step": 1818 }, { "epoch": 0.5008427642667951, "grad_norm": 0.37015634775161743, "learning_rate": 1e-05, "loss": 0.4687, "step": 1820 }, { "epoch": 0.5013931409308245, "grad_norm": 0.3528953790664673, "learning_rate": 1e-05, "loss": 0.4315, "step": 1822 }, { "epoch": 0.501943517594854, "grad_norm": 0.357120543718338, "learning_rate": 1e-05, "loss": 0.4423, "step": 1824 }, { "epoch": 0.5024938942588835, "grad_norm": 0.3655802607536316, "learning_rate": 1e-05, "loss": 0.4475, "step": 1826 }, { "epoch": 0.5030442709229128, "grad_norm": 0.3676040470600128, "learning_rate": 1e-05, "loss": 0.4345, "step": 1828 }, { "epoch": 0.5035946475869423, "grad_norm": 0.3427799940109253, "learning_rate": 1e-05, "loss": 0.4422, "step": 1830 }, { "epoch": 0.5041450242509717, "grad_norm": 0.3482607305049896, "learning_rate": 1e-05, "loss": 0.4347, "step": 1832 }, { "epoch": 0.5046954009150012, "grad_norm": 0.3690313398838043, "learning_rate": 1e-05, "loss": 0.4572, "step": 1834 }, { "epoch": 0.5052457775790307, "grad_norm": 0.351601243019104, "learning_rate": 1e-05, "loss": 0.4445, "step": 1836 }, { "epoch": 0.5057961542430601, "grad_norm": 0.3506658971309662, "learning_rate": 1e-05, "loss": 0.4482, "step": 1838 }, { "epoch": 0.5063465309070896, "grad_norm": 0.36706456542015076, "learning_rate": 1e-05, "loss": 0.4503, "step": 1840 }, { "epoch": 0.5068969075711189, "grad_norm": 0.36632585525512695, "learning_rate": 1e-05, "loss": 0.4385, "step": 1842 }, { "epoch": 0.5074472842351484, "grad_norm": 0.3675621747970581, "learning_rate": 1e-05, "loss": 0.4391, "step": 1844 }, { "epoch": 0.5079976608991779, "grad_norm": 0.3883734941482544, "learning_rate": 1e-05, "loss": 0.4435, "step": 1846 }, { "epoch": 0.5085480375632073, "grad_norm": 0.34348422288894653, "learning_rate": 1e-05, "loss": 0.4388, "step": 1848 }, { "epoch": 0.5090984142272368, "grad_norm": 0.36695536971092224, "learning_rate": 1e-05, "loss": 0.4473, "step": 1850 }, { "epoch": 0.5096487908912662, "grad_norm": 0.36929944157600403, "learning_rate": 1e-05, "loss": 0.4542, "step": 1852 }, { "epoch": 0.5101991675552957, "grad_norm": 0.3946716785430908, "learning_rate": 1e-05, "loss": 0.4399, "step": 1854 }, { "epoch": 0.5107495442193251, "grad_norm": 0.3619132936000824, "learning_rate": 1e-05, "loss": 0.4471, "step": 1856 }, { "epoch": 0.5112999208833545, "grad_norm": 0.34836745262145996, "learning_rate": 1e-05, "loss": 0.4392, "step": 1858 }, { "epoch": 0.511850297547384, "grad_norm": 0.37516769766807556, "learning_rate": 1e-05, "loss": 0.4579, "step": 1860 }, { "epoch": 0.5124006742114134, "grad_norm": 0.35800984501838684, "learning_rate": 1e-05, "loss": 0.4479, "step": 1862 }, { "epoch": 0.5129510508754429, "grad_norm": 0.3664796054363251, "learning_rate": 1e-05, "loss": 0.4556, "step": 1864 }, { "epoch": 0.5135014275394724, "grad_norm": 0.3633113503456116, "learning_rate": 1e-05, "loss": 0.4405, "step": 1866 }, { "epoch": 0.5140518042035017, "grad_norm": 0.3655359447002411, "learning_rate": 1e-05, "loss": 0.4486, "step": 1868 }, { "epoch": 0.5146021808675312, "grad_norm": 0.36135318875312805, "learning_rate": 1e-05, "loss": 0.4473, "step": 1870 }, { "epoch": 0.5151525575315606, "grad_norm": 0.4725627601146698, "learning_rate": 1e-05, "loss": 0.4579, "step": 1872 }, { "epoch": 0.5157029341955901, "grad_norm": 0.37844300270080566, "learning_rate": 1e-05, "loss": 0.4502, "step": 1874 }, { "epoch": 0.5162533108596196, "grad_norm": 0.35601717233657837, "learning_rate": 1e-05, "loss": 0.4392, "step": 1876 }, { "epoch": 0.516803687523649, "grad_norm": 0.3960351049900055, "learning_rate": 1e-05, "loss": 0.4519, "step": 1878 }, { "epoch": 0.5173540641876785, "grad_norm": 0.3775772154331207, "learning_rate": 1e-05, "loss": 0.4553, "step": 1880 }, { "epoch": 0.5179044408517078, "grad_norm": 0.3815532624721527, "learning_rate": 1e-05, "loss": 0.4479, "step": 1882 }, { "epoch": 0.5184548175157373, "grad_norm": 0.3661166727542877, "learning_rate": 1e-05, "loss": 0.4423, "step": 1884 }, { "epoch": 0.5190051941797668, "grad_norm": 0.3378327786922455, "learning_rate": 1e-05, "loss": 0.4419, "step": 1886 }, { "epoch": 0.5195555708437962, "grad_norm": 0.34638261795043945, "learning_rate": 1e-05, "loss": 0.4379, "step": 1888 }, { "epoch": 0.5201059475078257, "grad_norm": 0.35764721035957336, "learning_rate": 1e-05, "loss": 0.4389, "step": 1890 }, { "epoch": 0.5206563241718551, "grad_norm": 0.3674796223640442, "learning_rate": 1e-05, "loss": 0.4438, "step": 1892 }, { "epoch": 0.5212067008358846, "grad_norm": 0.34744736552238464, "learning_rate": 1e-05, "loss": 0.4317, "step": 1894 }, { "epoch": 0.521757077499914, "grad_norm": 0.39198940992355347, "learning_rate": 1e-05, "loss": 0.4406, "step": 1896 }, { "epoch": 0.5223074541639434, "grad_norm": 0.3545363247394562, "learning_rate": 1e-05, "loss": 0.4255, "step": 1898 }, { "epoch": 0.5228578308279729, "grad_norm": 0.3635193407535553, "learning_rate": 1e-05, "loss": 0.4521, "step": 1900 }, { "epoch": 0.5234082074920023, "grad_norm": 0.33844560384750366, "learning_rate": 1e-05, "loss": 0.4371, "step": 1902 }, { "epoch": 0.5239585841560318, "grad_norm": 0.34886521100997925, "learning_rate": 1e-05, "loss": 0.4328, "step": 1904 }, { "epoch": 0.5245089608200613, "grad_norm": 0.34973517060279846, "learning_rate": 1e-05, "loss": 0.4442, "step": 1906 }, { "epoch": 0.5250593374840907, "grad_norm": 0.35180777311325073, "learning_rate": 1e-05, "loss": 0.4575, "step": 1908 }, { "epoch": 0.5256097141481201, "grad_norm": 0.36237335205078125, "learning_rate": 1e-05, "loss": 0.4357, "step": 1910 }, { "epoch": 0.5261600908121495, "grad_norm": 0.3784085512161255, "learning_rate": 1e-05, "loss": 0.4559, "step": 1912 }, { "epoch": 0.526710467476179, "grad_norm": 0.3556850254535675, "learning_rate": 1e-05, "loss": 0.4563, "step": 1914 }, { "epoch": 0.5272608441402085, "grad_norm": 0.3620041310787201, "learning_rate": 1e-05, "loss": 0.4458, "step": 1916 }, { "epoch": 0.5278112208042379, "grad_norm": 0.3616819679737091, "learning_rate": 1e-05, "loss": 0.4304, "step": 1918 }, { "epoch": 0.5283615974682674, "grad_norm": 0.3651537597179413, "learning_rate": 1e-05, "loss": 0.4463, "step": 1920 }, { "epoch": 0.5289119741322967, "grad_norm": 0.3924584686756134, "learning_rate": 1e-05, "loss": 0.4418, "step": 1922 }, { "epoch": 0.5294623507963262, "grad_norm": 0.353217214345932, "learning_rate": 1e-05, "loss": 0.4437, "step": 1924 }, { "epoch": 0.5300127274603557, "grad_norm": 0.3897522985935211, "learning_rate": 1e-05, "loss": 0.4549, "step": 1926 }, { "epoch": 0.5305631041243851, "grad_norm": 0.36462587118148804, "learning_rate": 1e-05, "loss": 0.4247, "step": 1928 }, { "epoch": 0.5311134807884146, "grad_norm": 0.3874776363372803, "learning_rate": 1e-05, "loss": 0.4502, "step": 1930 }, { "epoch": 0.531663857452444, "grad_norm": 0.3533260226249695, "learning_rate": 1e-05, "loss": 0.4515, "step": 1932 }, { "epoch": 0.5322142341164735, "grad_norm": 0.3668268024921417, "learning_rate": 1e-05, "loss": 0.4474, "step": 1934 }, { "epoch": 0.5327646107805029, "grad_norm": 0.3501083254814148, "learning_rate": 1e-05, "loss": 0.4344, "step": 1936 }, { "epoch": 0.5333149874445323, "grad_norm": 0.3565337657928467, "learning_rate": 1e-05, "loss": 0.4412, "step": 1938 }, { "epoch": 0.5338653641085618, "grad_norm": 0.34048742055892944, "learning_rate": 1e-05, "loss": 0.4502, "step": 1940 }, { "epoch": 0.5344157407725912, "grad_norm": 0.35694393515586853, "learning_rate": 1e-05, "loss": 0.4532, "step": 1942 }, { "epoch": 0.5349661174366207, "grad_norm": 0.3527338206768036, "learning_rate": 1e-05, "loss": 0.4378, "step": 1944 }, { "epoch": 0.5355164941006502, "grad_norm": 0.3684084117412567, "learning_rate": 1e-05, "loss": 0.4562, "step": 1946 }, { "epoch": 0.5360668707646796, "grad_norm": 0.3584345281124115, "learning_rate": 1e-05, "loss": 0.4561, "step": 1948 }, { "epoch": 0.536617247428709, "grad_norm": 0.35685622692108154, "learning_rate": 1e-05, "loss": 0.4532, "step": 1950 }, { "epoch": 0.5371676240927384, "grad_norm": 0.36560460925102234, "learning_rate": 1e-05, "loss": 0.4529, "step": 1952 }, { "epoch": 0.5377180007567679, "grad_norm": 0.36613890528678894, "learning_rate": 1e-05, "loss": 0.4536, "step": 1954 }, { "epoch": 0.5382683774207974, "grad_norm": 0.3513580858707428, "learning_rate": 1e-05, "loss": 0.4496, "step": 1956 }, { "epoch": 0.5388187540848268, "grad_norm": 0.38372403383255005, "learning_rate": 1e-05, "loss": 0.4506, "step": 1958 }, { "epoch": 0.5393691307488563, "grad_norm": 0.35690757632255554, "learning_rate": 1e-05, "loss": 0.4371, "step": 1960 }, { "epoch": 0.5399195074128857, "grad_norm": 0.36706483364105225, "learning_rate": 1e-05, "loss": 0.4292, "step": 1962 }, { "epoch": 0.5404698840769151, "grad_norm": 0.35754841566085815, "learning_rate": 1e-05, "loss": 0.4543, "step": 1964 }, { "epoch": 0.5410202607409446, "grad_norm": 0.35544702410697937, "learning_rate": 1e-05, "loss": 0.4522, "step": 1966 }, { "epoch": 0.541570637404974, "grad_norm": 0.3689357042312622, "learning_rate": 1e-05, "loss": 0.4447, "step": 1968 }, { "epoch": 0.5421210140690035, "grad_norm": 0.35911116003990173, "learning_rate": 1e-05, "loss": 0.4253, "step": 1970 }, { "epoch": 0.5426713907330329, "grad_norm": 0.3458103537559509, "learning_rate": 1e-05, "loss": 0.4398, "step": 1972 }, { "epoch": 0.5432217673970624, "grad_norm": 0.3606932759284973, "learning_rate": 1e-05, "loss": 0.4486, "step": 1974 }, { "epoch": 0.5437721440610918, "grad_norm": 0.3759188652038574, "learning_rate": 1e-05, "loss": 0.4339, "step": 1976 }, { "epoch": 0.5443225207251212, "grad_norm": 0.3803597390651703, "learning_rate": 1e-05, "loss": 0.4575, "step": 1978 }, { "epoch": 0.5448728973891507, "grad_norm": 0.36220523715019226, "learning_rate": 1e-05, "loss": 0.4427, "step": 1980 }, { "epoch": 0.5454232740531801, "grad_norm": 0.36756813526153564, "learning_rate": 1e-05, "loss": 0.4297, "step": 1982 }, { "epoch": 0.5459736507172096, "grad_norm": 0.35930246114730835, "learning_rate": 1e-05, "loss": 0.4375, "step": 1984 }, { "epoch": 0.5465240273812391, "grad_norm": 0.38998985290527344, "learning_rate": 1e-05, "loss": 0.4331, "step": 1986 }, { "epoch": 0.5470744040452685, "grad_norm": 0.35975074768066406, "learning_rate": 1e-05, "loss": 0.4493, "step": 1988 }, { "epoch": 0.5476247807092979, "grad_norm": 0.3618590533733368, "learning_rate": 1e-05, "loss": 0.4431, "step": 1990 }, { "epoch": 0.5481751573733273, "grad_norm": 0.3768090009689331, "learning_rate": 1e-05, "loss": 0.4414, "step": 1992 }, { "epoch": 0.5487255340373568, "grad_norm": 0.3526524305343628, "learning_rate": 1e-05, "loss": 0.4349, "step": 1994 }, { "epoch": 0.5492759107013863, "grad_norm": 0.3426629900932312, "learning_rate": 1e-05, "loss": 0.4345, "step": 1996 }, { "epoch": 0.5498262873654157, "grad_norm": 0.3500785529613495, "learning_rate": 1e-05, "loss": 0.4415, "step": 1998 }, { "epoch": 0.5503766640294452, "grad_norm": 0.3602929413318634, "learning_rate": 1e-05, "loss": 0.4454, "step": 2000 }, { "epoch": 0.5503766640294452, "eval_merge_loss": 0.3855894207954407, "eval_merge_runtime": 600.0048, "eval_merge_samples_per_second": 56.253, "eval_merge_steps_per_second": 2.345, "step": 2000 }, { "epoch": 0.5503766640294452, "eval_new_aug_datas_filtered.json_loss": 0.5099759697914124, "eval_new_aug_datas_filtered.json_runtime": 10.3782, "eval_new_aug_datas_filtered.json_samples_per_second": 73.905, "eval_new_aug_datas_filtered.json_steps_per_second": 3.083, "step": 2000 }, { "epoch": 0.5503766640294452, "eval_sharegpt_gpt4.json_loss": 0.763576865196228, "eval_sharegpt_gpt4.json_runtime": 31.7204, "eval_sharegpt_gpt4.json_samples_per_second": 58.669, "eval_sharegpt_gpt4.json_steps_per_second": 2.459, "step": 2000 }, { "epoch": 0.5503766640294452, "eval_Table_GPT.json_loss": 0.055675260722637177, "eval_Table_GPT.json_runtime": 24.9781, "eval_Table_GPT.json_samples_per_second": 83.793, "eval_Table_GPT.json_steps_per_second": 3.523, "step": 2000 }, { "epoch": 0.5503766640294452, "eval_gpt_4o_200k.json_loss": 0.7919400334358215, "eval_gpt_4o_200k.json_runtime": 48.5207, "eval_gpt_4o_200k.json_samples_per_second": 129.45, "eval_gpt_4o_200k.json_steps_per_second": 5.4, "step": 2000 }, { "epoch": 0.5503766640294452, "eval_multi_turn_datas.json_loss": 0.321598082780838, "eval_multi_turn_datas.json_runtime": 75.7401, "eval_multi_turn_datas.json_samples_per_second": 52.839, "eval_multi_turn_datas.json_steps_per_second": 2.205, "step": 2000 }, { "epoch": 0.5503766640294452, "eval_table_python_code_datas.json_loss": 0.26337531208992004, "eval_table_python_code_datas.json_runtime": 43.1695, "eval_table_python_code_datas.json_samples_per_second": 50.012, "eval_table_python_code_datas.json_steps_per_second": 2.085, "step": 2000 }, { "epoch": 0.5503766640294452, "eval_tabular_llm_data.json_loss": 0.09393570572137833, "eval_tabular_llm_data.json_runtime": 8.5822, "eval_tabular_llm_data.json_samples_per_second": 28.664, "eval_tabular_llm_data.json_steps_per_second": 1.282, "step": 2000 }, { "epoch": 0.5503766640294452, "eval_python_code_critic_21k.json_loss": 0.5615730285644531, "eval_python_code_critic_21k.json_runtime": 3.2332, "eval_python_code_critic_21k.json_samples_per_second": 184.645, "eval_python_code_critic_21k.json_steps_per_second": 7.732, "step": 2000 }, { "epoch": 0.5503766640294452, "eval_all_merge_table_dataset.json_loss": 0.07384855300188065, "eval_all_merge_table_dataset.json_runtime": 23.3929, "eval_all_merge_table_dataset.json_samples_per_second": 30.437, "eval_all_merge_table_dataset.json_steps_per_second": 1.282, "step": 2000 }, { "epoch": 0.5503766640294452, "eval_code_feedback_multi_turn.json_loss": 0.5769618153572083, "eval_code_feedback_multi_turn.json_runtime": 32.4541, "eval_code_feedback_multi_turn.json_samples_per_second": 67.819, "eval_code_feedback_multi_turn.json_steps_per_second": 2.835, "step": 2000 }, { "epoch": 0.5503766640294452, "eval_ultrainteract_sft.json_loss": 0.41532665491104126, "eval_ultrainteract_sft.json_runtime": 8.6954, "eval_ultrainteract_sft.json_samples_per_second": 167.445, "eval_ultrainteract_sft.json_steps_per_second": 7.015, "step": 2000 }, { "epoch": 0.5503766640294452, "eval_synthetic_text_to_sql.json_loss": 0.09223779290914536, "eval_synthetic_text_to_sql.json_runtime": 0.1265, "eval_synthetic_text_to_sql.json_samples_per_second": 268.84, "eval_synthetic_text_to_sql.json_steps_per_second": 15.814, "step": 2000 }, { "epoch": 0.5503766640294452, "eval_sft_react_sql_datas.json_loss": 0.6254591941833496, "eval_sft_react_sql_datas.json_runtime": 7.8542, "eval_sft_react_sql_datas.json_samples_per_second": 39.979, "eval_sft_react_sql_datas.json_steps_per_second": 1.782, "step": 2000 }, { "epoch": 0.5503766640294452, "eval_all_merge_code.json_loss": 0.2845838665962219, "eval_all_merge_code.json_runtime": 0.3345, "eval_all_merge_code.json_samples_per_second": 188.319, "eval_all_merge_code.json_steps_per_second": 8.968, "step": 2000 }, { "epoch": 0.5503766640294452, "eval_magpie_datas.json_loss": 0.4300972521305084, "eval_magpie_datas.json_runtime": 2.2166, "eval_magpie_datas.json_samples_per_second": 77.598, "eval_magpie_datas.json_steps_per_second": 3.609, "step": 2000 }, { "epoch": 0.5503766640294452, "eval_train_data_for_qwen.json_loss": 0.0036769520957022905, "eval_train_data_for_qwen.json_runtime": 0.2431, "eval_train_data_for_qwen.json_samples_per_second": 41.14, "eval_train_data_for_qwen.json_steps_per_second": 4.114, "step": 2000 }, { "epoch": 0.5503766640294452, "eval_alpaca_cleaned.json_loss": 0.9104709625244141, "eval_alpaca_cleaned.json_runtime": 0.1148, "eval_alpaca_cleaned.json_samples_per_second": 235.266, "eval_alpaca_cleaned.json_steps_per_second": 17.427, "step": 2000 }, { "epoch": 0.5503766640294452, "eval_agent_instruct.json_loss": 0.220087930560112, "eval_agent_instruct.json_runtime": 0.5143, "eval_agent_instruct.json_samples_per_second": 93.334, "eval_agent_instruct.json_steps_per_second": 3.889, "step": 2000 }, { "epoch": 0.5503766640294452, "eval_MathInstruct.json_loss": 0.1989249587059021, "eval_MathInstruct.json_runtime": 0.3499, "eval_MathInstruct.json_samples_per_second": 162.904, "eval_MathInstruct.json_steps_per_second": 8.574, "step": 2000 }, { "epoch": 0.5503766640294452, "eval_tested_143k_python_alpaca.json_loss": 0.4425477683544159, "eval_tested_143k_python_alpaca.json_runtime": 0.3008, "eval_tested_143k_python_alpaca.json_samples_per_second": 113.029, "eval_tested_143k_python_alpaca.json_steps_per_second": 6.649, "step": 2000 }, { "epoch": 0.5503766640294452, "eval_xlam_function_calling_60k.json_loss": 0.008927595801651478, "eval_xlam_function_calling_60k.json_runtime": 0.1003, "eval_xlam_function_calling_60k.json_samples_per_second": 229.301, "eval_xlam_function_calling_60k.json_steps_per_second": 9.97, "step": 2000 }, { "epoch": 0.5503766640294452, "eval_alpaca_data_gpt4_chinese.json_loss": 1.5485728979110718, "eval_alpaca_data_gpt4_chinese.json_runtime": 0.0512, "eval_alpaca_data_gpt4_chinese.json_samples_per_second": 312.726, "eval_alpaca_data_gpt4_chinese.json_steps_per_second": 19.545, "step": 2000 }, { "epoch": 0.5503766640294452, "eval_alpaca_gpt4_zh.json_loss": 0.9768400192260742, "eval_alpaca_gpt4_zh.json_runtime": 0.0505, "eval_alpaca_gpt4_zh.json_samples_per_second": 217.931, "eval_alpaca_gpt4_zh.json_steps_per_second": 19.812, "step": 2000 }, { "epoch": 0.5503766640294452, "eval_codefeedback_filtered_instruction.json_loss": 0.587010383605957, "eval_codefeedback_filtered_instruction.json_runtime": 0.4876, "eval_codefeedback_filtered_instruction.json_samples_per_second": 41.015, "eval_codefeedback_filtered_instruction.json_steps_per_second": 2.051, "step": 2000 }, { "epoch": 0.5509270406934746, "grad_norm": 0.3626772463321686, "learning_rate": 1e-05, "loss": 0.4442, "step": 2002 }, { "epoch": 0.551477417357504, "grad_norm": 0.34878280758857727, "learning_rate": 1e-05, "loss": 0.4458, "step": 2004 }, { "epoch": 0.5520277940215335, "grad_norm": 0.35377946496009827, "learning_rate": 1e-05, "loss": 0.4273, "step": 2006 }, { "epoch": 0.5525781706855629, "grad_norm": 0.3649701774120331, "learning_rate": 1e-05, "loss": 0.4342, "step": 2008 }, { "epoch": 0.5531285473495924, "grad_norm": 0.34736165404319763, "learning_rate": 1e-05, "loss": 0.4298, "step": 2010 }, { "epoch": 0.5536789240136218, "grad_norm": 0.3697884678840637, "learning_rate": 1e-05, "loss": 0.4424, "step": 2012 }, { "epoch": 0.5542293006776513, "grad_norm": 0.40290403366088867, "learning_rate": 1e-05, "loss": 0.4388, "step": 2014 }, { "epoch": 0.5547796773416808, "grad_norm": 0.36797061562538147, "learning_rate": 1e-05, "loss": 0.4648, "step": 2016 }, { "epoch": 0.5553300540057101, "grad_norm": 0.35621124505996704, "learning_rate": 1e-05, "loss": 0.433, "step": 2018 }, { "epoch": 0.5558804306697396, "grad_norm": 0.3625437915325165, "learning_rate": 1e-05, "loss": 0.441, "step": 2020 }, { "epoch": 0.556430807333769, "grad_norm": 0.3642013370990753, "learning_rate": 1e-05, "loss": 0.4425, "step": 2022 }, { "epoch": 0.5569811839977985, "grad_norm": 0.36053115129470825, "learning_rate": 1e-05, "loss": 0.4422, "step": 2024 }, { "epoch": 0.557531560661828, "grad_norm": 0.36283549666404724, "learning_rate": 1e-05, "loss": 0.4338, "step": 2026 }, { "epoch": 0.5580819373258574, "grad_norm": 0.3758421540260315, "learning_rate": 1e-05, "loss": 0.439, "step": 2028 }, { "epoch": 0.5586323139898868, "grad_norm": 0.33730989694595337, "learning_rate": 1e-05, "loss": 0.4446, "step": 2030 }, { "epoch": 0.5591826906539162, "grad_norm": 0.36297255754470825, "learning_rate": 1e-05, "loss": 0.4358, "step": 2032 }, { "epoch": 0.5597330673179457, "grad_norm": 0.3534908890724182, "learning_rate": 1e-05, "loss": 0.4257, "step": 2034 }, { "epoch": 0.5602834439819752, "grad_norm": 0.3690515160560608, "learning_rate": 1e-05, "loss": 0.4383, "step": 2036 }, { "epoch": 0.5608338206460046, "grad_norm": 0.3638661503791809, "learning_rate": 1e-05, "loss": 0.4452, "step": 2038 }, { "epoch": 0.5613841973100341, "grad_norm": 0.3521392047405243, "learning_rate": 1e-05, "loss": 0.4342, "step": 2040 }, { "epoch": 0.5619345739740635, "grad_norm": 0.3569532632827759, "learning_rate": 1e-05, "loss": 0.4507, "step": 2042 }, { "epoch": 0.5624849506380929, "grad_norm": 0.37072595953941345, "learning_rate": 1e-05, "loss": 0.4354, "step": 2044 }, { "epoch": 0.5630353273021224, "grad_norm": 0.38489988446235657, "learning_rate": 1e-05, "loss": 0.4528, "step": 2046 }, { "epoch": 0.5635857039661518, "grad_norm": 0.38305357098579407, "learning_rate": 1e-05, "loss": 0.4428, "step": 2048 }, { "epoch": 0.5641360806301813, "grad_norm": 0.3491927981376648, "learning_rate": 1e-05, "loss": 0.4242, "step": 2050 }, { "epoch": 0.5646864572942107, "grad_norm": 0.35508430004119873, "learning_rate": 1e-05, "loss": 0.4556, "step": 2052 }, { "epoch": 0.5652368339582402, "grad_norm": 0.36298030614852905, "learning_rate": 1e-05, "loss": 0.4337, "step": 2054 }, { "epoch": 0.5657872106222697, "grad_norm": 0.3598901629447937, "learning_rate": 1e-05, "loss": 0.4378, "step": 2056 }, { "epoch": 0.566337587286299, "grad_norm": 0.3838946223258972, "learning_rate": 1e-05, "loss": 0.4346, "step": 2058 }, { "epoch": 0.5668879639503285, "grad_norm": 0.3986867666244507, "learning_rate": 1e-05, "loss": 0.45, "step": 2060 }, { "epoch": 0.5674383406143579, "grad_norm": 0.3509708344936371, "learning_rate": 1e-05, "loss": 0.4462, "step": 2062 }, { "epoch": 0.5679887172783874, "grad_norm": 0.35189950466156006, "learning_rate": 1e-05, "loss": 0.4307, "step": 2064 }, { "epoch": 0.5685390939424169, "grad_norm": 0.37416207790374756, "learning_rate": 1e-05, "loss": 0.4368, "step": 2066 }, { "epoch": 0.5690894706064463, "grad_norm": 0.3902382254600525, "learning_rate": 1e-05, "loss": 0.4278, "step": 2068 }, { "epoch": 0.5696398472704758, "grad_norm": 0.384260892868042, "learning_rate": 1e-05, "loss": 0.4449, "step": 2070 }, { "epoch": 0.5701902239345051, "grad_norm": 0.367347776889801, "learning_rate": 1e-05, "loss": 0.4397, "step": 2072 }, { "epoch": 0.5707406005985346, "grad_norm": 0.35011574625968933, "learning_rate": 1e-05, "loss": 0.4375, "step": 2074 }, { "epoch": 0.5712909772625641, "grad_norm": 0.3609907329082489, "learning_rate": 1e-05, "loss": 0.446, "step": 2076 }, { "epoch": 0.5718413539265935, "grad_norm": 0.3640425205230713, "learning_rate": 1e-05, "loss": 0.4453, "step": 2078 }, { "epoch": 0.572391730590623, "grad_norm": 0.3464198112487793, "learning_rate": 1e-05, "loss": 0.4489, "step": 2080 }, { "epoch": 0.5729421072546524, "grad_norm": 0.3741483688354492, "learning_rate": 1e-05, "loss": 0.4515, "step": 2082 }, { "epoch": 0.5734924839186818, "grad_norm": 0.37388619780540466, "learning_rate": 1e-05, "loss": 0.4632, "step": 2084 }, { "epoch": 0.5740428605827113, "grad_norm": 0.37237605452537537, "learning_rate": 1e-05, "loss": 0.4425, "step": 2086 }, { "epoch": 0.5745932372467407, "grad_norm": 0.35421323776245117, "learning_rate": 1e-05, "loss": 0.4474, "step": 2088 }, { "epoch": 0.5751436139107702, "grad_norm": 0.33015069365501404, "learning_rate": 1e-05, "loss": 0.43, "step": 2090 }, { "epoch": 0.5756939905747996, "grad_norm": 0.3670506179332733, "learning_rate": 1e-05, "loss": 0.4452, "step": 2092 }, { "epoch": 0.5762443672388291, "grad_norm": 0.3514888882637024, "learning_rate": 1e-05, "loss": 0.4287, "step": 2094 }, { "epoch": 0.5767947439028586, "grad_norm": 0.3714512288570404, "learning_rate": 1e-05, "loss": 0.4344, "step": 2096 }, { "epoch": 0.5773451205668879, "grad_norm": 0.35363397002220154, "learning_rate": 1e-05, "loss": 0.4408, "step": 2098 }, { "epoch": 0.5778954972309174, "grad_norm": 0.3529844582080841, "learning_rate": 1e-05, "loss": 0.4434, "step": 2100 }, { "epoch": 0.5784458738949468, "grad_norm": 0.3400002121925354, "learning_rate": 1e-05, "loss": 0.4443, "step": 2102 }, { "epoch": 0.5789962505589763, "grad_norm": 0.3620370328426361, "learning_rate": 1e-05, "loss": 0.4377, "step": 2104 }, { "epoch": 0.5795466272230058, "grad_norm": 0.3476988971233368, "learning_rate": 1e-05, "loss": 0.4321, "step": 2106 }, { "epoch": 0.5800970038870352, "grad_norm": 0.35739636421203613, "learning_rate": 1e-05, "loss": 0.4495, "step": 2108 }, { "epoch": 0.5806473805510647, "grad_norm": 0.3718028962612152, "learning_rate": 1e-05, "loss": 0.4391, "step": 2110 }, { "epoch": 0.581197757215094, "grad_norm": 0.35041627287864685, "learning_rate": 1e-05, "loss": 0.454, "step": 2112 }, { "epoch": 0.5817481338791235, "grad_norm": 0.36277493834495544, "learning_rate": 1e-05, "loss": 0.44, "step": 2114 }, { "epoch": 0.582298510543153, "grad_norm": 0.36685582995414734, "learning_rate": 1e-05, "loss": 0.4401, "step": 2116 }, { "epoch": 0.5828488872071824, "grad_norm": 0.33634135127067566, "learning_rate": 1e-05, "loss": 0.4338, "step": 2118 }, { "epoch": 0.5833992638712119, "grad_norm": 0.36546674370765686, "learning_rate": 1e-05, "loss": 0.4456, "step": 2120 }, { "epoch": 0.5839496405352413, "grad_norm": 0.361472487449646, "learning_rate": 1e-05, "loss": 0.4368, "step": 2122 }, { "epoch": 0.5845000171992708, "grad_norm": 0.36743828654289246, "learning_rate": 1e-05, "loss": 0.4464, "step": 2124 }, { "epoch": 0.5850503938633002, "grad_norm": 0.35304173827171326, "learning_rate": 1e-05, "loss": 0.4407, "step": 2126 }, { "epoch": 0.5856007705273296, "grad_norm": 0.35151979327201843, "learning_rate": 1e-05, "loss": 0.4532, "step": 2128 }, { "epoch": 0.5861511471913591, "grad_norm": 0.34761616587638855, "learning_rate": 1e-05, "loss": 0.444, "step": 2130 }, { "epoch": 0.5867015238553885, "grad_norm": 0.3763500452041626, "learning_rate": 1e-05, "loss": 0.4524, "step": 2132 }, { "epoch": 0.587251900519418, "grad_norm": 0.36489951610565186, "learning_rate": 1e-05, "loss": 0.4333, "step": 2134 }, { "epoch": 0.5878022771834475, "grad_norm": 0.38710853457450867, "learning_rate": 1e-05, "loss": 0.4517, "step": 2136 }, { "epoch": 0.5883526538474768, "grad_norm": 0.36153027415275574, "learning_rate": 1e-05, "loss": 0.438, "step": 2138 }, { "epoch": 0.5889030305115063, "grad_norm": 0.3907857835292816, "learning_rate": 1e-05, "loss": 0.4429, "step": 2140 }, { "epoch": 0.5894534071755357, "grad_norm": 0.3813617527484894, "learning_rate": 1e-05, "loss": 0.4392, "step": 2142 }, { "epoch": 0.5900037838395652, "grad_norm": 0.3563400208950043, "learning_rate": 1e-05, "loss": 0.434, "step": 2144 }, { "epoch": 0.5905541605035947, "grad_norm": 0.3556332290172577, "learning_rate": 1e-05, "loss": 0.4436, "step": 2146 }, { "epoch": 0.5911045371676241, "grad_norm": 0.3623802363872528, "learning_rate": 1e-05, "loss": 0.4378, "step": 2148 }, { "epoch": 0.5916549138316536, "grad_norm": 0.36329442262649536, "learning_rate": 1e-05, "loss": 0.4386, "step": 2150 }, { "epoch": 0.5922052904956829, "grad_norm": 0.3771746754646301, "learning_rate": 1e-05, "loss": 0.4494, "step": 2152 }, { "epoch": 0.5927556671597124, "grad_norm": 0.34596994519233704, "learning_rate": 1e-05, "loss": 0.4173, "step": 2154 }, { "epoch": 0.5933060438237419, "grad_norm": 0.36507177352905273, "learning_rate": 1e-05, "loss": 0.4254, "step": 2156 }, { "epoch": 0.5938564204877713, "grad_norm": 0.3519168794155121, "learning_rate": 1e-05, "loss": 0.4447, "step": 2158 }, { "epoch": 0.5944067971518008, "grad_norm": 0.35316991806030273, "learning_rate": 1e-05, "loss": 0.4622, "step": 2160 }, { "epoch": 0.5949571738158302, "grad_norm": 0.3529471158981323, "learning_rate": 1e-05, "loss": 0.4482, "step": 2162 }, { "epoch": 0.5955075504798597, "grad_norm": 0.3722255825996399, "learning_rate": 1e-05, "loss": 0.4454, "step": 2164 }, { "epoch": 0.596057927143889, "grad_norm": 0.3557456433773041, "learning_rate": 1e-05, "loss": 0.4435, "step": 2166 }, { "epoch": 0.5966083038079185, "grad_norm": 0.3348141610622406, "learning_rate": 1e-05, "loss": 0.436, "step": 2168 }, { "epoch": 0.597158680471948, "grad_norm": 0.38193532824516296, "learning_rate": 1e-05, "loss": 0.4543, "step": 2170 }, { "epoch": 0.5977090571359774, "grad_norm": 0.3672102391719818, "learning_rate": 1e-05, "loss": 0.4356, "step": 2172 }, { "epoch": 0.5982594338000069, "grad_norm": 0.37538838386535645, "learning_rate": 1e-05, "loss": 0.4442, "step": 2174 }, { "epoch": 0.5988098104640364, "grad_norm": 0.3512885272502899, "learning_rate": 1e-05, "loss": 0.4249, "step": 2176 }, { "epoch": 0.5993601871280658, "grad_norm": 0.4028591811656952, "learning_rate": 1e-05, "loss": 0.4495, "step": 2178 }, { "epoch": 0.5999105637920952, "grad_norm": 0.3539179861545563, "learning_rate": 1e-05, "loss": 0.4504, "step": 2180 }, { "epoch": 0.6004609404561246, "grad_norm": 0.34848934412002563, "learning_rate": 1e-05, "loss": 0.4348, "step": 2182 }, { "epoch": 0.6010113171201541, "grad_norm": 0.37469926476478577, "learning_rate": 1e-05, "loss": 0.4414, "step": 2184 }, { "epoch": 0.6015616937841836, "grad_norm": 0.3511207103729248, "learning_rate": 1e-05, "loss": 0.4489, "step": 2186 }, { "epoch": 0.602112070448213, "grad_norm": 0.3594874441623688, "learning_rate": 1e-05, "loss": 0.4429, "step": 2188 }, { "epoch": 0.6026624471122425, "grad_norm": 0.37694159150123596, "learning_rate": 1e-05, "loss": 0.4365, "step": 2190 }, { "epoch": 0.6032128237762718, "grad_norm": 0.3630627393722534, "learning_rate": 1e-05, "loss": 0.4449, "step": 2192 }, { "epoch": 0.6037632004403013, "grad_norm": 0.352230042219162, "learning_rate": 1e-05, "loss": 0.4382, "step": 2194 }, { "epoch": 0.6043135771043308, "grad_norm": 0.369757741689682, "learning_rate": 1e-05, "loss": 0.4443, "step": 2196 }, { "epoch": 0.6048639537683602, "grad_norm": 0.37120938301086426, "learning_rate": 1e-05, "loss": 0.454, "step": 2198 }, { "epoch": 0.6054143304323897, "grad_norm": 0.3475727140903473, "learning_rate": 1e-05, "loss": 0.4424, "step": 2200 }, { "epoch": 0.6054143304323897, "eval_merge_loss": 0.38126564025878906, "eval_merge_runtime": 600.3103, "eval_merge_samples_per_second": 56.224, "eval_merge_steps_per_second": 2.344, "step": 2200 }, { "epoch": 0.6054143304323897, "eval_new_aug_datas_filtered.json_loss": 0.5048007369041443, "eval_new_aug_datas_filtered.json_runtime": 10.3514, "eval_new_aug_datas_filtered.json_samples_per_second": 74.096, "eval_new_aug_datas_filtered.json_steps_per_second": 3.091, "step": 2200 }, { "epoch": 0.6054143304323897, "eval_sharegpt_gpt4.json_loss": 0.7578977346420288, "eval_sharegpt_gpt4.json_runtime": 31.6981, "eval_sharegpt_gpt4.json_samples_per_second": 58.71, "eval_sharegpt_gpt4.json_steps_per_second": 2.461, "step": 2200 }, { "epoch": 0.6054143304323897, "eval_Table_GPT.json_loss": 0.05305211618542671, "eval_Table_GPT.json_runtime": 25.0091, "eval_Table_GPT.json_samples_per_second": 83.69, "eval_Table_GPT.json_steps_per_second": 3.519, "step": 2200 }, { "epoch": 0.6054143304323897, "eval_gpt_4o_200k.json_loss": 0.7855507135391235, "eval_gpt_4o_200k.json_runtime": 48.5546, "eval_gpt_4o_200k.json_samples_per_second": 129.36, "eval_gpt_4o_200k.json_steps_per_second": 5.396, "step": 2200 }, { "epoch": 0.6054143304323897, "eval_multi_turn_datas.json_loss": 0.3139781355857849, "eval_multi_turn_datas.json_runtime": 75.6414, "eval_multi_turn_datas.json_samples_per_second": 52.908, "eval_multi_turn_datas.json_steps_per_second": 2.208, "step": 2200 }, { "epoch": 0.6054143304323897, "eval_table_python_code_datas.json_loss": 0.2603669762611389, "eval_table_python_code_datas.json_runtime": 43.0857, "eval_table_python_code_datas.json_samples_per_second": 50.109, "eval_table_python_code_datas.json_steps_per_second": 2.089, "step": 2200 }, { "epoch": 0.6054143304323897, "eval_tabular_llm_data.json_loss": 0.0890057235956192, "eval_tabular_llm_data.json_runtime": 8.5461, "eval_tabular_llm_data.json_samples_per_second": 28.785, "eval_tabular_llm_data.json_steps_per_second": 1.287, "step": 2200 }, { "epoch": 0.6054143304323897, "eval_python_code_critic_21k.json_loss": 0.5582770705223083, "eval_python_code_critic_21k.json_runtime": 3.2316, "eval_python_code_critic_21k.json_samples_per_second": 184.737, "eval_python_code_critic_21k.json_steps_per_second": 7.736, "step": 2200 }, { "epoch": 0.6054143304323897, "eval_all_merge_table_dataset.json_loss": 0.07120716571807861, "eval_all_merge_table_dataset.json_runtime": 23.3637, "eval_all_merge_table_dataset.json_samples_per_second": 30.475, "eval_all_merge_table_dataset.json_steps_per_second": 1.284, "step": 2200 }, { "epoch": 0.6054143304323897, "eval_code_feedback_multi_turn.json_loss": 0.5745006799697876, "eval_code_feedback_multi_turn.json_runtime": 32.5197, "eval_code_feedback_multi_turn.json_samples_per_second": 67.682, "eval_code_feedback_multi_turn.json_steps_per_second": 2.829, "step": 2200 }, { "epoch": 0.6054143304323897, "eval_ultrainteract_sft.json_loss": 0.41318273544311523, "eval_ultrainteract_sft.json_runtime": 8.6602, "eval_ultrainteract_sft.json_samples_per_second": 168.125, "eval_ultrainteract_sft.json_steps_per_second": 7.044, "step": 2200 }, { "epoch": 0.6054143304323897, "eval_synthetic_text_to_sql.json_loss": 0.09635543823242188, "eval_synthetic_text_to_sql.json_runtime": 0.1265, "eval_synthetic_text_to_sql.json_samples_per_second": 268.832, "eval_synthetic_text_to_sql.json_steps_per_second": 15.814, "step": 2200 }, { "epoch": 0.6054143304323897, "eval_sft_react_sql_datas.json_loss": 0.6216484904289246, "eval_sft_react_sql_datas.json_runtime": 7.8599, "eval_sft_react_sql_datas.json_samples_per_second": 39.949, "eval_sft_react_sql_datas.json_steps_per_second": 1.781, "step": 2200 }, { "epoch": 0.6054143304323897, "eval_all_merge_code.json_loss": 0.2849319279193878, "eval_all_merge_code.json_runtime": 0.3296, "eval_all_merge_code.json_samples_per_second": 191.112, "eval_all_merge_code.json_steps_per_second": 9.101, "step": 2200 }, { "epoch": 0.6054143304323897, "eval_magpie_datas.json_loss": 0.4269045293331146, "eval_magpie_datas.json_runtime": 2.2161, "eval_magpie_datas.json_samples_per_second": 77.615, "eval_magpie_datas.json_steps_per_second": 3.61, "step": 2200 }, { "epoch": 0.6054143304323897, "eval_train_data_for_qwen.json_loss": 0.005929525941610336, "eval_train_data_for_qwen.json_runtime": 0.2454, "eval_train_data_for_qwen.json_samples_per_second": 40.757, "eval_train_data_for_qwen.json_steps_per_second": 4.076, "step": 2200 }, { "epoch": 0.6054143304323897, "eval_alpaca_cleaned.json_loss": 0.9076781272888184, "eval_alpaca_cleaned.json_runtime": 0.1144, "eval_alpaca_cleaned.json_samples_per_second": 236.011, "eval_alpaca_cleaned.json_steps_per_second": 17.482, "step": 2200 }, { "epoch": 0.6054143304323897, "eval_agent_instruct.json_loss": 0.2231922596693039, "eval_agent_instruct.json_runtime": 0.5154, "eval_agent_instruct.json_samples_per_second": 93.136, "eval_agent_instruct.json_steps_per_second": 3.881, "step": 2200 }, { "epoch": 0.6054143304323897, "eval_MathInstruct.json_loss": 0.19876058399677277, "eval_MathInstruct.json_runtime": 0.3563, "eval_MathInstruct.json_samples_per_second": 159.969, "eval_MathInstruct.json_steps_per_second": 8.419, "step": 2200 }, { "epoch": 0.6054143304323897, "eval_tested_143k_python_alpaca.json_loss": 0.4431252181529999, "eval_tested_143k_python_alpaca.json_runtime": 0.3026, "eval_tested_143k_python_alpaca.json_samples_per_second": 112.374, "eval_tested_143k_python_alpaca.json_steps_per_second": 6.61, "step": 2200 }, { "epoch": 0.6054143304323897, "eval_xlam_function_calling_60k.json_loss": 0.00838847178965807, "eval_xlam_function_calling_60k.json_runtime": 0.1, "eval_xlam_function_calling_60k.json_samples_per_second": 230.081, "eval_xlam_function_calling_60k.json_steps_per_second": 10.004, "step": 2200 }, { "epoch": 0.6054143304323897, "eval_alpaca_data_gpt4_chinese.json_loss": 1.5384413003921509, "eval_alpaca_data_gpt4_chinese.json_runtime": 0.0514, "eval_alpaca_data_gpt4_chinese.json_samples_per_second": 311.198, "eval_alpaca_data_gpt4_chinese.json_steps_per_second": 19.45, "step": 2200 }, { "epoch": 0.6054143304323897, "eval_alpaca_gpt4_zh.json_loss": 0.969275712966919, "eval_alpaca_gpt4_zh.json_runtime": 0.0504, "eval_alpaca_gpt4_zh.json_samples_per_second": 218.311, "eval_alpaca_gpt4_zh.json_steps_per_second": 19.846, "step": 2200 }, { "epoch": 0.6054143304323897, "eval_codefeedback_filtered_instruction.json_loss": 0.5901365876197815, "eval_codefeedback_filtered_instruction.json_runtime": 0.4874, "eval_codefeedback_filtered_instruction.json_samples_per_second": 41.032, "eval_codefeedback_filtered_instruction.json_steps_per_second": 2.052, "step": 2200 }, { "epoch": 0.6059647070964191, "grad_norm": 0.37194857001304626, "learning_rate": 1e-05, "loss": 0.424, "step": 2202 }, { "epoch": 0.6065150837604486, "grad_norm": 0.36095818877220154, "learning_rate": 1e-05, "loss": 0.4344, "step": 2204 }, { "epoch": 0.607065460424478, "grad_norm": 0.36337706446647644, "learning_rate": 1e-05, "loss": 0.4446, "step": 2206 }, { "epoch": 0.6076158370885074, "grad_norm": 0.3500390946865082, "learning_rate": 1e-05, "loss": 0.4304, "step": 2208 }, { "epoch": 0.6081662137525369, "grad_norm": 0.3477112054824829, "learning_rate": 1e-05, "loss": 0.4346, "step": 2210 }, { "epoch": 0.6087165904165663, "grad_norm": 0.36322692036628723, "learning_rate": 1e-05, "loss": 0.4311, "step": 2212 }, { "epoch": 0.6092669670805958, "grad_norm": 0.37783941626548767, "learning_rate": 1e-05, "loss": 0.4389, "step": 2214 }, { "epoch": 0.6098173437446253, "grad_norm": 0.36018887162208557, "learning_rate": 1e-05, "loss": 0.4321, "step": 2216 }, { "epoch": 0.6103677204086547, "grad_norm": 0.34396857023239136, "learning_rate": 1e-05, "loss": 0.4349, "step": 2218 }, { "epoch": 0.6109180970726841, "grad_norm": 0.3611605167388916, "learning_rate": 1e-05, "loss": 0.4305, "step": 2220 }, { "epoch": 0.6114684737367135, "grad_norm": 0.339339941740036, "learning_rate": 1e-05, "loss": 0.4338, "step": 2222 }, { "epoch": 0.612018850400743, "grad_norm": 0.32705169916152954, "learning_rate": 1e-05, "loss": 0.4275, "step": 2224 }, { "epoch": 0.6125692270647725, "grad_norm": 0.3551005721092224, "learning_rate": 1e-05, "loss": 0.4365, "step": 2226 }, { "epoch": 0.6131196037288019, "grad_norm": 0.3826168179512024, "learning_rate": 1e-05, "loss": 0.4325, "step": 2228 }, { "epoch": 0.6136699803928314, "grad_norm": 0.376407653093338, "learning_rate": 1e-05, "loss": 0.4325, "step": 2230 }, { "epoch": 0.6142203570568608, "grad_norm": 0.3507418930530548, "learning_rate": 1e-05, "loss": 0.4315, "step": 2232 }, { "epoch": 0.6147707337208902, "grad_norm": 0.3515014946460724, "learning_rate": 1e-05, "loss": 0.4432, "step": 2234 }, { "epoch": 0.6153211103849197, "grad_norm": 0.37726324796676636, "learning_rate": 1e-05, "loss": 0.4389, "step": 2236 }, { "epoch": 0.6158714870489491, "grad_norm": 0.35043272376060486, "learning_rate": 1e-05, "loss": 0.4406, "step": 2238 }, { "epoch": 0.6164218637129786, "grad_norm": 0.3619838356971741, "learning_rate": 1e-05, "loss": 0.4381, "step": 2240 }, { "epoch": 0.616972240377008, "grad_norm": 0.3727911114692688, "learning_rate": 1e-05, "loss": 0.4261, "step": 2242 }, { "epoch": 0.6175226170410375, "grad_norm": 0.35618454217910767, "learning_rate": 1e-05, "loss": 0.4353, "step": 2244 }, { "epoch": 0.6180729937050669, "grad_norm": 0.3659394681453705, "learning_rate": 1e-05, "loss": 0.4281, "step": 2246 }, { "epoch": 0.6186233703690963, "grad_norm": 0.35864701867103577, "learning_rate": 1e-05, "loss": 0.4409, "step": 2248 }, { "epoch": 0.6191737470331258, "grad_norm": 0.36990123987197876, "learning_rate": 1e-05, "loss": 0.4424, "step": 2250 }, { "epoch": 0.6197241236971552, "grad_norm": 0.36422237753868103, "learning_rate": 1e-05, "loss": 0.4516, "step": 2252 }, { "epoch": 0.6202745003611847, "grad_norm": 0.34886521100997925, "learning_rate": 1e-05, "loss": 0.4299, "step": 2254 }, { "epoch": 0.6208248770252142, "grad_norm": 0.3683704137802124, "learning_rate": 1e-05, "loss": 0.4379, "step": 2256 }, { "epoch": 0.6213752536892436, "grad_norm": 0.3535701334476471, "learning_rate": 1e-05, "loss": 0.4292, "step": 2258 }, { "epoch": 0.621925630353273, "grad_norm": 0.370959997177124, "learning_rate": 1e-05, "loss": 0.4425, "step": 2260 }, { "epoch": 0.6224760070173024, "grad_norm": 0.3473008871078491, "learning_rate": 1e-05, "loss": 0.4289, "step": 2262 }, { "epoch": 0.6230263836813319, "grad_norm": 0.36245644092559814, "learning_rate": 1e-05, "loss": 0.4525, "step": 2264 }, { "epoch": 0.6235767603453614, "grad_norm": 0.37182751297950745, "learning_rate": 1e-05, "loss": 0.4438, "step": 2266 }, { "epoch": 0.6241271370093908, "grad_norm": 0.35843655467033386, "learning_rate": 1e-05, "loss": 0.4403, "step": 2268 }, { "epoch": 0.6246775136734203, "grad_norm": 0.3484828472137451, "learning_rate": 1e-05, "loss": 0.429, "step": 2270 }, { "epoch": 0.6252278903374497, "grad_norm": 0.35097572207450867, "learning_rate": 1e-05, "loss": 0.4435, "step": 2272 }, { "epoch": 0.6257782670014791, "grad_norm": 0.35911381244659424, "learning_rate": 1e-05, "loss": 0.435, "step": 2274 }, { "epoch": 0.6263286436655086, "grad_norm": 0.3544057309627533, "learning_rate": 1e-05, "loss": 0.4359, "step": 2276 }, { "epoch": 0.626879020329538, "grad_norm": 0.34516793489456177, "learning_rate": 1e-05, "loss": 0.4261, "step": 2278 }, { "epoch": 0.6274293969935675, "grad_norm": 0.3534994423389435, "learning_rate": 1e-05, "loss": 0.4539, "step": 2280 }, { "epoch": 0.6279797736575969, "grad_norm": 0.356238454580307, "learning_rate": 1e-05, "loss": 0.4321, "step": 2282 }, { "epoch": 0.6285301503216264, "grad_norm": 0.37285274267196655, "learning_rate": 1e-05, "loss": 0.4515, "step": 2284 }, { "epoch": 0.6290805269856559, "grad_norm": 0.3517172336578369, "learning_rate": 1e-05, "loss": 0.4268, "step": 2286 }, { "epoch": 0.6296309036496852, "grad_norm": 0.35732871294021606, "learning_rate": 1e-05, "loss": 0.4363, "step": 2288 }, { "epoch": 0.6301812803137147, "grad_norm": 0.3592797815799713, "learning_rate": 1e-05, "loss": 0.4424, "step": 2290 }, { "epoch": 0.6307316569777441, "grad_norm": 0.3233913481235504, "learning_rate": 1e-05, "loss": 0.421, "step": 2292 }, { "epoch": 0.6312820336417736, "grad_norm": 0.361591100692749, "learning_rate": 1e-05, "loss": 0.4299, "step": 2294 }, { "epoch": 0.6318324103058031, "grad_norm": 0.3468184173107147, "learning_rate": 1e-05, "loss": 0.4442, "step": 2296 }, { "epoch": 0.6323827869698325, "grad_norm": 0.4019412398338318, "learning_rate": 1e-05, "loss": 0.4453, "step": 2298 }, { "epoch": 0.6329331636338619, "grad_norm": 0.3713074326515198, "learning_rate": 1e-05, "loss": 0.435, "step": 2300 }, { "epoch": 0.6334835402978913, "grad_norm": 0.35839253664016724, "learning_rate": 1e-05, "loss": 0.4449, "step": 2302 }, { "epoch": 0.6340339169619208, "grad_norm": 0.33958542346954346, "learning_rate": 1e-05, "loss": 0.4433, "step": 2304 }, { "epoch": 0.6345842936259503, "grad_norm": 0.3750527501106262, "learning_rate": 1e-05, "loss": 0.4297, "step": 2306 }, { "epoch": 0.6351346702899797, "grad_norm": 0.35579168796539307, "learning_rate": 1e-05, "loss": 0.4307, "step": 2308 }, { "epoch": 0.6356850469540092, "grad_norm": 0.3424528241157532, "learning_rate": 1e-05, "loss": 0.4451, "step": 2310 }, { "epoch": 0.6362354236180386, "grad_norm": 0.3364480137825012, "learning_rate": 1e-05, "loss": 0.4251, "step": 2312 }, { "epoch": 0.636785800282068, "grad_norm": 0.35307276248931885, "learning_rate": 1e-05, "loss": 0.4221, "step": 2314 }, { "epoch": 0.6373361769460975, "grad_norm": 0.41354474425315857, "learning_rate": 1e-05, "loss": 0.4462, "step": 2316 }, { "epoch": 0.6378865536101269, "grad_norm": 0.37485471367836, "learning_rate": 1e-05, "loss": 0.4337, "step": 2318 }, { "epoch": 0.6384369302741564, "grad_norm": 0.344091534614563, "learning_rate": 1e-05, "loss": 0.43, "step": 2320 }, { "epoch": 0.6389873069381858, "grad_norm": 0.3772261440753937, "learning_rate": 1e-05, "loss": 0.4444, "step": 2322 }, { "epoch": 0.6395376836022153, "grad_norm": 0.35307928919792175, "learning_rate": 1e-05, "loss": 0.4332, "step": 2324 }, { "epoch": 0.6400880602662448, "grad_norm": 0.35815975069999695, "learning_rate": 1e-05, "loss": 0.4489, "step": 2326 }, { "epoch": 0.6406384369302741, "grad_norm": 0.3731154799461365, "learning_rate": 1e-05, "loss": 0.4355, "step": 2328 }, { "epoch": 0.6411888135943036, "grad_norm": 0.36875462532043457, "learning_rate": 1e-05, "loss": 0.4339, "step": 2330 }, { "epoch": 0.641739190258333, "grad_norm": 0.36913126707077026, "learning_rate": 1e-05, "loss": 0.4336, "step": 2332 }, { "epoch": 0.6422895669223625, "grad_norm": 0.35829678177833557, "learning_rate": 1e-05, "loss": 0.4438, "step": 2334 }, { "epoch": 0.642839943586392, "grad_norm": 0.36390239000320435, "learning_rate": 1e-05, "loss": 0.4405, "step": 2336 }, { "epoch": 0.6433903202504214, "grad_norm": 0.34786713123321533, "learning_rate": 1e-05, "loss": 0.451, "step": 2338 }, { "epoch": 0.6439406969144508, "grad_norm": 0.3522484600543976, "learning_rate": 1e-05, "loss": 0.4395, "step": 2340 }, { "epoch": 0.6444910735784802, "grad_norm": 0.36442965269088745, "learning_rate": 1e-05, "loss": 0.4204, "step": 2342 }, { "epoch": 0.6450414502425097, "grad_norm": 0.3635409474372864, "learning_rate": 1e-05, "loss": 0.4507, "step": 2344 }, { "epoch": 0.6455918269065392, "grad_norm": 0.35682952404022217, "learning_rate": 1e-05, "loss": 0.4333, "step": 2346 }, { "epoch": 0.6461422035705686, "grad_norm": 0.38101914525032043, "learning_rate": 1e-05, "loss": 0.4409, "step": 2348 }, { "epoch": 0.6466925802345981, "grad_norm": 0.37273916602134705, "learning_rate": 1e-05, "loss": 0.4386, "step": 2350 }, { "epoch": 0.6472429568986275, "grad_norm": 0.37394535541534424, "learning_rate": 1e-05, "loss": 0.4426, "step": 2352 }, { "epoch": 0.6477933335626569, "grad_norm": 0.3374865651130676, "learning_rate": 1e-05, "loss": 0.443, "step": 2354 }, { "epoch": 0.6483437102266864, "grad_norm": 0.34875357151031494, "learning_rate": 1e-05, "loss": 0.4135, "step": 2356 }, { "epoch": 0.6488940868907158, "grad_norm": 0.365508109331131, "learning_rate": 1e-05, "loss": 0.4455, "step": 2358 }, { "epoch": 0.6494444635547453, "grad_norm": 0.36924096941947937, "learning_rate": 1e-05, "loss": 0.4327, "step": 2360 }, { "epoch": 0.6499948402187747, "grad_norm": 0.3646699786186218, "learning_rate": 1e-05, "loss": 0.4324, "step": 2362 }, { "epoch": 0.6505452168828042, "grad_norm": 0.34241992235183716, "learning_rate": 1e-05, "loss": 0.4414, "step": 2364 }, { "epoch": 0.6510955935468337, "grad_norm": 0.3360735774040222, "learning_rate": 1e-05, "loss": 0.4228, "step": 2366 }, { "epoch": 0.651645970210863, "grad_norm": 0.3782423138618469, "learning_rate": 1e-05, "loss": 0.4366, "step": 2368 }, { "epoch": 0.6521963468748925, "grad_norm": 0.3839074373245239, "learning_rate": 1e-05, "loss": 0.4389, "step": 2370 }, { "epoch": 0.6527467235389219, "grad_norm": 0.3636200726032257, "learning_rate": 1e-05, "loss": 0.4418, "step": 2372 }, { "epoch": 0.6532971002029514, "grad_norm": 0.3629804253578186, "learning_rate": 1e-05, "loss": 0.4259, "step": 2374 }, { "epoch": 0.6538474768669809, "grad_norm": 0.3819858133792877, "learning_rate": 1e-05, "loss": 0.4348, "step": 2376 }, { "epoch": 0.6543978535310103, "grad_norm": 0.3597410321235657, "learning_rate": 1e-05, "loss": 0.428, "step": 2378 }, { "epoch": 0.6549482301950398, "grad_norm": 0.4084703326225281, "learning_rate": 1e-05, "loss": 0.4478, "step": 2380 }, { "epoch": 0.6554986068590691, "grad_norm": 0.35995879769325256, "learning_rate": 1e-05, "loss": 0.4356, "step": 2382 }, { "epoch": 0.6560489835230986, "grad_norm": 0.36047980189323425, "learning_rate": 1e-05, "loss": 0.4479, "step": 2384 }, { "epoch": 0.6565993601871281, "grad_norm": 0.3532986342906952, "learning_rate": 1e-05, "loss": 0.424, "step": 2386 }, { "epoch": 0.6571497368511575, "grad_norm": 0.3374999761581421, "learning_rate": 1e-05, "loss": 0.4338, "step": 2388 }, { "epoch": 0.657700113515187, "grad_norm": 0.34645605087280273, "learning_rate": 1e-05, "loss": 0.4257, "step": 2390 }, { "epoch": 0.6582504901792164, "grad_norm": 0.36470580101013184, "learning_rate": 1e-05, "loss": 0.4414, "step": 2392 }, { "epoch": 0.6588008668432458, "grad_norm": 0.3823862075805664, "learning_rate": 1e-05, "loss": 0.4306, "step": 2394 }, { "epoch": 0.6593512435072753, "grad_norm": 0.4070727229118347, "learning_rate": 1e-05, "loss": 0.4322, "step": 2396 }, { "epoch": 0.6599016201713047, "grad_norm": 0.37519609928131104, "learning_rate": 1e-05, "loss": 0.4248, "step": 2398 }, { "epoch": 0.6604519968353342, "grad_norm": 0.35447025299072266, "learning_rate": 1e-05, "loss": 0.4283, "step": 2400 }, { "epoch": 0.6604519968353342, "eval_merge_loss": 0.37715020775794983, "eval_merge_runtime": 600.5757, "eval_merge_samples_per_second": 56.199, "eval_merge_steps_per_second": 2.343, "step": 2400 }, { "epoch": 0.6604519968353342, "eval_new_aug_datas_filtered.json_loss": 0.5012194514274597, "eval_new_aug_datas_filtered.json_runtime": 10.4212, "eval_new_aug_datas_filtered.json_samples_per_second": 73.6, "eval_new_aug_datas_filtered.json_steps_per_second": 3.071, "step": 2400 }, { "epoch": 0.6604519968353342, "eval_sharegpt_gpt4.json_loss": 0.7534219026565552, "eval_sharegpt_gpt4.json_runtime": 31.7308, "eval_sharegpt_gpt4.json_samples_per_second": 58.65, "eval_sharegpt_gpt4.json_steps_per_second": 2.458, "step": 2400 }, { "epoch": 0.6604519968353342, "eval_Table_GPT.json_loss": 0.050881169736385345, "eval_Table_GPT.json_runtime": 24.9922, "eval_Table_GPT.json_samples_per_second": 83.746, "eval_Table_GPT.json_steps_per_second": 3.521, "step": 2400 }, { "epoch": 0.6604519968353342, "eval_gpt_4o_200k.json_loss": 0.7805712223052979, "eval_gpt_4o_200k.json_runtime": 48.518, "eval_gpt_4o_200k.json_samples_per_second": 129.457, "eval_gpt_4o_200k.json_steps_per_second": 5.4, "step": 2400 }, { "epoch": 0.6604519968353342, "eval_multi_turn_datas.json_loss": 0.3069368898868561, "eval_multi_turn_datas.json_runtime": 75.8513, "eval_multi_turn_datas.json_samples_per_second": 52.761, "eval_multi_turn_datas.json_steps_per_second": 2.202, "step": 2400 }, { "epoch": 0.6604519968353342, "eval_table_python_code_datas.json_loss": 0.2562294006347656, "eval_table_python_code_datas.json_runtime": 43.1545, "eval_table_python_code_datas.json_samples_per_second": 50.03, "eval_table_python_code_datas.json_steps_per_second": 2.086, "step": 2400 }, { "epoch": 0.6604519968353342, "eval_tabular_llm_data.json_loss": 0.09128429740667343, "eval_tabular_llm_data.json_runtime": 8.5524, "eval_tabular_llm_data.json_samples_per_second": 28.764, "eval_tabular_llm_data.json_steps_per_second": 1.286, "step": 2400 }, { "epoch": 0.6604519968353342, "eval_python_code_critic_21k.json_loss": 0.5555644631385803, "eval_python_code_critic_21k.json_runtime": 3.2271, "eval_python_code_critic_21k.json_samples_per_second": 184.994, "eval_python_code_critic_21k.json_steps_per_second": 7.747, "step": 2400 }, { "epoch": 0.6604519968353342, "eval_all_merge_table_dataset.json_loss": 0.07006299495697021, "eval_all_merge_table_dataset.json_runtime": 23.358, "eval_all_merge_table_dataset.json_samples_per_second": 30.482, "eval_all_merge_table_dataset.json_steps_per_second": 1.284, "step": 2400 }, { "epoch": 0.6604519968353342, "eval_code_feedback_multi_turn.json_loss": 0.5720005035400391, "eval_code_feedback_multi_turn.json_runtime": 32.5016, "eval_code_feedback_multi_turn.json_samples_per_second": 67.72, "eval_code_feedback_multi_turn.json_steps_per_second": 2.831, "step": 2400 }, { "epoch": 0.6604519968353342, "eval_ultrainteract_sft.json_loss": 0.4097177982330322, "eval_ultrainteract_sft.json_runtime": 8.6753, "eval_ultrainteract_sft.json_samples_per_second": 167.832, "eval_ultrainteract_sft.json_steps_per_second": 7.031, "step": 2400 }, { "epoch": 0.6604519968353342, "eval_synthetic_text_to_sql.json_loss": 0.09309177845716476, "eval_synthetic_text_to_sql.json_runtime": 0.1257, "eval_synthetic_text_to_sql.json_samples_per_second": 270.423, "eval_synthetic_text_to_sql.json_steps_per_second": 15.907, "step": 2400 }, { "epoch": 0.6604519968353342, "eval_sft_react_sql_datas.json_loss": 0.6212250590324402, "eval_sft_react_sql_datas.json_runtime": 7.859, "eval_sft_react_sql_datas.json_samples_per_second": 39.954, "eval_sft_react_sql_datas.json_steps_per_second": 1.781, "step": 2400 }, { "epoch": 0.6604519968353342, "eval_all_merge_code.json_loss": 0.28449881076812744, "eval_all_merge_code.json_runtime": 0.3298, "eval_all_merge_code.json_samples_per_second": 191.001, "eval_all_merge_code.json_steps_per_second": 9.095, "step": 2400 }, { "epoch": 0.6604519968353342, "eval_magpie_datas.json_loss": 0.426034539937973, "eval_magpie_datas.json_runtime": 2.2154, "eval_magpie_datas.json_samples_per_second": 77.638, "eval_magpie_datas.json_steps_per_second": 3.611, "step": 2400 }, { "epoch": 0.6604519968353342, "eval_train_data_for_qwen.json_loss": 0.005596214439719915, "eval_train_data_for_qwen.json_runtime": 0.2424, "eval_train_data_for_qwen.json_samples_per_second": 41.251, "eval_train_data_for_qwen.json_steps_per_second": 4.125, "step": 2400 }, { "epoch": 0.6604519968353342, "eval_alpaca_cleaned.json_loss": 0.9008170962333679, "eval_alpaca_cleaned.json_runtime": 0.1147, "eval_alpaca_cleaned.json_samples_per_second": 235.421, "eval_alpaca_cleaned.json_steps_per_second": 17.439, "step": 2400 }, { "epoch": 0.6604519968353342, "eval_agent_instruct.json_loss": 0.21443764865398407, "eval_agent_instruct.json_runtime": 0.5141, "eval_agent_instruct.json_samples_per_second": 93.36, "eval_agent_instruct.json_steps_per_second": 3.89, "step": 2400 }, { "epoch": 0.6604519968353342, "eval_MathInstruct.json_loss": 0.1956825852394104, "eval_MathInstruct.json_runtime": 0.3499, "eval_MathInstruct.json_samples_per_second": 162.885, "eval_MathInstruct.json_steps_per_second": 8.573, "step": 2400 }, { "epoch": 0.6604519968353342, "eval_tested_143k_python_alpaca.json_loss": 0.4434005320072174, "eval_tested_143k_python_alpaca.json_runtime": 0.3023, "eval_tested_143k_python_alpaca.json_samples_per_second": 112.46, "eval_tested_143k_python_alpaca.json_steps_per_second": 6.615, "step": 2400 }, { "epoch": 0.6604519968353342, "eval_xlam_function_calling_60k.json_loss": 0.009229443967342377, "eval_xlam_function_calling_60k.json_runtime": 0.1004, "eval_xlam_function_calling_60k.json_samples_per_second": 229.185, "eval_xlam_function_calling_60k.json_steps_per_second": 9.965, "step": 2400 }, { "epoch": 0.6604519968353342, "eval_alpaca_data_gpt4_chinese.json_loss": 1.5269618034362793, "eval_alpaca_data_gpt4_chinese.json_runtime": 0.0516, "eval_alpaca_data_gpt4_chinese.json_samples_per_second": 310.215, "eval_alpaca_data_gpt4_chinese.json_steps_per_second": 19.388, "step": 2400 }, { "epoch": 0.6604519968353342, "eval_alpaca_gpt4_zh.json_loss": 0.9699357151985168, "eval_alpaca_gpt4_zh.json_runtime": 0.0505, "eval_alpaca_gpt4_zh.json_samples_per_second": 217.964, "eval_alpaca_gpt4_zh.json_steps_per_second": 19.815, "step": 2400 }, { "epoch": 0.6604519968353342, "eval_codefeedback_filtered_instruction.json_loss": 0.5749525427818298, "eval_codefeedback_filtered_instruction.json_runtime": 0.4875, "eval_codefeedback_filtered_instruction.json_samples_per_second": 41.023, "eval_codefeedback_filtered_instruction.json_steps_per_second": 2.051, "step": 2400 }, { "epoch": 0.6610023734993636, "grad_norm": 0.38521307706832886, "learning_rate": 1e-05, "loss": 0.4408, "step": 2402 }, { "epoch": 0.6615527501633931, "grad_norm": 0.35963118076324463, "learning_rate": 1e-05, "loss": 0.4252, "step": 2404 }, { "epoch": 0.6621031268274226, "grad_norm": 0.34755435585975647, "learning_rate": 1e-05, "loss": 0.4319, "step": 2406 }, { "epoch": 0.6626535034914519, "grad_norm": 0.37133127450942993, "learning_rate": 1e-05, "loss": 0.4237, "step": 2408 }, { "epoch": 0.6632038801554814, "grad_norm": 0.35870301723480225, "learning_rate": 1e-05, "loss": 0.4388, "step": 2410 }, { "epoch": 0.6637542568195108, "grad_norm": 0.357415109872818, "learning_rate": 1e-05, "loss": 0.4322, "step": 2412 }, { "epoch": 0.6643046334835403, "grad_norm": 0.3610486090183258, "learning_rate": 1e-05, "loss": 0.434, "step": 2414 }, { "epoch": 0.6648550101475698, "grad_norm": 0.35058531165122986, "learning_rate": 1e-05, "loss": 0.4325, "step": 2416 }, { "epoch": 0.6654053868115992, "grad_norm": 0.3732353448867798, "learning_rate": 1e-05, "loss": 0.4266, "step": 2418 }, { "epoch": 0.6659557634756287, "grad_norm": 0.3728616535663605, "learning_rate": 1e-05, "loss": 0.4373, "step": 2420 }, { "epoch": 0.666506140139658, "grad_norm": 0.3697822093963623, "learning_rate": 1e-05, "loss": 0.4263, "step": 2422 }, { "epoch": 0.6670565168036875, "grad_norm": 0.34242671728134155, "learning_rate": 1e-05, "loss": 0.4234, "step": 2424 }, { "epoch": 0.667606893467717, "grad_norm": 0.34660401940345764, "learning_rate": 1e-05, "loss": 0.4438, "step": 2426 }, { "epoch": 0.6681572701317464, "grad_norm": 0.36335524916648865, "learning_rate": 1e-05, "loss": 0.447, "step": 2428 }, { "epoch": 0.6687076467957759, "grad_norm": 0.39879950881004333, "learning_rate": 1e-05, "loss": 0.4328, "step": 2430 }, { "epoch": 0.6692580234598053, "grad_norm": 0.3318917453289032, "learning_rate": 1e-05, "loss": 0.418, "step": 2432 }, { "epoch": 0.6698084001238348, "grad_norm": 0.3548910319805145, "learning_rate": 1e-05, "loss": 0.4297, "step": 2434 }, { "epoch": 0.6703587767878642, "grad_norm": 0.35431650280952454, "learning_rate": 1e-05, "loss": 0.4442, "step": 2436 }, { "epoch": 0.6709091534518936, "grad_norm": 0.3501831889152527, "learning_rate": 1e-05, "loss": 0.4231, "step": 2438 }, { "epoch": 0.6714595301159231, "grad_norm": 0.3664182424545288, "learning_rate": 1e-05, "loss": 0.4307, "step": 2440 }, { "epoch": 0.6720099067799525, "grad_norm": 0.36051392555236816, "learning_rate": 1e-05, "loss": 0.4348, "step": 2442 }, { "epoch": 0.672560283443982, "grad_norm": 0.38968268036842346, "learning_rate": 1e-05, "loss": 0.44, "step": 2444 }, { "epoch": 0.6731106601080115, "grad_norm": 0.34485840797424316, "learning_rate": 1e-05, "loss": 0.4387, "step": 2446 }, { "epoch": 0.6736610367720408, "grad_norm": 0.36389604210853577, "learning_rate": 1e-05, "loss": 0.4279, "step": 2448 }, { "epoch": 0.6742114134360703, "grad_norm": 0.3703545331954956, "learning_rate": 1e-05, "loss": 0.4498, "step": 2450 }, { "epoch": 0.6747617901000997, "grad_norm": 0.34628036618232727, "learning_rate": 1e-05, "loss": 0.4145, "step": 2452 }, { "epoch": 0.6753121667641292, "grad_norm": 0.3569451570510864, "learning_rate": 1e-05, "loss": 0.4308, "step": 2454 }, { "epoch": 0.6758625434281587, "grad_norm": 0.3471825122833252, "learning_rate": 1e-05, "loss": 0.4299, "step": 2456 }, { "epoch": 0.6764129200921881, "grad_norm": 0.37446585297584534, "learning_rate": 1e-05, "loss": 0.4417, "step": 2458 }, { "epoch": 0.6769632967562176, "grad_norm": 0.355708509683609, "learning_rate": 1e-05, "loss": 0.4306, "step": 2460 }, { "epoch": 0.6775136734202469, "grad_norm": 0.36398351192474365, "learning_rate": 1e-05, "loss": 0.4331, "step": 2462 }, { "epoch": 0.6780640500842764, "grad_norm": 0.38390782475471497, "learning_rate": 1e-05, "loss": 0.4421, "step": 2464 }, { "epoch": 0.6786144267483059, "grad_norm": 0.3586190938949585, "learning_rate": 1e-05, "loss": 0.4365, "step": 2466 }, { "epoch": 0.6791648034123353, "grad_norm": 0.33874934911727905, "learning_rate": 1e-05, "loss": 0.4346, "step": 2468 }, { "epoch": 0.6797151800763648, "grad_norm": 0.3699466586112976, "learning_rate": 1e-05, "loss": 0.4282, "step": 2470 }, { "epoch": 0.6802655567403942, "grad_norm": 0.35685962438583374, "learning_rate": 1e-05, "loss": 0.4386, "step": 2472 }, { "epoch": 0.6808159334044237, "grad_norm": 0.36509183049201965, "learning_rate": 1e-05, "loss": 0.4234, "step": 2474 }, { "epoch": 0.681366310068453, "grad_norm": 0.3677407503128052, "learning_rate": 1e-05, "loss": 0.4327, "step": 2476 }, { "epoch": 0.6819166867324825, "grad_norm": 0.361396849155426, "learning_rate": 1e-05, "loss": 0.4282, "step": 2478 }, { "epoch": 0.682467063396512, "grad_norm": 0.3637540936470032, "learning_rate": 1e-05, "loss": 0.4304, "step": 2480 }, { "epoch": 0.6830174400605414, "grad_norm": 0.38396722078323364, "learning_rate": 1e-05, "loss": 0.4326, "step": 2482 }, { "epoch": 0.6835678167245709, "grad_norm": 0.3760308623313904, "learning_rate": 1e-05, "loss": 0.4288, "step": 2484 }, { "epoch": 0.6841181933886004, "grad_norm": 0.36777281761169434, "learning_rate": 1e-05, "loss": 0.4435, "step": 2486 }, { "epoch": 0.6846685700526298, "grad_norm": 0.36967626214027405, "learning_rate": 1e-05, "loss": 0.4247, "step": 2488 }, { "epoch": 0.6852189467166592, "grad_norm": 0.37309199571609497, "learning_rate": 1e-05, "loss": 0.4514, "step": 2490 }, { "epoch": 0.6857693233806886, "grad_norm": 0.35478582978248596, "learning_rate": 1e-05, "loss": 0.436, "step": 2492 }, { "epoch": 0.6863197000447181, "grad_norm": 0.35142141580581665, "learning_rate": 1e-05, "loss": 0.4289, "step": 2494 }, { "epoch": 0.6868700767087476, "grad_norm": 0.37468215823173523, "learning_rate": 1e-05, "loss": 0.4363, "step": 2496 }, { "epoch": 0.687420453372777, "grad_norm": 0.3481496572494507, "learning_rate": 1e-05, "loss": 0.441, "step": 2498 }, { "epoch": 0.6879708300368065, "grad_norm": 0.34628838300704956, "learning_rate": 1e-05, "loss": 0.4425, "step": 2500 }, { "epoch": 0.6885212067008358, "grad_norm": 0.3759724497795105, "learning_rate": 1e-05, "loss": 0.4322, "step": 2502 }, { "epoch": 0.6890715833648653, "grad_norm": 0.37153902649879456, "learning_rate": 1e-05, "loss": 0.4412, "step": 2504 }, { "epoch": 0.6896219600288948, "grad_norm": 0.3601967990398407, "learning_rate": 1e-05, "loss": 0.4314, "step": 2506 }, { "epoch": 0.6901723366929242, "grad_norm": 0.3510344326496124, "learning_rate": 1e-05, "loss": 0.4261, "step": 2508 }, { "epoch": 0.6907227133569537, "grad_norm": 0.34007585048675537, "learning_rate": 1e-05, "loss": 0.4272, "step": 2510 }, { "epoch": 0.6912730900209831, "grad_norm": 0.34424078464508057, "learning_rate": 1e-05, "loss": 0.4253, "step": 2512 }, { "epoch": 0.6918234666850126, "grad_norm": 0.36498820781707764, "learning_rate": 1e-05, "loss": 0.434, "step": 2514 }, { "epoch": 0.692373843349042, "grad_norm": 0.3697148859500885, "learning_rate": 1e-05, "loss": 0.4358, "step": 2516 }, { "epoch": 0.6929242200130714, "grad_norm": 0.36114463210105896, "learning_rate": 1e-05, "loss": 0.4177, "step": 2518 }, { "epoch": 0.6934745966771009, "grad_norm": 0.3630925714969635, "learning_rate": 1e-05, "loss": 0.4438, "step": 2520 }, { "epoch": 0.6940249733411303, "grad_norm": 0.36949414014816284, "learning_rate": 1e-05, "loss": 0.4281, "step": 2522 }, { "epoch": 0.6945753500051598, "grad_norm": 0.36324694752693176, "learning_rate": 1e-05, "loss": 0.4253, "step": 2524 }, { "epoch": 0.6951257266691893, "grad_norm": 0.3471947908401489, "learning_rate": 1e-05, "loss": 0.4215, "step": 2526 }, { "epoch": 0.6956761033332187, "grad_norm": 0.33943814039230347, "learning_rate": 1e-05, "loss": 0.4546, "step": 2528 }, { "epoch": 0.6962264799972481, "grad_norm": 0.34675729274749756, "learning_rate": 1e-05, "loss": 0.4191, "step": 2530 }, { "epoch": 0.6967768566612775, "grad_norm": 0.3519613742828369, "learning_rate": 1e-05, "loss": 0.4272, "step": 2532 }, { "epoch": 0.697327233325307, "grad_norm": 0.3635639548301697, "learning_rate": 1e-05, "loss": 0.4489, "step": 2534 }, { "epoch": 0.6978776099893365, "grad_norm": 0.3636915385723114, "learning_rate": 1e-05, "loss": 0.4233, "step": 2536 }, { "epoch": 0.6984279866533659, "grad_norm": 0.36174023151397705, "learning_rate": 1e-05, "loss": 0.425, "step": 2538 }, { "epoch": 0.6989783633173954, "grad_norm": 0.35721176862716675, "learning_rate": 1e-05, "loss": 0.4279, "step": 2540 }, { "epoch": 0.6995287399814248, "grad_norm": 0.35394319891929626, "learning_rate": 1e-05, "loss": 0.4279, "step": 2542 }, { "epoch": 0.7000791166454542, "grad_norm": 0.37505972385406494, "learning_rate": 1e-05, "loss": 0.423, "step": 2544 }, { "epoch": 0.7006294933094837, "grad_norm": 0.3504476249217987, "learning_rate": 1e-05, "loss": 0.4212, "step": 2546 }, { "epoch": 0.7011798699735131, "grad_norm": 0.39700883626937866, "learning_rate": 1e-05, "loss": 0.4257, "step": 2548 }, { "epoch": 0.7017302466375426, "grad_norm": 0.36360886693000793, "learning_rate": 1e-05, "loss": 0.4276, "step": 2550 }, { "epoch": 0.702280623301572, "grad_norm": 0.36123448610305786, "learning_rate": 1e-05, "loss": 0.4266, "step": 2552 }, { "epoch": 0.7028309999656015, "grad_norm": 0.35183826088905334, "learning_rate": 1e-05, "loss": 0.421, "step": 2554 }, { "epoch": 0.7033813766296309, "grad_norm": 0.3557921350002289, "learning_rate": 1e-05, "loss": 0.4239, "step": 2556 }, { "epoch": 0.7039317532936603, "grad_norm": 0.35415929555892944, "learning_rate": 1e-05, "loss": 0.4216, "step": 2558 }, { "epoch": 0.7044821299576898, "grad_norm": 0.3662279546260834, "learning_rate": 1e-05, "loss": 0.4268, "step": 2560 }, { "epoch": 0.7050325066217192, "grad_norm": 0.35718172788619995, "learning_rate": 1e-05, "loss": 0.4213, "step": 2562 }, { "epoch": 0.7055828832857487, "grad_norm": 0.3595860004425049, "learning_rate": 1e-05, "loss": 0.4398, "step": 2564 }, { "epoch": 0.7061332599497782, "grad_norm": 0.3576621413230896, "learning_rate": 1e-05, "loss": 0.4263, "step": 2566 }, { "epoch": 0.7066836366138076, "grad_norm": 0.3699706792831421, "learning_rate": 1e-05, "loss": 0.4331, "step": 2568 }, { "epoch": 0.707234013277837, "grad_norm": 0.38423609733581543, "learning_rate": 1e-05, "loss": 0.436, "step": 2570 }, { "epoch": 0.7077843899418664, "grad_norm": 0.3747715651988983, "learning_rate": 1e-05, "loss": 0.4335, "step": 2572 }, { "epoch": 0.7083347666058959, "grad_norm": 0.3554603159427643, "learning_rate": 1e-05, "loss": 0.4236, "step": 2574 }, { "epoch": 0.7088851432699254, "grad_norm": 0.35446056723594666, "learning_rate": 1e-05, "loss": 0.4235, "step": 2576 }, { "epoch": 0.7094355199339548, "grad_norm": 0.3770659267902374, "learning_rate": 1e-05, "loss": 0.4344, "step": 2578 }, { "epoch": 0.7099858965979843, "grad_norm": 0.35676074028015137, "learning_rate": 1e-05, "loss": 0.4241, "step": 2580 }, { "epoch": 0.7105362732620137, "grad_norm": 0.3687559962272644, "learning_rate": 1e-05, "loss": 0.4329, "step": 2582 }, { "epoch": 0.7110866499260431, "grad_norm": 0.35311195254325867, "learning_rate": 1e-05, "loss": 0.4355, "step": 2584 }, { "epoch": 0.7116370265900726, "grad_norm": 0.3590395152568817, "learning_rate": 1e-05, "loss": 0.4213, "step": 2586 }, { "epoch": 0.712187403254102, "grad_norm": 0.3694981336593628, "learning_rate": 1e-05, "loss": 0.4344, "step": 2588 }, { "epoch": 0.7127377799181315, "grad_norm": 0.3516077399253845, "learning_rate": 1e-05, "loss": 0.4202, "step": 2590 }, { "epoch": 0.7132881565821609, "grad_norm": 0.38859254121780396, "learning_rate": 1e-05, "loss": 0.4179, "step": 2592 }, { "epoch": 0.7138385332461904, "grad_norm": 0.3825247883796692, "learning_rate": 1e-05, "loss": 0.4393, "step": 2594 }, { "epoch": 0.7143889099102199, "grad_norm": 0.36817750334739685, "learning_rate": 1e-05, "loss": 0.4341, "step": 2596 }, { "epoch": 0.7149392865742492, "grad_norm": 0.36351174116134644, "learning_rate": 1e-05, "loss": 0.4355, "step": 2598 }, { "epoch": 0.7154896632382787, "grad_norm": 0.3494237959384918, "learning_rate": 1e-05, "loss": 0.4176, "step": 2600 }, { "epoch": 0.7154896632382787, "eval_merge_loss": 0.3735547661781311, "eval_merge_runtime": 599.6483, "eval_merge_samples_per_second": 56.286, "eval_merge_steps_per_second": 2.346, "step": 2600 }, { "epoch": 0.7154896632382787, "eval_new_aug_datas_filtered.json_loss": 0.4953900873661041, "eval_new_aug_datas_filtered.json_runtime": 10.4567, "eval_new_aug_datas_filtered.json_samples_per_second": 73.35, "eval_new_aug_datas_filtered.json_steps_per_second": 3.06, "step": 2600 }, { "epoch": 0.7154896632382787, "eval_sharegpt_gpt4.json_loss": 0.748174786567688, "eval_sharegpt_gpt4.json_runtime": 31.7349, "eval_sharegpt_gpt4.json_samples_per_second": 58.642, "eval_sharegpt_gpt4.json_steps_per_second": 2.458, "step": 2600 }, { "epoch": 0.7154896632382787, "eval_Table_GPT.json_loss": 0.04870549216866493, "eval_Table_GPT.json_runtime": 25.0368, "eval_Table_GPT.json_samples_per_second": 83.597, "eval_Table_GPT.json_steps_per_second": 3.515, "step": 2600 }, { "epoch": 0.7154896632382787, "eval_gpt_4o_200k.json_loss": 0.775393009185791, "eval_gpt_4o_200k.json_runtime": 48.6152, "eval_gpt_4o_200k.json_samples_per_second": 129.198, "eval_gpt_4o_200k.json_steps_per_second": 5.389, "step": 2600 }, { "epoch": 0.7154896632382787, "eval_multi_turn_datas.json_loss": 0.29874685406684875, "eval_multi_turn_datas.json_runtime": 75.9064, "eval_multi_turn_datas.json_samples_per_second": 52.723, "eval_multi_turn_datas.json_steps_per_second": 2.2, "step": 2600 }, { "epoch": 0.7154896632382787, "eval_table_python_code_datas.json_loss": 0.2535416781902313, "eval_table_python_code_datas.json_runtime": 43.2787, "eval_table_python_code_datas.json_samples_per_second": 49.886, "eval_table_python_code_datas.json_steps_per_second": 2.08, "step": 2600 }, { "epoch": 0.7154896632382787, "eval_tabular_llm_data.json_loss": 0.08522781729698181, "eval_tabular_llm_data.json_runtime": 8.609, "eval_tabular_llm_data.json_samples_per_second": 28.575, "eval_tabular_llm_data.json_steps_per_second": 1.278, "step": 2600 }, { "epoch": 0.7154896632382787, "eval_python_code_critic_21k.json_loss": 0.5531289577484131, "eval_python_code_critic_21k.json_runtime": 3.2416, "eval_python_code_critic_21k.json_samples_per_second": 184.167, "eval_python_code_critic_21k.json_steps_per_second": 7.712, "step": 2600 }, { "epoch": 0.7154896632382787, "eval_all_merge_table_dataset.json_loss": 0.07141314446926117, "eval_all_merge_table_dataset.json_runtime": 23.4197, "eval_all_merge_table_dataset.json_samples_per_second": 30.402, "eval_all_merge_table_dataset.json_steps_per_second": 1.281, "step": 2600 }, { "epoch": 0.7154896632382787, "eval_code_feedback_multi_turn.json_loss": 0.5697857737541199, "eval_code_feedback_multi_turn.json_runtime": 32.4913, "eval_code_feedback_multi_turn.json_samples_per_second": 67.741, "eval_code_feedback_multi_turn.json_steps_per_second": 2.832, "step": 2600 }, { "epoch": 0.7154896632382787, "eval_ultrainteract_sft.json_loss": 0.406777024269104, "eval_ultrainteract_sft.json_runtime": 8.6553, "eval_ultrainteract_sft.json_samples_per_second": 168.22, "eval_ultrainteract_sft.json_steps_per_second": 7.048, "step": 2600 }, { "epoch": 0.7154896632382787, "eval_synthetic_text_to_sql.json_loss": 0.09255770593881607, "eval_synthetic_text_to_sql.json_runtime": 0.1264, "eval_synthetic_text_to_sql.json_samples_per_second": 268.887, "eval_synthetic_text_to_sql.json_steps_per_second": 15.817, "step": 2600 }, { "epoch": 0.7154896632382787, "eval_sft_react_sql_datas.json_loss": 0.6156443953514099, "eval_sft_react_sql_datas.json_runtime": 7.8669, "eval_sft_react_sql_datas.json_samples_per_second": 39.914, "eval_sft_react_sql_datas.json_steps_per_second": 1.78, "step": 2600 }, { "epoch": 0.7154896632382787, "eval_all_merge_code.json_loss": 0.2804557681083679, "eval_all_merge_code.json_runtime": 0.3331, "eval_all_merge_code.json_samples_per_second": 189.109, "eval_all_merge_code.json_steps_per_second": 9.005, "step": 2600 }, { "epoch": 0.7154896632382787, "eval_magpie_datas.json_loss": 0.42615047097206116, "eval_magpie_datas.json_runtime": 2.2188, "eval_magpie_datas.json_samples_per_second": 77.518, "eval_magpie_datas.json_steps_per_second": 3.605, "step": 2600 }, { "epoch": 0.7154896632382787, "eval_train_data_for_qwen.json_loss": 0.005531710106879473, "eval_train_data_for_qwen.json_runtime": 0.2446, "eval_train_data_for_qwen.json_samples_per_second": 40.888, "eval_train_data_for_qwen.json_steps_per_second": 4.089, "step": 2600 }, { "epoch": 0.7154896632382787, "eval_alpaca_cleaned.json_loss": 0.8993179202079773, "eval_alpaca_cleaned.json_runtime": 0.1158, "eval_alpaca_cleaned.json_samples_per_second": 233.205, "eval_alpaca_cleaned.json_steps_per_second": 17.274, "step": 2600 }, { "epoch": 0.7154896632382787, "eval_agent_instruct.json_loss": 0.20902203023433685, "eval_agent_instruct.json_runtime": 0.5148, "eval_agent_instruct.json_samples_per_second": 93.239, "eval_agent_instruct.json_steps_per_second": 3.885, "step": 2600 }, { "epoch": 0.7154896632382787, "eval_MathInstruct.json_loss": 0.20088934898376465, "eval_MathInstruct.json_runtime": 0.3521, "eval_MathInstruct.json_samples_per_second": 161.889, "eval_MathInstruct.json_steps_per_second": 8.52, "step": 2600 }, { "epoch": 0.7154896632382787, "eval_tested_143k_python_alpaca.json_loss": 0.44206199049949646, "eval_tested_143k_python_alpaca.json_runtime": 0.3013, "eval_tested_143k_python_alpaca.json_samples_per_second": 112.861, "eval_tested_143k_python_alpaca.json_steps_per_second": 6.639, "step": 2600 }, { "epoch": 0.7154896632382787, "eval_xlam_function_calling_60k.json_loss": 0.00838589109480381, "eval_xlam_function_calling_60k.json_runtime": 0.1004, "eval_xlam_function_calling_60k.json_samples_per_second": 229.101, "eval_xlam_function_calling_60k.json_steps_per_second": 9.961, "step": 2600 }, { "epoch": 0.7154896632382787, "eval_alpaca_data_gpt4_chinese.json_loss": 1.5224987268447876, "eval_alpaca_data_gpt4_chinese.json_runtime": 0.0517, "eval_alpaca_data_gpt4_chinese.json_samples_per_second": 309.243, "eval_alpaca_data_gpt4_chinese.json_steps_per_second": 19.328, "step": 2600 }, { "epoch": 0.7154896632382787, "eval_alpaca_gpt4_zh.json_loss": 0.9841532111167908, "eval_alpaca_gpt4_zh.json_runtime": 0.0501, "eval_alpaca_gpt4_zh.json_samples_per_second": 219.503, "eval_alpaca_gpt4_zh.json_steps_per_second": 19.955, "step": 2600 }, { "epoch": 0.7154896632382787, "eval_codefeedback_filtered_instruction.json_loss": 0.5787987112998962, "eval_codefeedback_filtered_instruction.json_runtime": 0.4863, "eval_codefeedback_filtered_instruction.json_samples_per_second": 41.126, "eval_codefeedback_filtered_instruction.json_steps_per_second": 2.056, "step": 2600 }, { "epoch": 0.7160400399023081, "grad_norm": 0.3617021143436432, "learning_rate": 1e-05, "loss": 0.4292, "step": 2602 }, { "epoch": 0.7165904165663376, "grad_norm": 0.39201030135154724, "learning_rate": 1e-05, "loss": 0.4565, "step": 2604 }, { "epoch": 0.7171407932303671, "grad_norm": 0.3617227077484131, "learning_rate": 1e-05, "loss": 0.4279, "step": 2606 }, { "epoch": 0.7176911698943965, "grad_norm": 0.3502630591392517, "learning_rate": 1e-05, "loss": 0.4191, "step": 2608 }, { "epoch": 0.7182415465584259, "grad_norm": 0.41853633522987366, "learning_rate": 1e-05, "loss": 0.4122, "step": 2610 }, { "epoch": 0.7187919232224553, "grad_norm": 0.35474300384521484, "learning_rate": 1e-05, "loss": 0.4225, "step": 2612 }, { "epoch": 0.7193422998864848, "grad_norm": 0.3673190772533417, "learning_rate": 1e-05, "loss": 0.4307, "step": 2614 }, { "epoch": 0.7198926765505143, "grad_norm": 0.383365273475647, "learning_rate": 1e-05, "loss": 0.4335, "step": 2616 }, { "epoch": 0.7204430532145437, "grad_norm": 0.35813844203948975, "learning_rate": 1e-05, "loss": 0.4462, "step": 2618 }, { "epoch": 0.7209934298785732, "grad_norm": 0.7552120685577393, "learning_rate": 1e-05, "loss": 0.4209, "step": 2620 }, { "epoch": 0.7215438065426026, "grad_norm": 0.365175724029541, "learning_rate": 1e-05, "loss": 0.441, "step": 2622 }, { "epoch": 0.722094183206632, "grad_norm": 0.3450736701488495, "learning_rate": 1e-05, "loss": 0.4302, "step": 2624 }, { "epoch": 0.7226445598706615, "grad_norm": 0.34044018387794495, "learning_rate": 1e-05, "loss": 0.4265, "step": 2626 }, { "epoch": 0.7231949365346909, "grad_norm": 0.36393091082572937, "learning_rate": 1e-05, "loss": 0.4226, "step": 2628 }, { "epoch": 0.7237453131987204, "grad_norm": 0.3462166488170624, "learning_rate": 1e-05, "loss": 0.4236, "step": 2630 }, { "epoch": 0.7242956898627498, "grad_norm": 0.4024192988872528, "learning_rate": 1e-05, "loss": 0.4377, "step": 2632 }, { "epoch": 0.7248460665267793, "grad_norm": 0.354809045791626, "learning_rate": 1e-05, "loss": 0.4245, "step": 2634 }, { "epoch": 0.7253964431908088, "grad_norm": 0.3701523244380951, "learning_rate": 1e-05, "loss": 0.438, "step": 2636 }, { "epoch": 0.7259468198548381, "grad_norm": 0.37080636620521545, "learning_rate": 1e-05, "loss": 0.4299, "step": 2638 }, { "epoch": 0.7264971965188676, "grad_norm": 0.3205287754535675, "learning_rate": 1e-05, "loss": 0.4193, "step": 2640 }, { "epoch": 0.727047573182897, "grad_norm": 0.3642041087150574, "learning_rate": 1e-05, "loss": 0.4259, "step": 2642 }, { "epoch": 0.7275979498469265, "grad_norm": 0.34573763608932495, "learning_rate": 1e-05, "loss": 0.438, "step": 2644 }, { "epoch": 0.728148326510956, "grad_norm": 0.3501754701137543, "learning_rate": 1e-05, "loss": 0.4184, "step": 2646 }, { "epoch": 0.7286987031749854, "grad_norm": 0.35315144062042236, "learning_rate": 1e-05, "loss": 0.4236, "step": 2648 }, { "epoch": 0.7292490798390149, "grad_norm": 0.36585912108421326, "learning_rate": 1e-05, "loss": 0.4205, "step": 2650 }, { "epoch": 0.7297994565030442, "grad_norm": 0.3684290051460266, "learning_rate": 1e-05, "loss": 0.4366, "step": 2652 }, { "epoch": 0.7303498331670737, "grad_norm": 0.3628571927547455, "learning_rate": 1e-05, "loss": 0.4205, "step": 2654 }, { "epoch": 0.7309002098311032, "grad_norm": 0.36779502034187317, "learning_rate": 1e-05, "loss": 0.4338, "step": 2656 }, { "epoch": 0.7314505864951326, "grad_norm": 0.3522249162197113, "learning_rate": 1e-05, "loss": 0.4268, "step": 2658 }, { "epoch": 0.7320009631591621, "grad_norm": 0.3840633034706116, "learning_rate": 1e-05, "loss": 0.425, "step": 2660 }, { "epoch": 0.7325513398231915, "grad_norm": 0.3498011529445648, "learning_rate": 1e-05, "loss": 0.4269, "step": 2662 }, { "epoch": 0.7331017164872209, "grad_norm": 0.36151036620140076, "learning_rate": 1e-05, "loss": 0.4215, "step": 2664 }, { "epoch": 0.7336520931512504, "grad_norm": 0.37008973956108093, "learning_rate": 1e-05, "loss": 0.4468, "step": 2666 }, { "epoch": 0.7342024698152798, "grad_norm": 0.3440816104412079, "learning_rate": 1e-05, "loss": 0.4349, "step": 2668 }, { "epoch": 0.7347528464793093, "grad_norm": 0.3912747800350189, "learning_rate": 1e-05, "loss": 0.4188, "step": 2670 }, { "epoch": 0.7353032231433387, "grad_norm": 0.3472096025943756, "learning_rate": 1e-05, "loss": 0.4344, "step": 2672 }, { "epoch": 0.7358535998073682, "grad_norm": 0.3477676510810852, "learning_rate": 1e-05, "loss": 0.4226, "step": 2674 }, { "epoch": 0.7364039764713977, "grad_norm": 0.3726285696029663, "learning_rate": 1e-05, "loss": 0.4263, "step": 2676 }, { "epoch": 0.736954353135427, "grad_norm": 0.3610732853412628, "learning_rate": 1e-05, "loss": 0.4272, "step": 2678 }, { "epoch": 0.7375047297994565, "grad_norm": 0.35711386799812317, "learning_rate": 1e-05, "loss": 0.4356, "step": 2680 }, { "epoch": 0.7380551064634859, "grad_norm": 0.36050212383270264, "learning_rate": 1e-05, "loss": 0.437, "step": 2682 }, { "epoch": 0.7386054831275154, "grad_norm": 0.33842894434928894, "learning_rate": 1e-05, "loss": 0.4136, "step": 2684 }, { "epoch": 0.7391558597915449, "grad_norm": 0.35878267884254456, "learning_rate": 1e-05, "loss": 0.4118, "step": 2686 }, { "epoch": 0.7397062364555743, "grad_norm": 0.3504185676574707, "learning_rate": 1e-05, "loss": 0.4157, "step": 2688 }, { "epoch": 0.7402566131196038, "grad_norm": 0.35226139426231384, "learning_rate": 1e-05, "loss": 0.4194, "step": 2690 }, { "epoch": 0.7408069897836331, "grad_norm": 0.3720513880252838, "learning_rate": 1e-05, "loss": 0.4225, "step": 2692 }, { "epoch": 0.7413573664476626, "grad_norm": 0.3444679081439972, "learning_rate": 1e-05, "loss": 0.433, "step": 2694 }, { "epoch": 0.741907743111692, "grad_norm": 0.3685862421989441, "learning_rate": 1e-05, "loss": 0.4139, "step": 2696 }, { "epoch": 0.7424581197757215, "grad_norm": 0.36269327998161316, "learning_rate": 1e-05, "loss": 0.4277, "step": 2698 }, { "epoch": 0.743008496439751, "grad_norm": 0.36458590626716614, "learning_rate": 1e-05, "loss": 0.4217, "step": 2700 }, { "epoch": 0.7435588731037804, "grad_norm": 0.3453613221645355, "learning_rate": 1e-05, "loss": 0.4174, "step": 2702 }, { "epoch": 0.7441092497678099, "grad_norm": 0.3562467098236084, "learning_rate": 1e-05, "loss": 0.4313, "step": 2704 }, { "epoch": 0.7446596264318392, "grad_norm": 0.3774909973144531, "learning_rate": 1e-05, "loss": 0.432, "step": 2706 }, { "epoch": 0.7452100030958687, "grad_norm": 0.3668104112148285, "learning_rate": 1e-05, "loss": 0.4236, "step": 2708 }, { "epoch": 0.7457603797598982, "grad_norm": 0.38669878244400024, "learning_rate": 1e-05, "loss": 0.4432, "step": 2710 }, { "epoch": 0.7463107564239276, "grad_norm": 0.3985156714916229, "learning_rate": 1e-05, "loss": 0.4422, "step": 2712 }, { "epoch": 0.7468611330879571, "grad_norm": 0.3647630512714386, "learning_rate": 1e-05, "loss": 0.4273, "step": 2714 }, { "epoch": 0.7474115097519866, "grad_norm": 0.37027841806411743, "learning_rate": 1e-05, "loss": 0.4166, "step": 2716 }, { "epoch": 0.7479618864160159, "grad_norm": 0.3770820200443268, "learning_rate": 1e-05, "loss": 0.4461, "step": 2718 }, { "epoch": 0.7485122630800454, "grad_norm": 0.35209086537361145, "learning_rate": 1e-05, "loss": 0.4473, "step": 2720 }, { "epoch": 0.7490626397440748, "grad_norm": 0.38394030928611755, "learning_rate": 1e-05, "loss": 0.4353, "step": 2722 }, { "epoch": 0.7496130164081043, "grad_norm": 0.3524518311023712, "learning_rate": 1e-05, "loss": 0.4277, "step": 2724 }, { "epoch": 0.7501633930721338, "grad_norm": 0.35822972655296326, "learning_rate": 1e-05, "loss": 0.4277, "step": 2726 }, { "epoch": 0.7507137697361632, "grad_norm": 0.3409929573535919, "learning_rate": 1e-05, "loss": 0.4172, "step": 2728 }, { "epoch": 0.7512641464001927, "grad_norm": 0.3534572422504425, "learning_rate": 1e-05, "loss": 0.431, "step": 2730 }, { "epoch": 0.751814523064222, "grad_norm": 0.3565024733543396, "learning_rate": 1e-05, "loss": 0.4297, "step": 2732 }, { "epoch": 0.7523648997282515, "grad_norm": 0.3499157130718231, "learning_rate": 1e-05, "loss": 0.4131, "step": 2734 }, { "epoch": 0.752915276392281, "grad_norm": 0.37271568179130554, "learning_rate": 1e-05, "loss": 0.4224, "step": 2736 }, { "epoch": 0.7534656530563104, "grad_norm": 0.38281935453414917, "learning_rate": 1e-05, "loss": 0.4366, "step": 2738 }, { "epoch": 0.7540160297203399, "grad_norm": 0.35982009768486023, "learning_rate": 1e-05, "loss": 0.4384, "step": 2740 }, { "epoch": 0.7545664063843693, "grad_norm": 0.3618968427181244, "learning_rate": 1e-05, "loss": 0.4484, "step": 2742 }, { "epoch": 0.7551167830483988, "grad_norm": 0.35112181305885315, "learning_rate": 1e-05, "loss": 0.4132, "step": 2744 }, { "epoch": 0.7556671597124281, "grad_norm": 0.35898518562316895, "learning_rate": 1e-05, "loss": 0.4234, "step": 2746 }, { "epoch": 0.7562175363764576, "grad_norm": 0.36049455404281616, "learning_rate": 1e-05, "loss": 0.4254, "step": 2748 }, { "epoch": 0.7567679130404871, "grad_norm": 0.3698630630970001, "learning_rate": 1e-05, "loss": 0.4387, "step": 2750 }, { "epoch": 0.7573182897045165, "grad_norm": 0.36196333169937134, "learning_rate": 1e-05, "loss": 0.4242, "step": 2752 }, { "epoch": 0.757868666368546, "grad_norm": 0.3553547263145447, "learning_rate": 1e-05, "loss": 0.4332, "step": 2754 }, { "epoch": 0.7584190430325755, "grad_norm": 0.36536121368408203, "learning_rate": 1e-05, "loss": 0.4123, "step": 2756 }, { "epoch": 0.7589694196966049, "grad_norm": 0.3394269049167633, "learning_rate": 1e-05, "loss": 0.4115, "step": 2758 }, { "epoch": 0.7595197963606343, "grad_norm": 0.35857659578323364, "learning_rate": 1e-05, "loss": 0.4174, "step": 2760 }, { "epoch": 0.7600701730246637, "grad_norm": 0.3676673173904419, "learning_rate": 1e-05, "loss": 0.4334, "step": 2762 }, { "epoch": 0.7606205496886932, "grad_norm": 0.35949233174324036, "learning_rate": 1e-05, "loss": 0.4345, "step": 2764 }, { "epoch": 0.7611709263527227, "grad_norm": 0.368569940328598, "learning_rate": 1e-05, "loss": 0.4241, "step": 2766 }, { "epoch": 0.7617213030167521, "grad_norm": 0.37473535537719727, "learning_rate": 1e-05, "loss": 0.4454, "step": 2768 }, { "epoch": 0.7622716796807816, "grad_norm": 0.34766483306884766, "learning_rate": 1e-05, "loss": 0.4193, "step": 2770 }, { "epoch": 0.7628220563448109, "grad_norm": 0.3594741225242615, "learning_rate": 1e-05, "loss": 0.4265, "step": 2772 }, { "epoch": 0.7633724330088404, "grad_norm": 0.35876014828681946, "learning_rate": 1e-05, "loss": 0.4401, "step": 2774 }, { "epoch": 0.7639228096728699, "grad_norm": 0.3698675036430359, "learning_rate": 1e-05, "loss": 0.4301, "step": 2776 }, { "epoch": 0.7644731863368993, "grad_norm": 0.3890196979045868, "learning_rate": 1e-05, "loss": 0.4312, "step": 2778 }, { "epoch": 0.7650235630009288, "grad_norm": 0.3495800793170929, "learning_rate": 1e-05, "loss": 0.4235, "step": 2780 }, { "epoch": 0.7655739396649582, "grad_norm": 0.3536211848258972, "learning_rate": 1e-05, "loss": 0.4319, "step": 2782 }, { "epoch": 0.7661243163289877, "grad_norm": 0.35744360089302063, "learning_rate": 1e-05, "loss": 0.419, "step": 2784 }, { "epoch": 0.766674692993017, "grad_norm": 0.35292670130729675, "learning_rate": 1e-05, "loss": 0.4428, "step": 2786 }, { "epoch": 0.7672250696570465, "grad_norm": 0.32827427983283997, "learning_rate": 1e-05, "loss": 0.4175, "step": 2788 }, { "epoch": 0.767775446321076, "grad_norm": 0.3385542929172516, "learning_rate": 1e-05, "loss": 0.4288, "step": 2790 }, { "epoch": 0.7683258229851054, "grad_norm": 0.3474958539009094, "learning_rate": 1e-05, "loss": 0.4424, "step": 2792 }, { "epoch": 0.7688761996491349, "grad_norm": 0.3551865816116333, "learning_rate": 1e-05, "loss": 0.4351, "step": 2794 }, { "epoch": 0.7694265763131644, "grad_norm": 0.3616306781768799, "learning_rate": 1e-05, "loss": 0.4481, "step": 2796 }, { "epoch": 0.7699769529771938, "grad_norm": 0.36132022738456726, "learning_rate": 1e-05, "loss": 0.4128, "step": 2798 }, { "epoch": 0.7705273296412232, "grad_norm": 0.3580198585987091, "learning_rate": 1e-05, "loss": 0.4242, "step": 2800 }, { "epoch": 0.7705273296412232, "eval_merge_loss": 0.3696165680885315, "eval_merge_runtime": 600.0202, "eval_merge_samples_per_second": 56.251, "eval_merge_steps_per_second": 2.345, "step": 2800 }, { "epoch": 0.7705273296412232, "eval_new_aug_datas_filtered.json_loss": 0.49126043915748596, "eval_new_aug_datas_filtered.json_runtime": 10.3252, "eval_new_aug_datas_filtered.json_samples_per_second": 74.285, "eval_new_aug_datas_filtered.json_steps_per_second": 3.099, "step": 2800 }, { "epoch": 0.7705273296412232, "eval_sharegpt_gpt4.json_loss": 0.7416729927062988, "eval_sharegpt_gpt4.json_runtime": 31.6069, "eval_sharegpt_gpt4.json_samples_per_second": 58.88, "eval_sharegpt_gpt4.json_steps_per_second": 2.468, "step": 2800 }, { "epoch": 0.7705273296412232, "eval_Table_GPT.json_loss": 0.04911120608448982, "eval_Table_GPT.json_runtime": 24.9282, "eval_Table_GPT.json_samples_per_second": 83.961, "eval_Table_GPT.json_steps_per_second": 3.53, "step": 2800 }, { "epoch": 0.7705273296412232, "eval_gpt_4o_200k.json_loss": 0.7679291367530823, "eval_gpt_4o_200k.json_runtime": 48.4021, "eval_gpt_4o_200k.json_samples_per_second": 129.767, "eval_gpt_4o_200k.json_steps_per_second": 5.413, "step": 2800 }, { "epoch": 0.7705273296412232, "eval_multi_turn_datas.json_loss": 0.2913420498371124, "eval_multi_turn_datas.json_runtime": 75.4573, "eval_multi_turn_datas.json_samples_per_second": 53.037, "eval_multi_turn_datas.json_steps_per_second": 2.213, "step": 2800 }, { "epoch": 0.7705273296412232, "eval_table_python_code_datas.json_loss": 0.25055599212646484, "eval_table_python_code_datas.json_runtime": 43.009, "eval_table_python_code_datas.json_samples_per_second": 50.199, "eval_table_python_code_datas.json_steps_per_second": 2.093, "step": 2800 }, { "epoch": 0.7705273296412232, "eval_tabular_llm_data.json_loss": 0.07946833223104477, "eval_tabular_llm_data.json_runtime": 8.5236, "eval_tabular_llm_data.json_samples_per_second": 28.861, "eval_tabular_llm_data.json_steps_per_second": 1.291, "step": 2800 }, { "epoch": 0.7705273296412232, "eval_python_code_critic_21k.json_loss": 0.5505719184875488, "eval_python_code_critic_21k.json_runtime": 3.2237, "eval_python_code_critic_21k.json_samples_per_second": 185.192, "eval_python_code_critic_21k.json_steps_per_second": 7.755, "step": 2800 }, { "epoch": 0.7705273296412232, "eval_all_merge_table_dataset.json_loss": 0.07032839208841324, "eval_all_merge_table_dataset.json_runtime": 23.2519, "eval_all_merge_table_dataset.json_samples_per_second": 30.621, "eval_all_merge_table_dataset.json_steps_per_second": 1.29, "step": 2800 }, { "epoch": 0.7705273296412232, "eval_code_feedback_multi_turn.json_loss": 0.5668665766716003, "eval_code_feedback_multi_turn.json_runtime": 32.3765, "eval_code_feedback_multi_turn.json_samples_per_second": 67.981, "eval_code_feedback_multi_turn.json_steps_per_second": 2.842, "step": 2800 }, { "epoch": 0.7705273296412232, "eval_ultrainteract_sft.json_loss": 0.405385285615921, "eval_ultrainteract_sft.json_runtime": 8.6576, "eval_ultrainteract_sft.json_samples_per_second": 168.176, "eval_ultrainteract_sft.json_steps_per_second": 7.046, "step": 2800 }, { "epoch": 0.7705273296412232, "eval_synthetic_text_to_sql.json_loss": 0.0894596055150032, "eval_synthetic_text_to_sql.json_runtime": 0.1263, "eval_synthetic_text_to_sql.json_samples_per_second": 269.263, "eval_synthetic_text_to_sql.json_steps_per_second": 15.839, "step": 2800 }, { "epoch": 0.7705273296412232, "eval_sft_react_sql_datas.json_loss": 0.6155156493186951, "eval_sft_react_sql_datas.json_runtime": 7.8457, "eval_sft_react_sql_datas.json_samples_per_second": 40.022, "eval_sft_react_sql_datas.json_steps_per_second": 1.784, "step": 2800 }, { "epoch": 0.7705273296412232, "eval_all_merge_code.json_loss": 0.2757679224014282, "eval_all_merge_code.json_runtime": 0.3332, "eval_all_merge_code.json_samples_per_second": 189.076, "eval_all_merge_code.json_steps_per_second": 9.004, "step": 2800 }, { "epoch": 0.7705273296412232, "eval_magpie_datas.json_loss": 0.42383918166160583, "eval_magpie_datas.json_runtime": 2.2093, "eval_magpie_datas.json_samples_per_second": 77.853, "eval_magpie_datas.json_steps_per_second": 3.621, "step": 2800 }, { "epoch": 0.7705273296412232, "eval_train_data_for_qwen.json_loss": 0.0028582699596881866, "eval_train_data_for_qwen.json_runtime": 0.244, "eval_train_data_for_qwen.json_samples_per_second": 40.988, "eval_train_data_for_qwen.json_steps_per_second": 4.099, "step": 2800 }, { "epoch": 0.7705273296412232, "eval_alpaca_cleaned.json_loss": 0.9000511169433594, "eval_alpaca_cleaned.json_runtime": 0.1144, "eval_alpaca_cleaned.json_samples_per_second": 235.991, "eval_alpaca_cleaned.json_steps_per_second": 17.481, "step": 2800 }, { "epoch": 0.7705273296412232, "eval_agent_instruct.json_loss": 0.21006985008716583, "eval_agent_instruct.json_runtime": 0.5133, "eval_agent_instruct.json_samples_per_second": 93.518, "eval_agent_instruct.json_steps_per_second": 3.897, "step": 2800 }, { "epoch": 0.7705273296412232, "eval_MathInstruct.json_loss": 0.19836944341659546, "eval_MathInstruct.json_runtime": 0.3623, "eval_MathInstruct.json_samples_per_second": 157.336, "eval_MathInstruct.json_steps_per_second": 8.281, "step": 2800 }, { "epoch": 0.7705273296412232, "eval_tested_143k_python_alpaca.json_loss": 0.44593295454978943, "eval_tested_143k_python_alpaca.json_runtime": 0.303, "eval_tested_143k_python_alpaca.json_samples_per_second": 112.196, "eval_tested_143k_python_alpaca.json_steps_per_second": 6.6, "step": 2800 }, { "epoch": 0.7705273296412232, "eval_xlam_function_calling_60k.json_loss": 0.0066245682537555695, "eval_xlam_function_calling_60k.json_runtime": 0.1016, "eval_xlam_function_calling_60k.json_samples_per_second": 226.385, "eval_xlam_function_calling_60k.json_steps_per_second": 9.843, "step": 2800 }, { "epoch": 0.7705273296412232, "eval_alpaca_data_gpt4_chinese.json_loss": 1.5253314971923828, "eval_alpaca_data_gpt4_chinese.json_runtime": 0.052, "eval_alpaca_data_gpt4_chinese.json_samples_per_second": 307.853, "eval_alpaca_data_gpt4_chinese.json_steps_per_second": 19.241, "step": 2800 }, { "epoch": 0.7705273296412232, "eval_alpaca_gpt4_zh.json_loss": 0.9524829983711243, "eval_alpaca_gpt4_zh.json_runtime": 0.0499, "eval_alpaca_gpt4_zh.json_samples_per_second": 220.602, "eval_alpaca_gpt4_zh.json_steps_per_second": 20.055, "step": 2800 }, { "epoch": 0.7705273296412232, "eval_codefeedback_filtered_instruction.json_loss": 0.5769651532173157, "eval_codefeedback_filtered_instruction.json_runtime": 0.4873, "eval_codefeedback_filtered_instruction.json_samples_per_second": 41.047, "eval_codefeedback_filtered_instruction.json_steps_per_second": 2.052, "step": 2800 }, { "epoch": 0.7710777063052526, "grad_norm": 0.3490790128707886, "learning_rate": 1e-05, "loss": 0.4279, "step": 2802 }, { "epoch": 0.7716280829692821, "grad_norm": 0.39200064539909363, "learning_rate": 1e-05, "loss": 0.4419, "step": 2804 }, { "epoch": 0.7721784596333116, "grad_norm": 0.36754128336906433, "learning_rate": 1e-05, "loss": 0.4298, "step": 2806 }, { "epoch": 0.772728836297341, "grad_norm": 0.3482655882835388, "learning_rate": 1e-05, "loss": 0.4249, "step": 2808 }, { "epoch": 0.7732792129613705, "grad_norm": 0.35949841141700745, "learning_rate": 1e-05, "loss": 0.4245, "step": 2810 }, { "epoch": 0.7738295896253999, "grad_norm": 0.3631410598754883, "learning_rate": 1e-05, "loss": 0.4221, "step": 2812 }, { "epoch": 0.7743799662894293, "grad_norm": 0.3531825542449951, "learning_rate": 1e-05, "loss": 0.415, "step": 2814 }, { "epoch": 0.7749303429534588, "grad_norm": 0.3741169571876526, "learning_rate": 1e-05, "loss": 0.421, "step": 2816 }, { "epoch": 0.7754807196174882, "grad_norm": 0.3431030511856079, "learning_rate": 1e-05, "loss": 0.4082, "step": 2818 }, { "epoch": 0.7760310962815177, "grad_norm": 0.35572293400764465, "learning_rate": 1e-05, "loss": 0.4279, "step": 2820 }, { "epoch": 0.7765814729455471, "grad_norm": 0.33715927600860596, "learning_rate": 1e-05, "loss": 0.4217, "step": 2822 }, { "epoch": 0.7771318496095766, "grad_norm": 0.3827720582485199, "learning_rate": 1e-05, "loss": 0.4195, "step": 2824 }, { "epoch": 0.777682226273606, "grad_norm": 0.34325775504112244, "learning_rate": 1e-05, "loss": 0.4359, "step": 2826 }, { "epoch": 0.7782326029376354, "grad_norm": 0.34917858242988586, "learning_rate": 1e-05, "loss": 0.4165, "step": 2828 }, { "epoch": 0.7787829796016649, "grad_norm": 0.3705228865146637, "learning_rate": 1e-05, "loss": 0.4234, "step": 2830 }, { "epoch": 0.7793333562656943, "grad_norm": 0.36879298090934753, "learning_rate": 1e-05, "loss": 0.4173, "step": 2832 }, { "epoch": 0.7798837329297238, "grad_norm": 0.35160768032073975, "learning_rate": 1e-05, "loss": 0.427, "step": 2834 }, { "epoch": 0.7804341095937533, "grad_norm": 0.35639581084251404, "learning_rate": 1e-05, "loss": 0.4342, "step": 2836 }, { "epoch": 0.7809844862577827, "grad_norm": 0.3821897804737091, "learning_rate": 1e-05, "loss": 0.4143, "step": 2838 }, { "epoch": 0.7815348629218121, "grad_norm": 0.35575130581855774, "learning_rate": 1e-05, "loss": 0.4052, "step": 2840 }, { "epoch": 0.7820852395858415, "grad_norm": 0.367026150226593, "learning_rate": 1e-05, "loss": 0.4507, "step": 2842 }, { "epoch": 0.782635616249871, "grad_norm": 0.35660848021507263, "learning_rate": 1e-05, "loss": 0.4112, "step": 2844 }, { "epoch": 0.7831859929139005, "grad_norm": 0.3623476028442383, "learning_rate": 1e-05, "loss": 0.4298, "step": 2846 }, { "epoch": 0.7837363695779299, "grad_norm": 0.36522987484931946, "learning_rate": 1e-05, "loss": 0.4197, "step": 2848 }, { "epoch": 0.7842867462419594, "grad_norm": 0.349153608083725, "learning_rate": 1e-05, "loss": 0.4179, "step": 2850 }, { "epoch": 0.7848371229059888, "grad_norm": 0.3868444263935089, "learning_rate": 1e-05, "loss": 0.4309, "step": 2852 }, { "epoch": 0.7853874995700182, "grad_norm": 0.3388199210166931, "learning_rate": 1e-05, "loss": 0.4255, "step": 2854 }, { "epoch": 0.7859378762340476, "grad_norm": 0.3848430812358856, "learning_rate": 1e-05, "loss": 0.4248, "step": 2856 }, { "epoch": 0.7864882528980771, "grad_norm": 0.34994250535964966, "learning_rate": 1e-05, "loss": 0.43, "step": 2858 }, { "epoch": 0.7870386295621066, "grad_norm": 0.3475828170776367, "learning_rate": 1e-05, "loss": 0.4245, "step": 2860 }, { "epoch": 0.787589006226136, "grad_norm": 0.3643713593482971, "learning_rate": 1e-05, "loss": 0.4285, "step": 2862 }, { "epoch": 0.7881393828901655, "grad_norm": 0.3819843828678131, "learning_rate": 1e-05, "loss": 0.4264, "step": 2864 }, { "epoch": 0.7886897595541948, "grad_norm": 0.3636263608932495, "learning_rate": 1e-05, "loss": 0.4354, "step": 2866 }, { "epoch": 0.7892401362182243, "grad_norm": 0.35367467999458313, "learning_rate": 1e-05, "loss": 0.4219, "step": 2868 }, { "epoch": 0.7897905128822538, "grad_norm": 0.33511704206466675, "learning_rate": 1e-05, "loss": 0.427, "step": 2870 }, { "epoch": 0.7903408895462832, "grad_norm": 0.3727225363254547, "learning_rate": 1e-05, "loss": 0.4325, "step": 2872 }, { "epoch": 0.7908912662103127, "grad_norm": 0.35963478684425354, "learning_rate": 1e-05, "loss": 0.4331, "step": 2874 }, { "epoch": 0.7914416428743422, "grad_norm": 0.3680688440799713, "learning_rate": 1e-05, "loss": 0.426, "step": 2876 }, { "epoch": 0.7919920195383716, "grad_norm": 0.3594858646392822, "learning_rate": 1e-05, "loss": 0.4251, "step": 2878 }, { "epoch": 0.792542396202401, "grad_norm": 0.3666832745075226, "learning_rate": 1e-05, "loss": 0.4148, "step": 2880 }, { "epoch": 0.7930927728664304, "grad_norm": 0.3594750761985779, "learning_rate": 1e-05, "loss": 0.424, "step": 2882 }, { "epoch": 0.7936431495304599, "grad_norm": 0.34796181321144104, "learning_rate": 1e-05, "loss": 0.4188, "step": 2884 }, { "epoch": 0.7941935261944894, "grad_norm": 0.3670448958873749, "learning_rate": 1e-05, "loss": 0.4184, "step": 2886 }, { "epoch": 0.7947439028585188, "grad_norm": 0.38206908106803894, "learning_rate": 1e-05, "loss": 0.4333, "step": 2888 }, { "epoch": 0.7952942795225483, "grad_norm": 0.3671881854534149, "learning_rate": 1e-05, "loss": 0.4117, "step": 2890 }, { "epoch": 0.7958446561865777, "grad_norm": 0.33647626638412476, "learning_rate": 1e-05, "loss": 0.4098, "step": 2892 }, { "epoch": 0.7963950328506071, "grad_norm": 0.3504905700683594, "learning_rate": 1e-05, "loss": 0.4227, "step": 2894 }, { "epoch": 0.7969454095146365, "grad_norm": 0.3571165204048157, "learning_rate": 1e-05, "loss": 0.4126, "step": 2896 }, { "epoch": 0.797495786178666, "grad_norm": 0.3529278337955475, "learning_rate": 1e-05, "loss": 0.4198, "step": 2898 }, { "epoch": 0.7980461628426955, "grad_norm": 0.3688133656978607, "learning_rate": 1e-05, "loss": 0.443, "step": 2900 }, { "epoch": 0.7985965395067249, "grad_norm": 0.37664586305618286, "learning_rate": 1e-05, "loss": 0.4345, "step": 2902 }, { "epoch": 0.7991469161707544, "grad_norm": 0.37368759512901306, "learning_rate": 1e-05, "loss": 0.4202, "step": 2904 }, { "epoch": 0.7996972928347839, "grad_norm": 0.3880954384803772, "learning_rate": 1e-05, "loss": 0.4234, "step": 2906 }, { "epoch": 0.8002476694988132, "grad_norm": 0.34263235330581665, "learning_rate": 1e-05, "loss": 0.4177, "step": 2908 }, { "epoch": 0.8007980461628427, "grad_norm": 0.37408214807510376, "learning_rate": 1e-05, "loss": 0.4366, "step": 2910 }, { "epoch": 0.8013484228268721, "grad_norm": 0.35213685035705566, "learning_rate": 1e-05, "loss": 0.411, "step": 2912 }, { "epoch": 0.8018987994909016, "grad_norm": 0.3545092046260834, "learning_rate": 1e-05, "loss": 0.4378, "step": 2914 }, { "epoch": 0.802449176154931, "grad_norm": 0.3618670701980591, "learning_rate": 1e-05, "loss": 0.4187, "step": 2916 }, { "epoch": 0.8029995528189605, "grad_norm": 0.3392831087112427, "learning_rate": 1e-05, "loss": 0.4305, "step": 2918 }, { "epoch": 0.8035499294829899, "grad_norm": 0.3700800836086273, "learning_rate": 1e-05, "loss": 0.4212, "step": 2920 }, { "epoch": 0.8041003061470193, "grad_norm": 0.35381945967674255, "learning_rate": 1e-05, "loss": 0.416, "step": 2922 }, { "epoch": 0.8046506828110488, "grad_norm": 0.3526875972747803, "learning_rate": 1e-05, "loss": 0.4287, "step": 2924 }, { "epoch": 0.8052010594750783, "grad_norm": 0.3656879663467407, "learning_rate": 1e-05, "loss": 0.4196, "step": 2926 }, { "epoch": 0.8057514361391077, "grad_norm": 0.3675120174884796, "learning_rate": 1e-05, "loss": 0.419, "step": 2928 }, { "epoch": 0.8063018128031372, "grad_norm": 0.34032610058784485, "learning_rate": 1e-05, "loss": 0.4301, "step": 2930 }, { "epoch": 0.8068521894671666, "grad_norm": 0.39022547006607056, "learning_rate": 1e-05, "loss": 0.4347, "step": 2932 }, { "epoch": 0.807402566131196, "grad_norm": 0.38301143050193787, "learning_rate": 1e-05, "loss": 0.4289, "step": 2934 }, { "epoch": 0.8079529427952254, "grad_norm": 0.34974217414855957, "learning_rate": 1e-05, "loss": 0.4233, "step": 2936 }, { "epoch": 0.8085033194592549, "grad_norm": 0.3554193377494812, "learning_rate": 1e-05, "loss": 0.4078, "step": 2938 }, { "epoch": 0.8090536961232844, "grad_norm": 0.3496205806732178, "learning_rate": 1e-05, "loss": 0.4241, "step": 2940 }, { "epoch": 0.8096040727873138, "grad_norm": 0.3549167513847351, "learning_rate": 1e-05, "loss": 0.4281, "step": 2942 }, { "epoch": 0.8101544494513433, "grad_norm": 0.3635149896144867, "learning_rate": 1e-05, "loss": 0.4307, "step": 2944 }, { "epoch": 0.8107048261153728, "grad_norm": 0.36100322008132935, "learning_rate": 1e-05, "loss": 0.4352, "step": 2946 }, { "epoch": 0.8112552027794021, "grad_norm": 0.36892169713974, "learning_rate": 1e-05, "loss": 0.4245, "step": 2948 }, { "epoch": 0.8118055794434316, "grad_norm": 0.34998342394828796, "learning_rate": 1e-05, "loss": 0.4214, "step": 2950 }, { "epoch": 0.812355956107461, "grad_norm": 0.36382123827934265, "learning_rate": 1e-05, "loss": 0.4342, "step": 2952 }, { "epoch": 0.8129063327714905, "grad_norm": 0.361068457365036, "learning_rate": 1e-05, "loss": 0.4198, "step": 2954 }, { "epoch": 0.81345670943552, "grad_norm": 0.36285367608070374, "learning_rate": 1e-05, "loss": 0.4297, "step": 2956 }, { "epoch": 0.8140070860995494, "grad_norm": 0.3376438319683075, "learning_rate": 1e-05, "loss": 0.3999, "step": 2958 }, { "epoch": 0.8145574627635789, "grad_norm": 0.35821884870529175, "learning_rate": 1e-05, "loss": 0.4283, "step": 2960 }, { "epoch": 0.8151078394276082, "grad_norm": 0.37185990810394287, "learning_rate": 1e-05, "loss": 0.4221, "step": 2962 }, { "epoch": 0.8156582160916377, "grad_norm": 0.3599165380001068, "learning_rate": 1e-05, "loss": 0.4222, "step": 2964 }, { "epoch": 0.8162085927556672, "grad_norm": 0.3599473237991333, "learning_rate": 1e-05, "loss": 0.4211, "step": 2966 }, { "epoch": 0.8167589694196966, "grad_norm": 0.3631754219532013, "learning_rate": 1e-05, "loss": 0.4273, "step": 2968 }, { "epoch": 0.8173093460837261, "grad_norm": 0.34736868739128113, "learning_rate": 1e-05, "loss": 0.4175, "step": 2970 }, { "epoch": 0.8178597227477555, "grad_norm": 0.34098127484321594, "learning_rate": 1e-05, "loss": 0.4297, "step": 2972 }, { "epoch": 0.8184100994117849, "grad_norm": 0.3562553822994232, "learning_rate": 1e-05, "loss": 0.4342, "step": 2974 }, { "epoch": 0.8189604760758143, "grad_norm": 0.3628046214580536, "learning_rate": 1e-05, "loss": 0.4146, "step": 2976 }, { "epoch": 0.8195108527398438, "grad_norm": 0.33993610739707947, "learning_rate": 1e-05, "loss": 0.4228, "step": 2978 }, { "epoch": 0.8200612294038733, "grad_norm": 0.35291528701782227, "learning_rate": 1e-05, "loss": 0.4179, "step": 2980 }, { "epoch": 0.8206116060679027, "grad_norm": 0.3480774164199829, "learning_rate": 1e-05, "loss": 0.4099, "step": 2982 }, { "epoch": 0.8211619827319322, "grad_norm": 0.36476173996925354, "learning_rate": 1e-05, "loss": 0.4153, "step": 2984 }, { "epoch": 0.8217123593959617, "grad_norm": 0.3587859869003296, "learning_rate": 1e-05, "loss": 0.4334, "step": 2986 }, { "epoch": 0.822262736059991, "grad_norm": 0.38419267535209656, "learning_rate": 1e-05, "loss": 0.4357, "step": 2988 }, { "epoch": 0.8228131127240205, "grad_norm": 0.3496173024177551, "learning_rate": 1e-05, "loss": 0.4156, "step": 2990 }, { "epoch": 0.8233634893880499, "grad_norm": 0.36481598019599915, "learning_rate": 1e-05, "loss": 0.4108, "step": 2992 }, { "epoch": 0.8239138660520794, "grad_norm": 0.36568546295166016, "learning_rate": 1e-05, "loss": 0.4329, "step": 2994 }, { "epoch": 0.8244642427161089, "grad_norm": 0.3675042390823364, "learning_rate": 1e-05, "loss": 0.4301, "step": 2996 }, { "epoch": 0.8250146193801383, "grad_norm": 0.3355284035205841, "learning_rate": 1e-05, "loss": 0.4162, "step": 2998 }, { "epoch": 0.8255649960441678, "grad_norm": 0.34280914068222046, "learning_rate": 1e-05, "loss": 0.4168, "step": 3000 }, { "epoch": 0.8255649960441678, "eval_merge_loss": 0.3659045696258545, "eval_merge_runtime": 599.8368, "eval_merge_samples_per_second": 56.269, "eval_merge_steps_per_second": 2.346, "step": 3000 }, { "epoch": 0.8255649960441678, "eval_new_aug_datas_filtered.json_loss": 0.48660770058631897, "eval_new_aug_datas_filtered.json_runtime": 10.3383, "eval_new_aug_datas_filtered.json_samples_per_second": 74.19, "eval_new_aug_datas_filtered.json_steps_per_second": 3.095, "step": 3000 }, { "epoch": 0.8255649960441678, "eval_sharegpt_gpt4.json_loss": 0.7358890175819397, "eval_sharegpt_gpt4.json_runtime": 31.7081, "eval_sharegpt_gpt4.json_samples_per_second": 58.692, "eval_sharegpt_gpt4.json_steps_per_second": 2.46, "step": 3000 }, { "epoch": 0.8255649960441678, "eval_Table_GPT.json_loss": 0.045936468988657, "eval_Table_GPT.json_runtime": 24.9946, "eval_Table_GPT.json_samples_per_second": 83.738, "eval_Table_GPT.json_steps_per_second": 3.521, "step": 3000 }, { "epoch": 0.8255649960441678, "eval_gpt_4o_200k.json_loss": 0.7624426484107971, "eval_gpt_4o_200k.json_runtime": 48.6264, "eval_gpt_4o_200k.json_samples_per_second": 129.169, "eval_gpt_4o_200k.json_steps_per_second": 5.388, "step": 3000 }, { "epoch": 0.8255649960441678, "eval_multi_turn_datas.json_loss": 0.2812780439853668, "eval_multi_turn_datas.json_runtime": 75.8593, "eval_multi_turn_datas.json_samples_per_second": 52.756, "eval_multi_turn_datas.json_steps_per_second": 2.201, "step": 3000 }, { "epoch": 0.8255649960441678, "eval_table_python_code_datas.json_loss": 0.24670127034187317, "eval_table_python_code_datas.json_runtime": 43.2305, "eval_table_python_code_datas.json_samples_per_second": 49.942, "eval_table_python_code_datas.json_steps_per_second": 2.082, "step": 3000 }, { "epoch": 0.8255649960441678, "eval_tabular_llm_data.json_loss": 0.08318436145782471, "eval_tabular_llm_data.json_runtime": 8.561, "eval_tabular_llm_data.json_samples_per_second": 28.735, "eval_tabular_llm_data.json_steps_per_second": 1.285, "step": 3000 }, { "epoch": 0.8255649960441678, "eval_python_code_critic_21k.json_loss": 0.5459744930267334, "eval_python_code_critic_21k.json_runtime": 3.2232, "eval_python_code_critic_21k.json_samples_per_second": 185.217, "eval_python_code_critic_21k.json_steps_per_second": 7.756, "step": 3000 }, { "epoch": 0.8255649960441678, "eval_all_merge_table_dataset.json_loss": 0.07661881297826767, "eval_all_merge_table_dataset.json_runtime": 23.3773, "eval_all_merge_table_dataset.json_samples_per_second": 30.457, "eval_all_merge_table_dataset.json_steps_per_second": 1.283, "step": 3000 }, { "epoch": 0.8255649960441678, "eval_code_feedback_multi_turn.json_loss": 0.5640604496002197, "eval_code_feedback_multi_turn.json_runtime": 32.4865, "eval_code_feedback_multi_turn.json_samples_per_second": 67.751, "eval_code_feedback_multi_turn.json_steps_per_second": 2.832, "step": 3000 }, { "epoch": 0.8255649960441678, "eval_ultrainteract_sft.json_loss": 0.40351128578186035, "eval_ultrainteract_sft.json_runtime": 8.6435, "eval_ultrainteract_sft.json_samples_per_second": 168.449, "eval_ultrainteract_sft.json_steps_per_second": 7.057, "step": 3000 }, { "epoch": 0.8255649960441678, "eval_synthetic_text_to_sql.json_loss": 0.09340357035398483, "eval_synthetic_text_to_sql.json_runtime": 0.1267, "eval_synthetic_text_to_sql.json_samples_per_second": 268.437, "eval_synthetic_text_to_sql.json_steps_per_second": 15.79, "step": 3000 }, { "epoch": 0.8255649960441678, "eval_sft_react_sql_datas.json_loss": 0.614182710647583, "eval_sft_react_sql_datas.json_runtime": 7.8427, "eval_sft_react_sql_datas.json_samples_per_second": 40.037, "eval_sft_react_sql_datas.json_steps_per_second": 1.785, "step": 3000 }, { "epoch": 0.8255649960441678, "eval_all_merge_code.json_loss": 0.2747681736946106, "eval_all_merge_code.json_runtime": 0.3335, "eval_all_merge_code.json_samples_per_second": 188.917, "eval_all_merge_code.json_steps_per_second": 8.996, "step": 3000 }, { "epoch": 0.8255649960441678, "eval_magpie_datas.json_loss": 0.42281365394592285, "eval_magpie_datas.json_runtime": 2.2171, "eval_magpie_datas.json_samples_per_second": 77.579, "eval_magpie_datas.json_steps_per_second": 3.608, "step": 3000 }, { "epoch": 0.8255649960441678, "eval_train_data_for_qwen.json_loss": 0.0027365919668227434, "eval_train_data_for_qwen.json_runtime": 0.2454, "eval_train_data_for_qwen.json_samples_per_second": 40.756, "eval_train_data_for_qwen.json_steps_per_second": 4.076, "step": 3000 }, { "epoch": 0.8255649960441678, "eval_alpaca_cleaned.json_loss": 0.9086716175079346, "eval_alpaca_cleaned.json_runtime": 0.1143, "eval_alpaca_cleaned.json_samples_per_second": 236.118, "eval_alpaca_cleaned.json_steps_per_second": 17.49, "step": 3000 }, { "epoch": 0.8255649960441678, "eval_agent_instruct.json_loss": 0.20960307121276855, "eval_agent_instruct.json_runtime": 0.5163, "eval_agent_instruct.json_samples_per_second": 92.971, "eval_agent_instruct.json_steps_per_second": 3.874, "step": 3000 }, { "epoch": 0.8255649960441678, "eval_MathInstruct.json_loss": 0.20019014179706573, "eval_MathInstruct.json_runtime": 0.3582, "eval_MathInstruct.json_samples_per_second": 159.116, "eval_MathInstruct.json_steps_per_second": 8.375, "step": 3000 }, { "epoch": 0.8255649960441678, "eval_tested_143k_python_alpaca.json_loss": 0.44821104407310486, "eval_tested_143k_python_alpaca.json_runtime": 0.3022, "eval_tested_143k_python_alpaca.json_samples_per_second": 112.526, "eval_tested_143k_python_alpaca.json_steps_per_second": 6.619, "step": 3000 }, { "epoch": 0.8255649960441678, "eval_xlam_function_calling_60k.json_loss": 0.008376230485737324, "eval_xlam_function_calling_60k.json_runtime": 0.1003, "eval_xlam_function_calling_60k.json_samples_per_second": 229.41, "eval_xlam_function_calling_60k.json_steps_per_second": 9.974, "step": 3000 }, { "epoch": 0.8255649960441678, "eval_alpaca_data_gpt4_chinese.json_loss": 1.513078212738037, "eval_alpaca_data_gpt4_chinese.json_runtime": 0.0516, "eval_alpaca_data_gpt4_chinese.json_samples_per_second": 310.009, "eval_alpaca_data_gpt4_chinese.json_steps_per_second": 19.376, "step": 3000 }, { "epoch": 0.8255649960441678, "eval_alpaca_gpt4_zh.json_loss": 0.9633126258850098, "eval_alpaca_gpt4_zh.json_runtime": 0.0499, "eval_alpaca_gpt4_zh.json_samples_per_second": 220.561, "eval_alpaca_gpt4_zh.json_steps_per_second": 20.051, "step": 3000 }, { "epoch": 0.8255649960441678, "eval_codefeedback_filtered_instruction.json_loss": 0.5788259506225586, "eval_codefeedback_filtered_instruction.json_runtime": 0.4854, "eval_codefeedback_filtered_instruction.json_samples_per_second": 41.202, "eval_codefeedback_filtered_instruction.json_steps_per_second": 2.06, "step": 3000 }, { "epoch": 0.8261153727081971, "grad_norm": 0.35386523604393005, "learning_rate": 1e-05, "loss": 0.4235, "step": 3002 }, { "epoch": 0.8266657493722266, "grad_norm": 0.35325145721435547, "learning_rate": 1e-05, "loss": 0.4296, "step": 3004 }, { "epoch": 0.827216126036256, "grad_norm": 0.35455331206321716, "learning_rate": 1e-05, "loss": 0.405, "step": 3006 }, { "epoch": 0.8277665027002855, "grad_norm": 0.37510380148887634, "learning_rate": 1e-05, "loss": 0.4288, "step": 3008 }, { "epoch": 0.828316879364315, "grad_norm": 0.356189489364624, "learning_rate": 1e-05, "loss": 0.4145, "step": 3010 }, { "epoch": 0.8288672560283444, "grad_norm": 0.36097854375839233, "learning_rate": 1e-05, "loss": 0.4247, "step": 3012 }, { "epoch": 0.8294176326923739, "grad_norm": 0.3489934802055359, "learning_rate": 1e-05, "loss": 0.422, "step": 3014 }, { "epoch": 0.8299680093564032, "grad_norm": 0.36287152767181396, "learning_rate": 1e-05, "loss": 0.4025, "step": 3016 }, { "epoch": 0.8305183860204327, "grad_norm": 0.3664880096912384, "learning_rate": 1e-05, "loss": 0.4278, "step": 3018 }, { "epoch": 0.8310687626844622, "grad_norm": 0.35230088233947754, "learning_rate": 1e-05, "loss": 0.4233, "step": 3020 }, { "epoch": 0.8316191393484916, "grad_norm": 0.3595122694969177, "learning_rate": 1e-05, "loss": 0.4173, "step": 3022 }, { "epoch": 0.8321695160125211, "grad_norm": 0.3618360757827759, "learning_rate": 1e-05, "loss": 0.4213, "step": 3024 }, { "epoch": 0.8327198926765506, "grad_norm": 0.3699500858783722, "learning_rate": 1e-05, "loss": 0.4267, "step": 3026 }, { "epoch": 0.8332702693405799, "grad_norm": 0.37343189120292664, "learning_rate": 1e-05, "loss": 0.4288, "step": 3028 }, { "epoch": 0.8338206460046094, "grad_norm": 0.34580445289611816, "learning_rate": 1e-05, "loss": 0.4232, "step": 3030 }, { "epoch": 0.8343710226686388, "grad_norm": 0.3410281836986542, "learning_rate": 1e-05, "loss": 0.4272, "step": 3032 }, { "epoch": 0.8349213993326683, "grad_norm": 0.37444379925727844, "learning_rate": 1e-05, "loss": 0.4264, "step": 3034 }, { "epoch": 0.8354717759966978, "grad_norm": 0.359546959400177, "learning_rate": 1e-05, "loss": 0.4243, "step": 3036 }, { "epoch": 0.8360221526607272, "grad_norm": 0.3611339032649994, "learning_rate": 1e-05, "loss": 0.4243, "step": 3038 }, { "epoch": 0.8365725293247567, "grad_norm": 0.3678295612335205, "learning_rate": 1e-05, "loss": 0.4089, "step": 3040 }, { "epoch": 0.837122905988786, "grad_norm": 0.37094810605049133, "learning_rate": 1e-05, "loss": 0.4191, "step": 3042 }, { "epoch": 0.8376732826528155, "grad_norm": 0.354481041431427, "learning_rate": 1e-05, "loss": 0.4235, "step": 3044 }, { "epoch": 0.838223659316845, "grad_norm": 0.3498587906360626, "learning_rate": 1e-05, "loss": 0.431, "step": 3046 }, { "epoch": 0.8387740359808744, "grad_norm": 0.35214436054229736, "learning_rate": 1e-05, "loss": 0.4132, "step": 3048 }, { "epoch": 0.8393244126449039, "grad_norm": 0.35119178891181946, "learning_rate": 1e-05, "loss": 0.4161, "step": 3050 }, { "epoch": 0.8398747893089333, "grad_norm": 0.3671429753303528, "learning_rate": 1e-05, "loss": 0.4276, "step": 3052 }, { "epoch": 0.8404251659729628, "grad_norm": 0.3626399636268616, "learning_rate": 1e-05, "loss": 0.417, "step": 3054 }, { "epoch": 0.8409755426369921, "grad_norm": 0.3819148540496826, "learning_rate": 1e-05, "loss": 0.4261, "step": 3056 }, { "epoch": 0.8415259193010216, "grad_norm": 0.3481554687023163, "learning_rate": 1e-05, "loss": 0.4339, "step": 3058 }, { "epoch": 0.8420762959650511, "grad_norm": 0.3603340983390808, "learning_rate": 1e-05, "loss": 0.406, "step": 3060 }, { "epoch": 0.8426266726290805, "grad_norm": 0.3565911650657654, "learning_rate": 1e-05, "loss": 0.4245, "step": 3062 }, { "epoch": 0.84317704929311, "grad_norm": 0.36305105686187744, "learning_rate": 1e-05, "loss": 0.4255, "step": 3064 }, { "epoch": 0.8437274259571395, "grad_norm": 0.33078432083129883, "learning_rate": 1e-05, "loss": 0.4045, "step": 3066 }, { "epoch": 0.8442778026211689, "grad_norm": 0.346562922000885, "learning_rate": 1e-05, "loss": 0.4279, "step": 3068 }, { "epoch": 0.8448281792851983, "grad_norm": 0.36170172691345215, "learning_rate": 1e-05, "loss": 0.4139, "step": 3070 }, { "epoch": 0.8453785559492277, "grad_norm": 0.360568106174469, "learning_rate": 1e-05, "loss": 0.4276, "step": 3072 }, { "epoch": 0.8459289326132572, "grad_norm": 0.38023245334625244, "learning_rate": 1e-05, "loss": 0.4317, "step": 3074 }, { "epoch": 0.8464793092772867, "grad_norm": 0.344732403755188, "learning_rate": 1e-05, "loss": 0.4109, "step": 3076 }, { "epoch": 0.8470296859413161, "grad_norm": 0.35157695412635803, "learning_rate": 1e-05, "loss": 0.4192, "step": 3078 }, { "epoch": 0.8475800626053456, "grad_norm": 0.36455512046813965, "learning_rate": 1e-05, "loss": 0.4247, "step": 3080 }, { "epoch": 0.8481304392693749, "grad_norm": 0.39768150448799133, "learning_rate": 1e-05, "loss": 0.4383, "step": 3082 }, { "epoch": 0.8486808159334044, "grad_norm": 0.38052836060523987, "learning_rate": 1e-05, "loss": 0.4199, "step": 3084 }, { "epoch": 0.8492311925974338, "grad_norm": 0.3625752925872803, "learning_rate": 1e-05, "loss": 0.4161, "step": 3086 }, { "epoch": 0.8497815692614633, "grad_norm": 0.3708571493625641, "learning_rate": 1e-05, "loss": 0.4402, "step": 3088 }, { "epoch": 0.8503319459254928, "grad_norm": 0.3581870496273041, "learning_rate": 1e-05, "loss": 0.4376, "step": 3090 }, { "epoch": 0.8508823225895222, "grad_norm": 0.33589842915534973, "learning_rate": 1e-05, "loss": 0.4144, "step": 3092 }, { "epoch": 0.8514326992535517, "grad_norm": 0.35838133096694946, "learning_rate": 1e-05, "loss": 0.4128, "step": 3094 }, { "epoch": 0.851983075917581, "grad_norm": 0.3660927712917328, "learning_rate": 1e-05, "loss": 0.42, "step": 3096 }, { "epoch": 0.8525334525816105, "grad_norm": 0.3606925904750824, "learning_rate": 1e-05, "loss": 0.4288, "step": 3098 }, { "epoch": 0.85308382924564, "grad_norm": 0.3437570333480835, "learning_rate": 1e-05, "loss": 0.4213, "step": 3100 }, { "epoch": 0.8536342059096694, "grad_norm": 0.35351496934890747, "learning_rate": 1e-05, "loss": 0.4238, "step": 3102 }, { "epoch": 0.8541845825736989, "grad_norm": 0.3595280051231384, "learning_rate": 1e-05, "loss": 0.4107, "step": 3104 }, { "epoch": 0.8547349592377284, "grad_norm": 0.3546600937843323, "learning_rate": 1e-05, "loss": 0.4105, "step": 3106 }, { "epoch": 0.8552853359017578, "grad_norm": 0.3654036819934845, "learning_rate": 1e-05, "loss": 0.4158, "step": 3108 }, { "epoch": 0.8558357125657872, "grad_norm": 0.3742349445819855, "learning_rate": 1e-05, "loss": 0.4217, "step": 3110 }, { "epoch": 0.8563860892298166, "grad_norm": 0.35527029633522034, "learning_rate": 1e-05, "loss": 0.414, "step": 3112 }, { "epoch": 0.8569364658938461, "grad_norm": 0.3408162295818329, "learning_rate": 1e-05, "loss": 0.4245, "step": 3114 }, { "epoch": 0.8574868425578756, "grad_norm": 0.3608722686767578, "learning_rate": 1e-05, "loss": 0.4346, "step": 3116 }, { "epoch": 0.858037219221905, "grad_norm": 0.36163628101348877, "learning_rate": 1e-05, "loss": 0.422, "step": 3118 }, { "epoch": 0.8585875958859345, "grad_norm": 0.35417988896369934, "learning_rate": 1e-05, "loss": 0.4101, "step": 3120 }, { "epoch": 0.8591379725499639, "grad_norm": 0.3626682162284851, "learning_rate": 1e-05, "loss": 0.4147, "step": 3122 }, { "epoch": 0.8596883492139933, "grad_norm": 0.34313321113586426, "learning_rate": 1e-05, "loss": 0.4215, "step": 3124 }, { "epoch": 0.8602387258780227, "grad_norm": 0.3839293122291565, "learning_rate": 1e-05, "loss": 0.4173, "step": 3126 }, { "epoch": 0.8607891025420522, "grad_norm": 0.3548083007335663, "learning_rate": 1e-05, "loss": 0.4153, "step": 3128 }, { "epoch": 0.8613394792060817, "grad_norm": 0.35141652822494507, "learning_rate": 1e-05, "loss": 0.4066, "step": 3130 }, { "epoch": 0.8618898558701111, "grad_norm": 0.3777351975440979, "learning_rate": 1e-05, "loss": 0.4128, "step": 3132 }, { "epoch": 0.8624402325341406, "grad_norm": 0.3580491840839386, "learning_rate": 1e-05, "loss": 0.4048, "step": 3134 }, { "epoch": 0.86299060919817, "grad_norm": 0.373532772064209, "learning_rate": 1e-05, "loss": 0.4173, "step": 3136 }, { "epoch": 0.8635409858621994, "grad_norm": 0.35365086793899536, "learning_rate": 1e-05, "loss": 0.4076, "step": 3138 }, { "epoch": 0.8640913625262289, "grad_norm": 0.3887852728366852, "learning_rate": 1e-05, "loss": 0.418, "step": 3140 }, { "epoch": 0.8646417391902583, "grad_norm": 0.35862478613853455, "learning_rate": 1e-05, "loss": 0.4234, "step": 3142 }, { "epoch": 0.8651921158542878, "grad_norm": 0.3472420275211334, "learning_rate": 1e-05, "loss": 0.4132, "step": 3144 }, { "epoch": 0.8657424925183173, "grad_norm": 0.344862163066864, "learning_rate": 1e-05, "loss": 0.41, "step": 3146 }, { "epoch": 0.8662928691823467, "grad_norm": 0.35329338908195496, "learning_rate": 1e-05, "loss": 0.4152, "step": 3148 }, { "epoch": 0.8668432458463761, "grad_norm": 0.3792724907398224, "learning_rate": 1e-05, "loss": 0.4307, "step": 3150 }, { "epoch": 0.8673936225104055, "grad_norm": 0.3611691892147064, "learning_rate": 1e-05, "loss": 0.4166, "step": 3152 }, { "epoch": 0.867943999174435, "grad_norm": 0.35675716400146484, "learning_rate": 1e-05, "loss": 0.4309, "step": 3154 }, { "epoch": 0.8684943758384644, "grad_norm": 0.37591055035591125, "learning_rate": 1e-05, "loss": 0.4174, "step": 3156 }, { "epoch": 0.8690447525024939, "grad_norm": 0.34695202112197876, "learning_rate": 1e-05, "loss": 0.4067, "step": 3158 }, { "epoch": 0.8695951291665234, "grad_norm": 0.36810246109962463, "learning_rate": 1e-05, "loss": 0.4236, "step": 3160 }, { "epoch": 0.8701455058305528, "grad_norm": 0.3910383880138397, "learning_rate": 1e-05, "loss": 0.4344, "step": 3162 }, { "epoch": 0.8706958824945822, "grad_norm": 0.3465210497379303, "learning_rate": 1e-05, "loss": 0.4156, "step": 3164 }, { "epoch": 0.8712462591586116, "grad_norm": 0.39839833974838257, "learning_rate": 1e-05, "loss": 0.417, "step": 3166 }, { "epoch": 0.8717966358226411, "grad_norm": 0.33419859409332275, "learning_rate": 1e-05, "loss": 0.4131, "step": 3168 }, { "epoch": 0.8723470124866706, "grad_norm": 0.3657875955104828, "learning_rate": 1e-05, "loss": 0.4243, "step": 3170 }, { "epoch": 0.8728973891507, "grad_norm": 0.35600635409355164, "learning_rate": 1e-05, "loss": 0.4251, "step": 3172 }, { "epoch": 0.8734477658147295, "grad_norm": 0.3642902374267578, "learning_rate": 1e-05, "loss": 0.4334, "step": 3174 }, { "epoch": 0.873998142478759, "grad_norm": 0.35452064871788025, "learning_rate": 1e-05, "loss": 0.4244, "step": 3176 }, { "epoch": 0.8745485191427883, "grad_norm": 0.372953861951828, "learning_rate": 1e-05, "loss": 0.421, "step": 3178 }, { "epoch": 0.8750988958068178, "grad_norm": 0.3428981900215149, "learning_rate": 1e-05, "loss": 0.4172, "step": 3180 }, { "epoch": 0.8756492724708472, "grad_norm": 0.36314892768859863, "learning_rate": 1e-05, "loss": 0.4008, "step": 3182 }, { "epoch": 0.8761996491348767, "grad_norm": 0.35167455673217773, "learning_rate": 1e-05, "loss": 0.4114, "step": 3184 }, { "epoch": 0.8767500257989062, "grad_norm": 0.3496149778366089, "learning_rate": 1e-05, "loss": 0.4053, "step": 3186 }, { "epoch": 0.8773004024629356, "grad_norm": 0.351510226726532, "learning_rate": 1e-05, "loss": 0.4173, "step": 3188 }, { "epoch": 0.877850779126965, "grad_norm": 0.35172203183174133, "learning_rate": 1e-05, "loss": 0.4202, "step": 3190 }, { "epoch": 0.8784011557909944, "grad_norm": 0.36200663447380066, "learning_rate": 1e-05, "loss": 0.3987, "step": 3192 }, { "epoch": 0.8789515324550239, "grad_norm": 0.36070528626441956, "learning_rate": 1e-05, "loss": 0.4202, "step": 3194 }, { "epoch": 0.8795019091190533, "grad_norm": 0.37506040930747986, "learning_rate": 1e-05, "loss": 0.4125, "step": 3196 }, { "epoch": 0.8800522857830828, "grad_norm": 0.3433153033256531, "learning_rate": 1e-05, "loss": 0.4185, "step": 3198 }, { "epoch": 0.8806026624471123, "grad_norm": 0.3672421872615814, "learning_rate": 1e-05, "loss": 0.4227, "step": 3200 }, { "epoch": 0.8806026624471123, "eval_merge_loss": 0.36182981729507446, "eval_merge_runtime": 600.3542, "eval_merge_samples_per_second": 56.22, "eval_merge_steps_per_second": 2.344, "step": 3200 }, { "epoch": 0.8806026624471123, "eval_new_aug_datas_filtered.json_loss": 0.4833287298679352, "eval_new_aug_datas_filtered.json_runtime": 10.3758, "eval_new_aug_datas_filtered.json_samples_per_second": 73.922, "eval_new_aug_datas_filtered.json_steps_per_second": 3.084, "step": 3200 }, { "epoch": 0.8806026624471123, "eval_sharegpt_gpt4.json_loss": 0.7305224537849426, "eval_sharegpt_gpt4.json_runtime": 31.7036, "eval_sharegpt_gpt4.json_samples_per_second": 58.7, "eval_sharegpt_gpt4.json_steps_per_second": 2.46, "step": 3200 }, { "epoch": 0.8806026624471123, "eval_Table_GPT.json_loss": 0.04232589527964592, "eval_Table_GPT.json_runtime": 24.994, "eval_Table_GPT.json_samples_per_second": 83.74, "eval_Table_GPT.json_steps_per_second": 3.521, "step": 3200 }, { "epoch": 0.8806026624471123, "eval_gpt_4o_200k.json_loss": 0.7571491003036499, "eval_gpt_4o_200k.json_runtime": 48.5629, "eval_gpt_4o_200k.json_samples_per_second": 129.337, "eval_gpt_4o_200k.json_steps_per_second": 5.395, "step": 3200 }, { "epoch": 0.8806026624471123, "eval_multi_turn_datas.json_loss": 0.2720319330692291, "eval_multi_turn_datas.json_runtime": 75.6646, "eval_multi_turn_datas.json_samples_per_second": 52.891, "eval_multi_turn_datas.json_steps_per_second": 2.207, "step": 3200 }, { "epoch": 0.8806026624471123, "eval_table_python_code_datas.json_loss": 0.24331320822238922, "eval_table_python_code_datas.json_runtime": 43.0385, "eval_table_python_code_datas.json_samples_per_second": 50.164, "eval_table_python_code_datas.json_steps_per_second": 2.091, "step": 3200 }, { "epoch": 0.8806026624471123, "eval_tabular_llm_data.json_loss": 0.0862693339586258, "eval_tabular_llm_data.json_runtime": 8.5454, "eval_tabular_llm_data.json_samples_per_second": 28.788, "eval_tabular_llm_data.json_steps_per_second": 1.287, "step": 3200 }, { "epoch": 0.8806026624471123, "eval_python_code_critic_21k.json_loss": 0.5425785183906555, "eval_python_code_critic_21k.json_runtime": 3.2194, "eval_python_code_critic_21k.json_samples_per_second": 185.437, "eval_python_code_critic_21k.json_steps_per_second": 7.765, "step": 3200 }, { "epoch": 0.8806026624471123, "eval_all_merge_table_dataset.json_loss": 0.06970688700675964, "eval_all_merge_table_dataset.json_runtime": 23.3201, "eval_all_merge_table_dataset.json_samples_per_second": 30.532, "eval_all_merge_table_dataset.json_steps_per_second": 1.286, "step": 3200 }, { "epoch": 0.8806026624471123, "eval_code_feedback_multi_turn.json_loss": 0.5619133114814758, "eval_code_feedback_multi_turn.json_runtime": 32.4257, "eval_code_feedback_multi_turn.json_samples_per_second": 67.878, "eval_code_feedback_multi_turn.json_steps_per_second": 2.837, "step": 3200 }, { "epoch": 0.8806026624471123, "eval_ultrainteract_sft.json_loss": 0.4016903042793274, "eval_ultrainteract_sft.json_runtime": 8.6472, "eval_ultrainteract_sft.json_samples_per_second": 168.378, "eval_ultrainteract_sft.json_steps_per_second": 7.054, "step": 3200 }, { "epoch": 0.8806026624471123, "eval_synthetic_text_to_sql.json_loss": 0.09171026945114136, "eval_synthetic_text_to_sql.json_runtime": 0.1264, "eval_synthetic_text_to_sql.json_samples_per_second": 268.914, "eval_synthetic_text_to_sql.json_steps_per_second": 15.818, "step": 3200 }, { "epoch": 0.8806026624471123, "eval_sft_react_sql_datas.json_loss": 0.6105172038078308, "eval_sft_react_sql_datas.json_runtime": 7.8425, "eval_sft_react_sql_datas.json_samples_per_second": 40.038, "eval_sft_react_sql_datas.json_steps_per_second": 1.785, "step": 3200 }, { "epoch": 0.8806026624471123, "eval_all_merge_code.json_loss": 0.264506459236145, "eval_all_merge_code.json_runtime": 0.3347, "eval_all_merge_code.json_samples_per_second": 188.208, "eval_all_merge_code.json_steps_per_second": 8.962, "step": 3200 }, { "epoch": 0.8806026624471123, "eval_magpie_datas.json_loss": 0.4236694872379303, "eval_magpie_datas.json_runtime": 2.213, "eval_magpie_datas.json_samples_per_second": 77.723, "eval_magpie_datas.json_steps_per_second": 3.615, "step": 3200 }, { "epoch": 0.8806026624471123, "eval_train_data_for_qwen.json_loss": 0.0027615067083388567, "eval_train_data_for_qwen.json_runtime": 0.2435, "eval_train_data_for_qwen.json_samples_per_second": 41.06, "eval_train_data_for_qwen.json_steps_per_second": 4.106, "step": 3200 }, { "epoch": 0.8806026624471123, "eval_alpaca_cleaned.json_loss": 0.9028782844543457, "eval_alpaca_cleaned.json_runtime": 0.1145, "eval_alpaca_cleaned.json_samples_per_second": 235.866, "eval_alpaca_cleaned.json_steps_per_second": 17.472, "step": 3200 }, { "epoch": 0.8806026624471123, "eval_agent_instruct.json_loss": 0.20418775081634521, "eval_agent_instruct.json_runtime": 0.5123, "eval_agent_instruct.json_samples_per_second": 93.693, "eval_agent_instruct.json_steps_per_second": 3.904, "step": 3200 }, { "epoch": 0.8806026624471123, "eval_MathInstruct.json_loss": 0.2024046629667282, "eval_MathInstruct.json_runtime": 0.35, "eval_MathInstruct.json_samples_per_second": 162.858, "eval_MathInstruct.json_steps_per_second": 8.571, "step": 3200 }, { "epoch": 0.8806026624471123, "eval_tested_143k_python_alpaca.json_loss": 0.4462108612060547, "eval_tested_143k_python_alpaca.json_runtime": 0.3037, "eval_tested_143k_python_alpaca.json_samples_per_second": 111.97, "eval_tested_143k_python_alpaca.json_steps_per_second": 6.586, "step": 3200 }, { "epoch": 0.8806026624471123, "eval_xlam_function_calling_60k.json_loss": 0.008976898156106472, "eval_xlam_function_calling_60k.json_runtime": 0.1004, "eval_xlam_function_calling_60k.json_samples_per_second": 229.083, "eval_xlam_function_calling_60k.json_steps_per_second": 9.96, "step": 3200 }, { "epoch": 0.8806026624471123, "eval_alpaca_data_gpt4_chinese.json_loss": 1.512216567993164, "eval_alpaca_data_gpt4_chinese.json_runtime": 0.0511, "eval_alpaca_data_gpt4_chinese.json_samples_per_second": 313.242, "eval_alpaca_data_gpt4_chinese.json_steps_per_second": 19.578, "step": 3200 }, { "epoch": 0.8806026624471123, "eval_alpaca_gpt4_zh.json_loss": 0.9716835021972656, "eval_alpaca_gpt4_zh.json_runtime": 0.0498, "eval_alpaca_gpt4_zh.json_samples_per_second": 220.769, "eval_alpaca_gpt4_zh.json_steps_per_second": 20.07, "step": 3200 }, { "epoch": 0.8806026624471123, "eval_codefeedback_filtered_instruction.json_loss": 0.5663765072822571, "eval_codefeedback_filtered_instruction.json_runtime": 0.4857, "eval_codefeedback_filtered_instruction.json_samples_per_second": 41.176, "eval_codefeedback_filtered_instruction.json_steps_per_second": 2.059, "step": 3200 } ], "logging_steps": 2, "max_steps": 3633, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 200, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 2.4654685450828094e+20, "train_batch_size": 3, "trial_name": null, "trial_params": null }