|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 3.0, |
|
"eval_steps": 500, |
|
"global_step": 1005, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.029850746268656716, |
|
"grad_norm": 3.2636772208309153, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7976, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.05970149253731343, |
|
"grad_norm": 4.198083473691167, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7201, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.08955223880597014, |
|
"grad_norm": 0.976830869365462, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7097, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.11940298507462686, |
|
"grad_norm": 1.0243806020799167, |
|
"learning_rate": 5e-06, |
|
"loss": 0.687, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.14925373134328357, |
|
"grad_norm": 1.0277760094413817, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6709, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.1791044776119403, |
|
"grad_norm": 0.8438731934295206, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6661, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.208955223880597, |
|
"grad_norm": 0.7599111846557922, |
|
"learning_rate": 5e-06, |
|
"loss": 0.661, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.23880597014925373, |
|
"grad_norm": 0.6979506430052831, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6481, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.26865671641791045, |
|
"grad_norm": 0.6344985723138951, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6494, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.29850746268656714, |
|
"grad_norm": 0.9496025763872573, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6351, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.3283582089552239, |
|
"grad_norm": 0.5640696000070055, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6292, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.3582089552238806, |
|
"grad_norm": 0.6244500377098101, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6246, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.3880597014925373, |
|
"grad_norm": 0.705148467511176, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6288, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.417910447761194, |
|
"grad_norm": 0.5782292155064868, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6208, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.44776119402985076, |
|
"grad_norm": 0.47879907688098855, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6307, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.47761194029850745, |
|
"grad_norm": 0.641221897829538, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6282, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.5074626865671642, |
|
"grad_norm": 0.5790307980225682, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6263, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.5373134328358209, |
|
"grad_norm": 0.5896556501658362, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6171, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.5671641791044776, |
|
"grad_norm": 0.6678853160055497, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6217, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.5970149253731343, |
|
"grad_norm": 0.4932224279411186, |
|
"learning_rate": 5e-06, |
|
"loss": 0.612, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.6268656716417911, |
|
"grad_norm": 0.59016513490961, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6207, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.6567164179104478, |
|
"grad_norm": 0.6628507303786519, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6205, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.6865671641791045, |
|
"grad_norm": 0.5728090788820762, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6198, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.7164179104477612, |
|
"grad_norm": 0.5909920699712947, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6195, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.746268656716418, |
|
"grad_norm": 0.5403630353274208, |
|
"learning_rate": 5e-06, |
|
"loss": 0.616, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.7761194029850746, |
|
"grad_norm": 0.4389359227878181, |
|
"learning_rate": 5e-06, |
|
"loss": 0.615, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.8059701492537313, |
|
"grad_norm": 0.5071059932681349, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6119, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.835820895522388, |
|
"grad_norm": 0.5232710694651, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6052, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.8656716417910447, |
|
"grad_norm": 0.6304632419292111, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6107, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.8955223880597015, |
|
"grad_norm": 0.4513885510016045, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6029, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.9253731343283582, |
|
"grad_norm": 0.46627673450975804, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6224, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.9552238805970149, |
|
"grad_norm": 0.4541172691354782, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6005, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.9850746268656716, |
|
"grad_norm": 0.5886900494680195, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6102, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 0.6099968552589417, |
|
"eval_runtime": 181.0403, |
|
"eval_samples_per_second": 49.84, |
|
"eval_steps_per_second": 0.392, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 1.0149253731343284, |
|
"grad_norm": 0.7366116445067696, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5859, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 1.044776119402985, |
|
"grad_norm": 0.571210239937034, |
|
"learning_rate": 5e-06, |
|
"loss": 0.56, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 1.0746268656716418, |
|
"grad_norm": 0.4668816015405423, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5592, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 1.1044776119402986, |
|
"grad_norm": 0.5628631007519552, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5657, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 1.1343283582089552, |
|
"grad_norm": 0.5929997737908812, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5563, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 1.164179104477612, |
|
"grad_norm": 0.5658505485409301, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5578, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 1.1940298507462686, |
|
"grad_norm": 0.7526049702677613, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5615, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 1.2238805970149254, |
|
"grad_norm": 0.5393989822533652, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5673, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 1.2537313432835822, |
|
"grad_norm": 0.5325263654739846, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5638, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 1.2835820895522387, |
|
"grad_norm": 0.5064885711552297, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5607, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 1.3134328358208955, |
|
"grad_norm": 0.5684568858118794, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5576, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 1.3432835820895521, |
|
"grad_norm": 0.5283953048651826, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5561, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 1.373134328358209, |
|
"grad_norm": 0.46950929831714766, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5602, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 1.4029850746268657, |
|
"grad_norm": 0.5385565397537747, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5581, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 1.4328358208955223, |
|
"grad_norm": 0.5659871694460666, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5617, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 1.462686567164179, |
|
"grad_norm": 0.546965476854921, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5553, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 1.4925373134328357, |
|
"grad_norm": 0.497538374377084, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5602, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 1.5223880597014925, |
|
"grad_norm": 0.5312343228790736, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5618, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 1.5522388059701493, |
|
"grad_norm": 0.5166406423583252, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5591, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 1.582089552238806, |
|
"grad_norm": 0.5255413189150826, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5604, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 1.6119402985074627, |
|
"grad_norm": 0.49551565255287583, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5618, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 1.6417910447761193, |
|
"grad_norm": 0.49267014615726396, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5585, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 1.671641791044776, |
|
"grad_norm": 0.5643970187910693, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5652, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 1.7014925373134329, |
|
"grad_norm": 0.5037470018499742, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5625, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 1.7313432835820897, |
|
"grad_norm": 0.48141660593931324, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5568, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 1.7611940298507462, |
|
"grad_norm": 0.5300295332006951, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5597, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 1.7910447761194028, |
|
"grad_norm": 0.5373472168427755, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5562, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 1.8208955223880596, |
|
"grad_norm": 0.5707503080671126, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5527, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 1.8507462686567164, |
|
"grad_norm": 0.5730054222686448, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5665, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 1.8805970149253732, |
|
"grad_norm": 0.5072679291608492, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5513, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 1.9104477611940298, |
|
"grad_norm": 0.5178695559305173, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5583, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 1.9402985074626866, |
|
"grad_norm": 0.5656735937545936, |
|
"learning_rate": 5e-06, |
|
"loss": 0.563, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 1.9701492537313432, |
|
"grad_norm": 0.4863140367653273, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5595, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 0.5328164667475485, |
|
"learning_rate": 5e-06, |
|
"loss": 0.553, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 0.6020517349243164, |
|
"eval_runtime": 181.5231, |
|
"eval_samples_per_second": 49.707, |
|
"eval_steps_per_second": 0.391, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 2.029850746268657, |
|
"grad_norm": 0.6224628520404539, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5057, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 2.0597014925373136, |
|
"grad_norm": 0.5148413250898433, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5083, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 2.08955223880597, |
|
"grad_norm": 0.5438872711269419, |
|
"learning_rate": 5e-06, |
|
"loss": 0.504, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 2.1194029850746268, |
|
"grad_norm": 0.7136622301499072, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5084, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 2.1492537313432836, |
|
"grad_norm": 0.565521915151828, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5141, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 2.1791044776119404, |
|
"grad_norm": 0.5319158435460484, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5107, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 2.208955223880597, |
|
"grad_norm": 0.5565738587496012, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5096, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 2.2388059701492535, |
|
"grad_norm": 0.4910069126256564, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5154, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 2.2686567164179103, |
|
"grad_norm": 0.48169824804919137, |
|
"learning_rate": 5e-06, |
|
"loss": 0.515, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 2.298507462686567, |
|
"grad_norm": 0.5716888321310263, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5059, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 2.328358208955224, |
|
"grad_norm": 0.5310388379236972, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5065, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 2.3582089552238807, |
|
"grad_norm": 0.5340384268973276, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5097, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 2.388059701492537, |
|
"grad_norm": 0.5612588485250506, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5092, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 2.417910447761194, |
|
"grad_norm": 0.5517492107037403, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5155, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 2.4477611940298507, |
|
"grad_norm": 0.5868418861130762, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5099, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 2.4776119402985075, |
|
"grad_norm": 0.5335016774082386, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4953, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 2.5074626865671643, |
|
"grad_norm": 0.47262015142721125, |
|
"learning_rate": 5e-06, |
|
"loss": 0.508, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 2.5373134328358207, |
|
"grad_norm": 0.4852270938780388, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5125, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 2.5671641791044775, |
|
"grad_norm": 0.5802563395388763, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5095, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 2.5970149253731343, |
|
"grad_norm": 0.503360426921311, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5189, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 2.626865671641791, |
|
"grad_norm": 0.49515721551631564, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5158, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 2.656716417910448, |
|
"grad_norm": 0.48219119082060224, |
|
"learning_rate": 5e-06, |
|
"loss": 0.509, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 2.6865671641791042, |
|
"grad_norm": 0.4921666373084123, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5099, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 2.716417910447761, |
|
"grad_norm": 0.547644485149717, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5056, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 2.746268656716418, |
|
"grad_norm": 0.5561203783171008, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5105, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 2.7761194029850746, |
|
"grad_norm": 0.5027190076200428, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5101, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 2.8059701492537314, |
|
"grad_norm": 0.5325189560701381, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5126, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 2.835820895522388, |
|
"grad_norm": 0.6188498248886457, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5125, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 2.8656716417910446, |
|
"grad_norm": 0.5315894494034671, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5101, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 2.8955223880597014, |
|
"grad_norm": 0.5715252097778681, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5138, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 2.925373134328358, |
|
"grad_norm": 0.4780679784022612, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5108, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 2.955223880597015, |
|
"grad_norm": 0.48072163191940603, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5103, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 2.9850746268656714, |
|
"grad_norm": 0.5364386007246811, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5156, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_loss": 0.6066075563430786, |
|
"eval_runtime": 181.24, |
|
"eval_samples_per_second": 49.785, |
|
"eval_steps_per_second": 0.392, |
|
"step": 1005 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"step": 1005, |
|
"total_flos": 1682993672355840.0, |
|
"train_loss": 0.5693152133505143, |
|
"train_runtime": 30325.6118, |
|
"train_samples_per_second": 16.959, |
|
"train_steps_per_second": 0.033 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 1005, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1682993672355840.0, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|