|
{ |
|
"best_metric": 0.8367530107498169, |
|
"best_model_checkpoint": "albert-base-v2-Malicious_URLs/checkpoint-51087", |
|
"epoch": 1.0, |
|
"global_step": 51087, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 1.9999608510971483e-05, |
|
"loss": 1.1706, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 1.9980425548574e-05, |
|
"loss": 1.1046, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 1.9960851097148004e-05, |
|
"loss": 0.9971, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 1.9941276645722004e-05, |
|
"loss": 1.0489, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 1.9921702194296007e-05, |
|
"loss": 0.9931, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 1.9902127742870007e-05, |
|
"loss": 1.0318, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.988255329144401e-05, |
|
"loss": 1.015, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.986297884001801e-05, |
|
"loss": 0.8838, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.9843404388592012e-05, |
|
"loss": 0.9699, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.9823829937166012e-05, |
|
"loss": 0.9734, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.9804255485740015e-05, |
|
"loss": 0.9944, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.9784681034314015e-05, |
|
"loss": 0.9173, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.9765106582888018e-05, |
|
"loss": 0.8308, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.9745532131462017e-05, |
|
"loss": 0.8602, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.9725957680036017e-05, |
|
"loss": 0.9536, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.970638322861002e-05, |
|
"loss": 0.9808, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 1.968680877718402e-05, |
|
"loss": 0.884, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 1.9667234325758023e-05, |
|
"loss": 0.9394, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 1.9647659874332022e-05, |
|
"loss": 0.9618, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 1.9628085422906025e-05, |
|
"loss": 0.9904, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 1.9608510971480025e-05, |
|
"loss": 0.9654, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 1.9588936520054028e-05, |
|
"loss": 0.8642, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 1.9569362068628028e-05, |
|
"loss": 0.949, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 1.954978761720203e-05, |
|
"loss": 0.8394, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 1.953021316577603e-05, |
|
"loss": 0.8666, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 1.9510638714350033e-05, |
|
"loss": 0.9656, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 1.9491064262924033e-05, |
|
"loss": 0.9162, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 1.9471489811498033e-05, |
|
"loss": 0.9125, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 1.9451915360072036e-05, |
|
"loss": 0.872, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 1.9432340908646035e-05, |
|
"loss": 0.9858, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 1.941276645722004e-05, |
|
"loss": 0.9059, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 1.9393192005794038e-05, |
|
"loss": 0.9647, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 1.937361755436804e-05, |
|
"loss": 0.9802, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 1.935404310294204e-05, |
|
"loss": 0.9604, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 1.9334468651516044e-05, |
|
"loss": 0.8774, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 1.9314894200090043e-05, |
|
"loss": 0.8881, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 1.9295319748664047e-05, |
|
"loss": 0.8775, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 1.9275745297238046e-05, |
|
"loss": 0.9427, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 1.925617084581205e-05, |
|
"loss": 0.8997, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 1.923659639438605e-05, |
|
"loss": 0.9101, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 1.9217021942960052e-05, |
|
"loss": 0.9247, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 1.919744749153405e-05, |
|
"loss": 0.8978, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 1.917787304010805e-05, |
|
"loss": 0.9436, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 1.9158298588682054e-05, |
|
"loss": 0.91, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 1.9138724137256054e-05, |
|
"loss": 0.965, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 1.9119149685830057e-05, |
|
"loss": 0.9488, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.9099575234404057e-05, |
|
"loss": 0.8534, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.908000078297806e-05, |
|
"loss": 0.9636, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.906042633155206e-05, |
|
"loss": 0.9581, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.9040851880126062e-05, |
|
"loss": 0.83, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.9021277428700062e-05, |
|
"loss": 0.9326, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.9001702977274065e-05, |
|
"loss": 0.9152, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.8982128525848065e-05, |
|
"loss": 0.9363, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.8962554074422068e-05, |
|
"loss": 0.8746, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.8942979622996067e-05, |
|
"loss": 0.8996, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.8923405171570067e-05, |
|
"loss": 0.9116, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.890383072014407e-05, |
|
"loss": 0.8015, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.888425626871807e-05, |
|
"loss": 0.926, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.8864681817292073e-05, |
|
"loss": 0.8506, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.8845107365866072e-05, |
|
"loss": 0.9913, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.8825532914440075e-05, |
|
"loss": 0.8355, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.8805958463014075e-05, |
|
"loss": 0.9649, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.8786384011588078e-05, |
|
"loss": 0.8837, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.8766809560162078e-05, |
|
"loss": 0.9546, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.874723510873608e-05, |
|
"loss": 0.8529, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.872766065731008e-05, |
|
"loss": 0.9503, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.8708086205884083e-05, |
|
"loss": 1.0059, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.8688511754458083e-05, |
|
"loss": 0.9493, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.8668937303032083e-05, |
|
"loss": 0.8838, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.8649362851606086e-05, |
|
"loss": 0.8728, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.8629788400180085e-05, |
|
"loss": 0.8237, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.861021394875409e-05, |
|
"loss": 0.9691, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.8590639497328088e-05, |
|
"loss": 0.8934, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.857106504590209e-05, |
|
"loss": 0.8628, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.855149059447609e-05, |
|
"loss": 0.9047, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.8531916143050094e-05, |
|
"loss": 0.9896, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.8512341691624093e-05, |
|
"loss": 0.8885, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.8492767240198096e-05, |
|
"loss": 0.9171, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.8473192788772096e-05, |
|
"loss": 0.9144, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.84536183373461e-05, |
|
"loss": 0.949, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.84340438859201e-05, |
|
"loss": 0.9595, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.84144694344941e-05, |
|
"loss": 0.9018, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.83948949830681e-05, |
|
"loss": 0.9434, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.83753205316421e-05, |
|
"loss": 1.0054, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.8355746080216104e-05, |
|
"loss": 0.8584, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.8336171628790104e-05, |
|
"loss": 0.8759, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.8316597177364107e-05, |
|
"loss": 0.7978, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.8297022725938106e-05, |
|
"loss": 0.8494, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.827744827451211e-05, |
|
"loss": 0.8714, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.825787382308611e-05, |
|
"loss": 0.9271, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.8238299371660112e-05, |
|
"loss": 0.9466, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.8218724920234112e-05, |
|
"loss": 0.8289, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.8199150468808115e-05, |
|
"loss": 0.9159, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.8179576017382114e-05, |
|
"loss": 0.8418, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.8160001565956114e-05, |
|
"loss": 0.9818, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.8140427114530117e-05, |
|
"loss": 0.8762, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.8120852663104117e-05, |
|
"loss": 0.9628, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.810127821167812e-05, |
|
"loss": 0.7778, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.808170376025212e-05, |
|
"loss": 0.8857, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.8062129308826123e-05, |
|
"loss": 0.9288, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.8042554857400122e-05, |
|
"loss": 0.9101, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.8022980405974125e-05, |
|
"loss": 0.9358, |
|
"step": 5050 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.8003405954548125e-05, |
|
"loss": 0.8956, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.7983831503122128e-05, |
|
"loss": 0.8506, |
|
"step": 5150 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.7964257051696128e-05, |
|
"loss": 0.8635, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.794468260027013e-05, |
|
"loss": 0.9267, |
|
"step": 5250 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.792510814884413e-05, |
|
"loss": 0.8688, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.790553369741813e-05, |
|
"loss": 0.8538, |
|
"step": 5350 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.7885959245992133e-05, |
|
"loss": 0.8628, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.7866384794566133e-05, |
|
"loss": 0.9528, |
|
"step": 5450 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.7846810343140136e-05, |
|
"loss": 0.9198, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.7827235891714135e-05, |
|
"loss": 0.8878, |
|
"step": 5550 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.7807661440288138e-05, |
|
"loss": 0.9425, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.7788086988862138e-05, |
|
"loss": 0.8977, |
|
"step": 5650 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.776851253743614e-05, |
|
"loss": 0.8967, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.774893808601014e-05, |
|
"loss": 0.9365, |
|
"step": 5750 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.7729363634584144e-05, |
|
"loss": 0.8376, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.7709789183158143e-05, |
|
"loss": 0.8587, |
|
"step": 5850 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.7690214731732146e-05, |
|
"loss": 0.8808, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.7670640280306146e-05, |
|
"loss": 0.8871, |
|
"step": 5950 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.765106582888015e-05, |
|
"loss": 0.8399, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.763149137745415e-05, |
|
"loss": 0.9007, |
|
"step": 6050 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.7611916926028148e-05, |
|
"loss": 0.9102, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.759234247460215e-05, |
|
"loss": 0.8689, |
|
"step": 6150 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.757276802317615e-05, |
|
"loss": 0.9715, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.7553193571750154e-05, |
|
"loss": 0.893, |
|
"step": 6250 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.7533619120324154e-05, |
|
"loss": 0.8597, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.7514044668898157e-05, |
|
"loss": 0.9172, |
|
"step": 6350 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.7494470217472156e-05, |
|
"loss": 0.9287, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.747489576604616e-05, |
|
"loss": 0.8687, |
|
"step": 6450 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.745532131462016e-05, |
|
"loss": 0.8804, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.7435746863194162e-05, |
|
"loss": 0.9483, |
|
"step": 6550 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.741617241176816e-05, |
|
"loss": 0.8749, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.7396597960342165e-05, |
|
"loss": 0.9452, |
|
"step": 6650 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.7377023508916164e-05, |
|
"loss": 0.9059, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.7357449057490164e-05, |
|
"loss": 0.8311, |
|
"step": 6750 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.7337874606064167e-05, |
|
"loss": 0.8671, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.7318300154638167e-05, |
|
"loss": 0.9169, |
|
"step": 6850 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.729872570321217e-05, |
|
"loss": 0.9448, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.727915125178617e-05, |
|
"loss": 0.8703, |
|
"step": 6950 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.7259576800360172e-05, |
|
"loss": 0.9438, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.7240002348934172e-05, |
|
"loss": 0.868, |
|
"step": 7050 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.7220427897508175e-05, |
|
"loss": 0.9868, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.7200853446082175e-05, |
|
"loss": 0.7904, |
|
"step": 7150 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.7181278994656178e-05, |
|
"loss": 0.8628, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.7161704543230177e-05, |
|
"loss": 0.8898, |
|
"step": 7250 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.714213009180418e-05, |
|
"loss": 0.9962, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.712255564037818e-05, |
|
"loss": 0.939, |
|
"step": 7350 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.710298118895218e-05, |
|
"loss": 0.7759, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.7083406737526183e-05, |
|
"loss": 0.8616, |
|
"step": 7450 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.7063832286100182e-05, |
|
"loss": 0.9495, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.7044257834674185e-05, |
|
"loss": 0.8752, |
|
"step": 7550 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.7024683383248185e-05, |
|
"loss": 0.8908, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.7005108931822188e-05, |
|
"loss": 0.8843, |
|
"step": 7650 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.6985534480396188e-05, |
|
"loss": 0.9643, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.696596002897019e-05, |
|
"loss": 0.8528, |
|
"step": 7750 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.694638557754419e-05, |
|
"loss": 0.8625, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.6926811126118193e-05, |
|
"loss": 0.9945, |
|
"step": 7850 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.6907236674692193e-05, |
|
"loss": 0.8112, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.6887662223266196e-05, |
|
"loss": 0.9049, |
|
"step": 7950 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.6868087771840196e-05, |
|
"loss": 0.8941, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.6848513320414195e-05, |
|
"loss": 0.8887, |
|
"step": 8050 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.68289388689882e-05, |
|
"loss": 0.941, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.6809364417562198e-05, |
|
"loss": 0.8806, |
|
"step": 8150 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.67897899661362e-05, |
|
"loss": 0.9605, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.67702155147102e-05, |
|
"loss": 0.8432, |
|
"step": 8250 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.6750641063284204e-05, |
|
"loss": 0.8381, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.6731066611858204e-05, |
|
"loss": 0.8558, |
|
"step": 8350 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.6711492160432207e-05, |
|
"loss": 0.9525, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.6691917709006206e-05, |
|
"loss": 0.8599, |
|
"step": 8450 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.667234325758021e-05, |
|
"loss": 0.9754, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.665276880615421e-05, |
|
"loss": 0.9006, |
|
"step": 8550 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.6633194354728212e-05, |
|
"loss": 0.8845, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.661361990330221e-05, |
|
"loss": 1.0112, |
|
"step": 8650 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.659404545187621e-05, |
|
"loss": 1.0344, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.6574471000450214e-05, |
|
"loss": 0.9505, |
|
"step": 8750 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.6554896549024214e-05, |
|
"loss": 0.8472, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.6535322097598217e-05, |
|
"loss": 0.923, |
|
"step": 8850 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.6515747646172217e-05, |
|
"loss": 0.9265, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.649617319474622e-05, |
|
"loss": 0.9428, |
|
"step": 8950 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.647659874332022e-05, |
|
"loss": 0.961, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.6457024291894222e-05, |
|
"loss": 0.8903, |
|
"step": 9050 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.6437449840468222e-05, |
|
"loss": 0.9757, |
|
"step": 9100 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.6417875389042225e-05, |
|
"loss": 0.9208, |
|
"step": 9150 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.6398300937616225e-05, |
|
"loss": 0.9723, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.6378726486190228e-05, |
|
"loss": 0.845, |
|
"step": 9250 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.6359152034764227e-05, |
|
"loss": 0.9029, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.633957758333823e-05, |
|
"loss": 0.8009, |
|
"step": 9350 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.632000313191223e-05, |
|
"loss": 0.9144, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.630042868048623e-05, |
|
"loss": 0.9191, |
|
"step": 9450 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.6280854229060233e-05, |
|
"loss": 0.9451, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.6261279777634232e-05, |
|
"loss": 0.9365, |
|
"step": 9550 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.6241705326208235e-05, |
|
"loss": 0.8914, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.6222130874782235e-05, |
|
"loss": 0.849, |
|
"step": 9650 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.6202556423356238e-05, |
|
"loss": 0.8415, |
|
"step": 9700 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.6182981971930238e-05, |
|
"loss": 0.9671, |
|
"step": 9750 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.616340752050424e-05, |
|
"loss": 0.8248, |
|
"step": 9800 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.614383306907824e-05, |
|
"loss": 0.8498, |
|
"step": 9850 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.6124258617652243e-05, |
|
"loss": 1.0332, |
|
"step": 9900 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.6104684166226243e-05, |
|
"loss": 0.9052, |
|
"step": 9950 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.6085109714800246e-05, |
|
"loss": 0.9069, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.6065535263374246e-05, |
|
"loss": 0.8651, |
|
"step": 10050 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.6045960811948245e-05, |
|
"loss": 0.9262, |
|
"step": 10100 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.602638636052225e-05, |
|
"loss": 0.9556, |
|
"step": 10150 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.6006811909096248e-05, |
|
"loss": 0.8846, |
|
"step": 10200 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.598723745767025e-05, |
|
"loss": 0.8435, |
|
"step": 10250 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.596766300624425e-05, |
|
"loss": 0.9585, |
|
"step": 10300 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.5948088554818254e-05, |
|
"loss": 0.8239, |
|
"step": 10350 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.5928514103392253e-05, |
|
"loss": 0.811, |
|
"step": 10400 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.5908939651966256e-05, |
|
"loss": 0.8884, |
|
"step": 10450 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.5889365200540256e-05, |
|
"loss": 0.8534, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.586979074911426e-05, |
|
"loss": 0.9667, |
|
"step": 10550 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.585021629768826e-05, |
|
"loss": 0.9295, |
|
"step": 10600 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.5830641846262262e-05, |
|
"loss": 0.9055, |
|
"step": 10650 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.581106739483626e-05, |
|
"loss": 0.8377, |
|
"step": 10700 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.579149294341026e-05, |
|
"loss": 0.9062, |
|
"step": 10750 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.5771918491984264e-05, |
|
"loss": 0.8655, |
|
"step": 10800 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.5752344040558264e-05, |
|
"loss": 0.8602, |
|
"step": 10850 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.5732769589132267e-05, |
|
"loss": 0.8707, |
|
"step": 10900 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.5713195137706266e-05, |
|
"loss": 0.8257, |
|
"step": 10950 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.569362068628027e-05, |
|
"loss": 0.8605, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.567404623485427e-05, |
|
"loss": 0.7885, |
|
"step": 11050 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.5654471783428272e-05, |
|
"loss": 0.8821, |
|
"step": 11100 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.5634897332002272e-05, |
|
"loss": 0.8238, |
|
"step": 11150 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.5615322880576275e-05, |
|
"loss": 0.9978, |
|
"step": 11200 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.5595748429150274e-05, |
|
"loss": 0.8357, |
|
"step": 11250 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.5576173977724278e-05, |
|
"loss": 0.9715, |
|
"step": 11300 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.5556599526298277e-05, |
|
"loss": 0.8519, |
|
"step": 11350 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.5537025074872277e-05, |
|
"loss": 0.7997, |
|
"step": 11400 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.551745062344628e-05, |
|
"loss": 0.7839, |
|
"step": 11450 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.549787617202028e-05, |
|
"loss": 0.7902, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.5478301720594283e-05, |
|
"loss": 0.9108, |
|
"step": 11550 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.5458727269168282e-05, |
|
"loss": 0.9531, |
|
"step": 11600 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.5439152817742285e-05, |
|
"loss": 0.9653, |
|
"step": 11650 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.5419578366316285e-05, |
|
"loss": 0.8278, |
|
"step": 11700 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.5400003914890288e-05, |
|
"loss": 0.9261, |
|
"step": 11750 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.5380429463464288e-05, |
|
"loss": 0.8742, |
|
"step": 11800 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.536085501203829e-05, |
|
"loss": 0.8484, |
|
"step": 11850 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.534128056061229e-05, |
|
"loss": 0.9005, |
|
"step": 11900 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.5321706109186293e-05, |
|
"loss": 0.9226, |
|
"step": 11950 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.5302131657760293e-05, |
|
"loss": 0.8965, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.5282557206334293e-05, |
|
"loss": 0.9745, |
|
"step": 12050 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.5262982754908296e-05, |
|
"loss": 0.9371, |
|
"step": 12100 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.5243408303482297e-05, |
|
"loss": 0.8983, |
|
"step": 12150 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.5223833852056298e-05, |
|
"loss": 0.9373, |
|
"step": 12200 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.52042594006303e-05, |
|
"loss": 0.854, |
|
"step": 12250 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.5184684949204301e-05, |
|
"loss": 0.8189, |
|
"step": 12300 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.5165110497778302e-05, |
|
"loss": 0.8548, |
|
"step": 12350 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.5145536046352302e-05, |
|
"loss": 0.8075, |
|
"step": 12400 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.5125961594926303e-05, |
|
"loss": 0.8479, |
|
"step": 12450 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.5106387143500305e-05, |
|
"loss": 0.7847, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.5086812692074306e-05, |
|
"loss": 0.859, |
|
"step": 12550 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.5067238240648307e-05, |
|
"loss": 0.9378, |
|
"step": 12600 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.5047663789222309e-05, |
|
"loss": 0.911, |
|
"step": 12650 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.502808933779631e-05, |
|
"loss": 0.9074, |
|
"step": 12700 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.5008514886370311e-05, |
|
"loss": 0.8419, |
|
"step": 12750 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.4988940434944313e-05, |
|
"loss": 0.8942, |
|
"step": 12800 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.4969365983518314e-05, |
|
"loss": 0.8709, |
|
"step": 12850 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.4949791532092315e-05, |
|
"loss": 0.9247, |
|
"step": 12900 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.4930217080666317e-05, |
|
"loss": 0.9094, |
|
"step": 12950 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.4910642629240318e-05, |
|
"loss": 0.8503, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.4891068177814318e-05, |
|
"loss": 0.9849, |
|
"step": 13050 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.4871493726388319e-05, |
|
"loss": 0.9683, |
|
"step": 13100 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.485191927496232e-05, |
|
"loss": 0.9421, |
|
"step": 13150 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.4832344823536322e-05, |
|
"loss": 0.8789, |
|
"step": 13200 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.4812770372110323e-05, |
|
"loss": 0.8922, |
|
"step": 13250 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.4793195920684324e-05, |
|
"loss": 0.8455, |
|
"step": 13300 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.4773621469258326e-05, |
|
"loss": 0.8357, |
|
"step": 13350 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.4754047017832327e-05, |
|
"loss": 0.8953, |
|
"step": 13400 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.4734472566406328e-05, |
|
"loss": 0.8378, |
|
"step": 13450 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.471489811498033e-05, |
|
"loss": 0.8787, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.4695323663554331e-05, |
|
"loss": 0.8001, |
|
"step": 13550 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.4675749212128332e-05, |
|
"loss": 0.9104, |
|
"step": 13600 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.4656174760702334e-05, |
|
"loss": 0.8789, |
|
"step": 13650 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.4636600309276333e-05, |
|
"loss": 0.8946, |
|
"step": 13700 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.4617025857850335e-05, |
|
"loss": 0.9366, |
|
"step": 13750 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.4597451406424336e-05, |
|
"loss": 0.9516, |
|
"step": 13800 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.4577876954998337e-05, |
|
"loss": 0.9444, |
|
"step": 13850 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.4558302503572339e-05, |
|
"loss": 0.8554, |
|
"step": 13900 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.453872805214634e-05, |
|
"loss": 0.9004, |
|
"step": 13950 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.4519153600720341e-05, |
|
"loss": 0.7618, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.4499579149294343e-05, |
|
"loss": 0.8806, |
|
"step": 14050 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.4480004697868344e-05, |
|
"loss": 0.901, |
|
"step": 14100 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.4460430246442345e-05, |
|
"loss": 0.8916, |
|
"step": 14150 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.4440855795016347e-05, |
|
"loss": 0.8715, |
|
"step": 14200 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.4421281343590348e-05, |
|
"loss": 0.8867, |
|
"step": 14250 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.440170689216435e-05, |
|
"loss": 0.8262, |
|
"step": 14300 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.438213244073835e-05, |
|
"loss": 0.8875, |
|
"step": 14350 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.436255798931235e-05, |
|
"loss": 0.8602, |
|
"step": 14400 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.4342983537886352e-05, |
|
"loss": 0.9087, |
|
"step": 14450 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.4323409086460353e-05, |
|
"loss": 0.8778, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.4303834635034355e-05, |
|
"loss": 0.8652, |
|
"step": 14550 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.4284260183608356e-05, |
|
"loss": 0.8563, |
|
"step": 14600 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.4264685732182357e-05, |
|
"loss": 0.8385, |
|
"step": 14650 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.4245111280756359e-05, |
|
"loss": 0.8856, |
|
"step": 14700 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.422553682933036e-05, |
|
"loss": 0.8771, |
|
"step": 14750 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.4205962377904361e-05, |
|
"loss": 0.8611, |
|
"step": 14800 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.4186387926478363e-05, |
|
"loss": 0.9059, |
|
"step": 14850 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.4166813475052364e-05, |
|
"loss": 0.9306, |
|
"step": 14900 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.4147239023626365e-05, |
|
"loss": 0.8578, |
|
"step": 14950 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.4127664572200367e-05, |
|
"loss": 0.8826, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.4108090120774366e-05, |
|
"loss": 0.9013, |
|
"step": 15050 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.4088515669348368e-05, |
|
"loss": 0.8291, |
|
"step": 15100 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.4068941217922369e-05, |
|
"loss": 0.9133, |
|
"step": 15150 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.404936676649637e-05, |
|
"loss": 0.7908, |
|
"step": 15200 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.4029792315070372e-05, |
|
"loss": 0.8334, |
|
"step": 15250 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.4010217863644373e-05, |
|
"loss": 0.8911, |
|
"step": 15300 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.3990643412218374e-05, |
|
"loss": 0.8854, |
|
"step": 15350 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.3971068960792376e-05, |
|
"loss": 0.9154, |
|
"step": 15400 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.3951494509366377e-05, |
|
"loss": 0.891, |
|
"step": 15450 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.3931920057940378e-05, |
|
"loss": 0.7739, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.391234560651438e-05, |
|
"loss": 0.8708, |
|
"step": 15550 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.3892771155088381e-05, |
|
"loss": 0.7749, |
|
"step": 15600 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.3873196703662382e-05, |
|
"loss": 0.9595, |
|
"step": 15650 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.3853622252236382e-05, |
|
"loss": 0.8669, |
|
"step": 15700 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.3834047800810383e-05, |
|
"loss": 0.9046, |
|
"step": 15750 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.3814473349384385e-05, |
|
"loss": 0.8351, |
|
"step": 15800 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.3794898897958386e-05, |
|
"loss": 0.9234, |
|
"step": 15850 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.3775324446532387e-05, |
|
"loss": 0.8685, |
|
"step": 15900 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.3755749995106389e-05, |
|
"loss": 0.8342, |
|
"step": 15950 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.373617554368039e-05, |
|
"loss": 0.8972, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.3716601092254391e-05, |
|
"loss": 0.8457, |
|
"step": 16050 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.3697026640828393e-05, |
|
"loss": 0.8288, |
|
"step": 16100 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.3677452189402394e-05, |
|
"loss": 0.966, |
|
"step": 16150 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.3657877737976395e-05, |
|
"loss": 0.9036, |
|
"step": 16200 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.3638303286550397e-05, |
|
"loss": 0.8774, |
|
"step": 16250 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.3618728835124398e-05, |
|
"loss": 0.9368, |
|
"step": 16300 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.35991543836984e-05, |
|
"loss": 0.8784, |
|
"step": 16350 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.3579579932272399e-05, |
|
"loss": 0.8421, |
|
"step": 16400 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.35600054808464e-05, |
|
"loss": 0.8555, |
|
"step": 16450 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.3540431029420402e-05, |
|
"loss": 0.8573, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.3520856577994403e-05, |
|
"loss": 0.8638, |
|
"step": 16550 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.3501282126568404e-05, |
|
"loss": 0.924, |
|
"step": 16600 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.3481707675142406e-05, |
|
"loss": 0.8371, |
|
"step": 16650 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.3462133223716407e-05, |
|
"loss": 0.8086, |
|
"step": 16700 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.3442558772290408e-05, |
|
"loss": 0.8677, |
|
"step": 16750 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.342298432086441e-05, |
|
"loss": 0.8288, |
|
"step": 16800 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.3403409869438411e-05, |
|
"loss": 0.7825, |
|
"step": 16850 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.3383835418012412e-05, |
|
"loss": 0.7759, |
|
"step": 16900 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.3364260966586414e-05, |
|
"loss": 0.9239, |
|
"step": 16950 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.3344686515160415e-05, |
|
"loss": 0.8655, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.3325112063734415e-05, |
|
"loss": 0.8756, |
|
"step": 17050 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.3305537612308416e-05, |
|
"loss": 0.8513, |
|
"step": 17100 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.3285963160882417e-05, |
|
"loss": 0.8829, |
|
"step": 17150 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.3266388709456419e-05, |
|
"loss": 0.7919, |
|
"step": 17200 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.324681425803042e-05, |
|
"loss": 0.7929, |
|
"step": 17250 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.3227239806604421e-05, |
|
"loss": 0.9112, |
|
"step": 17300 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.3207665355178423e-05, |
|
"loss": 0.8695, |
|
"step": 17350 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.3188090903752424e-05, |
|
"loss": 0.8948, |
|
"step": 17400 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.3168516452326425e-05, |
|
"loss": 0.8628, |
|
"step": 17450 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.3148942000900427e-05, |
|
"loss": 0.848, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.3129367549474428e-05, |
|
"loss": 0.9755, |
|
"step": 17550 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.310979309804843e-05, |
|
"loss": 0.9757, |
|
"step": 17600 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.3090218646622431e-05, |
|
"loss": 0.873, |
|
"step": 17650 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.307064419519643e-05, |
|
"loss": 0.8937, |
|
"step": 17700 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.3051069743770432e-05, |
|
"loss": 0.8326, |
|
"step": 17750 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.3031495292344433e-05, |
|
"loss": 0.8749, |
|
"step": 17800 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.3011920840918435e-05, |
|
"loss": 0.8842, |
|
"step": 17850 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.2992346389492436e-05, |
|
"loss": 0.8405, |
|
"step": 17900 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.2972771938066437e-05, |
|
"loss": 0.8236, |
|
"step": 17950 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.2953197486640439e-05, |
|
"loss": 0.9123, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.293362303521444e-05, |
|
"loss": 0.868, |
|
"step": 18050 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.2914048583788441e-05, |
|
"loss": 0.8031, |
|
"step": 18100 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.2894474132362443e-05, |
|
"loss": 0.855, |
|
"step": 18150 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.2874899680936444e-05, |
|
"loss": 0.8214, |
|
"step": 18200 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.2855325229510445e-05, |
|
"loss": 0.7972, |
|
"step": 18250 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.2835750778084447e-05, |
|
"loss": 0.9112, |
|
"step": 18300 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.2816176326658448e-05, |
|
"loss": 0.9455, |
|
"step": 18350 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.2796601875232448e-05, |
|
"loss": 0.8682, |
|
"step": 18400 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.2777027423806449e-05, |
|
"loss": 0.8356, |
|
"step": 18450 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.275745297238045e-05, |
|
"loss": 0.9152, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.2737878520954452e-05, |
|
"loss": 0.8404, |
|
"step": 18550 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.2718304069528453e-05, |
|
"loss": 0.8722, |
|
"step": 18600 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.2698729618102454e-05, |
|
"loss": 0.7857, |
|
"step": 18650 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.2679155166676456e-05, |
|
"loss": 0.9105, |
|
"step": 18700 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.2659580715250457e-05, |
|
"loss": 0.8208, |
|
"step": 18750 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.2640006263824458e-05, |
|
"loss": 0.8793, |
|
"step": 18800 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.262043181239846e-05, |
|
"loss": 0.8134, |
|
"step": 18850 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.2600857360972461e-05, |
|
"loss": 0.9198, |
|
"step": 18900 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.2581282909546462e-05, |
|
"loss": 0.8148, |
|
"step": 18950 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.2561708458120464e-05, |
|
"loss": 0.8317, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.2542134006694463e-05, |
|
"loss": 0.8664, |
|
"step": 19050 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.2522559555268465e-05, |
|
"loss": 0.9378, |
|
"step": 19100 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.2502985103842466e-05, |
|
"loss": 0.8184, |
|
"step": 19150 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.2483410652416467e-05, |
|
"loss": 0.8996, |
|
"step": 19200 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.2463836200990469e-05, |
|
"loss": 0.9632, |
|
"step": 19250 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.244426174956447e-05, |
|
"loss": 0.8904, |
|
"step": 19300 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.2424687298138471e-05, |
|
"loss": 0.8046, |
|
"step": 19350 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.2405112846712473e-05, |
|
"loss": 0.8562, |
|
"step": 19400 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.2385538395286474e-05, |
|
"loss": 0.8907, |
|
"step": 19450 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.2365963943860475e-05, |
|
"loss": 0.8435, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.2346389492434477e-05, |
|
"loss": 0.8589, |
|
"step": 19550 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.2326815041008478e-05, |
|
"loss": 0.9204, |
|
"step": 19600 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.230724058958248e-05, |
|
"loss": 0.886, |
|
"step": 19650 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.228766613815648e-05, |
|
"loss": 0.8723, |
|
"step": 19700 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.226809168673048e-05, |
|
"loss": 0.9037, |
|
"step": 19750 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.2248517235304482e-05, |
|
"loss": 0.8555, |
|
"step": 19800 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.2228942783878483e-05, |
|
"loss": 0.8601, |
|
"step": 19850 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.2209368332452484e-05, |
|
"loss": 0.818, |
|
"step": 19900 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.2189793881026486e-05, |
|
"loss": 0.8717, |
|
"step": 19950 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.2170219429600487e-05, |
|
"loss": 0.8883, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.2150644978174488e-05, |
|
"loss": 0.814, |
|
"step": 20050 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.213107052674849e-05, |
|
"loss": 0.9047, |
|
"step": 20100 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.2111496075322491e-05, |
|
"loss": 0.9227, |
|
"step": 20150 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.2091921623896492e-05, |
|
"loss": 0.971, |
|
"step": 20200 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.2072347172470494e-05, |
|
"loss": 0.8127, |
|
"step": 20250 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.2052772721044495e-05, |
|
"loss": 0.8432, |
|
"step": 20300 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.2033198269618496e-05, |
|
"loss": 0.8525, |
|
"step": 20350 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.2013623818192496e-05, |
|
"loss": 0.8678, |
|
"step": 20400 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.1994049366766497e-05, |
|
"loss": 0.8416, |
|
"step": 20450 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.1974474915340499e-05, |
|
"loss": 0.8313, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.19549004639145e-05, |
|
"loss": 0.8867, |
|
"step": 20550 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.1935326012488501e-05, |
|
"loss": 0.8573, |
|
"step": 20600 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.1915751561062503e-05, |
|
"loss": 0.8349, |
|
"step": 20650 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.1896177109636504e-05, |
|
"loss": 0.8767, |
|
"step": 20700 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.1876602658210506e-05, |
|
"loss": 0.9189, |
|
"step": 20750 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.1857028206784507e-05, |
|
"loss": 0.8946, |
|
"step": 20800 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.1837453755358508e-05, |
|
"loss": 0.7432, |
|
"step": 20850 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.181787930393251e-05, |
|
"loss": 0.9358, |
|
"step": 20900 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.1798304852506511e-05, |
|
"loss": 0.8968, |
|
"step": 20950 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.1778730401080512e-05, |
|
"loss": 0.7498, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.1759155949654512e-05, |
|
"loss": 0.8436, |
|
"step": 21050 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.1739581498228513e-05, |
|
"loss": 0.8336, |
|
"step": 21100 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.1720007046802515e-05, |
|
"loss": 0.8154, |
|
"step": 21150 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.1700432595376516e-05, |
|
"loss": 0.8352, |
|
"step": 21200 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.1680858143950517e-05, |
|
"loss": 0.84, |
|
"step": 21250 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.1661283692524519e-05, |
|
"loss": 0.9231, |
|
"step": 21300 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.164170924109852e-05, |
|
"loss": 0.8819, |
|
"step": 21350 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.1622134789672521e-05, |
|
"loss": 0.8829, |
|
"step": 21400 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.1602560338246523e-05, |
|
"loss": 0.9224, |
|
"step": 21450 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.1582985886820524e-05, |
|
"loss": 0.8652, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.1563411435394525e-05, |
|
"loss": 0.8843, |
|
"step": 21550 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.1543836983968527e-05, |
|
"loss": 0.8513, |
|
"step": 21600 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.1524262532542528e-05, |
|
"loss": 0.9332, |
|
"step": 21650 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.150468808111653e-05, |
|
"loss": 0.8782, |
|
"step": 21700 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.1485113629690529e-05, |
|
"loss": 0.8213, |
|
"step": 21750 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.146553917826453e-05, |
|
"loss": 0.9054, |
|
"step": 21800 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.1445964726838532e-05, |
|
"loss": 0.913, |
|
"step": 21850 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.1426390275412533e-05, |
|
"loss": 0.9431, |
|
"step": 21900 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.1406815823986534e-05, |
|
"loss": 0.889, |
|
"step": 21950 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.1387241372560536e-05, |
|
"loss": 0.8428, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.1367666921134537e-05, |
|
"loss": 0.8754, |
|
"step": 22050 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.1348092469708538e-05, |
|
"loss": 0.8986, |
|
"step": 22100 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.132851801828254e-05, |
|
"loss": 0.7835, |
|
"step": 22150 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.1308943566856541e-05, |
|
"loss": 0.9178, |
|
"step": 22200 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.1289369115430542e-05, |
|
"loss": 0.8806, |
|
"step": 22250 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.1269794664004544e-05, |
|
"loss": 0.8896, |
|
"step": 22300 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.1250220212578545e-05, |
|
"loss": 0.8652, |
|
"step": 22350 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.1230645761152545e-05, |
|
"loss": 0.959, |
|
"step": 22400 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.1211071309726546e-05, |
|
"loss": 0.9101, |
|
"step": 22450 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.1191496858300547e-05, |
|
"loss": 0.9111, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.1171922406874549e-05, |
|
"loss": 0.8679, |
|
"step": 22550 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.115234795544855e-05, |
|
"loss": 0.9192, |
|
"step": 22600 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.1132773504022551e-05, |
|
"loss": 0.869, |
|
"step": 22650 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.1113199052596553e-05, |
|
"loss": 0.8669, |
|
"step": 22700 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.1093624601170554e-05, |
|
"loss": 0.7818, |
|
"step": 22750 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.1074050149744555e-05, |
|
"loss": 0.8594, |
|
"step": 22800 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.1054475698318557e-05, |
|
"loss": 0.8431, |
|
"step": 22850 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.1034901246892558e-05, |
|
"loss": 0.8806, |
|
"step": 22900 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.101532679546656e-05, |
|
"loss": 0.9083, |
|
"step": 22950 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.099575234404056e-05, |
|
"loss": 0.8489, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.097617789261456e-05, |
|
"loss": 0.9215, |
|
"step": 23050 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.0956603441188562e-05, |
|
"loss": 0.8501, |
|
"step": 23100 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.0937028989762563e-05, |
|
"loss": 0.8669, |
|
"step": 23150 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.0917454538336564e-05, |
|
"loss": 0.8635, |
|
"step": 23200 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.0897880086910566e-05, |
|
"loss": 0.8314, |
|
"step": 23250 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.0878305635484567e-05, |
|
"loss": 0.9476, |
|
"step": 23300 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.0858731184058568e-05, |
|
"loss": 0.9073, |
|
"step": 23350 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.083915673263257e-05, |
|
"loss": 0.8649, |
|
"step": 23400 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.0819582281206571e-05, |
|
"loss": 0.8369, |
|
"step": 23450 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.0800007829780572e-05, |
|
"loss": 0.8857, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.0780433378354574e-05, |
|
"loss": 0.7556, |
|
"step": 23550 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.0760858926928575e-05, |
|
"loss": 0.8351, |
|
"step": 23600 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.0741284475502576e-05, |
|
"loss": 0.8619, |
|
"step": 23650 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.0721710024076578e-05, |
|
"loss": 0.8475, |
|
"step": 23700 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.0702135572650577e-05, |
|
"loss": 0.8963, |
|
"step": 23750 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.0682561121224579e-05, |
|
"loss": 0.8072, |
|
"step": 23800 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.066298666979858e-05, |
|
"loss": 0.7855, |
|
"step": 23850 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.0643412218372582e-05, |
|
"loss": 0.9233, |
|
"step": 23900 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.0623837766946583e-05, |
|
"loss": 0.8856, |
|
"step": 23950 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.0604263315520584e-05, |
|
"loss": 0.8831, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.0584688864094586e-05, |
|
"loss": 0.7711, |
|
"step": 24050 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.0565114412668587e-05, |
|
"loss": 0.8804, |
|
"step": 24100 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.0545539961242588e-05, |
|
"loss": 0.8166, |
|
"step": 24150 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.052596550981659e-05, |
|
"loss": 0.88, |
|
"step": 24200 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.0506391058390591e-05, |
|
"loss": 0.8698, |
|
"step": 24250 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.0486816606964592e-05, |
|
"loss": 0.8684, |
|
"step": 24300 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.0467242155538594e-05, |
|
"loss": 0.93, |
|
"step": 24350 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.0447667704112593e-05, |
|
"loss": 0.8597, |
|
"step": 24400 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.0428093252686595e-05, |
|
"loss": 0.9301, |
|
"step": 24450 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.0408518801260596e-05, |
|
"loss": 0.8598, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.0388944349834597e-05, |
|
"loss": 0.9186, |
|
"step": 24550 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.0369369898408599e-05, |
|
"loss": 0.8626, |
|
"step": 24600 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.03497954469826e-05, |
|
"loss": 0.8402, |
|
"step": 24650 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.0330220995556601e-05, |
|
"loss": 0.8344, |
|
"step": 24700 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.0310646544130603e-05, |
|
"loss": 0.7338, |
|
"step": 24750 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.0291072092704604e-05, |
|
"loss": 0.924, |
|
"step": 24800 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.0271497641278605e-05, |
|
"loss": 0.8656, |
|
"step": 24850 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.0251923189852607e-05, |
|
"loss": 0.8263, |
|
"step": 24900 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.0232348738426608e-05, |
|
"loss": 0.8556, |
|
"step": 24950 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.021277428700061e-05, |
|
"loss": 0.8331, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.0193199835574609e-05, |
|
"loss": 0.7991, |
|
"step": 25050 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.017362538414861e-05, |
|
"loss": 0.7974, |
|
"step": 25100 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.0154050932722612e-05, |
|
"loss": 0.8658, |
|
"step": 25150 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.0134476481296613e-05, |
|
"loss": 0.8685, |
|
"step": 25200 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.0114902029870614e-05, |
|
"loss": 0.9491, |
|
"step": 25250 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 1.0095327578444616e-05, |
|
"loss": 0.8939, |
|
"step": 25300 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 1.0075753127018617e-05, |
|
"loss": 0.8762, |
|
"step": 25350 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 1.0056178675592618e-05, |
|
"loss": 0.8753, |
|
"step": 25400 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 1.003660422416662e-05, |
|
"loss": 0.7322, |
|
"step": 25450 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 1.0017029772740621e-05, |
|
"loss": 0.9321, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 9.99745532131462e-06, |
|
"loss": 0.8506, |
|
"step": 25550 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 9.977880869888622e-06, |
|
"loss": 0.8747, |
|
"step": 25600 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 9.958306418462623e-06, |
|
"loss": 0.8376, |
|
"step": 25650 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 9.938731967036625e-06, |
|
"loss": 0.8633, |
|
"step": 25700 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 9.919157515610626e-06, |
|
"loss": 0.843, |
|
"step": 25750 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 9.899583064184627e-06, |
|
"loss": 0.9029, |
|
"step": 25800 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 9.880008612758629e-06, |
|
"loss": 0.9153, |
|
"step": 25850 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 9.86043416133263e-06, |
|
"loss": 0.8513, |
|
"step": 25900 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 9.84085970990663e-06, |
|
"loss": 0.8539, |
|
"step": 25950 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 9.821285258480631e-06, |
|
"loss": 0.8501, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 9.801710807054632e-06, |
|
"loss": 0.931, |
|
"step": 26050 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 9.782136355628634e-06, |
|
"loss": 0.8488, |
|
"step": 26100 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 9.762561904202635e-06, |
|
"loss": 0.9098, |
|
"step": 26150 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 9.742987452776636e-06, |
|
"loss": 0.8358, |
|
"step": 26200 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 9.723413001350638e-06, |
|
"loss": 0.8837, |
|
"step": 26250 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 9.703838549924639e-06, |
|
"loss": 0.7796, |
|
"step": 26300 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.68426409849864e-06, |
|
"loss": 0.8227, |
|
"step": 26350 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.664689647072642e-06, |
|
"loss": 0.8286, |
|
"step": 26400 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.645115195646643e-06, |
|
"loss": 0.9235, |
|
"step": 26450 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.625540744220644e-06, |
|
"loss": 0.845, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.605966292794646e-06, |
|
"loss": 0.7694, |
|
"step": 26550 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.586391841368647e-06, |
|
"loss": 0.9169, |
|
"step": 26600 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.566817389942647e-06, |
|
"loss": 0.9425, |
|
"step": 26650 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.547242938516648e-06, |
|
"loss": 0.7728, |
|
"step": 26700 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.52766848709065e-06, |
|
"loss": 0.9274, |
|
"step": 26750 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.50809403566465e-06, |
|
"loss": 0.8754, |
|
"step": 26800 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.488519584238652e-06, |
|
"loss": 0.8206, |
|
"step": 26850 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.468945132812653e-06, |
|
"loss": 0.8349, |
|
"step": 26900 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.449370681386655e-06, |
|
"loss": 0.7908, |
|
"step": 26950 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.429796229960656e-06, |
|
"loss": 0.776, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.410221778534658e-06, |
|
"loss": 0.8717, |
|
"step": 27050 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.390647327108659e-06, |
|
"loss": 0.8673, |
|
"step": 27100 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.37107287568266e-06, |
|
"loss": 0.8936, |
|
"step": 27150 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.351498424256662e-06, |
|
"loss": 0.8782, |
|
"step": 27200 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.331923972830663e-06, |
|
"loss": 0.7767, |
|
"step": 27250 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.312349521404663e-06, |
|
"loss": 0.8948, |
|
"step": 27300 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.292775069978664e-06, |
|
"loss": 0.8819, |
|
"step": 27350 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.273200618552665e-06, |
|
"loss": 0.8244, |
|
"step": 27400 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.253626167126667e-06, |
|
"loss": 0.8453, |
|
"step": 27450 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.234051715700668e-06, |
|
"loss": 0.8921, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.21447726427467e-06, |
|
"loss": 0.7972, |
|
"step": 27550 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.19490281284867e-06, |
|
"loss": 0.8651, |
|
"step": 27600 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.175328361422672e-06, |
|
"loss": 0.9117, |
|
"step": 27650 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.155753909996673e-06, |
|
"loss": 0.8853, |
|
"step": 27700 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.136179458570675e-06, |
|
"loss": 0.8501, |
|
"step": 27750 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.116605007144676e-06, |
|
"loss": 0.8758, |
|
"step": 27800 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 9.097030555718677e-06, |
|
"loss": 0.846, |
|
"step": 27850 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 9.077456104292679e-06, |
|
"loss": 0.9889, |
|
"step": 27900 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 9.057881652866678e-06, |
|
"loss": 0.8388, |
|
"step": 27950 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 9.03830720144068e-06, |
|
"loss": 0.8852, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 9.018732750014681e-06, |
|
"loss": 0.8383, |
|
"step": 28050 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 8.999158298588682e-06, |
|
"loss": 0.8078, |
|
"step": 28100 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 8.979583847162684e-06, |
|
"loss": 0.8822, |
|
"step": 28150 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 8.960009395736685e-06, |
|
"loss": 0.8824, |
|
"step": 28200 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 8.940434944310686e-06, |
|
"loss": 0.8647, |
|
"step": 28250 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 8.920860492884688e-06, |
|
"loss": 0.8348, |
|
"step": 28300 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 8.901286041458689e-06, |
|
"loss": 0.7925, |
|
"step": 28350 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 8.88171159003269e-06, |
|
"loss": 0.9533, |
|
"step": 28400 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 8.862137138606692e-06, |
|
"loss": 0.8684, |
|
"step": 28450 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 8.842562687180693e-06, |
|
"loss": 0.8617, |
|
"step": 28500 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 8.822988235754694e-06, |
|
"loss": 0.8674, |
|
"step": 28550 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 8.803413784328696e-06, |
|
"loss": 0.8993, |
|
"step": 28600 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 8.783839332902695e-06, |
|
"loss": 0.8651, |
|
"step": 28650 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 8.764264881476697e-06, |
|
"loss": 0.9059, |
|
"step": 28700 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 8.744690430050698e-06, |
|
"loss": 0.8669, |
|
"step": 28750 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 8.7251159786247e-06, |
|
"loss": 0.8995, |
|
"step": 28800 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 8.7055415271987e-06, |
|
"loss": 0.7663, |
|
"step": 28850 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 8.685967075772702e-06, |
|
"loss": 0.9201, |
|
"step": 28900 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 8.666392624346703e-06, |
|
"loss": 0.8251, |
|
"step": 28950 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 8.646818172920705e-06, |
|
"loss": 0.8798, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 8.627243721494706e-06, |
|
"loss": 0.8492, |
|
"step": 29050 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 8.607669270068707e-06, |
|
"loss": 0.8379, |
|
"step": 29100 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 8.588094818642709e-06, |
|
"loss": 0.8279, |
|
"step": 29150 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 8.56852036721671e-06, |
|
"loss": 0.9321, |
|
"step": 29200 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 8.548945915790711e-06, |
|
"loss": 0.8259, |
|
"step": 29250 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 8.529371464364711e-06, |
|
"loss": 0.945, |
|
"step": 29300 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 8.509797012938712e-06, |
|
"loss": 0.8367, |
|
"step": 29350 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 8.490222561512714e-06, |
|
"loss": 0.7596, |
|
"step": 29400 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 8.470648110086715e-06, |
|
"loss": 0.9031, |
|
"step": 29450 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 8.451073658660716e-06, |
|
"loss": 0.9436, |
|
"step": 29500 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 8.431499207234718e-06, |
|
"loss": 0.8308, |
|
"step": 29550 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 8.411924755808719e-06, |
|
"loss": 0.7965, |
|
"step": 29600 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 8.39235030438272e-06, |
|
"loss": 0.8116, |
|
"step": 29650 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 8.372775852956722e-06, |
|
"loss": 0.894, |
|
"step": 29700 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 8.353201401530723e-06, |
|
"loss": 0.8788, |
|
"step": 29750 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 8.333626950104724e-06, |
|
"loss": 0.8463, |
|
"step": 29800 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 8.314052498678726e-06, |
|
"loss": 0.7681, |
|
"step": 29850 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 8.294478047252727e-06, |
|
"loss": 0.8351, |
|
"step": 29900 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 8.274903595826728e-06, |
|
"loss": 0.8549, |
|
"step": 29950 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 8.255329144400728e-06, |
|
"loss": 0.9201, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 8.23575469297473e-06, |
|
"loss": 0.8351, |
|
"step": 30050 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 8.21618024154873e-06, |
|
"loss": 0.8675, |
|
"step": 30100 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 8.196605790122732e-06, |
|
"loss": 0.7949, |
|
"step": 30150 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 8.177031338696733e-06, |
|
"loss": 0.8228, |
|
"step": 30200 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 8.157456887270735e-06, |
|
"loss": 0.861, |
|
"step": 30250 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 8.137882435844736e-06, |
|
"loss": 0.9081, |
|
"step": 30300 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 8.118307984418738e-06, |
|
"loss": 0.7959, |
|
"step": 30350 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 8.098733532992739e-06, |
|
"loss": 0.8647, |
|
"step": 30400 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 8.07915908156674e-06, |
|
"loss": 0.8884, |
|
"step": 30450 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 8.059584630140742e-06, |
|
"loss": 0.7761, |
|
"step": 30500 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 8.040010178714743e-06, |
|
"loss": 0.9423, |
|
"step": 30550 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 8.020435727288744e-06, |
|
"loss": 0.8885, |
|
"step": 30600 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 8.000861275862744e-06, |
|
"loss": 0.9188, |
|
"step": 30650 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 7.981286824436745e-06, |
|
"loss": 0.7685, |
|
"step": 30700 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 7.961712373010747e-06, |
|
"loss": 0.8876, |
|
"step": 30750 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 7.942137921584748e-06, |
|
"loss": 0.8314, |
|
"step": 30800 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 7.92256347015875e-06, |
|
"loss": 0.963, |
|
"step": 30850 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 7.90298901873275e-06, |
|
"loss": 0.8584, |
|
"step": 30900 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 7.883414567306752e-06, |
|
"loss": 0.8097, |
|
"step": 30950 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 7.863840115880753e-06, |
|
"loss": 0.868, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 7.844265664454755e-06, |
|
"loss": 0.8494, |
|
"step": 31050 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 7.824691213028756e-06, |
|
"loss": 0.7916, |
|
"step": 31100 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 7.805116761602757e-06, |
|
"loss": 0.9308, |
|
"step": 31150 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 7.785542310176759e-06, |
|
"loss": 0.7219, |
|
"step": 31200 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 7.76596785875076e-06, |
|
"loss": 0.8988, |
|
"step": 31250 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 7.74639340732476e-06, |
|
"loss": 0.7423, |
|
"step": 31300 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 7.726818955898761e-06, |
|
"loss": 0.7179, |
|
"step": 31350 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 7.707244504472762e-06, |
|
"loss": 0.9416, |
|
"step": 31400 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 7.687670053046764e-06, |
|
"loss": 0.8769, |
|
"step": 31450 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 7.668095601620765e-06, |
|
"loss": 0.992, |
|
"step": 31500 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 7.648521150194766e-06, |
|
"loss": 0.8055, |
|
"step": 31550 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 7.628946698768768e-06, |
|
"loss": 0.8286, |
|
"step": 31600 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 7.609372247342769e-06, |
|
"loss": 0.7925, |
|
"step": 31650 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 7.58979779591677e-06, |
|
"loss": 0.812, |
|
"step": 31700 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 7.570223344490772e-06, |
|
"loss": 0.8058, |
|
"step": 31750 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 7.550648893064772e-06, |
|
"loss": 0.8939, |
|
"step": 31800 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 7.5310744416387735e-06, |
|
"loss": 0.9144, |
|
"step": 31850 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 7.511499990212775e-06, |
|
"loss": 0.856, |
|
"step": 31900 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 7.491925538786776e-06, |
|
"loss": 0.8356, |
|
"step": 31950 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 7.4723510873607775e-06, |
|
"loss": 0.8551, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 7.452776635934779e-06, |
|
"loss": 0.8674, |
|
"step": 32050 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 7.43320218450878e-06, |
|
"loss": 0.8289, |
|
"step": 32100 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 7.413627733082781e-06, |
|
"loss": 0.9137, |
|
"step": 32150 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 7.394053281656782e-06, |
|
"loss": 0.8694, |
|
"step": 32200 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 7.374478830230783e-06, |
|
"loss": 0.8175, |
|
"step": 32250 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 7.354904378804785e-06, |
|
"loss": 0.8521, |
|
"step": 32300 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 7.335329927378786e-06, |
|
"loss": 0.8094, |
|
"step": 32350 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 7.315755475952787e-06, |
|
"loss": 0.8094, |
|
"step": 32400 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 7.296181024526789e-06, |
|
"loss": 0.8146, |
|
"step": 32450 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 7.276606573100789e-06, |
|
"loss": 0.8444, |
|
"step": 32500 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 7.257032121674791e-06, |
|
"loss": 0.8016, |
|
"step": 32550 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 7.237457670248792e-06, |
|
"loss": 0.8408, |
|
"step": 32600 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 7.217883218822793e-06, |
|
"loss": 0.8744, |
|
"step": 32650 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 7.198308767396795e-06, |
|
"loss": 0.8412, |
|
"step": 32700 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 7.178734315970796e-06, |
|
"loss": 0.8446, |
|
"step": 32750 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 7.1591598645447964e-06, |
|
"loss": 0.8469, |
|
"step": 32800 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 7.139585413118798e-06, |
|
"loss": 0.7875, |
|
"step": 32850 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 7.120010961692799e-06, |
|
"loss": 0.8713, |
|
"step": 32900 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 7.1004365102668005e-06, |
|
"loss": 0.8614, |
|
"step": 32950 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 7.080862058840802e-06, |
|
"loss": 0.8067, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 7.061287607414803e-06, |
|
"loss": 0.8323, |
|
"step": 33050 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 7.0417131559888045e-06, |
|
"loss": 0.8127, |
|
"step": 33100 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 7.022138704562805e-06, |
|
"loss": 0.9393, |
|
"step": 33150 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 7.002564253136806e-06, |
|
"loss": 0.8109, |
|
"step": 33200 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 6.982989801710808e-06, |
|
"loss": 0.871, |
|
"step": 33250 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 6.963415350284809e-06, |
|
"loss": 0.7685, |
|
"step": 33300 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 6.94384089885881e-06, |
|
"loss": 0.8988, |
|
"step": 33350 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 6.924266447432812e-06, |
|
"loss": 0.8079, |
|
"step": 33400 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 6.904691996006813e-06, |
|
"loss": 0.8664, |
|
"step": 33450 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 6.8851175445808135e-06, |
|
"loss": 0.8659, |
|
"step": 33500 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 6.865543093154815e-06, |
|
"loss": 0.8087, |
|
"step": 33550 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 6.845968641728816e-06, |
|
"loss": 0.9701, |
|
"step": 33600 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 6.8263941903028175e-06, |
|
"loss": 0.8431, |
|
"step": 33650 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 6.806819738876819e-06, |
|
"loss": 0.8664, |
|
"step": 33700 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 6.78724528745082e-06, |
|
"loss": 0.8538, |
|
"step": 33750 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 6.767670836024821e-06, |
|
"loss": 0.9135, |
|
"step": 33800 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 6.748096384598822e-06, |
|
"loss": 0.8, |
|
"step": 33850 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 6.728521933172823e-06, |
|
"loss": 0.8769, |
|
"step": 33900 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 6.708947481746825e-06, |
|
"loss": 0.8763, |
|
"step": 33950 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 6.689373030320826e-06, |
|
"loss": 0.8174, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 6.669798578894827e-06, |
|
"loss": 0.7703, |
|
"step": 34050 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 6.650224127468829e-06, |
|
"loss": 0.9214, |
|
"step": 34100 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 6.630649676042829e-06, |
|
"loss": 0.8517, |
|
"step": 34150 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 6.611075224616831e-06, |
|
"loss": 0.8129, |
|
"step": 34200 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 6.591500773190832e-06, |
|
"loss": 0.8221, |
|
"step": 34250 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 6.571926321764833e-06, |
|
"loss": 0.8089, |
|
"step": 34300 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 6.552351870338835e-06, |
|
"loss": 0.9105, |
|
"step": 34350 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 6.532777418912836e-06, |
|
"loss": 0.7871, |
|
"step": 34400 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 6.513202967486837e-06, |
|
"loss": 0.7993, |
|
"step": 34450 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 6.493628516060838e-06, |
|
"loss": 0.7592, |
|
"step": 34500 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 6.474054064634839e-06, |
|
"loss": 0.8226, |
|
"step": 34550 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 6.4544796132088405e-06, |
|
"loss": 0.8362, |
|
"step": 34600 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 6.434905161782842e-06, |
|
"loss": 0.8218, |
|
"step": 34650 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 6.415330710356843e-06, |
|
"loss": 0.7943, |
|
"step": 34700 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 6.3957562589308445e-06, |
|
"loss": 0.9096, |
|
"step": 34750 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 6.376181807504845e-06, |
|
"loss": 0.8132, |
|
"step": 34800 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 6.356607356078846e-06, |
|
"loss": 0.8449, |
|
"step": 34850 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 6.337032904652848e-06, |
|
"loss": 0.8221, |
|
"step": 34900 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 6.317458453226849e-06, |
|
"loss": 0.8582, |
|
"step": 34950 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 6.29788400180085e-06, |
|
"loss": 0.869, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 6.278309550374852e-06, |
|
"loss": 0.8262, |
|
"step": 35050 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 6.258735098948853e-06, |
|
"loss": 0.8728, |
|
"step": 35100 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 6.2391606475228535e-06, |
|
"loss": 0.9178, |
|
"step": 35150 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 6.219586196096855e-06, |
|
"loss": 0.7946, |
|
"step": 35200 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 6.200011744670856e-06, |
|
"loss": 0.8229, |
|
"step": 35250 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 6.1804372932448575e-06, |
|
"loss": 0.8999, |
|
"step": 35300 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 6.160862841818859e-06, |
|
"loss": 0.9206, |
|
"step": 35350 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 6.14128839039286e-06, |
|
"loss": 0.8875, |
|
"step": 35400 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 6.1217139389668616e-06, |
|
"loss": 0.8835, |
|
"step": 35450 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 6.102139487540862e-06, |
|
"loss": 0.8897, |
|
"step": 35500 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 6.082565036114863e-06, |
|
"loss": 0.9601, |
|
"step": 35550 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 6.062990584688865e-06, |
|
"loss": 0.7805, |
|
"step": 35600 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 6.043416133262866e-06, |
|
"loss": 0.9185, |
|
"step": 35650 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 6.0238416818368674e-06, |
|
"loss": 0.7948, |
|
"step": 35700 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 6.004267230410869e-06, |
|
"loss": 0.8508, |
|
"step": 35750 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 5.984692778984869e-06, |
|
"loss": 0.8683, |
|
"step": 35800 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 5.965118327558871e-06, |
|
"loss": 0.889, |
|
"step": 35850 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 5.945543876132872e-06, |
|
"loss": 0.8705, |
|
"step": 35900 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 5.925969424706873e-06, |
|
"loss": 0.8509, |
|
"step": 35950 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 5.906394973280875e-06, |
|
"loss": 0.8239, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 5.886820521854876e-06, |
|
"loss": 0.817, |
|
"step": 36050 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 5.867246070428877e-06, |
|
"loss": 0.8169, |
|
"step": 36100 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 5.847671619002878e-06, |
|
"loss": 0.8845, |
|
"step": 36150 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 5.828097167576879e-06, |
|
"loss": 0.7943, |
|
"step": 36200 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 5.8085227161508805e-06, |
|
"loss": 0.8947, |
|
"step": 36250 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 5.788948264724882e-06, |
|
"loss": 0.9, |
|
"step": 36300 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 5.769373813298883e-06, |
|
"loss": 0.8847, |
|
"step": 36350 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 5.7497993618728845e-06, |
|
"loss": 0.868, |
|
"step": 36400 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 5.730224910446886e-06, |
|
"loss": 0.7603, |
|
"step": 36450 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 5.710650459020886e-06, |
|
"loss": 0.8632, |
|
"step": 36500 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 5.691076007594888e-06, |
|
"loss": 0.8426, |
|
"step": 36550 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 5.671501556168889e-06, |
|
"loss": 0.939, |
|
"step": 36600 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 5.65192710474289e-06, |
|
"loss": 0.8623, |
|
"step": 36650 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 5.632352653316892e-06, |
|
"loss": 0.8796, |
|
"step": 36700 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 5.612778201890893e-06, |
|
"loss": 0.9162, |
|
"step": 36750 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 5.5932037504648935e-06, |
|
"loss": 0.7553, |
|
"step": 36800 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 5.573629299038895e-06, |
|
"loss": 0.8613, |
|
"step": 36850 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 5.554054847612896e-06, |
|
"loss": 0.9001, |
|
"step": 36900 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 5.5344803961868976e-06, |
|
"loss": 0.9094, |
|
"step": 36950 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 5.514905944760899e-06, |
|
"loss": 0.8644, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 5.4953314933349e-06, |
|
"loss": 0.8831, |
|
"step": 37050 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 5.4757570419089016e-06, |
|
"loss": 0.8597, |
|
"step": 37100 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 5.456182590482902e-06, |
|
"loss": 0.7675, |
|
"step": 37150 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 5.436608139056903e-06, |
|
"loss": 0.8921, |
|
"step": 37200 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 5.417033687630905e-06, |
|
"loss": 0.8159, |
|
"step": 37250 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 5.397459236204906e-06, |
|
"loss": 0.8174, |
|
"step": 37300 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 5.3778847847789074e-06, |
|
"loss": 0.8421, |
|
"step": 37350 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 5.358310333352909e-06, |
|
"loss": 0.8831, |
|
"step": 37400 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 5.33873588192691e-06, |
|
"loss": 0.9144, |
|
"step": 37450 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 5.319161430500911e-06, |
|
"loss": 0.7922, |
|
"step": 37500 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 5.299586979074912e-06, |
|
"loss": 0.8456, |
|
"step": 37550 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 5.280012527648913e-06, |
|
"loss": 0.8568, |
|
"step": 37600 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 5.260438076222915e-06, |
|
"loss": 0.8747, |
|
"step": 37650 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 5.240863624796916e-06, |
|
"loss": 0.8253, |
|
"step": 37700 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 5.221289173370917e-06, |
|
"loss": 0.9293, |
|
"step": 37750 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 5.201714721944918e-06, |
|
"loss": 0.7685, |
|
"step": 37800 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 5.182140270518919e-06, |
|
"loss": 0.8248, |
|
"step": 37850 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 5.1625658190929205e-06, |
|
"loss": 0.828, |
|
"step": 37900 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 5.142991367666922e-06, |
|
"loss": 0.8755, |
|
"step": 37950 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 5.123416916240923e-06, |
|
"loss": 0.7804, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 5.1038424648149245e-06, |
|
"loss": 0.8399, |
|
"step": 38050 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 5.084268013388926e-06, |
|
"loss": 0.936, |
|
"step": 38100 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 5.064693561962926e-06, |
|
"loss": 0.8735, |
|
"step": 38150 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 5.045119110536928e-06, |
|
"loss": 0.7677, |
|
"step": 38200 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 5.025544659110929e-06, |
|
"loss": 0.8416, |
|
"step": 38250 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 5.00597020768493e-06, |
|
"loss": 0.7239, |
|
"step": 38300 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 4.986395756258931e-06, |
|
"loss": 0.8729, |
|
"step": 38350 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 4.966821304832932e-06, |
|
"loss": 0.8779, |
|
"step": 38400 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 4.9472468534069335e-06, |
|
"loss": 0.9235, |
|
"step": 38450 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 4.927672401980935e-06, |
|
"loss": 0.8315, |
|
"step": 38500 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 4.908097950554936e-06, |
|
"loss": 0.7789, |
|
"step": 38550 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 4.888523499128937e-06, |
|
"loss": 0.8934, |
|
"step": 38600 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 4.868949047702938e-06, |
|
"loss": 0.8634, |
|
"step": 38650 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 4.849374596276939e-06, |
|
"loss": 0.7719, |
|
"step": 38700 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 4.829800144850941e-06, |
|
"loss": 0.8202, |
|
"step": 38750 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 4.810225693424942e-06, |
|
"loss": 0.8192, |
|
"step": 38800 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 4.790651241998943e-06, |
|
"loss": 0.8286, |
|
"step": 38850 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 4.771076790572945e-06, |
|
"loss": 0.8023, |
|
"step": 38900 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 4.751502339146945e-06, |
|
"loss": 0.8284, |
|
"step": 38950 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 4.731927887720947e-06, |
|
"loss": 0.7514, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 4.712353436294948e-06, |
|
"loss": 0.8086, |
|
"step": 39050 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 4.692778984868949e-06, |
|
"loss": 0.8498, |
|
"step": 39100 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 4.673204533442951e-06, |
|
"loss": 0.8595, |
|
"step": 39150 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 4.653630082016952e-06, |
|
"loss": 0.7855, |
|
"step": 39200 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 4.6340556305909524e-06, |
|
"loss": 0.85, |
|
"step": 39250 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 4.614481179164954e-06, |
|
"loss": 0.9688, |
|
"step": 39300 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 4.594906727738955e-06, |
|
"loss": 0.7548, |
|
"step": 39350 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 4.5753322763129565e-06, |
|
"loss": 0.8934, |
|
"step": 39400 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 4.555757824886958e-06, |
|
"loss": 0.8086, |
|
"step": 39450 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 4.536183373460959e-06, |
|
"loss": 0.8638, |
|
"step": 39500 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 4.5166089220349605e-06, |
|
"loss": 0.8957, |
|
"step": 39550 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 4.497034470608961e-06, |
|
"loss": 0.935, |
|
"step": 39600 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 4.477460019182962e-06, |
|
"loss": 0.8048, |
|
"step": 39650 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 4.457885567756964e-06, |
|
"loss": 0.8327, |
|
"step": 39700 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 4.438311116330965e-06, |
|
"loss": 0.7151, |
|
"step": 39750 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 4.418736664904966e-06, |
|
"loss": 0.8875, |
|
"step": 39800 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 4.399162213478968e-06, |
|
"loss": 0.8725, |
|
"step": 39850 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 4.379587762052969e-06, |
|
"loss": 0.9833, |
|
"step": 39900 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 4.3600133106269695e-06, |
|
"loss": 0.8513, |
|
"step": 39950 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 4.340438859200971e-06, |
|
"loss": 0.8536, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 4.320864407774972e-06, |
|
"loss": 0.8258, |
|
"step": 40050 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 4.3012899563489735e-06, |
|
"loss": 0.822, |
|
"step": 40100 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 4.281715504922975e-06, |
|
"loss": 0.9229, |
|
"step": 40150 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 4.262141053496976e-06, |
|
"loss": 0.8064, |
|
"step": 40200 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 4.2425666020709776e-06, |
|
"loss": 0.8638, |
|
"step": 40250 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 4.222992150644978e-06, |
|
"loss": 0.8485, |
|
"step": 40300 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 4.203417699218979e-06, |
|
"loss": 0.874, |
|
"step": 40350 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 4.183843247792981e-06, |
|
"loss": 0.8052, |
|
"step": 40400 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 4.164268796366982e-06, |
|
"loss": 0.8778, |
|
"step": 40450 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 4.1446943449409834e-06, |
|
"loss": 0.8941, |
|
"step": 40500 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 4.125119893514985e-06, |
|
"loss": 0.8616, |
|
"step": 40550 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 4.105545442088985e-06, |
|
"loss": 0.8288, |
|
"step": 40600 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 4.085970990662987e-06, |
|
"loss": 0.8596, |
|
"step": 40650 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 4.066396539236988e-06, |
|
"loss": 0.7813, |
|
"step": 40700 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 4.046822087810989e-06, |
|
"loss": 0.8639, |
|
"step": 40750 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 4.027247636384991e-06, |
|
"loss": 0.7813, |
|
"step": 40800 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 4.007673184958992e-06, |
|
"loss": 0.8996, |
|
"step": 40850 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 3.988098733532993e-06, |
|
"loss": 0.8716, |
|
"step": 40900 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 3.968524282106994e-06, |
|
"loss": 0.7635, |
|
"step": 40950 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 3.948949830680995e-06, |
|
"loss": 0.9362, |
|
"step": 41000 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 3.9293753792549965e-06, |
|
"loss": 0.8802, |
|
"step": 41050 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 3.909800927828998e-06, |
|
"loss": 0.8532, |
|
"step": 41100 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 3.890226476402999e-06, |
|
"loss": 0.9293, |
|
"step": 41150 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 3.8706520249770005e-06, |
|
"loss": 0.8303, |
|
"step": 41200 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 3.851077573551002e-06, |
|
"loss": 0.8075, |
|
"step": 41250 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 3.831503122125002e-06, |
|
"loss": 0.8403, |
|
"step": 41300 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 3.8119286706990037e-06, |
|
"loss": 0.7709, |
|
"step": 41350 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 3.792354219273005e-06, |
|
"loss": 0.8533, |
|
"step": 41400 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 3.7727797678470064e-06, |
|
"loss": 0.8405, |
|
"step": 41450 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 3.7532053164210077e-06, |
|
"loss": 0.9128, |
|
"step": 41500 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 3.7336308649950086e-06, |
|
"loss": 0.8026, |
|
"step": 41550 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 3.71405641356901e-06, |
|
"loss": 0.9514, |
|
"step": 41600 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 3.6944819621430113e-06, |
|
"loss": 0.7858, |
|
"step": 41650 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 3.6749075107170122e-06, |
|
"loss": 0.8624, |
|
"step": 41700 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 3.6553330592910136e-06, |
|
"loss": 0.8648, |
|
"step": 41750 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 3.635758607865015e-06, |
|
"loss": 0.8852, |
|
"step": 41800 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 3.616184156439016e-06, |
|
"loss": 0.8459, |
|
"step": 41850 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 3.596609705013017e-06, |
|
"loss": 0.8714, |
|
"step": 41900 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 3.5770352535870185e-06, |
|
"loss": 0.939, |
|
"step": 41950 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 3.55746080216102e-06, |
|
"loss": 0.8739, |
|
"step": 42000 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 3.5378863507350208e-06, |
|
"loss": 0.8866, |
|
"step": 42050 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 3.518311899309022e-06, |
|
"loss": 0.8766, |
|
"step": 42100 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 3.4987374478830234e-06, |
|
"loss": 0.9094, |
|
"step": 42150 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 3.4791629964570244e-06, |
|
"loss": 0.8963, |
|
"step": 42200 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 3.4595885450310257e-06, |
|
"loss": 0.7177, |
|
"step": 42250 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 3.440014093605027e-06, |
|
"loss": 0.8551, |
|
"step": 42300 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 3.420439642179028e-06, |
|
"loss": 0.8674, |
|
"step": 42350 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 3.4008651907530293e-06, |
|
"loss": 0.8375, |
|
"step": 42400 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 3.3812907393270306e-06, |
|
"loss": 0.8921, |
|
"step": 42450 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 3.361716287901032e-06, |
|
"loss": 0.8845, |
|
"step": 42500 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 3.342141836475033e-06, |
|
"loss": 0.8361, |
|
"step": 42550 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 3.3225673850490342e-06, |
|
"loss": 0.8161, |
|
"step": 42600 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 3.3029929336230356e-06, |
|
"loss": 0.8561, |
|
"step": 42650 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 3.2834184821970365e-06, |
|
"loss": 0.8798, |
|
"step": 42700 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 3.263844030771038e-06, |
|
"loss": 0.8713, |
|
"step": 42750 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 3.244269579345039e-06, |
|
"loss": 0.8094, |
|
"step": 42800 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 3.22469512791904e-06, |
|
"loss": 0.8569, |
|
"step": 42850 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 3.2051206764930414e-06, |
|
"loss": 0.7974, |
|
"step": 42900 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 3.1855462250670428e-06, |
|
"loss": 0.7707, |
|
"step": 42950 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 3.165971773641044e-06, |
|
"loss": 0.8129, |
|
"step": 43000 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 3.146397322215045e-06, |
|
"loss": 0.8381, |
|
"step": 43050 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 3.1268228707890464e-06, |
|
"loss": 0.9326, |
|
"step": 43100 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 3.1072484193630477e-06, |
|
"loss": 0.9263, |
|
"step": 43150 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 3.0876739679370486e-06, |
|
"loss": 0.8431, |
|
"step": 43200 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 3.06809951651105e-06, |
|
"loss": 0.7909, |
|
"step": 43250 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 3.0485250650850513e-06, |
|
"loss": 0.832, |
|
"step": 43300 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 3.0289506136590522e-06, |
|
"loss": 0.9167, |
|
"step": 43350 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 3.0093761622330536e-06, |
|
"loss": 0.9032, |
|
"step": 43400 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 2.989801710807055e-06, |
|
"loss": 0.8083, |
|
"step": 43450 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 2.9702272593810562e-06, |
|
"loss": 0.8541, |
|
"step": 43500 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 2.950652807955057e-06, |
|
"loss": 0.787, |
|
"step": 43550 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 2.9310783565290585e-06, |
|
"loss": 0.879, |
|
"step": 43600 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 2.91150390510306e-06, |
|
"loss": 0.8119, |
|
"step": 43650 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 2.8919294536770608e-06, |
|
"loss": 0.86, |
|
"step": 43700 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 2.872355002251062e-06, |
|
"loss": 0.8162, |
|
"step": 43750 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 2.8527805508250634e-06, |
|
"loss": 0.8034, |
|
"step": 43800 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 2.8332060993990644e-06, |
|
"loss": 0.7883, |
|
"step": 43850 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 2.8136316479730657e-06, |
|
"loss": 0.8927, |
|
"step": 43900 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 2.794057196547067e-06, |
|
"loss": 0.8585, |
|
"step": 43950 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 2.7744827451210684e-06, |
|
"loss": 0.8768, |
|
"step": 44000 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 2.7549082936950693e-06, |
|
"loss": 0.826, |
|
"step": 44050 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 2.7353338422690706e-06, |
|
"loss": 0.8907, |
|
"step": 44100 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 2.715759390843072e-06, |
|
"loss": 0.8869, |
|
"step": 44150 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 2.696184939417073e-06, |
|
"loss": 0.8397, |
|
"step": 44200 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 2.6766104879910742e-06, |
|
"loss": 0.9316, |
|
"step": 44250 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 2.6570360365650756e-06, |
|
"loss": 0.8169, |
|
"step": 44300 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 2.6374615851390765e-06, |
|
"loss": 0.841, |
|
"step": 44350 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 2.617887133713078e-06, |
|
"loss": 0.8966, |
|
"step": 44400 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 2.598312682287079e-06, |
|
"loss": 0.8448, |
|
"step": 44450 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 2.5787382308610805e-06, |
|
"loss": 0.7451, |
|
"step": 44500 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 2.5591637794350814e-06, |
|
"loss": 0.921, |
|
"step": 44550 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 2.5395893280090828e-06, |
|
"loss": 0.8679, |
|
"step": 44600 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 2.520014876583084e-06, |
|
"loss": 0.94, |
|
"step": 44650 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 2.500440425157085e-06, |
|
"loss": 0.7933, |
|
"step": 44700 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 2.4808659737310864e-06, |
|
"loss": 0.8309, |
|
"step": 44750 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 2.4612915223050877e-06, |
|
"loss": 0.8401, |
|
"step": 44800 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 2.4417170708790886e-06, |
|
"loss": 0.8334, |
|
"step": 44850 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 2.42214261945309e-06, |
|
"loss": 0.885, |
|
"step": 44900 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 2.4025681680270913e-06, |
|
"loss": 0.8044, |
|
"step": 44950 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 2.3829937166010927e-06, |
|
"loss": 0.8739, |
|
"step": 45000 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 2.3634192651750936e-06, |
|
"loss": 0.8277, |
|
"step": 45050 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 2.343844813749095e-06, |
|
"loss": 0.7745, |
|
"step": 45100 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 2.3242703623230963e-06, |
|
"loss": 0.8381, |
|
"step": 45150 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 2.304695910897097e-06, |
|
"loss": 0.8805, |
|
"step": 45200 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.2851214594710985e-06, |
|
"loss": 0.8687, |
|
"step": 45250 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.2655470080451e-06, |
|
"loss": 0.986, |
|
"step": 45300 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.2459725566191008e-06, |
|
"loss": 0.8102, |
|
"step": 45350 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.226398105193102e-06, |
|
"loss": 0.8374, |
|
"step": 45400 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.2068236537671035e-06, |
|
"loss": 0.8183, |
|
"step": 45450 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.187249202341105e-06, |
|
"loss": 0.7379, |
|
"step": 45500 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.1676747509151057e-06, |
|
"loss": 0.8932, |
|
"step": 45550 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.148100299489107e-06, |
|
"loss": 0.8039, |
|
"step": 45600 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.1285258480631084e-06, |
|
"loss": 0.8697, |
|
"step": 45650 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.1089513966371093e-06, |
|
"loss": 0.8854, |
|
"step": 45700 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.0893769452111106e-06, |
|
"loss": 0.7589, |
|
"step": 45750 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.069802493785112e-06, |
|
"loss": 0.8185, |
|
"step": 45800 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.050228042359113e-06, |
|
"loss": 0.8476, |
|
"step": 45850 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.0306535909331142e-06, |
|
"loss": 0.8286, |
|
"step": 45900 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.0110791395071156e-06, |
|
"loss": 0.8612, |
|
"step": 45950 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 1.991504688081117e-06, |
|
"loss": 0.8429, |
|
"step": 46000 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 1.971930236655118e-06, |
|
"loss": 0.7646, |
|
"step": 46050 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 1.952355785229119e-06, |
|
"loss": 0.8977, |
|
"step": 46100 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 1.9327813338031205e-06, |
|
"loss": 0.8532, |
|
"step": 46150 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 1.9132068823771214e-06, |
|
"loss": 0.7703, |
|
"step": 46200 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.8936324309511228e-06, |
|
"loss": 0.8496, |
|
"step": 46250 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.8740579795251241e-06, |
|
"loss": 0.7815, |
|
"step": 46300 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.8544835280991253e-06, |
|
"loss": 0.8324, |
|
"step": 46350 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.8349090766731264e-06, |
|
"loss": 0.793, |
|
"step": 46400 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.8153346252471277e-06, |
|
"loss": 0.852, |
|
"step": 46450 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.7957601738211289e-06, |
|
"loss": 0.8608, |
|
"step": 46500 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.7761857223951302e-06, |
|
"loss": 0.8912, |
|
"step": 46550 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.7566112709691313e-06, |
|
"loss": 0.821, |
|
"step": 46600 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.7370368195431325e-06, |
|
"loss": 1.0388, |
|
"step": 46650 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.7174623681171338e-06, |
|
"loss": 0.91, |
|
"step": 46700 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.697887916691135e-06, |
|
"loss": 0.8276, |
|
"step": 46750 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.6783134652651363e-06, |
|
"loss": 0.8285, |
|
"step": 46800 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.6587390138391374e-06, |
|
"loss": 0.8882, |
|
"step": 46850 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.6391645624131385e-06, |
|
"loss": 0.8399, |
|
"step": 46900 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.6195901109871399e-06, |
|
"loss": 0.8796, |
|
"step": 46950 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.600015659561141e-06, |
|
"loss": 0.8012, |
|
"step": 47000 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.5804412081351423e-06, |
|
"loss": 0.8199, |
|
"step": 47050 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.5608667567091435e-06, |
|
"loss": 0.8194, |
|
"step": 47100 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.5412923052831446e-06, |
|
"loss": 0.8026, |
|
"step": 47150 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.521717853857146e-06, |
|
"loss": 0.9046, |
|
"step": 47200 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.502143402431147e-06, |
|
"loss": 0.8155, |
|
"step": 47250 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.4825689510051484e-06, |
|
"loss": 0.8619, |
|
"step": 47300 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.4629944995791495e-06, |
|
"loss": 0.8269, |
|
"step": 47350 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.4434200481531507e-06, |
|
"loss": 0.8113, |
|
"step": 47400 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.423845596727152e-06, |
|
"loss": 0.8725, |
|
"step": 47450 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.4042711453011531e-06, |
|
"loss": 0.8423, |
|
"step": 47500 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.3846966938751545e-06, |
|
"loss": 0.8607, |
|
"step": 47550 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.3651222424491556e-06, |
|
"loss": 0.7522, |
|
"step": 47600 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.3455477910231567e-06, |
|
"loss": 0.804, |
|
"step": 47650 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.325973339597158e-06, |
|
"loss": 0.8359, |
|
"step": 47700 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.3063988881711592e-06, |
|
"loss": 0.8405, |
|
"step": 47750 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.2868244367451605e-06, |
|
"loss": 0.8168, |
|
"step": 47800 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.2672499853191617e-06, |
|
"loss": 0.7954, |
|
"step": 47850 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.2476755338931628e-06, |
|
"loss": 0.8689, |
|
"step": 47900 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.228101082467164e-06, |
|
"loss": 0.8532, |
|
"step": 47950 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.2085266310411653e-06, |
|
"loss": 0.846, |
|
"step": 48000 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.1889521796151664e-06, |
|
"loss": 0.8195, |
|
"step": 48050 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.1693777281891675e-06, |
|
"loss": 0.8346, |
|
"step": 48100 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.1498032767631689e-06, |
|
"loss": 0.8103, |
|
"step": 48150 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.13022882533717e-06, |
|
"loss": 0.9013, |
|
"step": 48200 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.1106543739111713e-06, |
|
"loss": 0.8704, |
|
"step": 48250 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.0910799224851725e-06, |
|
"loss": 0.7644, |
|
"step": 48300 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.0715054710591736e-06, |
|
"loss": 0.8247, |
|
"step": 48350 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.051931019633175e-06, |
|
"loss": 0.7922, |
|
"step": 48400 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.032356568207176e-06, |
|
"loss": 0.8569, |
|
"step": 48450 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.0127821167811774e-06, |
|
"loss": 0.8725, |
|
"step": 48500 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 9.932076653551785e-07, |
|
"loss": 0.8302, |
|
"step": 48550 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 9.736332139291797e-07, |
|
"loss": 0.8937, |
|
"step": 48600 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 9.54058762503181e-07, |
|
"loss": 0.8193, |
|
"step": 48650 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 9.344843110771821e-07, |
|
"loss": 0.8036, |
|
"step": 48700 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 9.149098596511834e-07, |
|
"loss": 0.7537, |
|
"step": 48750 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 8.953354082251846e-07, |
|
"loss": 0.7915, |
|
"step": 48800 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 8.757609567991858e-07, |
|
"loss": 0.8179, |
|
"step": 48850 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 8.56186505373187e-07, |
|
"loss": 0.7896, |
|
"step": 48900 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 8.366120539471882e-07, |
|
"loss": 0.8691, |
|
"step": 48950 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 8.170376025211894e-07, |
|
"loss": 0.8881, |
|
"step": 49000 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 7.974631510951907e-07, |
|
"loss": 0.8348, |
|
"step": 49050 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 7.778886996691919e-07, |
|
"loss": 0.831, |
|
"step": 49100 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 7.58314248243193e-07, |
|
"loss": 0.9022, |
|
"step": 49150 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 7.387397968171943e-07, |
|
"loss": 0.8517, |
|
"step": 49200 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 7.191653453911955e-07, |
|
"loss": 0.8352, |
|
"step": 49250 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 6.995908939651967e-07, |
|
"loss": 0.8007, |
|
"step": 49300 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 6.80016442539198e-07, |
|
"loss": 0.8099, |
|
"step": 49350 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 6.604419911131991e-07, |
|
"loss": 0.8561, |
|
"step": 49400 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 6.408675396872003e-07, |
|
"loss": 0.893, |
|
"step": 49450 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 6.212930882612015e-07, |
|
"loss": 0.785, |
|
"step": 49500 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 6.017186368352027e-07, |
|
"loss": 0.915, |
|
"step": 49550 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 5.821441854092039e-07, |
|
"loss": 0.7017, |
|
"step": 49600 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 5.625697339832052e-07, |
|
"loss": 0.8437, |
|
"step": 49650 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 5.429952825572064e-07, |
|
"loss": 0.8002, |
|
"step": 49700 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 5.234208311312075e-07, |
|
"loss": 0.8034, |
|
"step": 49750 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 5.038463797052088e-07, |
|
"loss": 0.7989, |
|
"step": 49800 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 4.8427192827921e-07, |
|
"loss": 0.8462, |
|
"step": 49850 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 4.6469747685321123e-07, |
|
"loss": 0.7838, |
|
"step": 49900 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 4.451230254272124e-07, |
|
"loss": 0.826, |
|
"step": 49950 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 4.2554857400121365e-07, |
|
"loss": 0.9047, |
|
"step": 50000 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 4.059741225752149e-07, |
|
"loss": 0.8526, |
|
"step": 50050 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 3.8639967114921607e-07, |
|
"loss": 0.8042, |
|
"step": 50100 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 3.668252197232173e-07, |
|
"loss": 0.9519, |
|
"step": 50150 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 3.472507682972185e-07, |
|
"loss": 0.8923, |
|
"step": 50200 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 3.276763168712197e-07, |
|
"loss": 0.7782, |
|
"step": 50250 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 3.0810186544522095e-07, |
|
"loss": 0.8607, |
|
"step": 50300 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 2.8852741401922214e-07, |
|
"loss": 0.8376, |
|
"step": 50350 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 2.6895296259322337e-07, |
|
"loss": 0.842, |
|
"step": 50400 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 2.4937851116722455e-07, |
|
"loss": 0.878, |
|
"step": 50450 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 2.298040597412258e-07, |
|
"loss": 0.8581, |
|
"step": 50500 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 2.10229608315227e-07, |
|
"loss": 0.8958, |
|
"step": 50550 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.9065515688922818e-07, |
|
"loss": 0.92, |
|
"step": 50600 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.7108070546322939e-07, |
|
"loss": 0.9209, |
|
"step": 50650 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.5150625403723062e-07, |
|
"loss": 0.9095, |
|
"step": 50700 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.3193180261123183e-07, |
|
"loss": 0.8469, |
|
"step": 50750 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.1235735118523305e-07, |
|
"loss": 0.8987, |
|
"step": 50800 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 9.278289975923425e-08, |
|
"loss": 0.8769, |
|
"step": 50850 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 7.320844833323547e-08, |
|
"loss": 0.8072, |
|
"step": 50900 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 5.3633996907236675e-08, |
|
"loss": 0.7988, |
|
"step": 50950 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 3.405954548123789e-08, |
|
"loss": 0.7943, |
|
"step": 51000 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 1.4485094055239104e-08, |
|
"loss": 0.7839, |
|
"step": 51050 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_Macro F1": 0.45208930263938163, |
|
"eval_Macro Precision": 0.5507686024916876, |
|
"eval_Macro Recall": 0.4294052669214516, |
|
"eval_Micro F1": 0.7266960098561323, |
|
"eval_Micro Precision": 0.7266960098561323, |
|
"eval_Micro Recall": 0.7266960098561323, |
|
"eval_Weighted F1": 0.6482219016896026, |
|
"eval_Weighted Precision": 0.6261988330407515, |
|
"eval_Weighted Recall": 0.7266960098561323, |
|
"eval_accuracy": 0.7266960098561323, |
|
"eval_loss": 0.8367530107498169, |
|
"eval_runtime": 1309.1607, |
|
"eval_samples_per_second": 153.76, |
|
"eval_steps_per_second": 19.22, |
|
"step": 51087 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"step": 51087, |
|
"total_flos": 1927062258615528.0, |
|
"train_loss": 0.8681267702094373, |
|
"train_runtime": 9176.1966, |
|
"train_samples_per_second": 44.538, |
|
"train_steps_per_second": 5.567 |
|
} |
|
], |
|
"max_steps": 51087, |
|
"num_train_epochs": 1, |
|
"total_flos": 1927062258615528.0, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|