|
{ |
|
"best_metric": 21.3524811218986, |
|
"best_model_checkpoint": "tamil_models/whisper-medium-ta_alldata_multigpu/checkpoint-8100", |
|
"epoch": 2.9779411764705883, |
|
"global_step": 8100, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 2.8562520089993833e-06, |
|
"loss": 1.5919, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 4.805947323355403e-06, |
|
"loss": 0.7788, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 5.712504017998767e-06, |
|
"loss": 0.5415, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 6.309637647321797e-06, |
|
"loss": 0.403, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 6.755642637711422e-06, |
|
"loss": 0.3221, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 7.111770744693953e-06, |
|
"loss": 0.2477, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 7.4082398170603955e-06, |
|
"loss": 0.2032, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 7.662199332354785e-06, |
|
"loss": 0.1913, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 7.884325004773166e-06, |
|
"loss": 0.1803, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 8.0817159331721e-06, |
|
"loss": 0.1755, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"eval_loss": 0.20166015625, |
|
"eval_runtime": 284.2904, |
|
"eval_samples_per_second": 9.744, |
|
"eval_steps_per_second": 0.077, |
|
"eval_wer": 39.73840345199569, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 8.259332961677815e-06, |
|
"loss": 0.1628, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 8.420779589879445e-06, |
|
"loss": 0.1622, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 8.56875602699815e-06, |
|
"loss": 0.1555, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 8.705337952067442e-06, |
|
"loss": 0.1506, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 8.832155400481049e-06, |
|
"loss": 0.1489, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 8.95051185954238e-06, |
|
"loss": 0.1453, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.061466059049972e-06, |
|
"loss": 0.143, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.16588965632118e-06, |
|
"loss": 0.1402, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.264508868515998e-06, |
|
"loss": 0.1386, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.357935131416414e-06, |
|
"loss": 0.1374, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"eval_loss": 0.1690673828125, |
|
"eval_runtime": 294.2297, |
|
"eval_samples_per_second": 9.414, |
|
"eval_steps_per_second": 0.075, |
|
"eval_wer": 36.370685005393746, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.446688082452126e-06, |
|
"loss": 0.1323, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.531213064471803e-06, |
|
"loss": 0.1343, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.611894646710806e-06, |
|
"loss": 0.1317, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.689067203278456e-06, |
|
"loss": 0.125, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.76302328564421e-06, |
|
"loss": 0.1323, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.834020319129184e-06, |
|
"loss": 0.1286, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.902286010551205e-06, |
|
"loss": 0.1225, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.968022753693337e-06, |
|
"loss": 0.1282, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.998556998557e-06, |
|
"loss": 0.1244, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.994949494949497e-06, |
|
"loss": 0.1231, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"eval_loss": 0.1517333984375, |
|
"eval_runtime": 240.2543, |
|
"eval_samples_per_second": 11.529, |
|
"eval_steps_per_second": 0.092, |
|
"eval_wer": 33.087243797195256, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.991341991341992e-06, |
|
"loss": 0.1244, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.987734487734489e-06, |
|
"loss": 0.1163, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.984126984126986e-06, |
|
"loss": 0.1194, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.980519480519481e-06, |
|
"loss": 0.1191, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.976911976911978e-06, |
|
"loss": 0.115, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.973304473304473e-06, |
|
"loss": 0.1119, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.96969696969697e-06, |
|
"loss": 0.1171, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.966089466089467e-06, |
|
"loss": 0.1159, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.962481962481964e-06, |
|
"loss": 0.1134, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.95887445887446e-06, |
|
"loss": 0.116, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"eval_loss": 0.1417236328125, |
|
"eval_runtime": 254.6878, |
|
"eval_samples_per_second": 10.876, |
|
"eval_steps_per_second": 0.086, |
|
"eval_wer": 31.47586299892125, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.955266955266956e-06, |
|
"loss": 0.1126, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.951659451659453e-06, |
|
"loss": 0.1114, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.94805194805195e-06, |
|
"loss": 0.1116, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.944444444444445e-06, |
|
"loss": 0.111, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.940836940836942e-06, |
|
"loss": 0.1122, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.937229437229437e-06, |
|
"loss": 0.1081, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.933621933621934e-06, |
|
"loss": 0.1116, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.93001443001443e-06, |
|
"loss": 0.1067, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.926406926406928e-06, |
|
"loss": 0.1075, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.922799422799425e-06, |
|
"loss": 0.1073, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"eval_loss": 0.134765625, |
|
"eval_runtime": 864.1715, |
|
"eval_samples_per_second": 3.205, |
|
"eval_steps_per_second": 0.025, |
|
"eval_wer": 30.818500539374327, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.91919191919192e-06, |
|
"loss": 0.1085, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.915584415584417e-06, |
|
"loss": 0.1079, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.911976911976914e-06, |
|
"loss": 0.1068, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.908369408369409e-06, |
|
"loss": 0.1096, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.904761904761906e-06, |
|
"loss": 0.104, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.901154401154402e-06, |
|
"loss": 0.1055, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.897546897546898e-06, |
|
"loss": 0.1014, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.893939393939395e-06, |
|
"loss": 0.1086, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.890331890331891e-06, |
|
"loss": 0.106, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.886724386724388e-06, |
|
"loss": 0.1037, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"eval_loss": 0.1295166015625, |
|
"eval_runtime": 258.1898, |
|
"eval_samples_per_second": 10.729, |
|
"eval_steps_per_second": 0.085, |
|
"eval_wer": 29.72289644012945, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.883116883116885e-06, |
|
"loss": 0.1027, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.87950937950938e-06, |
|
"loss": 0.1015, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.875901875901877e-06, |
|
"loss": 0.1002, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.872294372294373e-06, |
|
"loss": 0.1027, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.86868686868687e-06, |
|
"loss": 0.102, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.865079365079366e-06, |
|
"loss": 0.0981, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.861471861471862e-06, |
|
"loss": 0.0992, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.857864357864358e-06, |
|
"loss": 0.1041, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.854256854256855e-06, |
|
"loss": 0.0979, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.850649350649352e-06, |
|
"loss": 0.0997, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"eval_loss": 0.1251220703125, |
|
"eval_runtime": 253.4804, |
|
"eval_samples_per_second": 10.928, |
|
"eval_steps_per_second": 0.087, |
|
"eval_wer": 29.26442826321467, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.847041847041849e-06, |
|
"loss": 0.0963, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.843434343434344e-06, |
|
"loss": 0.0971, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.839826839826841e-06, |
|
"loss": 0.0997, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.836219336219336e-06, |
|
"loss": 0.0979, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.832611832611833e-06, |
|
"loss": 0.099, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.82900432900433e-06, |
|
"loss": 0.0974, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.825396825396825e-06, |
|
"loss": 0.0955, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.821789321789322e-06, |
|
"loss": 0.0953, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.81818181818182e-06, |
|
"loss": 0.0965, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.814574314574316e-06, |
|
"loss": 0.0982, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"eval_loss": 0.12176513671875, |
|
"eval_runtime": 245.3351, |
|
"eval_samples_per_second": 11.291, |
|
"eval_steps_per_second": 0.09, |
|
"eval_wer": 28.56324163969795, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.810966810966811e-06, |
|
"loss": 0.0949, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.807359307359308e-06, |
|
"loss": 0.0963, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.803751803751805e-06, |
|
"loss": 0.0973, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.8001443001443e-06, |
|
"loss": 0.0965, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.796536796536797e-06, |
|
"loss": 0.0938, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.792929292929294e-06, |
|
"loss": 0.0939, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.789321789321791e-06, |
|
"loss": 0.0963, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.785714285714286e-06, |
|
"loss": 0.0978, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.782106782106783e-06, |
|
"loss": 0.0923, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.77849927849928e-06, |
|
"loss": 0.098, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"eval_loss": 0.1192626953125, |
|
"eval_runtime": 520.2114, |
|
"eval_samples_per_second": 5.325, |
|
"eval_steps_per_second": 0.042, |
|
"eval_wer": 28.030609492988134, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.774891774891775e-06, |
|
"loss": 0.0914, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.771284271284272e-06, |
|
"loss": 0.0944, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.767676767676767e-06, |
|
"loss": 0.0966, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.764069264069264e-06, |
|
"loss": 0.0903, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.760461760461761e-06, |
|
"loss": 0.0952, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.756854256854258e-06, |
|
"loss": 0.0913, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 9.753246753246755e-06, |
|
"loss": 0.0943, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 9.74963924963925e-06, |
|
"loss": 0.0939, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 9.746031746031747e-06, |
|
"loss": 0.0946, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 9.742424242424244e-06, |
|
"loss": 0.0941, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"eval_loss": 0.11590576171875, |
|
"eval_runtime": 241.569, |
|
"eval_samples_per_second": 11.467, |
|
"eval_steps_per_second": 0.091, |
|
"eval_wer": 27.4608953613808, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 9.738816738816739e-06, |
|
"loss": 0.0898, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 9.735209235209236e-06, |
|
"loss": 0.0899, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 9.731601731601731e-06, |
|
"loss": 0.093, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 9.727994227994228e-06, |
|
"loss": 0.0897, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 9.724386724386725e-06, |
|
"loss": 0.095, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 9.720779220779222e-06, |
|
"loss": 0.0965, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 9.717171717171719e-06, |
|
"loss": 0.0911, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 9.713564213564214e-06, |
|
"loss": 0.0903, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 9.70995670995671e-06, |
|
"loss": 0.0952, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 9.706349206349208e-06, |
|
"loss": 0.0941, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"eval_loss": 0.11358642578125, |
|
"eval_runtime": 255.0901, |
|
"eval_samples_per_second": 10.859, |
|
"eval_steps_per_second": 0.086, |
|
"eval_wer": 26.965345199568503, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 9.702741702741703e-06, |
|
"loss": 0.0898, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 9.6991341991342e-06, |
|
"loss": 0.0917, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 9.695526695526695e-06, |
|
"loss": 0.0895, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 9.691919191919192e-06, |
|
"loss": 0.0915, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 9.688311688311689e-06, |
|
"loss": 0.089, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 9.684704184704186e-06, |
|
"loss": 0.0887, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 9.681096681096683e-06, |
|
"loss": 0.0909, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 9.67748917748918e-06, |
|
"loss": 0.0897, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 9.673881673881675e-06, |
|
"loss": 0.0917, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 9.670274170274172e-06, |
|
"loss": 0.0868, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"eval_loss": 0.11199951171875, |
|
"eval_runtime": 247.8554, |
|
"eval_samples_per_second": 11.176, |
|
"eval_steps_per_second": 0.089, |
|
"eval_wer": 27.164239482200646, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 9.666666666666667e-06, |
|
"loss": 0.091, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 9.663059163059164e-06, |
|
"loss": 0.0903, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 9.65945165945166e-06, |
|
"loss": 0.0856, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 9.655844155844156e-06, |
|
"loss": 0.0907, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 9.652236652236653e-06, |
|
"loss": 0.0886, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 9.64862914862915e-06, |
|
"loss": 0.0841, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 9.645021645021646e-06, |
|
"loss": 0.0855, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 9.641414141414143e-06, |
|
"loss": 0.0843, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 9.637806637806638e-06, |
|
"loss": 0.0901, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 9.634199134199135e-06, |
|
"loss": 0.0899, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"eval_loss": 0.10968017578125, |
|
"eval_runtime": 245.0559, |
|
"eval_samples_per_second": 11.304, |
|
"eval_steps_per_second": 0.09, |
|
"eval_wer": 26.581040992448756, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 9.63059163059163e-06, |
|
"loss": 0.0857, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 9.626984126984127e-06, |
|
"loss": 0.0879, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 9.623376623376624e-06, |
|
"loss": 0.0909, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 9.61976911976912e-06, |
|
"loss": 0.0859, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 9.616161616161616e-06, |
|
"loss": 0.0844, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 9.612554112554113e-06, |
|
"loss": 0.0878, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 9.60894660894661e-06, |
|
"loss": 0.0867, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 9.605339105339107e-06, |
|
"loss": 0.0885, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 9.601731601731602e-06, |
|
"loss": 0.0856, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 9.5981240981241e-06, |
|
"loss": 0.0888, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"eval_loss": 0.1080322265625, |
|
"eval_runtime": 243.3367, |
|
"eval_samples_per_second": 11.383, |
|
"eval_steps_per_second": 0.09, |
|
"eval_wer": 26.007955771305287, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.594516594516594e-06, |
|
"loss": 0.0872, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.590909090909091e-06, |
|
"loss": 0.0861, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.587301587301588e-06, |
|
"loss": 0.0852, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.583694083694083e-06, |
|
"loss": 0.0866, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.58008658008658e-06, |
|
"loss": 0.0842, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.576479076479077e-06, |
|
"loss": 0.0861, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.572871572871574e-06, |
|
"loss": 0.086, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.569264069264071e-06, |
|
"loss": 0.0831, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 9.565656565656566e-06, |
|
"loss": 0.0858, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 9.562049062049063e-06, |
|
"loss": 0.0871, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"eval_loss": 0.10699462890625, |
|
"eval_runtime": 701.1447, |
|
"eval_samples_per_second": 3.951, |
|
"eval_steps_per_second": 0.031, |
|
"eval_wer": 25.974244875943903, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 9.558441558441558e-06, |
|
"loss": 0.0805, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 9.554834054834055e-06, |
|
"loss": 0.0857, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 9.551226551226552e-06, |
|
"loss": 0.0855, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 9.547619047619049e-06, |
|
"loss": 0.0855, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 9.544011544011544e-06, |
|
"loss": 0.0872, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 9.540404040404041e-06, |
|
"loss": 0.0866, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 9.536796536796538e-06, |
|
"loss": 0.0852, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 9.533189033189035e-06, |
|
"loss": 0.0878, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 9.52958152958153e-06, |
|
"loss": 0.0855, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 9.525974025974027e-06, |
|
"loss": 0.0848, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"eval_loss": 0.10595703125, |
|
"eval_runtime": 254.3389, |
|
"eval_samples_per_second": 10.891, |
|
"eval_steps_per_second": 0.086, |
|
"eval_wer": 25.589940668824163, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 9.522366522366522e-06, |
|
"loss": 0.0893, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 9.518759018759019e-06, |
|
"loss": 0.0822, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 9.515151515151516e-06, |
|
"loss": 0.0813, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 9.511544011544013e-06, |
|
"loss": 0.0844, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 9.507936507936508e-06, |
|
"loss": 0.0823, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 9.504329004329005e-06, |
|
"loss": 0.0868, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 9.500721500721502e-06, |
|
"loss": 0.0851, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 9.497113997113999e-06, |
|
"loss": 0.0818, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 9.493506493506494e-06, |
|
"loss": 0.0835, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 9.48989898989899e-06, |
|
"loss": 0.0835, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"eval_loss": 0.1044921875, |
|
"eval_runtime": 1330.4645, |
|
"eval_samples_per_second": 2.082, |
|
"eval_steps_per_second": 0.017, |
|
"eval_wer": 25.573085221143472, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 9.486291486291486e-06, |
|
"loss": 0.0841, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 9.482683982683983e-06, |
|
"loss": 0.0858, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 9.47907647907648e-06, |
|
"loss": 0.0845, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 9.475468975468977e-06, |
|
"loss": 0.0854, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 9.471861471861472e-06, |
|
"loss": 0.083, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 9.468253968253969e-06, |
|
"loss": 0.0834, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 9.464646464646466e-06, |
|
"loss": 0.0811, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 9.461038961038963e-06, |
|
"loss": 0.0882, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 9.457431457431458e-06, |
|
"loss": 0.0829, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 9.453823953823955e-06, |
|
"loss": 0.08, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"eval_loss": 0.10369873046875, |
|
"eval_runtime": 244.3089, |
|
"eval_samples_per_second": 11.338, |
|
"eval_steps_per_second": 0.09, |
|
"eval_wer": 25.13484358144552, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 9.45021645021645e-06, |
|
"loss": 0.0853, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 9.446608946608947e-06, |
|
"loss": 0.0826, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 9.443001443001444e-06, |
|
"loss": 0.0828, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 9.43939393939394e-06, |
|
"loss": 0.0797, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 9.435786435786437e-06, |
|
"loss": 0.0814, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 9.432178932178933e-06, |
|
"loss": 0.0827, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 9.42857142857143e-06, |
|
"loss": 0.0789, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 9.424963924963926e-06, |
|
"loss": 0.0838, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 9.421356421356422e-06, |
|
"loss": 0.0853, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 9.417748917748919e-06, |
|
"loss": 0.0819, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"eval_loss": 0.10162353515625, |
|
"eval_runtime": 705.4564, |
|
"eval_samples_per_second": 3.927, |
|
"eval_steps_per_second": 0.031, |
|
"eval_wer": 24.976402373247033, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 9.414141414141414e-06, |
|
"loss": 0.0821, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 9.41053391053391e-06, |
|
"loss": 0.0813, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 9.406926406926408e-06, |
|
"loss": 0.0814, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 9.403318903318904e-06, |
|
"loss": 0.0856, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 9.399711399711401e-06, |
|
"loss": 0.0842, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 9.396103896103896e-06, |
|
"loss": 0.0806, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 9.392496392496393e-06, |
|
"loss": 0.0794, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 9.38888888888889e-06, |
|
"loss": 0.084, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 9.385281385281385e-06, |
|
"loss": 0.0853, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 9.381673881673882e-06, |
|
"loss": 0.0801, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"eval_loss": 0.100830078125, |
|
"eval_runtime": 693.4811, |
|
"eval_samples_per_second": 3.994, |
|
"eval_steps_per_second": 0.032, |
|
"eval_wer": 24.96628910463862, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 9.378066378066378e-06, |
|
"loss": 0.0807, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 9.374458874458874e-06, |
|
"loss": 0.0807, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 9.370851370851371e-06, |
|
"loss": 0.08, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 9.367243867243868e-06, |
|
"loss": 0.0829, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 9.363636363636365e-06, |
|
"loss": 0.0805, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 9.36002886002886e-06, |
|
"loss": 0.0807, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 9.356421356421357e-06, |
|
"loss": 0.0826, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 9.352813852813854e-06, |
|
"loss": 0.0796, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 9.34920634920635e-06, |
|
"loss": 0.0786, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 9.345598845598846e-06, |
|
"loss": 0.0793, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"eval_loss": 0.09967041015625, |
|
"eval_runtime": 817.1155, |
|
"eval_samples_per_second": 3.39, |
|
"eval_steps_per_second": 0.027, |
|
"eval_wer": 24.504449838187703, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 9.341991341991343e-06, |
|
"loss": 0.0808, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 9.338383838383838e-06, |
|
"loss": 0.0814, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 9.334776334776335e-06, |
|
"loss": 0.0832, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 9.331168831168832e-06, |
|
"loss": 0.0812, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 9.327561327561329e-06, |
|
"loss": 0.0776, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 9.323953823953826e-06, |
|
"loss": 0.0807, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 9.320346320346321e-06, |
|
"loss": 0.0831, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 9.316738816738818e-06, |
|
"loss": 0.0795, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 9.313131313131313e-06, |
|
"loss": 0.079, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 9.30952380952381e-06, |
|
"loss": 0.081, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"eval_loss": 0.09979248046875, |
|
"eval_runtime": 515.1218, |
|
"eval_samples_per_second": 5.377, |
|
"eval_steps_per_second": 0.043, |
|
"eval_wer": 24.514563106796118, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 9.305916305916307e-06, |
|
"loss": 0.083, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 9.302308802308802e-06, |
|
"loss": 0.0823, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 9.298701298701299e-06, |
|
"loss": 0.0809, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 9.295093795093796e-06, |
|
"loss": 0.0795, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 9.291486291486293e-06, |
|
"loss": 0.0786, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 9.28787878787879e-06, |
|
"loss": 0.0838, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 9.284271284271285e-06, |
|
"loss": 0.0788, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 9.280663780663782e-06, |
|
"loss": 0.0788, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 9.277056277056277e-06, |
|
"loss": 0.0765, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 9.273448773448774e-06, |
|
"loss": 0.079, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"eval_loss": 0.0982666015625, |
|
"eval_runtime": 568.3193, |
|
"eval_samples_per_second": 4.874, |
|
"eval_steps_per_second": 0.039, |
|
"eval_wer": 24.733683926645092, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 9.26984126984127e-06, |
|
"loss": 0.0772, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 9.266233766233766e-06, |
|
"loss": 0.0808, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 9.262626262626263e-06, |
|
"loss": 0.0786, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 9.25901875901876e-06, |
|
"loss": 0.0787, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 9.255411255411257e-06, |
|
"loss": 0.0798, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 9.251803751803754e-06, |
|
"loss": 0.0808, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 9.248196248196249e-06, |
|
"loss": 0.0753, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 9.244588744588746e-06, |
|
"loss": 0.0801, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 9.240981240981241e-06, |
|
"loss": 0.0794, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 9.237373737373738e-06, |
|
"loss": 0.0758, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"eval_loss": 0.09796142578125, |
|
"eval_runtime": 249.0453, |
|
"eval_samples_per_second": 11.122, |
|
"eval_steps_per_second": 0.088, |
|
"eval_wer": 24.747168284789645, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 9.233766233766235e-06, |
|
"loss": 0.0798, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 9.230158730158732e-06, |
|
"loss": 0.082, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 9.226551226551227e-06, |
|
"loss": 0.0796, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 9.222943722943724e-06, |
|
"loss": 0.0787, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 9.21933621933622e-06, |
|
"loss": 0.075, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 9.215728715728717e-06, |
|
"loss": 0.0778, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 9.212121212121213e-06, |
|
"loss": 0.077, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 9.20851370851371e-06, |
|
"loss": 0.0747, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 9.204906204906205e-06, |
|
"loss": 0.075, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 9.201298701298702e-06, |
|
"loss": 0.0806, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"eval_loss": 0.0966796875, |
|
"eval_runtime": 487.9447, |
|
"eval_samples_per_second": 5.677, |
|
"eval_steps_per_second": 0.045, |
|
"eval_wer": 23.77629449838188, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 9.197691197691199e-06, |
|
"loss": 0.0786, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 9.194083694083695e-06, |
|
"loss": 0.0789, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 9.19047619047619e-06, |
|
"loss": 0.0765, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 9.186868686868688e-06, |
|
"loss": 0.0769, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 9.183261183261184e-06, |
|
"loss": 0.0773, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 9.179653679653681e-06, |
|
"loss": 0.076, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 9.176046176046177e-06, |
|
"loss": 0.0751, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 9.172438672438673e-06, |
|
"loss": 0.0764, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 9.168831168831169e-06, |
|
"loss": 0.0767, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 9.165223665223666e-06, |
|
"loss": 0.077, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"eval_loss": 0.095703125, |
|
"eval_runtime": 770.8972, |
|
"eval_samples_per_second": 3.593, |
|
"eval_steps_per_second": 0.029, |
|
"eval_wer": 24.029126213592235, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 9.161616161616162e-06, |
|
"loss": 0.0777, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 9.15800865800866e-06, |
|
"loss": 0.0773, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 9.154401154401155e-06, |
|
"loss": 0.0801, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 9.150793650793651e-06, |
|
"loss": 0.0799, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 9.147186147186148e-06, |
|
"loss": 0.0778, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 9.143578643578645e-06, |
|
"loss": 0.0757, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 9.13997113997114e-06, |
|
"loss": 0.0797, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 9.136363636363637e-06, |
|
"loss": 0.0756, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 9.132756132756132e-06, |
|
"loss": 0.0759, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 9.12914862914863e-06, |
|
"loss": 0.0797, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"eval_loss": 0.09466552734375, |
|
"eval_runtime": 554.2007, |
|
"eval_samples_per_second": 4.998, |
|
"eval_steps_per_second": 0.04, |
|
"eval_wer": 23.678532901833872, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 9.125541125541126e-06, |
|
"loss": 0.0775, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 9.121933621933623e-06, |
|
"loss": 0.0774, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 9.11832611832612e-06, |
|
"loss": 0.0691, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 9.114718614718615e-06, |
|
"loss": 0.067, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 9.111111111111112e-06, |
|
"loss": 0.0684, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 9.107503607503609e-06, |
|
"loss": 0.0688, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 9.103896103896104e-06, |
|
"loss": 0.0667, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 9.100288600288601e-06, |
|
"loss": 0.0664, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 9.096681096681096e-06, |
|
"loss": 0.0654, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 9.093073593073593e-06, |
|
"loss": 0.0697, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"eval_loss": 0.09423828125, |
|
"eval_runtime": 378.3675, |
|
"eval_samples_per_second": 7.321, |
|
"eval_steps_per_second": 0.058, |
|
"eval_wer": 23.570658036677454, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 9.08946608946609e-06, |
|
"loss": 0.0675, |
|
"step": 2810 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 9.085858585858587e-06, |
|
"loss": 0.0684, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 9.082251082251084e-06, |
|
"loss": 0.0697, |
|
"step": 2830 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 9.078643578643579e-06, |
|
"loss": 0.0678, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 9.075036075036076e-06, |
|
"loss": 0.0673, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 9.071428571428573e-06, |
|
"loss": 0.0691, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 9.067821067821068e-06, |
|
"loss": 0.0657, |
|
"step": 2870 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 9.064213564213565e-06, |
|
"loss": 0.067, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 9.06060606060606e-06, |
|
"loss": 0.0695, |
|
"step": 2890 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 9.056998556998557e-06, |
|
"loss": 0.0685, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"eval_loss": 0.09454345703125, |
|
"eval_runtime": 252.9239, |
|
"eval_samples_per_second": 10.952, |
|
"eval_steps_per_second": 0.087, |
|
"eval_wer": 23.5571736785329, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 9.053391053391054e-06, |
|
"loss": 0.0688, |
|
"step": 2910 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 9.049783549783551e-06, |
|
"loss": 0.0657, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 9.046176046176048e-06, |
|
"loss": 0.068, |
|
"step": 2930 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 9.042568542568543e-06, |
|
"loss": 0.0692, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 9.03896103896104e-06, |
|
"loss": 0.0681, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 9.035353535353537e-06, |
|
"loss": 0.0699, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 9.031746031746032e-06, |
|
"loss": 0.0725, |
|
"step": 2970 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 9.028138528138529e-06, |
|
"loss": 0.0665, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 9.024531024531026e-06, |
|
"loss": 0.0666, |
|
"step": 2990 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 9.020923520923521e-06, |
|
"loss": 0.0685, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"eval_loss": 0.094970703125, |
|
"eval_runtime": 448.4694, |
|
"eval_samples_per_second": 6.177, |
|
"eval_steps_per_second": 0.049, |
|
"eval_wer": 23.722357065803667, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 9.017316017316018e-06, |
|
"loss": 0.0692, |
|
"step": 3010 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 9.013708513708515e-06, |
|
"loss": 0.067, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 9.010101010101012e-06, |
|
"loss": 0.0689, |
|
"step": 3030 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 9.006493506493509e-06, |
|
"loss": 0.0692, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 9.002886002886004e-06, |
|
"loss": 0.0696, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 8.9992784992785e-06, |
|
"loss": 0.0676, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 8.995670995670996e-06, |
|
"loss": 0.0696, |
|
"step": 3070 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 8.992063492063493e-06, |
|
"loss": 0.0702, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 8.98845598845599e-06, |
|
"loss": 0.0669, |
|
"step": 3090 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 8.984848484848485e-06, |
|
"loss": 0.0669, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"eval_loss": 0.0938720703125, |
|
"eval_runtime": 367.7769, |
|
"eval_samples_per_second": 7.532, |
|
"eval_steps_per_second": 0.06, |
|
"eval_wer": 23.59762675296656, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 8.981240981240982e-06, |
|
"loss": 0.0677, |
|
"step": 3110 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 8.977633477633479e-06, |
|
"loss": 0.0684, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 8.974025974025975e-06, |
|
"loss": 0.0677, |
|
"step": 3130 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 8.970418470418472e-06, |
|
"loss": 0.0658, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 8.966810966810968e-06, |
|
"loss": 0.0689, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 8.963203463203464e-06, |
|
"loss": 0.0653, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 8.95959595959596e-06, |
|
"loss": 0.0668, |
|
"step": 3170 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 8.955988455988457e-06, |
|
"loss": 0.066, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 8.952380952380953e-06, |
|
"loss": 0.0668, |
|
"step": 3190 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 8.948773448773449e-06, |
|
"loss": 0.0678, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"eval_loss": 0.09356689453125, |
|
"eval_runtime": 288.5179, |
|
"eval_samples_per_second": 9.601, |
|
"eval_steps_per_second": 0.076, |
|
"eval_wer": 23.415587918015103, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 8.945165945165946e-06, |
|
"loss": 0.0679, |
|
"step": 3210 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 8.941558441558442e-06, |
|
"loss": 0.0677, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 8.93795093795094e-06, |
|
"loss": 0.0694, |
|
"step": 3230 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 8.934343434343436e-06, |
|
"loss": 0.0678, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 8.930735930735931e-06, |
|
"loss": 0.0682, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 8.927128427128428e-06, |
|
"loss": 0.0674, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 8.923520923520924e-06, |
|
"loss": 0.0688, |
|
"step": 3270 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 8.91991341991342e-06, |
|
"loss": 0.0684, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 8.916305916305917e-06, |
|
"loss": 0.0659, |
|
"step": 3290 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 8.912698412698414e-06, |
|
"loss": 0.0692, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"eval_loss": 0.09393310546875, |
|
"eval_runtime": 478.9048, |
|
"eval_samples_per_second": 5.784, |
|
"eval_steps_per_second": 0.046, |
|
"eval_wer": 23.358279395900755, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 8.90909090909091e-06, |
|
"loss": 0.0698, |
|
"step": 3310 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 8.905483405483406e-06, |
|
"loss": 0.069, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 8.901875901875903e-06, |
|
"loss": 0.0668, |
|
"step": 3330 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 8.8982683982684e-06, |
|
"loss": 0.0663, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 8.894660894660895e-06, |
|
"loss": 0.0676, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 8.891053391053392e-06, |
|
"loss": 0.0657, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 8.887445887445887e-06, |
|
"loss": 0.0667, |
|
"step": 3370 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 8.883838383838384e-06, |
|
"loss": 0.0657, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 8.880230880230881e-06, |
|
"loss": 0.0647, |
|
"step": 3390 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 8.876623376623378e-06, |
|
"loss": 0.0664, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"eval_loss": 0.09320068359375, |
|
"eval_runtime": 804.4129, |
|
"eval_samples_per_second": 3.444, |
|
"eval_steps_per_second": 0.027, |
|
"eval_wer": 23.52683387270766, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 8.873015873015873e-06, |
|
"loss": 0.0668, |
|
"step": 3410 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 8.86940836940837e-06, |
|
"loss": 0.0662, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 8.865800865800867e-06, |
|
"loss": 0.0656, |
|
"step": 3430 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 8.862193362193364e-06, |
|
"loss": 0.0668, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 8.85858585858586e-06, |
|
"loss": 0.0688, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 8.854978354978356e-06, |
|
"loss": 0.0667, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 8.851370851370851e-06, |
|
"loss": 0.0703, |
|
"step": 3470 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 8.847763347763348e-06, |
|
"loss": 0.0703, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 8.844155844155845e-06, |
|
"loss": 0.0678, |
|
"step": 3490 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 8.840548340548342e-06, |
|
"loss": 0.0684, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"eval_loss": 0.09197998046875, |
|
"eval_runtime": 577.6791, |
|
"eval_samples_per_second": 4.795, |
|
"eval_steps_per_second": 0.038, |
|
"eval_wer": 23.43918554476807, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 8.836940836940837e-06, |
|
"loss": 0.0679, |
|
"step": 3510 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 8.833333333333334e-06, |
|
"loss": 0.068, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 8.829725829725831e-06, |
|
"loss": 0.0673, |
|
"step": 3530 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 8.826118326118328e-06, |
|
"loss": 0.0646, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 8.822510822510823e-06, |
|
"loss": 0.0667, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 8.81890331890332e-06, |
|
"loss": 0.0664, |
|
"step": 3560 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 8.815295815295815e-06, |
|
"loss": 0.0669, |
|
"step": 3570 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 8.811688311688312e-06, |
|
"loss": 0.0681, |
|
"step": 3580 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 8.808080808080809e-06, |
|
"loss": 0.0689, |
|
"step": 3590 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 8.804473304473306e-06, |
|
"loss": 0.0681, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"eval_loss": 0.0921630859375, |
|
"eval_runtime": 316.0563, |
|
"eval_samples_per_second": 8.764, |
|
"eval_steps_per_second": 0.07, |
|
"eval_wer": 22.82564724919094, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 8.800865800865803e-06, |
|
"loss": 0.0683, |
|
"step": 3610 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 8.797258297258298e-06, |
|
"loss": 0.0662, |
|
"step": 3620 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 8.793650793650795e-06, |
|
"loss": 0.0705, |
|
"step": 3630 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 8.79004329004329e-06, |
|
"loss": 0.0668, |
|
"step": 3640 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 8.786435786435787e-06, |
|
"loss": 0.0649, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 8.782828282828284e-06, |
|
"loss": 0.0676, |
|
"step": 3660 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 8.779220779220779e-06, |
|
"loss": 0.067, |
|
"step": 3670 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 8.775613275613276e-06, |
|
"loss": 0.068, |
|
"step": 3680 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 8.772005772005773e-06, |
|
"loss": 0.0656, |
|
"step": 3690 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 8.76839826839827e-06, |
|
"loss": 0.0668, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"eval_loss": 0.09124755859375, |
|
"eval_runtime": 443.766, |
|
"eval_samples_per_second": 6.242, |
|
"eval_steps_per_second": 0.05, |
|
"eval_wer": 22.866100323624593, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 8.764790764790767e-06, |
|
"loss": 0.0656, |
|
"step": 3710 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 8.761183261183262e-06, |
|
"loss": 0.0672, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 8.757575757575759e-06, |
|
"loss": 0.0691, |
|
"step": 3730 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 8.753968253968254e-06, |
|
"loss": 0.0658, |
|
"step": 3740 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 8.75036075036075e-06, |
|
"loss": 0.067, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 8.746753246753248e-06, |
|
"loss": 0.0659, |
|
"step": 3760 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 8.743145743145743e-06, |
|
"loss": 0.0644, |
|
"step": 3770 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 8.73953823953824e-06, |
|
"loss": 0.0664, |
|
"step": 3780 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 8.735930735930737e-06, |
|
"loss": 0.0674, |
|
"step": 3790 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 8.732323232323234e-06, |
|
"loss": 0.0678, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"eval_loss": 0.09130859375, |
|
"eval_runtime": 455.9937, |
|
"eval_samples_per_second": 6.075, |
|
"eval_steps_per_second": 0.048, |
|
"eval_wer": 23.580771305285868, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 8.72871572871573e-06, |
|
"loss": 0.0658, |
|
"step": 3810 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 8.725108225108226e-06, |
|
"loss": 0.0646, |
|
"step": 3820 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 8.721500721500722e-06, |
|
"loss": 0.0649, |
|
"step": 3830 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 8.717893217893218e-06, |
|
"loss": 0.067, |
|
"step": 3840 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 8.714285714285715e-06, |
|
"loss": 0.0656, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 8.710678210678211e-06, |
|
"loss": 0.0685, |
|
"step": 3860 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 8.707070707070707e-06, |
|
"loss": 0.0652, |
|
"step": 3870 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 8.703463203463204e-06, |
|
"loss": 0.0686, |
|
"step": 3880 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 8.6998556998557e-06, |
|
"loss": 0.0669, |
|
"step": 3890 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 8.696248196248197e-06, |
|
"loss": 0.064, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"eval_loss": 0.09088134765625, |
|
"eval_runtime": 323.8053, |
|
"eval_samples_per_second": 8.555, |
|
"eval_steps_per_second": 0.068, |
|
"eval_wer": 23.15601402373247, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 8.692640692640694e-06, |
|
"loss": 0.0667, |
|
"step": 3910 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 8.68903318903319e-06, |
|
"loss": 0.0635, |
|
"step": 3920 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 8.685425685425686e-06, |
|
"loss": 0.067, |
|
"step": 3930 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 8.681818181818182e-06, |
|
"loss": 0.0644, |
|
"step": 3940 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 8.678210678210678e-06, |
|
"loss": 0.0629, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 8.674603174603175e-06, |
|
"loss": 0.0667, |
|
"step": 3960 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 8.670995670995672e-06, |
|
"loss": 0.0664, |
|
"step": 3970 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 8.667388167388167e-06, |
|
"loss": 0.0658, |
|
"step": 3980 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 8.663780663780664e-06, |
|
"loss": 0.0682, |
|
"step": 3990 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 8.660173160173161e-06, |
|
"loss": 0.0663, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"eval_loss": 0.09112548828125, |
|
"eval_runtime": 454.1676, |
|
"eval_samples_per_second": 6.099, |
|
"eval_steps_per_second": 0.048, |
|
"eval_wer": 23.826860841423947, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 8.657287157287158e-06, |
|
"loss": 0.0664, |
|
"step": 4010 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 8.653679653679655e-06, |
|
"loss": 0.0658, |
|
"step": 4020 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 8.650072150072152e-06, |
|
"loss": 0.0678, |
|
"step": 4030 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 8.646464646464647e-06, |
|
"loss": 0.0699, |
|
"step": 4040 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 8.642857142857144e-06, |
|
"loss": 0.0662, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 8.639249639249639e-06, |
|
"loss": 0.0677, |
|
"step": 4060 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 8.635642135642136e-06, |
|
"loss": 0.0651, |
|
"step": 4070 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 8.632034632034633e-06, |
|
"loss": 0.0681, |
|
"step": 4080 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 8.62842712842713e-06, |
|
"loss": 0.0658, |
|
"step": 4090 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 8.624819624819626e-06, |
|
"loss": 0.0657, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"eval_loss": 0.0904541015625, |
|
"eval_runtime": 483.7028, |
|
"eval_samples_per_second": 5.727, |
|
"eval_steps_per_second": 0.045, |
|
"eval_wer": 22.562702265372167, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 8.621212121212122e-06, |
|
"loss": 0.0668, |
|
"step": 4110 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 8.617604617604618e-06, |
|
"loss": 0.066, |
|
"step": 4120 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 8.613997113997115e-06, |
|
"loss": 0.0646, |
|
"step": 4130 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 8.61038961038961e-06, |
|
"loss": 0.0657, |
|
"step": 4140 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 8.606782106782107e-06, |
|
"loss": 0.0658, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 8.603174603174604e-06, |
|
"loss": 0.0666, |
|
"step": 4160 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 8.5995670995671e-06, |
|
"loss": 0.0677, |
|
"step": 4170 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 8.595959595959596e-06, |
|
"loss": 0.0667, |
|
"step": 4180 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 8.592352092352093e-06, |
|
"loss": 0.0671, |
|
"step": 4190 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 8.58874458874459e-06, |
|
"loss": 0.0679, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"eval_loss": 0.090576171875, |
|
"eval_runtime": 527.5145, |
|
"eval_samples_per_second": 5.251, |
|
"eval_steps_per_second": 0.042, |
|
"eval_wer": 22.808791801510246, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 8.585137085137087e-06, |
|
"loss": 0.0684, |
|
"step": 4210 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 8.581529581529582e-06, |
|
"loss": 0.0677, |
|
"step": 4220 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 8.57792207792208e-06, |
|
"loss": 0.0646, |
|
"step": 4230 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 8.574314574314574e-06, |
|
"loss": 0.068, |
|
"step": 4240 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 8.570707070707071e-06, |
|
"loss": 0.0668, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 8.567099567099568e-06, |
|
"loss": 0.0656, |
|
"step": 4260 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 8.563492063492063e-06, |
|
"loss": 0.0654, |
|
"step": 4270 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 8.55988455988456e-06, |
|
"loss": 0.0648, |
|
"step": 4280 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 8.556277056277057e-06, |
|
"loss": 0.067, |
|
"step": 4290 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 8.552669552669554e-06, |
|
"loss": 0.0675, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"eval_loss": 0.090087890625, |
|
"eval_runtime": 250.2835, |
|
"eval_samples_per_second": 11.067, |
|
"eval_steps_per_second": 0.088, |
|
"eval_wer": 22.515507011866234, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 8.549062049062051e-06, |
|
"loss": 0.0657, |
|
"step": 4310 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 8.545454545454546e-06, |
|
"loss": 0.0648, |
|
"step": 4320 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 8.541847041847043e-06, |
|
"loss": 0.0679, |
|
"step": 4330 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 8.538239538239538e-06, |
|
"loss": 0.0682, |
|
"step": 4340 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 8.534632034632035e-06, |
|
"loss": 0.0693, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 8.531024531024532e-06, |
|
"loss": 0.0637, |
|
"step": 4360 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 8.527417027417027e-06, |
|
"loss": 0.068, |
|
"step": 4370 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 8.523809523809524e-06, |
|
"loss": 0.0688, |
|
"step": 4380 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 8.520202020202021e-06, |
|
"loss": 0.0643, |
|
"step": 4390 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 8.516594516594518e-06, |
|
"loss": 0.0673, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"eval_loss": 0.0897216796875, |
|
"eval_runtime": 333.8313, |
|
"eval_samples_per_second": 8.298, |
|
"eval_steps_per_second": 0.066, |
|
"eval_wer": 22.43797195253506, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 8.512987012987015e-06, |
|
"loss": 0.0634, |
|
"step": 4410 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 8.50937950937951e-06, |
|
"loss": 0.0683, |
|
"step": 4420 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 8.505772005772007e-06, |
|
"loss": 0.0673, |
|
"step": 4430 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 8.502164502164502e-06, |
|
"loss": 0.0677, |
|
"step": 4440 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 8.498556998556999e-06, |
|
"loss": 0.0676, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 8.494949494949496e-06, |
|
"loss": 0.0686, |
|
"step": 4460 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 8.491341991341993e-06, |
|
"loss": 0.0678, |
|
"step": 4470 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 8.487734487734488e-06, |
|
"loss": 0.0663, |
|
"step": 4480 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 8.484126984126985e-06, |
|
"loss": 0.0659, |
|
"step": 4490 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 8.480519480519482e-06, |
|
"loss": 0.0639, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"eval_loss": 0.08935546875, |
|
"eval_runtime": 1011.4705, |
|
"eval_samples_per_second": 2.739, |
|
"eval_steps_per_second": 0.022, |
|
"eval_wer": 22.55596008629989, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 8.476911976911979e-06, |
|
"loss": 0.0664, |
|
"step": 4510 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 8.473304473304474e-06, |
|
"loss": 0.0662, |
|
"step": 4520 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 8.46969696969697e-06, |
|
"loss": 0.067, |
|
"step": 4530 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 8.466089466089466e-06, |
|
"loss": 0.0661, |
|
"step": 4540 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 8.462481962481963e-06, |
|
"loss": 0.0655, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 8.45887445887446e-06, |
|
"loss": 0.0655, |
|
"step": 4560 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 8.455266955266957e-06, |
|
"loss": 0.0641, |
|
"step": 4570 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 8.451659451659452e-06, |
|
"loss": 0.0658, |
|
"step": 4580 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 8.448051948051949e-06, |
|
"loss": 0.0646, |
|
"step": 4590 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 8.444444444444446e-06, |
|
"loss": 0.0675, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"eval_loss": 0.088623046875, |
|
"eval_runtime": 296.69, |
|
"eval_samples_per_second": 9.336, |
|
"eval_steps_per_second": 0.074, |
|
"eval_wer": 22.431229773462782, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 8.440836940836943e-06, |
|
"loss": 0.0646, |
|
"step": 4610 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 8.437229437229438e-06, |
|
"loss": 0.0634, |
|
"step": 4620 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 8.433621933621935e-06, |
|
"loss": 0.0665, |
|
"step": 4630 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 8.43001443001443e-06, |
|
"loss": 0.0659, |
|
"step": 4640 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 8.426406926406927e-06, |
|
"loss": 0.0661, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 8.422799422799424e-06, |
|
"loss": 0.0665, |
|
"step": 4660 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 8.41919191919192e-06, |
|
"loss": 0.0633, |
|
"step": 4670 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 8.415584415584416e-06, |
|
"loss": 0.0644, |
|
"step": 4680 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 8.411976911976913e-06, |
|
"loss": 0.0632, |
|
"step": 4690 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 8.40836940836941e-06, |
|
"loss": 0.0621, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"eval_loss": 0.08856201171875, |
|
"eval_runtime": 349.513, |
|
"eval_samples_per_second": 7.925, |
|
"eval_steps_per_second": 0.063, |
|
"eval_wer": 22.518878101402372, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 8.404761904761905e-06, |
|
"loss": 0.0652, |
|
"step": 4710 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 8.401154401154402e-06, |
|
"loss": 0.0629, |
|
"step": 4720 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 8.397546897546899e-06, |
|
"loss": 0.0657, |
|
"step": 4730 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 8.393939393939394e-06, |
|
"loss": 0.0645, |
|
"step": 4740 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 8.39033189033189e-06, |
|
"loss": 0.0667, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 8.386724386724387e-06, |
|
"loss": 0.0634, |
|
"step": 4760 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 8.383116883116884e-06, |
|
"loss": 0.0666, |
|
"step": 4770 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 8.379509379509381e-06, |
|
"loss": 0.0663, |
|
"step": 4780 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 8.375901875901876e-06, |
|
"loss": 0.0686, |
|
"step": 4790 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 8.372294372294373e-06, |
|
"loss": 0.0658, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"eval_loss": 0.08856201171875, |
|
"eval_runtime": 691.7141, |
|
"eval_samples_per_second": 4.005, |
|
"eval_steps_per_second": 0.032, |
|
"eval_wer": 22.249190938511326, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 8.368686868686869e-06, |
|
"loss": 0.0649, |
|
"step": 4810 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 8.365079365079365e-06, |
|
"loss": 0.064, |
|
"step": 4820 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 8.361471861471862e-06, |
|
"loss": 0.0639, |
|
"step": 4830 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 8.357864357864358e-06, |
|
"loss": 0.0658, |
|
"step": 4840 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 8.354256854256854e-06, |
|
"loss": 0.0712, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 8.350649350649351e-06, |
|
"loss": 0.0664, |
|
"step": 4860 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 8.347041847041848e-06, |
|
"loss": 0.0611, |
|
"step": 4870 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 8.343434343434345e-06, |
|
"loss": 0.0675, |
|
"step": 4880 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 8.33982683982684e-06, |
|
"loss": 0.0659, |
|
"step": 4890 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 8.336219336219337e-06, |
|
"loss": 0.0654, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"eval_loss": 0.08856201171875, |
|
"eval_runtime": 583.1687, |
|
"eval_samples_per_second": 4.75, |
|
"eval_steps_per_second": 0.038, |
|
"eval_wer": 22.090749730312837, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 8.332611832611832e-06, |
|
"loss": 0.0679, |
|
"step": 4910 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 8.32900432900433e-06, |
|
"loss": 0.0667, |
|
"step": 4920 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 8.325396825396826e-06, |
|
"loss": 0.0661, |
|
"step": 4930 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 8.321789321789321e-06, |
|
"loss": 0.0635, |
|
"step": 4940 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 8.318181818181818e-06, |
|
"loss": 0.0645, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 8.314574314574315e-06, |
|
"loss": 0.0663, |
|
"step": 4960 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 8.310966810966812e-06, |
|
"loss": 0.0646, |
|
"step": 4970 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 8.307359307359309e-06, |
|
"loss": 0.0631, |
|
"step": 4980 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 8.303751803751804e-06, |
|
"loss": 0.0659, |
|
"step": 4990 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 8.300144300144301e-06, |
|
"loss": 0.0639, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"eval_loss": 0.0875244140625, |
|
"eval_runtime": 638.4901, |
|
"eval_samples_per_second": 4.338, |
|
"eval_steps_per_second": 0.034, |
|
"eval_wer": 22.249190938511326, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 8.297258297258298e-06, |
|
"loss": 0.0649, |
|
"step": 5010 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 8.293650793650794e-06, |
|
"loss": 0.0653, |
|
"step": 5020 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 8.29004329004329e-06, |
|
"loss": 0.0672, |
|
"step": 5030 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 8.286435786435787e-06, |
|
"loss": 0.0646, |
|
"step": 5040 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 8.282828282828283e-06, |
|
"loss": 0.0639, |
|
"step": 5050 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 8.27922077922078e-06, |
|
"loss": 0.0671, |
|
"step": 5060 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 8.275613275613277e-06, |
|
"loss": 0.0695, |
|
"step": 5070 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 8.272005772005772e-06, |
|
"loss": 0.0654, |
|
"step": 5080 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 8.26839826839827e-06, |
|
"loss": 0.0654, |
|
"step": 5090 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 8.264790764790766e-06, |
|
"loss": 0.0631, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"eval_loss": 0.08734130859375, |
|
"eval_runtime": 259.0875, |
|
"eval_samples_per_second": 10.691, |
|
"eval_steps_per_second": 0.085, |
|
"eval_wer": 22.239077669902912, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 8.261183261183261e-06, |
|
"loss": 0.0656, |
|
"step": 5110 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 8.257575757575758e-06, |
|
"loss": 0.0644, |
|
"step": 5120 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 8.253968253968254e-06, |
|
"loss": 0.0634, |
|
"step": 5130 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 8.25036075036075e-06, |
|
"loss": 0.0659, |
|
"step": 5140 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 8.246753246753247e-06, |
|
"loss": 0.0656, |
|
"step": 5150 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 8.243145743145744e-06, |
|
"loss": 0.0633, |
|
"step": 5160 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 8.239538239538241e-06, |
|
"loss": 0.0665, |
|
"step": 5170 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 8.235930735930736e-06, |
|
"loss": 0.0652, |
|
"step": 5180 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 8.232323232323233e-06, |
|
"loss": 0.0626, |
|
"step": 5190 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 8.22871572871573e-06, |
|
"loss": 0.0645, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"eval_loss": 0.0869140625, |
|
"eval_runtime": 493.2946, |
|
"eval_samples_per_second": 5.615, |
|
"eval_steps_per_second": 0.045, |
|
"eval_wer": 22.31324163969795, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 8.225108225108225e-06, |
|
"loss": 0.0644, |
|
"step": 5210 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 8.221500721500722e-06, |
|
"loss": 0.0649, |
|
"step": 5220 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 8.217893217893217e-06, |
|
"loss": 0.0658, |
|
"step": 5230 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 8.214285714285714e-06, |
|
"loss": 0.066, |
|
"step": 5240 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 8.210678210678211e-06, |
|
"loss": 0.0634, |
|
"step": 5250 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 8.207070707070708e-06, |
|
"loss": 0.065, |
|
"step": 5260 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 8.203463203463205e-06, |
|
"loss": 0.0633, |
|
"step": 5270 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 8.199855699855702e-06, |
|
"loss": 0.0628, |
|
"step": 5280 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 8.196248196248197e-06, |
|
"loss": 0.065, |
|
"step": 5290 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 8.192640692640694e-06, |
|
"loss": 0.0665, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"eval_loss": 0.0872802734375, |
|
"eval_runtime": 673.3399, |
|
"eval_samples_per_second": 4.114, |
|
"eval_steps_per_second": 0.033, |
|
"eval_wer": 22.198624595469255, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 8.18903318903319e-06, |
|
"loss": 0.0622, |
|
"step": 5310 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 8.185425685425686e-06, |
|
"loss": 0.0666, |
|
"step": 5320 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 8.181818181818183e-06, |
|
"loss": 0.0666, |
|
"step": 5330 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 8.178210678210678e-06, |
|
"loss": 0.0631, |
|
"step": 5340 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 8.174603174603175e-06, |
|
"loss": 0.0632, |
|
"step": 5350 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 8.170995670995672e-06, |
|
"loss": 0.0623, |
|
"step": 5360 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 8.167388167388169e-06, |
|
"loss": 0.0657, |
|
"step": 5370 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 8.163780663780666e-06, |
|
"loss": 0.0651, |
|
"step": 5380 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 8.160173160173161e-06, |
|
"loss": 0.0655, |
|
"step": 5390 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 8.156565656565658e-06, |
|
"loss": 0.0641, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"eval_loss": 0.0867919921875, |
|
"eval_runtime": 648.7299, |
|
"eval_samples_per_second": 4.27, |
|
"eval_steps_per_second": 0.034, |
|
"eval_wer": 22.228964401294498, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 8.152958152958153e-06, |
|
"loss": 0.0657, |
|
"step": 5410 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 8.14935064935065e-06, |
|
"loss": 0.065, |
|
"step": 5420 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 8.145743145743147e-06, |
|
"loss": 0.0652, |
|
"step": 5430 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 8.142135642135642e-06, |
|
"loss": 0.064, |
|
"step": 5440 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 8.138528138528139e-06, |
|
"loss": 0.0557, |
|
"step": 5450 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 8.134920634920636e-06, |
|
"loss": 0.0552, |
|
"step": 5460 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 8.131313131313133e-06, |
|
"loss": 0.0555, |
|
"step": 5470 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 8.12770562770563e-06, |
|
"loss": 0.0545, |
|
"step": 5480 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 8.124098124098125e-06, |
|
"loss": 0.0565, |
|
"step": 5490 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 8.120490620490622e-06, |
|
"loss": 0.0558, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"eval_loss": 0.0870361328125, |
|
"eval_runtime": 693.0613, |
|
"eval_samples_per_second": 3.997, |
|
"eval_steps_per_second": 0.032, |
|
"eval_wer": 22.063781014023732, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 8.116883116883117e-06, |
|
"loss": 0.0541, |
|
"step": 5510 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 8.113275613275614e-06, |
|
"loss": 0.0544, |
|
"step": 5520 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 8.10966810966811e-06, |
|
"loss": 0.0532, |
|
"step": 5530 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 8.106060606060606e-06, |
|
"loss": 0.054, |
|
"step": 5540 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 8.102453102453103e-06, |
|
"loss": 0.0555, |
|
"step": 5550 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 8.0988455988456e-06, |
|
"loss": 0.0543, |
|
"step": 5560 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 8.095238095238097e-06, |
|
"loss": 0.0558, |
|
"step": 5570 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 8.091630591630593e-06, |
|
"loss": 0.0523, |
|
"step": 5580 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 8.088023088023089e-06, |
|
"loss": 0.0546, |
|
"step": 5590 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 8.084415584415586e-06, |
|
"loss": 0.0556, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"eval_loss": 0.08795166015625, |
|
"eval_runtime": 712.0699, |
|
"eval_samples_per_second": 3.89, |
|
"eval_steps_per_second": 0.031, |
|
"eval_wer": 21.96264832793959, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 8.08080808080808e-06, |
|
"loss": 0.0568, |
|
"step": 5610 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 8.077200577200578e-06, |
|
"loss": 0.053, |
|
"step": 5620 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 8.073593073593075e-06, |
|
"loss": 0.0541, |
|
"step": 5630 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 8.069985569985571e-06, |
|
"loss": 0.0538, |
|
"step": 5640 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 8.066378066378067e-06, |
|
"loss": 0.053, |
|
"step": 5650 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 8.062770562770564e-06, |
|
"loss": 0.055, |
|
"step": 5660 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 8.05916305916306e-06, |
|
"loss": 0.0539, |
|
"step": 5670 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 8.055555555555557e-06, |
|
"loss": 0.056, |
|
"step": 5680 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 8.051948051948052e-06, |
|
"loss": 0.0557, |
|
"step": 5690 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 8.04834054834055e-06, |
|
"loss": 0.0562, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"eval_loss": 0.08782958984375, |
|
"eval_runtime": 270.8387, |
|
"eval_samples_per_second": 10.227, |
|
"eval_steps_per_second": 0.081, |
|
"eval_wer": 22.10423408845739, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 8.044733044733045e-06, |
|
"loss": 0.0554, |
|
"step": 5710 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 8.041125541125541e-06, |
|
"loss": 0.0524, |
|
"step": 5720 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 8.037518037518038e-06, |
|
"loss": 0.0554, |
|
"step": 5730 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 8.033910533910535e-06, |
|
"loss": 0.0561, |
|
"step": 5740 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 8.03030303030303e-06, |
|
"loss": 0.0535, |
|
"step": 5750 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 8.026695526695527e-06, |
|
"loss": 0.0548, |
|
"step": 5760 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 8.023088023088024e-06, |
|
"loss": 0.0528, |
|
"step": 5770 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 8.019480519480521e-06, |
|
"loss": 0.057, |
|
"step": 5780 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 8.015873015873016e-06, |
|
"loss": 0.0579, |
|
"step": 5790 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 8.012265512265513e-06, |
|
"loss": 0.0547, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"eval_loss": 0.0888671875, |
|
"eval_runtime": 251.3262, |
|
"eval_samples_per_second": 11.022, |
|
"eval_steps_per_second": 0.088, |
|
"eval_wer": 22.05029665587918, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 8.008658008658008e-06, |
|
"loss": 0.0538, |
|
"step": 5810 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 8.005050505050505e-06, |
|
"loss": 0.0565, |
|
"step": 5820 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 8.001443001443002e-06, |
|
"loss": 0.0551, |
|
"step": 5830 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 7.997835497835499e-06, |
|
"loss": 0.0537, |
|
"step": 5840 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 7.994227994227994e-06, |
|
"loss": 0.0545, |
|
"step": 5850 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 7.990620490620491e-06, |
|
"loss": 0.0547, |
|
"step": 5860 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 7.987012987012988e-06, |
|
"loss": 0.0549, |
|
"step": 5870 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 7.983405483405483e-06, |
|
"loss": 0.0533, |
|
"step": 5880 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 7.97979797979798e-06, |
|
"loss": 0.0535, |
|
"step": 5890 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 7.976190476190477e-06, |
|
"loss": 0.0553, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"eval_loss": 0.08807373046875, |
|
"eval_runtime": 528.7421, |
|
"eval_samples_per_second": 5.239, |
|
"eval_steps_per_second": 0.042, |
|
"eval_wer": 22.218851132686083, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 7.972582972582972e-06, |
|
"loss": 0.0531, |
|
"step": 5910 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 7.96897546897547e-06, |
|
"loss": 0.0553, |
|
"step": 5920 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 7.965367965367966e-06, |
|
"loss": 0.0548, |
|
"step": 5930 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 7.961760461760463e-06, |
|
"loss": 0.0533, |
|
"step": 5940 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 7.95815295815296e-06, |
|
"loss": 0.0531, |
|
"step": 5950 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 7.954545454545455e-06, |
|
"loss": 0.0536, |
|
"step": 5960 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 7.950937950937952e-06, |
|
"loss": 0.0564, |
|
"step": 5970 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 7.947330447330447e-06, |
|
"loss": 0.0542, |
|
"step": 5980 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 7.943722943722944e-06, |
|
"loss": 0.0551, |
|
"step": 5990 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 7.940115440115441e-06, |
|
"loss": 0.0547, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"eval_loss": 0.08795166015625, |
|
"eval_runtime": 556.2644, |
|
"eval_samples_per_second": 4.98, |
|
"eval_steps_per_second": 0.04, |
|
"eval_wer": 21.98287486515642, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 7.936868686868688e-06, |
|
"loss": 0.0552, |
|
"step": 6010 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 7.933621933621934e-06, |
|
"loss": 0.0556, |
|
"step": 6020 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 7.930014430014431e-06, |
|
"loss": 0.0536, |
|
"step": 6030 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 7.926406926406926e-06, |
|
"loss": 0.056, |
|
"step": 6040 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 7.922799422799423e-06, |
|
"loss": 0.0556, |
|
"step": 6050 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 7.91919191919192e-06, |
|
"loss": 0.0551, |
|
"step": 6060 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 7.915584415584417e-06, |
|
"loss": 0.0552, |
|
"step": 6070 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 7.911976911976912e-06, |
|
"loss": 0.0558, |
|
"step": 6080 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 7.90836940836941e-06, |
|
"loss": 0.0531, |
|
"step": 6090 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 7.904761904761904e-06, |
|
"loss": 0.0544, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"eval_loss": 0.08709716796875, |
|
"eval_runtime": 263.4486, |
|
"eval_samples_per_second": 10.514, |
|
"eval_steps_per_second": 0.084, |
|
"eval_wer": 22.110976267529665, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 7.901154401154401e-06, |
|
"loss": 0.0523, |
|
"step": 6110 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 7.897546897546898e-06, |
|
"loss": 0.0552, |
|
"step": 6120 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 7.893939393939395e-06, |
|
"loss": 0.0545, |
|
"step": 6130 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 7.890331890331892e-06, |
|
"loss": 0.0554, |
|
"step": 6140 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 7.886724386724387e-06, |
|
"loss": 0.0555, |
|
"step": 6150 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 7.883116883116884e-06, |
|
"loss": 0.0544, |
|
"step": 6160 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 7.879509379509381e-06, |
|
"loss": 0.0523, |
|
"step": 6170 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 7.875901875901876e-06, |
|
"loss": 0.0547, |
|
"step": 6180 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 7.872294372294373e-06, |
|
"loss": 0.0536, |
|
"step": 6190 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 7.868686868686868e-06, |
|
"loss": 0.0573, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"eval_loss": 0.08660888671875, |
|
"eval_runtime": 367.0765, |
|
"eval_samples_per_second": 7.546, |
|
"eval_steps_per_second": 0.06, |
|
"eval_wer": 21.635652642934197, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 7.865079365079365e-06, |
|
"loss": 0.0541, |
|
"step": 6210 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 7.861471861471862e-06, |
|
"loss": 0.053, |
|
"step": 6220 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 7.857864357864359e-06, |
|
"loss": 0.0558, |
|
"step": 6230 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 7.854256854256856e-06, |
|
"loss": 0.0543, |
|
"step": 6240 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 7.850649350649351e-06, |
|
"loss": 0.0527, |
|
"step": 6250 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 7.847041847041848e-06, |
|
"loss": 0.0535, |
|
"step": 6260 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 7.843434343434345e-06, |
|
"loss": 0.0537, |
|
"step": 6270 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 7.83982683982684e-06, |
|
"loss": 0.0563, |
|
"step": 6280 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 7.836219336219337e-06, |
|
"loss": 0.0555, |
|
"step": 6290 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 7.832611832611832e-06, |
|
"loss": 0.0562, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"eval_loss": 0.08721923828125, |
|
"eval_runtime": 637.4835, |
|
"eval_samples_per_second": 4.345, |
|
"eval_steps_per_second": 0.035, |
|
"eval_wer": 21.97950377562028, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 7.829004329004329e-06, |
|
"loss": 0.0571, |
|
"step": 6310 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 7.825396825396826e-06, |
|
"loss": 0.0566, |
|
"step": 6320 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 7.821789321789323e-06, |
|
"loss": 0.0547, |
|
"step": 6330 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 7.81818181818182e-06, |
|
"loss": 0.0557, |
|
"step": 6340 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 7.814574314574315e-06, |
|
"loss": 0.0564, |
|
"step": 6350 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 7.810966810966812e-06, |
|
"loss": 0.0552, |
|
"step": 6360 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 7.807359307359309e-06, |
|
"loss": 0.0539, |
|
"step": 6370 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 7.803751803751804e-06, |
|
"loss": 0.0536, |
|
"step": 6380 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 7.8001443001443e-06, |
|
"loss": 0.0543, |
|
"step": 6390 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 7.796536796536796e-06, |
|
"loss": 0.0551, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"eval_loss": 0.0872802734375, |
|
"eval_runtime": 686.505, |
|
"eval_samples_per_second": 4.035, |
|
"eval_steps_per_second": 0.032, |
|
"eval_wer": 21.976132686084142, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 7.792929292929293e-06, |
|
"loss": 0.0549, |
|
"step": 6410 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 7.78932178932179e-06, |
|
"loss": 0.055, |
|
"step": 6420 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 7.785714285714287e-06, |
|
"loss": 0.0545, |
|
"step": 6430 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 7.782106782106784e-06, |
|
"loss": 0.0532, |
|
"step": 6440 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 7.77849927849928e-06, |
|
"loss": 0.0558, |
|
"step": 6450 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 7.774891774891776e-06, |
|
"loss": 0.0538, |
|
"step": 6460 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 7.771284271284273e-06, |
|
"loss": 0.0534, |
|
"step": 6470 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 7.767676767676768e-06, |
|
"loss": 0.0516, |
|
"step": 6480 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 7.764069264069265e-06, |
|
"loss": 0.0537, |
|
"step": 6490 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 7.760461760461762e-06, |
|
"loss": 0.0556, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"eval_loss": 0.08697509765625, |
|
"eval_runtime": 549.3253, |
|
"eval_samples_per_second": 5.043, |
|
"eval_steps_per_second": 0.04, |
|
"eval_wer": 22.01321467098166, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 7.756854256854257e-06, |
|
"loss": 0.0535, |
|
"step": 6510 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 7.753246753246754e-06, |
|
"loss": 0.0539, |
|
"step": 6520 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 7.74963924963925e-06, |
|
"loss": 0.0531, |
|
"step": 6530 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 7.746031746031747e-06, |
|
"loss": 0.0546, |
|
"step": 6540 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 7.742424242424244e-06, |
|
"loss": 0.0552, |
|
"step": 6550 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 7.73881673881674e-06, |
|
"loss": 0.0565, |
|
"step": 6560 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 7.735209235209236e-06, |
|
"loss": 0.0551, |
|
"step": 6570 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 7.731601731601732e-06, |
|
"loss": 0.0551, |
|
"step": 6580 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 7.727994227994229e-06, |
|
"loss": 0.0534, |
|
"step": 6590 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 7.724386724386725e-06, |
|
"loss": 0.0558, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"eval_loss": 0.0870361328125, |
|
"eval_runtime": 256.8196, |
|
"eval_samples_per_second": 10.786, |
|
"eval_steps_per_second": 0.086, |
|
"eval_wer": 21.672734627831716, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 7.72077922077922e-06, |
|
"loss": 0.0564, |
|
"step": 6610 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 7.717171717171717e-06, |
|
"loss": 0.0572, |
|
"step": 6620 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 7.713564213564214e-06, |
|
"loss": 0.0569, |
|
"step": 6630 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 7.709956709956711e-06, |
|
"loss": 0.0538, |
|
"step": 6640 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 7.706349206349208e-06, |
|
"loss": 0.0541, |
|
"step": 6650 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 7.702741702741703e-06, |
|
"loss": 0.0542, |
|
"step": 6660 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 7.6991341991342e-06, |
|
"loss": 0.0532, |
|
"step": 6670 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 7.695526695526695e-06, |
|
"loss": 0.0552, |
|
"step": 6680 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 7.691919191919192e-06, |
|
"loss": 0.0553, |
|
"step": 6690 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 7.68831168831169e-06, |
|
"loss": 0.0565, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"eval_loss": 0.08868408203125, |
|
"eval_runtime": 462.3577, |
|
"eval_samples_per_second": 5.991, |
|
"eval_steps_per_second": 0.048, |
|
"eval_wer": 22.646979503775622, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 7.684704184704186e-06, |
|
"loss": 0.0555, |
|
"step": 6710 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 7.681096681096681e-06, |
|
"loss": 0.0566, |
|
"step": 6720 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 7.677489177489178e-06, |
|
"loss": 0.0541, |
|
"step": 6730 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 7.673881673881675e-06, |
|
"loss": 0.055, |
|
"step": 6740 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 7.670274170274172e-06, |
|
"loss": 0.055, |
|
"step": 6750 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 7.666666666666667e-06, |
|
"loss": 0.0544, |
|
"step": 6760 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 7.663059163059164e-06, |
|
"loss": 0.0553, |
|
"step": 6770 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 7.65945165945166e-06, |
|
"loss": 0.0538, |
|
"step": 6780 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 7.655844155844156e-06, |
|
"loss": 0.0567, |
|
"step": 6790 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 7.652236652236653e-06, |
|
"loss": 0.0549, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"eval_loss": 0.08734130859375, |
|
"eval_runtime": 650.8788, |
|
"eval_samples_per_second": 4.256, |
|
"eval_steps_per_second": 0.034, |
|
"eval_wer": 21.794093851132686, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 7.64862914862915e-06, |
|
"loss": 0.0559, |
|
"step": 6810 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 7.645021645021645e-06, |
|
"loss": 0.0557, |
|
"step": 6820 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 7.641414141414142e-06, |
|
"loss": 0.0532, |
|
"step": 6830 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 7.637806637806639e-06, |
|
"loss": 0.0537, |
|
"step": 6840 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 7.634199134199136e-06, |
|
"loss": 0.0572, |
|
"step": 6850 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 7.630591630591631e-06, |
|
"loss": 0.0559, |
|
"step": 6860 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 7.626984126984127e-06, |
|
"loss": 0.0558, |
|
"step": 6870 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 7.623376623376624e-06, |
|
"loss": 0.0549, |
|
"step": 6880 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 7.61976911976912e-06, |
|
"loss": 0.0533, |
|
"step": 6890 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 7.616161616161617e-06, |
|
"loss": 0.0559, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"eval_loss": 0.08636474609375, |
|
"eval_runtime": 684.4605, |
|
"eval_samples_per_second": 4.047, |
|
"eval_steps_per_second": 0.032, |
|
"eval_wer": 22.185140237324703, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 7.612554112554114e-06, |
|
"loss": 0.0555, |
|
"step": 6910 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 7.608946608946609e-06, |
|
"loss": 0.0552, |
|
"step": 6920 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 7.605339105339106e-06, |
|
"loss": 0.0545, |
|
"step": 6930 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 7.601731601731602e-06, |
|
"loss": 0.0522, |
|
"step": 6940 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 7.598124098124099e-06, |
|
"loss": 0.0545, |
|
"step": 6950 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 7.594516594516596e-06, |
|
"loss": 0.0562, |
|
"step": 6960 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 7.590909090909091e-06, |
|
"loss": 0.0544, |
|
"step": 6970 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 7.587301587301588e-06, |
|
"loss": 0.0545, |
|
"step": 6980 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 7.583694083694084e-06, |
|
"loss": 0.0544, |
|
"step": 6990 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 7.580086580086581e-06, |
|
"loss": 0.0552, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"eval_loss": 0.0867919921875, |
|
"eval_runtime": 359.998, |
|
"eval_samples_per_second": 7.694, |
|
"eval_steps_per_second": 0.061, |
|
"eval_wer": 21.85477346278317, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 7.576479076479078e-06, |
|
"loss": 0.0548, |
|
"step": 7010 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 7.573593073593074e-06, |
|
"loss": 0.0534, |
|
"step": 7020 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 7.56998556998557e-06, |
|
"loss": 0.055, |
|
"step": 7030 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 7.566378066378067e-06, |
|
"loss": 0.0547, |
|
"step": 7040 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 7.562770562770564e-06, |
|
"loss": 0.0539, |
|
"step": 7050 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 7.559163059163059e-06, |
|
"loss": 0.0554, |
|
"step": 7060 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 7.555555555555556e-06, |
|
"loss": 0.0587, |
|
"step": 7070 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 7.551948051948052e-06, |
|
"loss": 0.0554, |
|
"step": 7080 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 7.548340548340549e-06, |
|
"loss": 0.0551, |
|
"step": 7090 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 7.544733044733046e-06, |
|
"loss": 0.0546, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"eval_loss": 0.08642578125, |
|
"eval_runtime": 281.6212, |
|
"eval_samples_per_second": 9.836, |
|
"eval_steps_per_second": 0.078, |
|
"eval_wer": 21.794093851132686, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 7.541125541125541e-06, |
|
"loss": 0.0537, |
|
"step": 7110 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 7.537518037518038e-06, |
|
"loss": 0.0562, |
|
"step": 7120 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 7.533910533910534e-06, |
|
"loss": 0.0545, |
|
"step": 7130 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 7.530303030303031e-06, |
|
"loss": 0.0555, |
|
"step": 7140 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 7.526695526695528e-06, |
|
"loss": 0.0557, |
|
"step": 7150 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 7.523088023088023e-06, |
|
"loss": 0.0525, |
|
"step": 7160 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 7.51948051948052e-06, |
|
"loss": 0.0566, |
|
"step": 7170 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 7.515873015873016e-06, |
|
"loss": 0.0545, |
|
"step": 7180 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 7.512265512265513e-06, |
|
"loss": 0.0562, |
|
"step": 7190 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 7.50865800865801e-06, |
|
"loss": 0.0549, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"eval_loss": 0.08612060546875, |
|
"eval_runtime": 784.1899, |
|
"eval_samples_per_second": 3.532, |
|
"eval_steps_per_second": 0.028, |
|
"eval_wer": 22.357065803667744, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 7.505050505050505e-06, |
|
"loss": 0.0548, |
|
"step": 7210 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 7.501443001443002e-06, |
|
"loss": 0.056, |
|
"step": 7220 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 7.497835497835498e-06, |
|
"loss": 0.0544, |
|
"step": 7230 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 7.494227994227995e-06, |
|
"loss": 0.054, |
|
"step": 7240 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 7.490620490620492e-06, |
|
"loss": 0.0558, |
|
"step": 7250 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 7.487012987012988e-06, |
|
"loss": 0.0549, |
|
"step": 7260 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 7.483405483405484e-06, |
|
"loss": 0.0555, |
|
"step": 7270 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 7.47979797979798e-06, |
|
"loss": 0.0559, |
|
"step": 7280 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 7.476190476190477e-06, |
|
"loss": 0.0574, |
|
"step": 7290 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 7.472582972582974e-06, |
|
"loss": 0.0567, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"eval_loss": 0.0860595703125, |
|
"eval_runtime": 523.0183, |
|
"eval_samples_per_second": 5.296, |
|
"eval_steps_per_second": 0.042, |
|
"eval_wer": 21.709816612729234, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 7.46897546897547e-06, |
|
"loss": 0.0541, |
|
"step": 7310 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 7.465367965367966e-06, |
|
"loss": 0.0547, |
|
"step": 7320 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 7.461760461760462e-06, |
|
"loss": 0.0558, |
|
"step": 7330 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 7.458152958152959e-06, |
|
"loss": 0.0536, |
|
"step": 7340 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 7.454545454545456e-06, |
|
"loss": 0.0539, |
|
"step": 7350 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 7.450937950937952e-06, |
|
"loss": 0.0583, |
|
"step": 7360 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 7.447330447330448e-06, |
|
"loss": 0.0541, |
|
"step": 7370 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 7.443722943722944e-06, |
|
"loss": 0.0545, |
|
"step": 7380 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 7.440115440115441e-06, |
|
"loss": 0.0528, |
|
"step": 7390 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 7.4365079365079376e-06, |
|
"loss": 0.0548, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"eval_loss": 0.0858154296875, |
|
"eval_runtime": 282.1778, |
|
"eval_samples_per_second": 9.817, |
|
"eval_steps_per_second": 0.078, |
|
"eval_wer": 21.83791801510248, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 7.432900432900434e-06, |
|
"loss": 0.0563, |
|
"step": 7410 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 7.42929292929293e-06, |
|
"loss": 0.0534, |
|
"step": 7420 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 7.425685425685426e-06, |
|
"loss": 0.0571, |
|
"step": 7430 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 7.422077922077923e-06, |
|
"loss": 0.0548, |
|
"step": 7440 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 7.4184704184704195e-06, |
|
"loss": 0.0545, |
|
"step": 7450 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 7.4148629148629155e-06, |
|
"loss": 0.0542, |
|
"step": 7460 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 7.411255411255412e-06, |
|
"loss": 0.0547, |
|
"step": 7470 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 7.407647907647908e-06, |
|
"loss": 0.0567, |
|
"step": 7480 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 7.4040404040404045e-06, |
|
"loss": 0.0563, |
|
"step": 7490 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 7.400432900432901e-06, |
|
"loss": 0.055, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"eval_loss": 0.08563232421875, |
|
"eval_runtime": 382.1657, |
|
"eval_samples_per_second": 7.248, |
|
"eval_steps_per_second": 0.058, |
|
"eval_wer": 21.986245954692556, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 7.3968253968253975e-06, |
|
"loss": 0.0533, |
|
"step": 7510 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 7.3932178932178935e-06, |
|
"loss": 0.0546, |
|
"step": 7520 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 7.3896103896103896e-06, |
|
"loss": 0.0551, |
|
"step": 7530 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 7.3860028860028865e-06, |
|
"loss": 0.0537, |
|
"step": 7540 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 7.382395382395383e-06, |
|
"loss": 0.0547, |
|
"step": 7550 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 7.378787878787879e-06, |
|
"loss": 0.0556, |
|
"step": 7560 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 7.375180375180376e-06, |
|
"loss": 0.0547, |
|
"step": 7570 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 7.3715728715728715e-06, |
|
"loss": 0.0539, |
|
"step": 7580 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 7.367965367965368e-06, |
|
"loss": 0.0554, |
|
"step": 7590 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 7.364357864357865e-06, |
|
"loss": 0.0541, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"eval_loss": 0.08575439453125, |
|
"eval_runtime": 628.8874, |
|
"eval_samples_per_second": 4.405, |
|
"eval_steps_per_second": 0.035, |
|
"eval_wer": 21.810949298813377, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 7.360750360750361e-06, |
|
"loss": 0.0561, |
|
"step": 7610 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 7.357142857142858e-06, |
|
"loss": 0.0545, |
|
"step": 7620 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 7.353535353535353e-06, |
|
"loss": 0.0544, |
|
"step": 7630 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 7.34992784992785e-06, |
|
"loss": 0.0543, |
|
"step": 7640 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 7.346320346320347e-06, |
|
"loss": 0.0549, |
|
"step": 7650 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 7.342712842712843e-06, |
|
"loss": 0.0536, |
|
"step": 7660 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 7.33910533910534e-06, |
|
"loss": 0.056, |
|
"step": 7670 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 7.335497835497835e-06, |
|
"loss": 0.0545, |
|
"step": 7680 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 7.331890331890332e-06, |
|
"loss": 0.054, |
|
"step": 7690 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 7.328282828282829e-06, |
|
"loss": 0.0554, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"eval_loss": 0.08563232421875, |
|
"eval_runtime": 906.4345, |
|
"eval_samples_per_second": 3.056, |
|
"eval_steps_per_second": 0.024, |
|
"eval_wer": 21.85477346278317, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 7.324675324675325e-06, |
|
"loss": 0.0578, |
|
"step": 7710 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 7.321067821067822e-06, |
|
"loss": 0.0541, |
|
"step": 7720 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 7.317460317460317e-06, |
|
"loss": 0.0557, |
|
"step": 7730 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 7.313852813852814e-06, |
|
"loss": 0.0536, |
|
"step": 7740 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 7.310245310245311e-06, |
|
"loss": 0.0538, |
|
"step": 7750 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 7.306637806637807e-06, |
|
"loss": 0.0533, |
|
"step": 7760 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 7.303030303030304e-06, |
|
"loss": 0.0549, |
|
"step": 7770 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 7.299422799422799e-06, |
|
"loss": 0.0543, |
|
"step": 7780 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 7.295815295815296e-06, |
|
"loss": 0.0561, |
|
"step": 7790 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 7.292207792207793e-06, |
|
"loss": 0.0534, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"eval_loss": 0.08599853515625, |
|
"eval_runtime": 594.8444, |
|
"eval_samples_per_second": 4.657, |
|
"eval_steps_per_second": 0.037, |
|
"eval_wer": 21.39630528586839, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 7.288600288600289e-06, |
|
"loss": 0.0573, |
|
"step": 7810 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 7.284992784992786e-06, |
|
"loss": 0.0568, |
|
"step": 7820 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 7.281385281385281e-06, |
|
"loss": 0.0554, |
|
"step": 7830 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 7.277777777777778e-06, |
|
"loss": 0.0552, |
|
"step": 7840 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 7.274170274170275e-06, |
|
"loss": 0.0531, |
|
"step": 7850 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 7.270562770562771e-06, |
|
"loss": 0.0539, |
|
"step": 7860 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 7.266955266955268e-06, |
|
"loss": 0.0555, |
|
"step": 7870 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 7.263347763347765e-06, |
|
"loss": 0.0538, |
|
"step": 7880 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 7.25974025974026e-06, |
|
"loss": 0.0552, |
|
"step": 7890 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 7.256132756132757e-06, |
|
"loss": 0.0544, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"eval_loss": 0.08575439453125, |
|
"eval_runtime": 276.18, |
|
"eval_samples_per_second": 10.03, |
|
"eval_steps_per_second": 0.08, |
|
"eval_wer": 21.419902912621357, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 7.252525252525253e-06, |
|
"loss": 0.0526, |
|
"step": 7910 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 7.24891774891775e-06, |
|
"loss": 0.0554, |
|
"step": 7920 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 7.245310245310247e-06, |
|
"loss": 0.0576, |
|
"step": 7930 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 7.241702741702742e-06, |
|
"loss": 0.0552, |
|
"step": 7940 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 7.238095238095239e-06, |
|
"loss": 0.0557, |
|
"step": 7950 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 7.234487734487735e-06, |
|
"loss": 0.0517, |
|
"step": 7960 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 7.230880230880232e-06, |
|
"loss": 0.0537, |
|
"step": 7970 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 7.227272727272729e-06, |
|
"loss": 0.0576, |
|
"step": 7980 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 7.223665223665224e-06, |
|
"loss": 0.0547, |
|
"step": 7990 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 7.220057720057721e-06, |
|
"loss": 0.0556, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"eval_loss": 0.085205078125, |
|
"eval_runtime": 353.3226, |
|
"eval_samples_per_second": 7.84, |
|
"eval_steps_per_second": 0.062, |
|
"eval_wer": 21.64576591154261, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 7.216450216450217e-06, |
|
"loss": 0.0547, |
|
"step": 8010 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 7.213564213564213e-06, |
|
"loss": 0.0531, |
|
"step": 8020 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 7.20995670995671e-06, |
|
"loss": 0.0545, |
|
"step": 8030 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 7.206349206349207e-06, |
|
"loss": 0.0566, |
|
"step": 8040 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 7.202741702741703e-06, |
|
"loss": 0.0556, |
|
"step": 8050 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 7.1991341991342e-06, |
|
"loss": 0.0538, |
|
"step": 8060 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 7.195526695526695e-06, |
|
"loss": 0.0523, |
|
"step": 8070 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 7.191919191919192e-06, |
|
"loss": 0.0556, |
|
"step": 8080 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 7.188311688311689e-06, |
|
"loss": 0.0569, |
|
"step": 8090 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 7.184704184704185e-06, |
|
"loss": 0.0544, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"eval_loss": 0.08526611328125, |
|
"eval_runtime": 716.9844, |
|
"eval_samples_per_second": 3.863, |
|
"eval_steps_per_second": 0.031, |
|
"eval_wer": 21.3524811218986, |
|
"step": 8100 |
|
} |
|
], |
|
"max_steps": 28000, |
|
"num_train_epochs": 11, |
|
"total_flos": 2.1159834348111245e+21, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|