{ "best_metric": 0.5737143158912659, "best_model_checkpoint": "/scratch/skscla001/speech/results/whisper-medium-toigen-female-model/checkpoint-400", "epoch": 8.0, "eval_steps": 200, "global_step": 1000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.20080321285140562, "grad_norm": 61.12749481201172, "learning_rate": 4.2000000000000006e-07, "loss": 6.6534, "step": 25 }, { "epoch": 0.40160642570281124, "grad_norm": 43.05365753173828, "learning_rate": 9.200000000000001e-07, "loss": 5.1469, "step": 50 }, { "epoch": 0.6024096385542169, "grad_norm": 38.88462448120117, "learning_rate": 1.42e-06, "loss": 3.6082, "step": 75 }, { "epoch": 0.8032128514056225, "grad_norm": 33.215606689453125, "learning_rate": 1.9200000000000003e-06, "loss": 2.5276, "step": 100 }, { "epoch": 1.0, "grad_norm": 24.052446365356445, "learning_rate": 2.42e-06, "loss": 2.0062, "step": 125 }, { "epoch": 1.2008032128514057, "grad_norm": 31.749963760375977, "learning_rate": 2.92e-06, "loss": 1.5108, "step": 150 }, { "epoch": 1.4016064257028114, "grad_norm": 26.5882625579834, "learning_rate": 3.4200000000000007e-06, "loss": 1.35, "step": 175 }, { "epoch": 1.6024096385542168, "grad_norm": 24.725419998168945, "learning_rate": 3.920000000000001e-06, "loss": 1.3083, "step": 200 }, { "epoch": 1.6024096385542168, "eval_loss": 0.6686530113220215, "eval_runtime": 103.3468, "eval_samples_per_second": 2.138, "eval_steps_per_second": 0.542, "eval_wer": 0.5309236947791165, "step": 200 }, { "epoch": 1.8032128514056225, "grad_norm": 23.32423973083496, "learning_rate": 4.42e-06, "loss": 1.2009, "step": 225 }, { "epoch": 2.0, "grad_norm": 23.683210372924805, "learning_rate": 4.92e-06, "loss": 1.2218, "step": 250 }, { "epoch": 2.2008032128514055, "grad_norm": 17.880393981933594, "learning_rate": 5.420000000000001e-06, "loss": 0.7736, "step": 275 }, { "epoch": 2.4016064257028114, "grad_norm": 16.88402557373047, "learning_rate": 5.92e-06, "loss": 0.676, "step": 300 }, { "epoch": 2.602409638554217, "grad_norm": 19.675437927246094, "learning_rate": 6.42e-06, "loss": 0.744, "step": 325 }, { "epoch": 2.8032128514056227, "grad_norm": 24.291088104248047, "learning_rate": 6.92e-06, "loss": 0.7339, "step": 350 }, { "epoch": 3.0, "grad_norm": 11.682101249694824, "learning_rate": 7.420000000000001e-06, "loss": 0.7098, "step": 375 }, { "epoch": 3.2008032128514055, "grad_norm": 10.226051330566406, "learning_rate": 7.92e-06, "loss": 0.3381, "step": 400 }, { "epoch": 3.2008032128514055, "eval_loss": 0.5737143158912659, "eval_runtime": 102.7558, "eval_samples_per_second": 2.151, "eval_steps_per_second": 0.545, "eval_wer": 0.44457831325301206, "step": 400 }, { "epoch": 3.4016064257028114, "grad_norm": 16.29640769958496, "learning_rate": 8.42e-06, "loss": 0.3465, "step": 425 }, { "epoch": 3.602409638554217, "grad_norm": 16.958227157592773, "learning_rate": 8.920000000000001e-06, "loss": 0.3707, "step": 450 }, { "epoch": 3.8032128514056227, "grad_norm": 20.855695724487305, "learning_rate": 9.42e-06, "loss": 0.445, "step": 475 }, { "epoch": 4.0, "grad_norm": 10.269837379455566, "learning_rate": 9.920000000000002e-06, "loss": 0.4447, "step": 500 }, { "epoch": 4.2008032128514055, "grad_norm": 15.100114822387695, "learning_rate": 9.953333333333333e-06, "loss": 0.1814, "step": 525 }, { "epoch": 4.401606425702811, "grad_norm": 13.603893280029297, "learning_rate": 9.89777777777778e-06, "loss": 0.2124, "step": 550 }, { "epoch": 4.602409638554217, "grad_norm": 13.522416114807129, "learning_rate": 9.842222222222223e-06, "loss": 0.2537, "step": 575 }, { "epoch": 4.803212851405623, "grad_norm": 17.89769172668457, "learning_rate": 9.786666666666667e-06, "loss": 0.2639, "step": 600 }, { "epoch": 4.803212851405623, "eval_loss": 0.6059707999229431, "eval_runtime": 102.8339, "eval_samples_per_second": 2.149, "eval_steps_per_second": 0.545, "eval_wer": 0.42971887550200805, "step": 600 }, { "epoch": 5.0, "grad_norm": 6.560233116149902, "learning_rate": 9.731111111111113e-06, "loss": 0.2428, "step": 625 }, { "epoch": 5.2008032128514055, "grad_norm": 7.097219467163086, "learning_rate": 9.675555555555555e-06, "loss": 0.1016, "step": 650 }, { "epoch": 5.401606425702811, "grad_norm": 14.023282051086426, "learning_rate": 9.620000000000001e-06, "loss": 0.1447, "step": 675 }, { "epoch": 5.602409638554217, "grad_norm": 8.226017951965332, "learning_rate": 9.564444444444445e-06, "loss": 0.1111, "step": 700 }, { "epoch": 5.803212851405623, "grad_norm": 11.734825134277344, "learning_rate": 9.508888888888889e-06, "loss": 0.1351, "step": 725 }, { "epoch": 6.0, "grad_norm": 9.586465835571289, "learning_rate": 9.453333333333335e-06, "loss": 0.1444, "step": 750 }, { "epoch": 6.2008032128514055, "grad_norm": 7.064974784851074, "learning_rate": 9.397777777777779e-06, "loss": 0.0732, "step": 775 }, { "epoch": 6.401606425702811, "grad_norm": 7.27360725402832, "learning_rate": 9.342222222222223e-06, "loss": 0.0831, "step": 800 }, { "epoch": 6.401606425702811, "eval_loss": 0.6426535844802856, "eval_runtime": 103.0495, "eval_samples_per_second": 2.145, "eval_steps_per_second": 0.543, "eval_wer": 0.40321285140562246, "step": 800 }, { "epoch": 6.602409638554217, "grad_norm": 5.486606597900391, "learning_rate": 9.286666666666667e-06, "loss": 0.0764, "step": 825 }, { "epoch": 6.803212851405623, "grad_norm": 11.913525581359863, "learning_rate": 9.231111111111111e-06, "loss": 0.0753, "step": 850 }, { "epoch": 7.0, "grad_norm": 4.659058094024658, "learning_rate": 9.175555555555557e-06, "loss": 0.1132, "step": 875 }, { "epoch": 7.2008032128514055, "grad_norm": 2.5287861824035645, "learning_rate": 9.12e-06, "loss": 0.0583, "step": 900 }, { "epoch": 7.401606425702811, "grad_norm": 3.8375866413116455, "learning_rate": 9.064444444444447e-06, "loss": 0.0513, "step": 925 }, { "epoch": 7.602409638554217, "grad_norm": 10.924978256225586, "learning_rate": 9.008888888888889e-06, "loss": 0.0674, "step": 950 }, { "epoch": 7.803212851405623, "grad_norm": 7.799922943115234, "learning_rate": 8.953333333333335e-06, "loss": 0.0612, "step": 975 }, { "epoch": 8.0, "grad_norm": 0.7825481295585632, "learning_rate": 8.897777777777779e-06, "loss": 0.0776, "step": 1000 }, { "epoch": 8.0, "eval_loss": 0.6705393195152283, "eval_runtime": 102.2471, "eval_samples_per_second": 2.161, "eval_steps_per_second": 0.548, "eval_wer": 0.42088353413654617, "step": 1000 }, { "epoch": 8.0, "step": 1000, "total_flos": 8.1240150638592e+18, "train_loss": 0.8661212004423141, "train_runtime": 2219.5582, "train_samples_per_second": 18.022, "train_steps_per_second": 2.253 } ], "logging_steps": 25, "max_steps": 5000, "num_input_tokens_seen": 0, "num_train_epochs": 41, "save_steps": 200, "stateful_callbacks": { "EarlyStoppingCallback": { "args": { "early_stopping_patience": 3, "early_stopping_threshold": 0.0 }, "attributes": { "early_stopping_patience_counter": 3 } }, "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 8.1240150638592e+18, "train_batch_size": 4, "trial_name": null, "trial_params": null }