|
{ |
|
"best_metric": 1.2918329238891602, |
|
"best_model_checkpoint": "nrshoudi/wav2vec-arabic-V2-5/checkpoint-1640", |
|
"epoch": 20.0, |
|
"eval_steps": 500, |
|
"global_step": 1640, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 5e-05, |
|
"loss": 25.2164, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 9.322426795959473, |
|
"eval_per": 1.0, |
|
"eval_runtime": 118.9285, |
|
"eval_samples_per_second": 6.836, |
|
"eval_steps_per_second": 3.422, |
|
"eval_wer": 1.0, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 0.0001, |
|
"loss": 5.4912, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 3.420063018798828, |
|
"eval_per": 1.0, |
|
"eval_runtime": 116.7666, |
|
"eval_samples_per_second": 6.963, |
|
"eval_steps_per_second": 3.486, |
|
"eval_wer": 1.0, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 9.444444444444444e-05, |
|
"loss": 3.2902, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_loss": 3.2383553981781006, |
|
"eval_per": 1.0, |
|
"eval_runtime": 117.3546, |
|
"eval_samples_per_second": 6.928, |
|
"eval_steps_per_second": 3.468, |
|
"eval_wer": 1.0, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"learning_rate": 8.888888888888889e-05, |
|
"loss": 3.2447, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_loss": 3.207124948501587, |
|
"eval_per": 1.0, |
|
"eval_runtime": 116.8797, |
|
"eval_samples_per_second": 6.956, |
|
"eval_steps_per_second": 3.482, |
|
"eval_wer": 1.0, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"learning_rate": 8.333333333333334e-05, |
|
"loss": 3.2259, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_loss": 3.224609375, |
|
"eval_per": 1.0, |
|
"eval_runtime": 117.5167, |
|
"eval_samples_per_second": 6.918, |
|
"eval_steps_per_second": 3.463, |
|
"eval_wer": 1.0, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"learning_rate": 7.777777777777778e-05, |
|
"loss": 3.2254, |
|
"step": 492 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_loss": 3.205623149871826, |
|
"eval_per": 1.0, |
|
"eval_runtime": 117.6839, |
|
"eval_samples_per_second": 6.908, |
|
"eval_steps_per_second": 3.458, |
|
"eval_wer": 1.0, |
|
"step": 492 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"learning_rate": 7.222222222222222e-05, |
|
"loss": 3.2116, |
|
"step": 574 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"eval_loss": 3.2067959308624268, |
|
"eval_per": 1.0, |
|
"eval_runtime": 117.4968, |
|
"eval_samples_per_second": 6.919, |
|
"eval_steps_per_second": 3.464, |
|
"eval_wer": 1.0, |
|
"step": 574 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"learning_rate": 6.666666666666667e-05, |
|
"loss": 3.1886, |
|
"step": 656 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_loss": 3.28141188621521, |
|
"eval_per": 1.0, |
|
"eval_runtime": 117.8914, |
|
"eval_samples_per_second": 6.896, |
|
"eval_steps_per_second": 3.452, |
|
"eval_wer": 1.0, |
|
"step": 656 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"learning_rate": 6.111111111111112e-05, |
|
"loss": 3.1499, |
|
"step": 738 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"eval_loss": 3.115420341491699, |
|
"eval_per": 1.0, |
|
"eval_runtime": 117.2629, |
|
"eval_samples_per_second": 6.933, |
|
"eval_steps_per_second": 3.471, |
|
"eval_wer": 1.0, |
|
"step": 738 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"learning_rate": 5.555555555555556e-05, |
|
"loss": 3.0862, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_loss": 3.0622904300689697, |
|
"eval_per": 1.0, |
|
"eval_runtime": 117.8128, |
|
"eval_samples_per_second": 6.901, |
|
"eval_steps_per_second": 3.455, |
|
"eval_wer": 1.0, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"learning_rate": 5e-05, |
|
"loss": 2.9718, |
|
"step": 902 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"eval_loss": 2.9403610229492188, |
|
"eval_per": 0.9982568548007161, |
|
"eval_runtime": 117.9864, |
|
"eval_samples_per_second": 6.891, |
|
"eval_steps_per_second": 3.45, |
|
"eval_wer": 0.9973403014325043, |
|
"step": 902 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"learning_rate": 4.4444444444444447e-05, |
|
"loss": 2.81, |
|
"step": 984 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"eval_loss": 2.7654075622558594, |
|
"eval_per": 0.9858663902760765, |
|
"eval_runtime": 118.1242, |
|
"eval_samples_per_second": 6.883, |
|
"eval_steps_per_second": 3.446, |
|
"eval_wer": 0.982082030703187, |
|
"step": 984 |
|
}, |
|
{ |
|
"epoch": 13.0, |
|
"learning_rate": 3.888888888888889e-05, |
|
"loss": 2.612, |
|
"step": 1066 |
|
}, |
|
{ |
|
"epoch": 13.0, |
|
"eval_loss": 2.6607749462127686, |
|
"eval_per": 0.9784933572034298, |
|
"eval_runtime": 117.5231, |
|
"eval_samples_per_second": 6.918, |
|
"eval_steps_per_second": 3.463, |
|
"eval_wer": 0.9750828239466194, |
|
"step": 1066 |
|
}, |
|
{ |
|
"epoch": 14.0, |
|
"learning_rate": 3.3333333333333335e-05, |
|
"loss": 2.4167, |
|
"step": 1148 |
|
}, |
|
{ |
|
"epoch": 14.0, |
|
"eval_loss": 2.2972512245178223, |
|
"eval_per": 0.9152925657212853, |
|
"eval_runtime": 117.8908, |
|
"eval_samples_per_second": 6.896, |
|
"eval_steps_per_second": 3.452, |
|
"eval_wer": 0.9051374177593207, |
|
"step": 1148 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"learning_rate": 2.777777777777778e-05, |
|
"loss": 2.1839, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"eval_loss": 2.0348501205444336, |
|
"eval_per": 0.7780552153019882, |
|
"eval_runtime": 118.2007, |
|
"eval_samples_per_second": 6.878, |
|
"eval_steps_per_second": 3.443, |
|
"eval_wer": 0.767626335681956, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"learning_rate": 2.2222222222222223e-05, |
|
"loss": 1.8957, |
|
"step": 1312 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"eval_loss": 1.767521858215332, |
|
"eval_per": 0.6950438141901442, |
|
"eval_runtime": 118.4157, |
|
"eval_samples_per_second": 6.866, |
|
"eval_steps_per_second": 3.437, |
|
"eval_wer": 0.6932480985488312, |
|
"step": 1312 |
|
}, |
|
{ |
|
"epoch": 17.0, |
|
"learning_rate": 1.6666666666666667e-05, |
|
"loss": 1.6773, |
|
"step": 1394 |
|
}, |
|
{ |
|
"epoch": 17.0, |
|
"eval_loss": 1.5461149215698242, |
|
"eval_per": 0.5707387166682371, |
|
"eval_runtime": 118.0574, |
|
"eval_samples_per_second": 6.886, |
|
"eval_steps_per_second": 3.447, |
|
"eval_wer": 0.5751014884979703, |
|
"step": 1394 |
|
}, |
|
{ |
|
"epoch": 18.0, |
|
"learning_rate": 1.1111111111111112e-05, |
|
"loss": 1.5059, |
|
"step": 1476 |
|
}, |
|
{ |
|
"epoch": 18.0, |
|
"eval_loss": 1.419967532157898, |
|
"eval_per": 0.4164703665316122, |
|
"eval_runtime": 117.8196, |
|
"eval_samples_per_second": 6.9, |
|
"eval_steps_per_second": 3.454, |
|
"eval_wer": 0.438103681582754, |
|
"step": 1476 |
|
}, |
|
{ |
|
"epoch": 19.0, |
|
"learning_rate": 5.555555555555556e-06, |
|
"loss": 1.4219, |
|
"step": 1558 |
|
}, |
|
{ |
|
"epoch": 19.0, |
|
"eval_loss": 1.3218241930007935, |
|
"eval_per": 0.4048572505417884, |
|
"eval_runtime": 117.824, |
|
"eval_samples_per_second": 6.9, |
|
"eval_steps_per_second": 3.454, |
|
"eval_wer": 0.4255517707993094, |
|
"step": 1558 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"learning_rate": 0.0, |
|
"loss": 1.3273, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"eval_loss": 1.2918329238891602, |
|
"eval_per": 0.39406859511919345, |
|
"eval_runtime": 117.9622, |
|
"eval_samples_per_second": 6.892, |
|
"eval_steps_per_second": 3.45, |
|
"eval_wer": 0.41351313517801314, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"step": 1640, |
|
"total_flos": 5.333828756305947e+17, |
|
"train_loss": 3.8576244959017125, |
|
"train_runtime": 3468.4696, |
|
"train_samples_per_second": 0.94, |
|
"train_steps_per_second": 0.473 |
|
} |
|
], |
|
"logging_steps": 500, |
|
"max_steps": 1640, |
|
"num_train_epochs": 20, |
|
"save_steps": 500, |
|
"total_flos": 5.333828756305947e+17, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|