metadata
language:
- hu
license: apache-2.0
base_model: openai/whisper-tiny
tags:
- generated_from_trainer
datasets:
- fleurs
metrics:
- wer
model-index:
- name: Whisper Tiny Hu CV18
results:
- task:
name: Automatic Speech Recognition
type: automatic-speech-recognition
dataset:
name: Common Voice 18.0
type: fleurs
config: hu_hu
split: None
args: hu_hu
metrics:
- name: Wer
type: wer
value: 46.19517239639821
Whisper Tiny Hu CV18
This model is a fine-tuned version of openai/whisper-tiny on the Common Voice 18.0 dataset. It achieves the following results on the evaluation set:
- Loss: 1.2308
- Wer Ortho: 51.3968
- Wer: 46.1952
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 7.5e-05
- train_batch_size: 64
- eval_batch_size: 32
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 500
- num_epochs: 5
- mixed_precision_training: Native AMP
Training results
Training Loss | Epoch | Step | Validation Loss | Wer Ortho | Wer |
---|---|---|---|---|---|
0.5967 | 0.1723 | 250 | 1.2539 | 69.9246 | 66.9344 |
0.4456 | 0.3446 | 500 | 1.2200 | 65.5249 | 61.7108 |
0.3713 | 0.5169 | 750 | 1.1422 | 61.8665 | 58.0019 |
0.3337 | 0.6892 | 1000 | 1.1139 | 60.4332 | 55.7999 |
0.2829 | 0.8615 | 1250 | 1.1074 | 59.6528 | 56.2502 |
0.1931 | 1.0338 | 1500 | 1.1087 | 58.2686 | 53.8969 |
0.1855 | 1.2061 | 1750 | 1.1643 | 57.5828 | 52.8577 |
0.1827 | 1.3784 | 2000 | 1.1136 | 58.2951 | 54.2260 |
0.177 | 1.5507 | 2250 | 1.1326 | 57.4353 | 52.1628 |
0.1686 | 1.7229 | 2500 | 1.0970 | 54.8396 | 49.9067 |
0.1654 | 1.8952 | 2750 | 1.0957 | 56.3953 | 51.6975 |
0.0886 | 2.0675 | 3000 | 1.1150 | 53.5349 | 48.7805 |
0.0966 | 2.2398 | 3250 | 1.1417 | 54.4060 | 49.0113 |
0.0921 | 2.4121 | 3500 | 1.1387 | 53.9975 | 48.6001 |
0.0968 | 2.5844 | 3750 | 1.1587 | 53.8147 | 49.2660 |
0.0968 | 2.7567 | 4000 | 1.1459 | 52.8176 | 48.4438 |
0.086 | 2.9290 | 4250 | 1.1298 | 52.4784 | 47.4702 |
0.0456 | 3.1013 | 4500 | 1.1714 | 52.6663 | 47.3920 |
0.0487 | 3.2736 | 4750 | 1.1730 | 52.9524 | 48.1499 |
0.0475 | 3.4459 | 5000 | 1.1945 | 52.7898 | 47.3668 |
0.0442 | 3.6182 | 5250 | 1.2042 | 52.3410 | 47.3037 |
0.0434 | 3.7905 | 5500 | 1.1851 | 53.0205 | 47.7262 |
0.0438 | 3.9628 | 5750 | 1.1912 | 52.5869 | 48.0541 |
0.0211 | 4.1351 | 6000 | 1.2191 | 52.3562 | 47.9923 |
0.0198 | 4.3074 | 6250 | 1.2203 | 51.6136 | 46.6290 |
0.0185 | 4.4797 | 6500 | 1.2287 | 52.0297 | 46.4537 |
0.0196 | 4.6520 | 6750 | 1.2363 | 51.7183 | 46.2696 |
0.0196 | 4.8243 | 7000 | 1.2320 | 51.5594 | 46.0817 |
0.0178 | 4.9966 | 7250 | 1.2308 | 51.3968 | 46.1952 |
Framework versions
- Transformers 4.44.0
- Pytorch 2.3.0+cu121
- Datasets 2.21.0
- Tokenizers 0.19.1