xezpeleta commited on
Commit
6ed3801
·
verified ·
1 Parent(s): b84f667

Model save

Browse files
Files changed (2) hide show
  1. README.md +33 -49
  2. generation_config.json +1 -1
README.md CHANGED
@@ -1,43 +1,25 @@
1
  ---
2
  library_name: transformers
3
- language:
4
- - eu
5
  license: apache-2.0
6
  base_model: openai/whisper-large-v3
7
  tags:
8
- - whisper-event
9
  - generated_from_trainer
10
- datasets:
11
- - mozilla-foundation/common_voice_17_0
12
  metrics:
13
  - wer
14
  model-index:
15
- - name: Whisper Large Basque
16
- results:
17
- - task:
18
- name: Automatic Speech Recognition
19
- type: automatic-speech-recognition
20
- dataset:
21
- name: mozilla-foundation/common_voice_17_0 eu
22
- type: mozilla-foundation/common_voice_17_0
23
- config: eu
24
- split: test
25
- args: eu
26
- metrics:
27
- - name: Wer
28
- type: wer
29
- value: 7.215361500971087
30
  ---
31
 
32
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
33
  should probably proofread and complete it, then remove this comment. -->
34
 
35
- # Whisper Large Basque
36
 
37
- This model is a fine-tuned version of [openai/whisper-large-v3](https://huggingface.co/openai/whisper-large-v3) on the mozilla-foundation/common_voice_17_0 eu dataset.
38
  It achieves the following results on the evaluation set:
39
- - Loss: 0.1259
40
- - Wer: 7.2154
41
 
42
  ## Model description
43
 
@@ -62,37 +44,39 @@ The following hyperparameters were used during training:
62
  - seed: 42
63
  - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
64
  - lr_scheduler_type: linear
65
- - lr_scheduler_warmup_steps: 500
66
  - training_steps: 10000
67
  - mixed_precision_training: Native AMP
68
 
69
  ### Training results
70
 
71
- | Training Loss | Epoch | Step | Validation Loss | Wer |
72
- |:-------------:|:------:|:----:|:---------------:|:-------:|
73
- | 0.2208 | 0.05 | 500 | 0.2592 | 20.6915 |
74
- | 0.1489 | 0.1 | 1000 | 0.1971 | 14.6827 |
75
- | 0.1973 | 0.15 | 1500 | 0.1747 | 12.3777 |
76
- | 0.1353 | 1.0296 | 2000 | 0.1527 | 10.7195 |
77
- | 0.1065 | 1.0796 | 2500 | 0.1456 | 9.8694 |
78
- | 0.106 | 1.1296 | 3000 | 0.1362 | 9.0925 |
79
- | 0.0718 | 2.0092 | 3500 | 0.1326 | 8.5428 |
80
- | 0.0683 | 2.0592 | 4000 | 0.1343 | 8.4851 |
81
- | 0.0482 | 2.1092 | 4500 | 0.1336 | 8.1049 |
82
- | 0.0548 | 2.1592 | 5000 | 0.1316 | 7.9244 |
83
- | 0.0282 | 3.0388 | 5500 | 0.1391 | 7.8182 |
84
- | 0.025 | 3.0888 | 6000 | 0.1425 | 7.9409 |
85
- | 0.0274 | 3.1388 | 6500 | 0.1391 | 7.7311 |
86
- | 0.0155 | 4.0184 | 7000 | 0.1492 | 7.6972 |
87
- | 0.0189 | 4.0684 | 7500 | 0.1517 | 7.6569 |
88
- | 0.0139 | 4.1184 | 8000 | 0.1539 | 7.6267 |
89
- | 0.0141 | 4.1684 | 8500 | 0.1550 | 7.5424 |
90
- | 0.0368 | 5.048 | 9000 | 0.1259 | 7.2154 |
 
 
91
 
92
 
93
  ### Framework versions
94
 
95
- - Transformers 4.46.0.dev0
96
- - Pytorch 2.4.1+cu121
97
- - Datasets 3.0.2.dev0
98
- - Tokenizers 0.20.0
 
1
  ---
2
  library_name: transformers
 
 
3
  license: apache-2.0
4
  base_model: openai/whisper-large-v3
5
  tags:
 
6
  - generated_from_trainer
 
 
7
  metrics:
8
  - wer
9
  model-index:
10
+ - name: openai/whisper-large-v3
11
+ results: []
 
 
 
 
 
 
 
 
 
 
 
 
 
12
  ---
13
 
14
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
  should probably proofread and complete it, then remove this comment. -->
16
 
17
+ # openai/whisper-large-v3
18
 
19
+ This model is a fine-tuned version of [openai/whisper-large-v3](https://huggingface.co/openai/whisper-large-v3) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 0.1834
22
+ - Wer: 8.6426
23
 
24
  ## Model description
25
 
 
44
  - seed: 42
45
  - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
46
  - lr_scheduler_type: linear
47
+ - lr_scheduler_warmup_steps: 1000
48
  - training_steps: 10000
49
  - mixed_precision_training: Native AMP
50
 
51
  ### Training results
52
 
53
+ | Training Loss | Epoch | Step | Validation Loss | Wer |
54
+ |:-------------:|:-----:|:-----:|:---------------:|:-------:|
55
+ | 0.2854 | 0.05 | 500 | 0.3763 | 24.9836 |
56
+ | 0.1425 | 0.1 | 1000 | 0.3326 | 19.8654 |
57
+ | 0.2196 | 0.15 | 1500 | 0.2802 | 16.2475 |
58
+ | 0.2338 | 0.2 | 2000 | 0.2536 | 14.6116 |
59
+ | 0.1383 | 0.25 | 2500 | 0.2451 | 12.8961 |
60
+ | 0.0848 | 0.3 | 3000 | 0.2280 | 12.2464 |
61
+ | 0.0854 | 0.35 | 3500 | 0.2152 | 11.4144 |
62
+ | 0.1304 | 0.4 | 4000 | 0.2097 | 11.1433 |
63
+ | 0.1328 | 0.45 | 4500 | 0.2055 | 10.6011 |
64
+ | 0.0737 | 0.5 | 5000 | 0.2079 | 10.5357 |
65
+ | 0.0804 | 0.55 | 5500 | 0.2133 | 10.1150 |
66
+ | 0.0964 | 0.6 | 6000 | 0.1988 | 9.4606 |
67
+ | 0.0811 | 0.65 | 6500 | 0.2019 | 9.4933 |
68
+ | 0.0677 | 0.7 | 7000 | 0.1916 | 8.9231 |
69
+ | 0.1114 | 0.75 | 7500 | 0.2029 | 9.3250 |
70
+ | 0.1142 | 0.8 | 8000 | 0.1895 | 8.9978 |
71
+ | 0.0466 | 0.85 | 8500 | 0.1936 | 8.8576 |
72
+ | 0.0664 | 0.9 | 9000 | 0.1876 | 8.9698 |
73
+ | 0.0759 | 0.95 | 9500 | 0.1827 | 8.8202 |
74
+ | 0.0555 | 1.0 | 10000 | 0.1834 | 8.6426 |
75
 
76
 
77
  ### Framework versions
78
 
79
+ - Transformers 4.49.0.dev0
80
+ - Pytorch 2.6.0+cu124
81
+ - Datasets 3.2.1.dev0
82
+ - Tokenizers 0.21.0
generation_config.json CHANGED
@@ -261,5 +261,5 @@
261
  "transcribe": 50360,
262
  "translate": 50359
263
  },
264
- "transformers_version": "4.46.0.dev0"
265
  }
 
261
  "transcribe": 50360,
262
  "translate": 50359
263
  },
264
+ "transformers_version": "4.49.0.dev0"
265
  }