aleni commited on
Commit
6c4572e
·
verified ·
1 Parent(s): 44b8d22

End of training

Browse files
Files changed (2) hide show
  1. README.md +8 -44
  2. generation_config.json +2 -2
README.md CHANGED
@@ -1,30 +1,14 @@
1
  ---
 
2
  language:
3
  - vi
4
  license: apache-2.0
5
  base_model: openai/whisper-small
6
  tags:
7
  - generated_from_trainer
8
- datasets:
9
- - common_voice_17_0
10
- metrics:
11
- - wer
12
  model-index:
13
  - name: Whisper small vi - Ox
14
- results:
15
- - task:
16
- name: Automatic Speech Recognition
17
- type: automatic-speech-recognition
18
- dataset:
19
- name: common_voice_17_0
20
- type: common_voice_17_0
21
- config: vi
22
- split: test
23
- args: vi
24
- metrics:
25
- - name: Wer
26
- type: wer
27
- value: 31.26665341022072
28
  ---
29
 
30
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -32,10 +16,7 @@ should probably proofread and complete it, then remove this comment. -->
32
 
33
  # Whisper small vi - Ox
34
 
35
- This model is a fine-tuned version of [openai/whisper-small](https://huggingface.co/openai/whisper-small) on the common_voice_17_0 dataset.
36
- It achieves the following results on the evaluation set:
37
- - Loss: 1.0138
38
- - Wer: 31.2667
39
 
40
  ## Model description
41
 
@@ -58,7 +39,7 @@ The following hyperparameters were used during training:
58
  - train_batch_size: 16
59
  - eval_batch_size: 8
60
  - seed: 42
61
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
62
  - lr_scheduler_type: linear
63
  - lr_scheduler_warmup_steps: 500
64
  - num_epochs: 3.0
@@ -66,28 +47,11 @@ The following hyperparameters were used during training:
66
 
67
  ### Training results
68
 
69
- | Training Loss | Epoch | Step | Validation Loss | Wer |
70
- |:-------------:|:-----:|:-----:|:---------------:|:-------:|
71
- | 0.2276 | 0.08 | 1000 | 0.7506 | 29.8509 |
72
- | 0.1768 | 0.16 | 2000 | 0.8114 | 31.2189 |
73
- | 0.1828 | 0.24 | 3000 | 0.8569 | 31.2985 |
74
- | 0.1632 | 0.32 | 4000 | 0.8523 | 31.9268 |
75
- | 0.1566 | 0.4 | 5000 | 0.9062 | 31.9149 |
76
- | 0.1532 | 0.48 | 6000 | 0.8914 | 31.4496 |
77
- | 0.1593 | 0.56 | 7000 | 0.9342 | 31.9825 |
78
- | 0.1411 | 0.64 | 8000 | 0.9412 | 32.0302 |
79
- | 0.1531 | 0.72 | 9000 | 0.9456 | 31.6206 |
80
- | 0.1246 | 0.8 | 10000 | 0.9452 | 31.7240 |
81
- | 0.1336 | 0.88 | 11000 | 0.9622 | 31.1195 |
82
- | 0.1392 | 0.96 | 12000 | 0.9638 | 31.3939 |
83
- | 0.0725 | 1.04 | 13000 | 1.0032 | 31.5649 |
84
- | 0.0838 | 1.12 | 14000 | 1.0346 | 31.7916 |
85
- | 0.0766 | 1.2 | 15000 | 1.0138 | 31.2667 |
86
 
87
 
88
  ### Framework versions
89
 
90
- - Transformers 4.39.3
91
- - Pytorch 2.4.1
92
- - Datasets 3.0.1
93
- - Tokenizers 0.15.2
 
1
  ---
2
+ library_name: transformers
3
  language:
4
  - vi
5
  license: apache-2.0
6
  base_model: openai/whisper-small
7
  tags:
8
  - generated_from_trainer
 
 
 
 
9
  model-index:
10
  - name: Whisper small vi - Ox
11
+ results: []
 
 
 
 
 
 
 
 
 
 
 
 
 
12
  ---
13
 
14
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
16
 
17
  # Whisper small vi - Ox
18
 
19
+ This model is a fine-tuned version of [openai/whisper-small](https://huggingface.co/openai/whisper-small) on an unknown dataset.
 
 
 
20
 
21
  ## Model description
22
 
 
39
  - train_batch_size: 16
40
  - eval_batch_size: 8
41
  - seed: 42
42
+ - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
43
  - lr_scheduler_type: linear
44
  - lr_scheduler_warmup_steps: 500
45
  - num_epochs: 3.0
 
47
 
48
  ### Training results
49
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
50
 
51
 
52
  ### Framework versions
53
 
54
+ - Transformers 4.47.0.dev0
55
+ - Pytorch 2.5.1+cu121
56
+ - Datasets 3.1.0
57
+ - Tokenizers 0.20.3
generation_config.json CHANGED
@@ -156,7 +156,7 @@
156
  "no_timestamps_token_id": 50363,
157
  "pad_token_id": 50257,
158
  "prev_sot_token_id": 50361,
159
- "return_timestamps": false,
160
  "suppress_tokens": [
161
  1,
162
  2,
@@ -252,5 +252,5 @@
252
  "transcribe": 50359,
253
  "translate": 50358
254
  },
255
- "transformers_version": "4.39.3"
256
  }
 
156
  "no_timestamps_token_id": 50363,
157
  "pad_token_id": 50257,
158
  "prev_sot_token_id": 50361,
159
+ "return_timestamps": true,
160
  "suppress_tokens": [
161
  1,
162
  2,
 
252
  "transcribe": 50359,
253
  "translate": 50358
254
  },
255
+ "transformers_version": "4.47.0.dev0"
256
  }