kiranpantha commited on
Commit
0ab54b0
·
verified ·
1 Parent(s): 612a7b5

End of training

Browse files
Files changed (3) hide show
  1. README.md +33 -26
  2. generation_config.json +1 -1
  3. model.safetensors +1 -1
README.md CHANGED
@@ -7,36 +7,37 @@ base_model: openai/whisper-tiny
7
  tags:
8
  - generated_from_trainer
9
  datasets:
10
- - openslr/openslr
11
  metrics:
12
  - wer
13
  model-index:
14
- - name: Whisper Medium - Kiran Pantha
15
  results:
16
  - task:
17
  name: Automatic Speech Recognition
18
  type: automatic-speech-recognition
19
  dataset:
20
  name: OpenSLR54
21
- type: openslr/openslr
22
  config: default
23
  split: test
24
  args: 'config: ne, split: test'
25
  metrics:
26
  - name: Wer
27
  type: wer
28
- value: 53.889856134884994
29
  ---
30
 
31
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
32
  should probably proofread and complete it, then remove this comment. -->
33
 
34
- # Whisper Medium - Kiran Pantha
35
 
36
  This model is a fine-tuned version of [openai/whisper-tiny](https://huggingface.co/openai/whisper-tiny) on the OpenSLR54 dataset.
37
  It achieves the following results on the evaluation set:
38
- - Loss: 0.2728
39
- - Wer: 53.8899
 
40
 
41
  ## Model description
42
 
@@ -56,10 +57,10 @@ More information needed
56
 
57
  The following hyperparameters were used during training:
58
  - learning_rate: 1e-05
59
- - train_batch_size: 8
60
- - eval_batch_size: 4
61
  - seed: 42
62
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
63
  - lr_scheduler_type: linear
64
  - lr_scheduler_warmup_steps: 500
65
  - training_steps: 5000
@@ -67,23 +68,29 @@ The following hyperparameters were used during training:
67
 
68
  ### Training results
69
 
70
- | Training Loss | Epoch | Step | Validation Loss | Wer |
71
- |:-------------:|:------:|:----:|:---------------:|:-------:|
72
- | 0.6656 | 0.1200 | 500 | 0.6245 | 85.4211 |
73
- | 0.4586 | 0.2399 | 1000 | 0.4490 | 73.6369 |
74
- | 0.3772 | 0.3599 | 1500 | 0.3930 | 68.5879 |
75
- | 0.3437 | 0.4798 | 2000 | 0.3498 | 63.7222 |
76
- | 0.3214 | 0.5998 | 2500 | 0.3279 | 61.2297 |
77
- | 0.3186 | 0.7198 | 3000 | 0.3095 | 59.3696 |
78
- | 0.2965 | 0.8397 | 3500 | 0.2930 | 56.9504 |
79
- | 0.2759 | 0.9597 | 4000 | 0.2825 | 56.0249 |
80
- | 0.2474 | 1.0797 | 4500 | 0.2758 | 54.5588 |
81
- | 0.2195 | 1.1996 | 5000 | 0.2728 | 53.8899 |
 
 
 
 
 
 
82
 
83
 
84
  ### Framework versions
85
 
86
- - Transformers 4.44.2
87
- - Pytorch 2.4.0+cu121
88
- - Datasets 2.21.0
89
- - Tokenizers 0.19.1
 
7
  tags:
8
  - generated_from_trainer
9
  datasets:
10
+ - kiranpantha/OpenSLR54-Balanced-Nepali
11
  metrics:
12
  - wer
13
  model-index:
14
+ - name: Whisper Tiny Nepali - Kiran Pantha
15
  results:
16
  - task:
17
  name: Automatic Speech Recognition
18
  type: automatic-speech-recognition
19
  dataset:
20
  name: OpenSLR54
21
+ type: kiranpantha/OpenSLR54-Balanced-Nepali
22
  config: default
23
  split: test
24
  args: 'config: ne, split: test'
25
  metrics:
26
  - name: Wer
27
  type: wer
28
+ value: 53.726851851851855
29
  ---
30
 
31
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
32
  should probably proofread and complete it, then remove this comment. -->
33
 
34
+ # Whisper Tiny Nepali - Kiran Pantha
35
 
36
  This model is a fine-tuned version of [openai/whisper-tiny](https://huggingface.co/openai/whisper-tiny) on the OpenSLR54 dataset.
37
  It achieves the following results on the evaluation set:
38
+ - Loss: 0.2933
39
+ - Wer: 53.7269
40
+ - Cer: 16.1186
41
 
42
  ## Model description
43
 
 
57
 
58
  The following hyperparameters were used during training:
59
  - learning_rate: 1e-05
60
+ - train_batch_size: 16
61
+ - eval_batch_size: 8
62
  - seed: 42
63
+ - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
64
  - lr_scheduler_type: linear
65
  - lr_scheduler_warmup_steps: 500
66
  - training_steps: 5000
 
68
 
69
  ### Training results
70
 
71
+ | Training Loss | Epoch | Step | Validation Loss | Wer | Cer |
72
+ |:-------------:|:------:|:----:|:---------------:|:-------:|:-------:|
73
+ | 0.8115 | 0.3597 | 300 | 0.7467 | 92.9167 | 34.9897 |
74
+ | 0.4976 | 0.7194 | 600 | 0.4963 | 79.2130 | 26.2625 |
75
+ | 0.3874 | 1.0791 | 900 | 0.4198 | 71.5046 | 22.6696 |
76
+ | 0.3422 | 1.4388 | 1200 | 0.3797 | 67.5926 | 20.8896 |
77
+ | 0.3179 | 1.7986 | 1500 | 0.3467 | 63.9120 | 19.3959 |
78
+ | 0.2451 | 2.1583 | 1800 | 0.3299 | 62.1528 | 18.6950 |
79
+ | 0.2167 | 2.5180 | 2100 | 0.3224 | 60.6713 | 18.3977 |
80
+ | 0.2428 | 2.8777 | 2400 | 0.3085 | 59.6528 | 17.6196 |
81
+ | 0.1862 | 3.2374 | 2700 | 0.3057 | 57.6620 | 16.9113 |
82
+ | 0.1795 | 3.5971 | 3000 | 0.3007 | 57.5231 | 16.7792 |
83
+ | 0.1758 | 3.9568 | 3300 | 0.2935 | 55.8565 | 16.5297 |
84
+ | 0.1496 | 4.3165 | 3600 | 0.2960 | 55.8796 | 16.3792 |
85
+ | 0.156 | 4.6763 | 3900 | 0.2940 | 55.4398 | 16.4819 |
86
+ | 0.1235 | 5.0360 | 4200 | 0.2915 | 54.4444 | 16.0085 |
87
+ | 0.1311 | 5.3957 | 4500 | 0.2936 | 54.4676 | 16.2801 |
88
+ | 0.1136 | 5.7554 | 4800 | 0.2933 | 53.7269 | 16.1186 |
89
 
90
 
91
  ### Framework versions
92
 
93
+ - Transformers 4.46.3
94
+ - Pytorch 2.5.1+cxx11.abi
95
+ - Datasets 3.2.0
96
+ - Tokenizers 0.20.3
generation_config.json CHANGED
@@ -236,5 +236,5 @@
236
  "transcribe": 50359,
237
  "translate": 50358
238
  },
239
- "transformers_version": "4.44.2"
240
  }
 
236
  "transcribe": 50359,
237
  "translate": 50358
238
  },
239
+ "transformers_version": "4.46.3"
240
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:76b1053899a0943922fe951bf4819818f19af8b3f108ab84a9273fd160006f31
3
  size 151061672
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2162682de6d2ed6eb8ba1592473f6ae17a43491ac49579ab10a948fd86482f6e
3
  size 151061672