fsicoli commited on
Commit
2a26cc1
1 Parent(s): ec363bb

Model save

Browse files
README.md CHANGED
@@ -1,4 +1,5 @@
1
  ---
 
2
  license: apache-2.0
3
  base_model: openai/whisper-large-v3
4
  tags:
@@ -22,7 +23,7 @@ model-index:
22
  metrics:
23
  - name: Wer
24
  type: wer
25
- value: 0.10366752081998719
26
  ---
27
 
28
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -32,8 +33,8 @@ should probably proofread and complete it, then remove this comment. -->
32
 
33
  This model is a fine-tuned version of [openai/whisper-large-v3](https://huggingface.co/openai/whisper-large-v3) on the common_voice_18_0 dataset.
34
  It achieves the following results on the evaluation set:
35
- - Loss: 0.1486
36
- - Wer: 0.1037
37
 
38
  ## Model description
39
 
@@ -56,24 +57,33 @@ The following hyperparameters were used during training:
56
  - train_batch_size: 8
57
  - eval_batch_size: 8
58
  - seed: 42
59
- - gradient_accumulation_steps: 32
60
- - total_train_batch_size: 256
61
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
62
  - lr_scheduler_type: linear
63
  - lr_scheduler_warmup_steps: 1000
64
- - num_epochs: 2.0
65
  - mixed_precision_training: Native AMP
66
 
67
  ### Training results
68
 
69
  | Training Loss | Epoch | Step | Validation Loss | Wer |
70
  |:-------------:|:------:|:----:|:---------------:|:------:|
71
- | 0.13 | 0.9998 | 691 | 0.1486 | 0.1037 |
 
 
 
 
 
 
 
 
 
72
 
73
 
74
  ### Framework versions
75
 
76
- - Transformers 4.44.0.dev0
77
  - Pytorch 2.4.0+cu124
78
  - Datasets 2.18.1.dev0
79
  - Tokenizers 0.19.1
 
1
  ---
2
+ library_name: transformers
3
  license: apache-2.0
4
  base_model: openai/whisper-large-v3
5
  tags:
 
23
  metrics:
24
  - name: Wer
25
  type: wer
26
+ value: 0.10419602818705957
27
  ---
28
 
29
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
33
 
34
  This model is a fine-tuned version of [openai/whisper-large-v3](https://huggingface.co/openai/whisper-large-v3) on the common_voice_18_0 dataset.
35
  It achieves the following results on the evaluation set:
36
+ - Loss: 0.2671
37
+ - Wer: 0.1042
38
 
39
  ## Model description
40
 
 
57
  - train_batch_size: 8
58
  - eval_batch_size: 8
59
  - seed: 42
60
+ - gradient_accumulation_steps: 4
61
+ - total_train_batch_size: 32
62
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
63
  - lr_scheduler_type: linear
64
  - lr_scheduler_warmup_steps: 1000
65
+ - num_epochs: 10.0
66
  - mixed_precision_training: Native AMP
67
 
68
  ### Training results
69
 
70
  | Training Loss | Epoch | Step | Validation Loss | Wer |
71
  |:-------------:|:------:|:----:|:---------------:|:------:|
72
+ | 0.1388 | 0.9996 | 691 | 0.1501 | 0.1074 |
73
+ | 0.108 | 1.9993 | 1382 | 0.1619 | 0.1153 |
74
+ | 0.091 | 2.9989 | 2073 | 0.1697 | 0.1124 |
75
+ | 0.0461 | 4.0 | 2765 | 0.1764 | 0.1120 |
76
+ | 0.0264 | 4.9996 | 3456 | 0.2024 | 0.1133 |
77
+ | 0.0203 | 5.9993 | 4147 | 0.2200 | 0.1099 |
78
+ | 0.0129 | 6.9989 | 4838 | 0.2277 | 0.1114 |
79
+ | 0.0091 | 8.0 | 5530 | 0.2552 | 0.1067 |
80
+ | 0.0063 | 8.9996 | 6221 | 0.2565 | 0.1054 |
81
+ | 0.0019 | 9.9964 | 6910 | 0.2671 | 0.1042 |
82
 
83
 
84
  ### Framework versions
85
 
86
+ - Transformers 4.45.0.dev0
87
  - Pytorch 2.4.0+cu124
88
  - Datasets 2.18.1.dev0
89
  - Tokenizers 0.19.1
generation_config.json CHANGED
@@ -48,16 +48,6 @@
48
  "bos_token_id": 50257,
49
  "decoder_start_token_id": 50258,
50
  "eos_token_id": 50257,
51
- "forced_decoder_ids": [
52
- [
53
- 1,
54
- null
55
- ],
56
- [
57
- 2,
58
- 50360
59
- ]
60
- ],
61
  "is_multilingual": true,
62
  "lang_to_id": {
63
  "<|af|>": 50327,
@@ -161,7 +151,7 @@
161
  "<|yue|>": 50358,
162
  "<|zh|>": 50260
163
  },
164
- "language": "<|pt|>",
165
  "max_initial_timestamp_index": 50,
166
  "max_length": 448,
167
  "no_timestamps_token_id": 50364,
@@ -258,9 +248,10 @@
258
  50362,
259
  50363
260
  ],
 
261
  "task_to_id": {
262
  "transcribe": 50360,
263
  "translate": 50359
264
  },
265
- "transformers_version": "4.44.0.dev0"
266
  }
 
48
  "bos_token_id": 50257,
49
  "decoder_start_token_id": 50258,
50
  "eos_token_id": 50257,
 
 
 
 
 
 
 
 
 
 
51
  "is_multilingual": true,
52
  "lang_to_id": {
53
  "<|af|>": 50327,
 
151
  "<|yue|>": 50358,
152
  "<|zh|>": 50260
153
  },
154
+ "language": "portuguese",
155
  "max_initial_timestamp_index": 50,
156
  "max_length": 448,
157
  "no_timestamps_token_id": 50364,
 
248
  50362,
249
  50363
250
  ],
251
+ "task": "transcribe",
252
  "task_to_id": {
253
  "transcribe": 50360,
254
  "translate": 50359
255
  },
256
+ "transformers_version": "4.45.0.dev0"
257
  }
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:524e5d905ac451a424fcd908db2e019cc0c8127f75d9ddc60f72eb9e8c820905
3
  size 4993448880
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d0d8543f8a343cf6bd7d5acea5735e038153f225305548ad9147e66d092130fd
3
  size 4993448880
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dffabf943db6878e36e9ec8c9f8d318d8be564694a0294ca3c9ccdaccb18bae2
3
  size 1180663192
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:86eb946aa1f18d3940a43139960cc047edc7ac2addae69ac7012130eb797baa6
3
  size 1180663192
runs/Aug20_11-34-00_DITEC2014063010/events.out.tfevents.1724164511.DITEC2014063010.27248.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:094f803784c837c850ecbc36ee536ab49a222e38a170277aa75622860688921d
3
- size 67173
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cd31673c00dee3b5d966fc2e0d3187041e316f31dc7b87e1207763f5907c9fab
3
+ size 67527