TanelAlumae commited on
Commit
d4edeeb
·
1 Parent(s): c6a6ae4

Model files, fixed

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "openai/whisper-large-v3",
3
  "activation_dropout": 0.0,
4
  "activation_function": "gelu",
5
  "apply_spec_augment": false,
@@ -9,7 +9,7 @@
9
  "attention_dropout": 0.0,
10
  "begin_suppress_tokens": [
11
  220,
12
- 50257
13
  ],
14
  "bos_token_id": 50257,
15
  "classifier_proj_size": 256,
@@ -17,7 +17,7 @@
17
  "decoder_attention_heads": 20,
18
  "decoder_ffn_dim": 5120,
19
  "decoder_layerdrop": 0.0,
20
- "decoder_layers": 32,
21
  "decoder_start_token_id": 50258,
22
  "dropout": 0.0,
23
  "encoder_attention_heads": 20,
@@ -33,17 +33,16 @@
33
  "mask_time_length": 10,
34
  "mask_time_min_masks": 2,
35
  "mask_time_prob": 0.05,
36
- "max_length": 448,
37
  "max_source_positions": 1500,
38
  "max_target_positions": 448,
39
  "median_filter_width": 7,
40
  "model_type": "whisper",
41
  "num_hidden_layers": 32,
42
  "num_mel_bins": 128,
43
- "pad_token_id": 50256,
44
  "scale_embedding": false,
45
  "torch_dtype": "float16",
46
- "transformers_version": "4.39.3",
47
  "use_cache": true,
48
  "use_weighted_layer_sum": false,
49
  "vocab_size": 51866
 
1
  {
2
+ "_name_or_path": "openai/whisper-large-v3-turbo",
3
  "activation_dropout": 0.0,
4
  "activation_function": "gelu",
5
  "apply_spec_augment": false,
 
9
  "attention_dropout": 0.0,
10
  "begin_suppress_tokens": [
11
  220,
12
+ 50256
13
  ],
14
  "bos_token_id": 50257,
15
  "classifier_proj_size": 256,
 
17
  "decoder_attention_heads": 20,
18
  "decoder_ffn_dim": 5120,
19
  "decoder_layerdrop": 0.0,
20
+ "decoder_layers": 4,
21
  "decoder_start_token_id": 50258,
22
  "dropout": 0.0,
23
  "encoder_attention_heads": 20,
 
33
  "mask_time_length": 10,
34
  "mask_time_min_masks": 2,
35
  "mask_time_prob": 0.05,
 
36
  "max_source_positions": 1500,
37
  "max_target_positions": 448,
38
  "median_filter_width": 7,
39
  "model_type": "whisper",
40
  "num_hidden_layers": 32,
41
  "num_mel_bins": 128,
42
+ "pad_token_id": 50257,
43
  "scale_embedding": false,
44
  "torch_dtype": "float16",
45
+ "transformers_version": "4.47.1",
46
  "use_cache": true,
47
  "use_weighted_layer_sum": false,
48
  "vocab_size": 51866
generation_config.json CHANGED
@@ -1,5 +1,30 @@
1
  {
2
- "alignment_heads": null,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  "begin_suppress_tokens": [
4
  220,
5
  50257
@@ -14,7 +39,7 @@
14
  ],
15
  [
16
  2,
17
- 50359
18
  ]
19
  ],
20
  "is_multilingual": true,
@@ -117,13 +142,14 @@
117
  "<|vi|>": 50278,
118
  "<|yi|>": 50335,
119
  "<|yo|>": 50325,
 
120
  "<|zh|>": 50260
121
  },
122
  "max_initial_timestamp_index": 50,
123
  "max_length": 448,
124
- "no_timestamps_token_id": 50363,
125
  "pad_token_id": 50257,
126
- "prev_sot_token_id": 50361,
127
  "return_timestamps": false,
128
  "suppress_tokens": [
129
  1,
@@ -209,15 +235,15 @@
209
  49870,
210
  50254,
211
  50258,
212
- 50358,
213
  50359,
214
  50360,
215
  50361,
216
- 50362
 
217
  ],
218
  "task_to_id": {
219
- "transcribe": 50359,
220
- "translate": 50358
221
  },
222
- "transformers_version": "4.39.3"
223
  }
 
1
  {
2
+ "alignment_heads": [
3
+ [
4
+ 2,
5
+ 4
6
+ ],
7
+ [
8
+ 2,
9
+ 11
10
+ ],
11
+ [
12
+ 3,
13
+ 3
14
+ ],
15
+ [
16
+ 3,
17
+ 6
18
+ ],
19
+ [
20
+ 3,
21
+ 11
22
+ ],
23
+ [
24
+ 3,
25
+ 14
26
+ ]
27
+ ],
28
  "begin_suppress_tokens": [
29
  220,
30
  50257
 
39
  ],
40
  [
41
  2,
42
+ 50360
43
  ]
44
  ],
45
  "is_multilingual": true,
 
142
  "<|vi|>": 50278,
143
  "<|yi|>": 50335,
144
  "<|yo|>": 50325,
145
+ "<|yue|>": 50358,
146
  "<|zh|>": 50260
147
  },
148
  "max_initial_timestamp_index": 50,
149
  "max_length": 448,
150
+ "no_timestamps_token_id": 50364,
151
  "pad_token_id": 50257,
152
+ "prev_sot_token_id": 50362,
153
  "return_timestamps": false,
154
  "suppress_tokens": [
155
  1,
 
235
  49870,
236
  50254,
237
  50258,
 
238
  50359,
239
  50360,
240
  50361,
241
+ 50362,
242
+ 50363
243
  ],
244
  "task_to_id": {
245
+ "transcribe": 50360,
246
+ "translate": 50359
247
  },
248
+ "transformers_version": "4.47.1"
249
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9ef797c6dbb3256d71d68055f51074587e8b30e863b6e4be23fa77a293c14971
3
- size 3235581408
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e096126d1f60b49040516e5c65719b5cadaa9c2113f961384e9b0790e04733f
3
+ size 1750601920
tokenizer_config.json CHANGED
@@ -12987,6 +12987,7 @@
12987
  "clean_up_tokenization_spaces": true,
12988
  "eos_token": "<|endoftext|>",
12989
  "errors": "replace",
 
12990
  "model_max_length": 1000000000000000019884624838656,
12991
  "pad_token": "<|endoftext|>",
12992
  "processor_class": "WhisperProcessor",
 
12987
  "clean_up_tokenization_spaces": true,
12988
  "eos_token": "<|endoftext|>",
12989
  "errors": "replace",
12990
+ "extra_special_tokens": {},
12991
  "model_max_length": 1000000000000000019884624838656,
12992
  "pad_token": "<|endoftext|>",
12993
  "processor_class": "WhisperProcessor",