add readme file
Browse files- .ipynb_checkpoints/added_tokens-checkpoint.json +0 -1
- .ipynb_checkpoints/config-checkpoint.json +0 -108
- .ipynb_checkpoints/preprocessor_config-checkpoint.json +0 -9
- .ipynb_checkpoints/special_tokens_map-checkpoint.json +0 -1
- .ipynb_checkpoints/tokenizer_config-checkpoint.json +0 -1
- .ipynb_checkpoints/vocab-checkpoint.json +0 -1
- README.md +57 -0
- vocab.json +1 -1
.ipynb_checkpoints/added_tokens-checkpoint.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"<s>": 42, "</s>": 43}
|
|
|
|
.ipynb_checkpoints/config-checkpoint.json
DELETED
@@ -1,108 +0,0 @@
|
|
1 |
-
{
|
2 |
-
"_name_or_path": "facebook/wav2vec2-xls-r-300m",
|
3 |
-
"activation_dropout": 0.0,
|
4 |
-
"adapter_kernel_size": 3,
|
5 |
-
"adapter_stride": 2,
|
6 |
-
"add_adapter": false,
|
7 |
-
"apply_spec_augment": true,
|
8 |
-
"architectures": [
|
9 |
-
"Wav2Vec2ForCTC"
|
10 |
-
],
|
11 |
-
"attention_dropout": 0.1,
|
12 |
-
"bos_token_id": 1,
|
13 |
-
"classifier_proj_size": 256,
|
14 |
-
"codevector_dim": 768,
|
15 |
-
"contrastive_logits_temperature": 0.1,
|
16 |
-
"conv_bias": true,
|
17 |
-
"conv_dim": [
|
18 |
-
512,
|
19 |
-
512,
|
20 |
-
512,
|
21 |
-
512,
|
22 |
-
512,
|
23 |
-
512,
|
24 |
-
512
|
25 |
-
],
|
26 |
-
"conv_kernel": [
|
27 |
-
10,
|
28 |
-
3,
|
29 |
-
3,
|
30 |
-
3,
|
31 |
-
3,
|
32 |
-
2,
|
33 |
-
2
|
34 |
-
],
|
35 |
-
"conv_stride": [
|
36 |
-
5,
|
37 |
-
2,
|
38 |
-
2,
|
39 |
-
2,
|
40 |
-
2,
|
41 |
-
2,
|
42 |
-
2
|
43 |
-
],
|
44 |
-
"ctc_loss_reduction": "mean",
|
45 |
-
"ctc_zero_infinity": false,
|
46 |
-
"diversity_loss_weight": 0.1,
|
47 |
-
"do_stable_layer_norm": true,
|
48 |
-
"eos_token_id": 2,
|
49 |
-
"feat_extract_activation": "gelu",
|
50 |
-
"feat_extract_dropout": 0.0,
|
51 |
-
"feat_extract_norm": "layer",
|
52 |
-
"feat_proj_dropout": 0.0,
|
53 |
-
"feat_quantizer_dropout": 0.0,
|
54 |
-
"final_dropout": 0.0,
|
55 |
-
"gradient_checkpointing": false,
|
56 |
-
"hidden_act": "gelu",
|
57 |
-
"hidden_dropout": 0.1,
|
58 |
-
"hidden_size": 1024,
|
59 |
-
"initializer_range": 0.02,
|
60 |
-
"intermediate_size": 4096,
|
61 |
-
"layer_norm_eps": 1e-05,
|
62 |
-
"layerdrop": 0.0,
|
63 |
-
"mask_feature_length": 64,
|
64 |
-
"mask_feature_min_masks": 0,
|
65 |
-
"mask_feature_prob": 0.25,
|
66 |
-
"mask_time_length": 10,
|
67 |
-
"mask_time_min_masks": 2,
|
68 |
-
"mask_time_prob": 0.75,
|
69 |
-
"model_type": "wav2vec2",
|
70 |
-
"num_adapter_layers": 3,
|
71 |
-
"num_attention_heads": 16,
|
72 |
-
"num_codevector_groups": 2,
|
73 |
-
"num_codevectors_per_group": 320,
|
74 |
-
"num_conv_pos_embedding_groups": 16,
|
75 |
-
"num_conv_pos_embeddings": 128,
|
76 |
-
"num_feat_extract_layers": 7,
|
77 |
-
"num_hidden_layers": 24,
|
78 |
-
"num_negatives": 100,
|
79 |
-
"output_hidden_size": 1024,
|
80 |
-
"pad_token_id": 41,
|
81 |
-
"proj_codevector_dim": 768,
|
82 |
-
"tdnn_dilation": [
|
83 |
-
1,
|
84 |
-
2,
|
85 |
-
3,
|
86 |
-
1,
|
87 |
-
1
|
88 |
-
],
|
89 |
-
"tdnn_dim": [
|
90 |
-
512,
|
91 |
-
512,
|
92 |
-
512,
|
93 |
-
512,
|
94 |
-
1500
|
95 |
-
],
|
96 |
-
"tdnn_kernel": [
|
97 |
-
5,
|
98 |
-
3,
|
99 |
-
3,
|
100 |
-
1,
|
101 |
-
1
|
102 |
-
],
|
103 |
-
"torch_dtype": "float32",
|
104 |
-
"transformers_version": "4.17.0.dev0",
|
105 |
-
"use_weighted_layer_sum": false,
|
106 |
-
"vocab_size": 44,
|
107 |
-
"xvector_output_dim": 512
|
108 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
.ipynb_checkpoints/preprocessor_config-checkpoint.json
DELETED
@@ -1,9 +0,0 @@
|
|
1 |
-
{
|
2 |
-
"do_normalize": true,
|
3 |
-
"feature_extractor_type": "Wav2Vec2FeatureExtractor",
|
4 |
-
"feature_size": 1,
|
5 |
-
"padding_side": "right",
|
6 |
-
"padding_value": 0.0,
|
7 |
-
"return_attention_mask": true,
|
8 |
-
"sampling_rate": 16000
|
9 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
.ipynb_checkpoints/special_tokens_map-checkpoint.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"bos_token": "<s>", "eos_token": "</s>", "unk_token": "[UNK]", "pad_token": "[PAD]", "additional_special_tokens": [{"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}]}
|
|
|
|
.ipynb_checkpoints/tokenizer_config-checkpoint.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"unk_token": "[UNK]", "bos_token": "<s>", "eos_token": "</s>", "pad_token": "[PAD]", "do_lower_case": false, "word_delimiter_token": "|", "special_tokens_map_file": null, "tokenizer_file": null, "name_or_path": "/workspace/container_0/model/", "tokenizer_class": "Wav2Vec2CTCTokenizer"}
|
|
|
|
.ipynb_checkpoints/vocab-checkpoint.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"aa": 1, "ae": 2, "ah": 3, "aw": 4, "ay": 5, "b": 6, "ch": 7, "d": 8, "dh": 9, "dx": 10, "eh": 11, "er": 12, "ey": 13, "f": 14, "g": 15, "h#": 16, "hh": 17, "ih": 18, "iy": 19, "jh": 20, "k": 21, "l": 22, "m": 23, "n": 24, "ng": 25, "ow": 26, "oy": 27, "p": 28, "r": 29, "s": 30, "sh": 31, "t": 32, "th": 33, "uh": 34, "uw": 35, "v": 36, "w": 37, "y": 38, "z": 39, "|": 0, "[UNK]": 40, "[PAD]": 41}
|
|
|
|
README.md
ADDED
@@ -0,0 +1,57 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
language:
|
3 |
+
- en
|
4 |
+
license: apache-2.0
|
5 |
+
tags:
|
6 |
+
- automatic-speech-recognition
|
7 |
+
- pytorch
|
8 |
+
- transformers
|
9 |
+
- en
|
10 |
+
- generated_from_trainer
|
11 |
+
model-index:
|
12 |
+
- name: model
|
13 |
+
results: []
|
14 |
+
---
|
15 |
+
|
16 |
+
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
17 |
+
should probably proofread and complete it, then remove this comment. -->
|
18 |
+
|
19 |
+
# model
|
20 |
+
|
21 |
+
This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the Timit dataset.
|
22 |
+
|
23 |
+
## Model description
|
24 |
+
|
25 |
+
More information needed
|
26 |
+
|
27 |
+
## Intended uses & limitations
|
28 |
+
|
29 |
+
More information needed
|
30 |
+
|
31 |
+
## Training and evaluation data
|
32 |
+
|
33 |
+
More information needed
|
34 |
+
|
35 |
+
## Training procedure
|
36 |
+
|
37 |
+
### Training hyperparameters
|
38 |
+
|
39 |
+
The following hyperparameters were used during training:
|
40 |
+
- learning_rate: 3e-05
|
41 |
+
- train_batch_size: 8
|
42 |
+
- eval_batch_size: 8
|
43 |
+
- seed: 42
|
44 |
+
- gradient_accumulation_steps: 4
|
45 |
+
- total_train_batch_size: 32
|
46 |
+
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
47 |
+
- lr_scheduler_type: linear
|
48 |
+
- lr_scheduler_warmup_steps: 2000
|
49 |
+
- training_steps: 10000
|
50 |
+
- mixed_precision_training: Native AMP
|
51 |
+
|
52 |
+
### Framework versions
|
53 |
+
|
54 |
+
- Transformers 4.17.0.dev0
|
55 |
+
- Pytorch 1.10.2+cu102
|
56 |
+
- Datasets 1.18.2.dev0
|
57 |
+
- Tokenizers 0.11.0
|
vocab.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"\u0251": 1, "\u00e6": 2, "\u0259": 3, "a\u028a": 4, "a\u026a": 5, "b": 6, "\u02a7": 7, "d": 8, "\u00f0": 9, "dx": 10, "\u025b": 11, "\u0259r": 12, "e\u026a": 13, "f": 14, "g": 15, "": 16, "h": 17, "\u026a": 18, "i": 19, "\u02a4": 20, "k": 21, "l": 22, "m": 23, "n": 24, "\u014b": 25, "o\u028a": 26, "\u0254\u026a": 27, "p": 28, "r": 29, "s": 30, "\u0283": 31, "t": 32, "\u03b8": 33, "\u028a": 34, "u": 35, "v": 36, "w": 37, "j": 38, "z": 39, "|": 0, "[UNK]": 40, "[PAD]": 41}
|
|
|
1 |
+
{"\u0251": 1, "\u00e6": 2, "\u0259": 3, "a\u028a": 4, "a\u026a": 5, "b": 6, "\u02a7": 7, "d": 8, "\u00f0": 9, "dx": 10, "\u025b": 11, "\u0259r": 12, "e\u026a": 13, "f": 14, "g": 15, " ": 16, "h": 17, "\u026a": 18, "i": 19, "\u02a4": 20, "k": 21, "l": 22, "m": 23, "n": 24, "\u014b": 25, "o\u028a": 26, "\u0254\u026a": 27, "p": 28, "r": 29, "s": 30, "\u0283": 31, "t": 32, "\u03b8": 33, "\u028a": 34, "u": 35, "v": 36, "w": 37, "j": 38, "z": 39, "|": 0, "[UNK]": 40, "[PAD]": 41}
|