Kavin1701 commited on
Commit
0775c5e
·
verified ·
1 Parent(s): 0c8736e

End of training

Browse files
README.md CHANGED
@@ -1,39 +1,19 @@
1
  ---
2
  license: apache-2.0
3
- base_model: facebook/wav2vec2-xls-r-300m
4
  tags:
5
  - generated_from_trainer
6
- datasets:
7
- - common_voice_12_0
8
- metrics:
9
- - wer
10
  model-index:
11
- - name: wav2vec2-large-xls-r-300m-tamil2-colab
12
- results:
13
- - task:
14
- name: Automatic Speech Recognition
15
- type: automatic-speech-recognition
16
- dataset:
17
- name: common_voice_12_0
18
- type: common_voice_12_0
19
- config: ta
20
- split: test[:5%]
21
- args: ta
22
- metrics:
23
- - name: Wer
24
- type: wer
25
- value: 1.0
26
  ---
27
 
28
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
29
  should probably proofread and complete it, then remove this comment. -->
30
 
31
- # wav2vec2-large-xls-r-300m-tamil2-colab
32
 
33
- This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the common_voice_12_0 dataset.
34
- It achieves the following results on the evaluation set:
35
- - Loss: 3.1835
36
- - Wer: 1.0
37
 
38
  ## Model description
39
 
@@ -53,29 +33,19 @@ More information needed
53
 
54
  The following hyperparameters were used during training:
55
  - learning_rate: 0.0003
56
- - train_batch_size: 16
57
  - eval_batch_size: 8
58
  - seed: 42
59
  - gradient_accumulation_steps: 2
60
- - total_train_batch_size: 32
61
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
62
  - lr_scheduler_type: linear
63
  - lr_scheduler_warmup_steps: 50
64
- - num_epochs: 5
65
  - mixed_precision_training: Native AMP
66
 
67
  ### Training results
68
 
69
- | Training Loss | Epoch | Step | Validation Loss | Wer |
70
- |:-------------:|:------:|:----:|:---------------:|:---:|
71
- | 10.7536 | 0.5970 | 40 | 4.9828 | 1.0 |
72
- | 3.5054 | 1.1940 | 80 | 3.3975 | 1.0 |
73
- | 3.2253 | 1.7910 | 120 | 3.3687 | 1.0 |
74
- | 3.185 | 2.3881 | 160 | 3.2591 | 1.0 |
75
- | 3.1631 | 2.9851 | 200 | 3.2295 | 1.0 |
76
- | 3.1451 | 3.5821 | 240 | 3.2064 | 1.0 |
77
- | 3.1281 | 4.1791 | 280 | 3.1945 | 1.0 |
78
- | 3.1231 | 4.7761 | 320 | 3.1835 | 1.0 |
79
 
80
 
81
  ### Framework versions
 
1
  ---
2
  license: apache-2.0
3
+ base_model: Amrrs/wav2vec2-large-xlsr-53-tamil
4
  tags:
5
  - generated_from_trainer
 
 
 
 
6
  model-index:
7
+ - name: wav2vec2-large-xls-r-300m-Tamil2-colab
8
+ results: []
 
 
 
 
 
 
 
 
 
 
 
 
 
9
  ---
10
 
11
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
  should probably proofread and complete it, then remove this comment. -->
13
 
14
+ # wav2vec2-large-xls-r-300m-Tamil2-colab
15
 
16
+ This model is a fine-tuned version of [Amrrs/wav2vec2-large-xlsr-53-tamil](https://huggingface.co/Amrrs/wav2vec2-large-xlsr-53-tamil) on the None dataset.
 
 
 
17
 
18
  ## Model description
19
 
 
33
 
34
  The following hyperparameters were used during training:
35
  - learning_rate: 0.0003
36
+ - train_batch_size: 8
37
  - eval_batch_size: 8
38
  - seed: 42
39
  - gradient_accumulation_steps: 2
40
+ - total_train_batch_size: 16
41
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
42
  - lr_scheduler_type: linear
43
  - lr_scheduler_warmup_steps: 50
44
+ - num_epochs: 2
45
  - mixed_precision_training: Native AMP
46
 
47
  ### Training results
48
 
 
 
 
 
 
 
 
 
 
 
49
 
50
 
51
  ### Framework versions
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "facebook/wav2vec2-xls-r-300m",
3
  "activation_dropout": 0.0,
4
  "adapter_attn_dim": null,
5
  "adapter_kernel_size": 3,
@@ -9,10 +9,10 @@
9
  "architectures": [
10
  "Wav2Vec2ForCTC"
11
  ],
12
- "attention_dropout": 0.2,
13
  "bos_token_id": 1,
14
  "classifier_proj_size": 256,
15
- "codevector_dim": 768,
16
  "contrastive_logits_temperature": 0.1,
17
  "conv_bias": true,
18
  "conv_dim": [
@@ -54,18 +54,26 @@
54
  "feat_quantizer_dropout": 0.0,
55
  "final_dropout": 0.0,
56
  "hidden_act": "gelu",
57
- "hidden_dropout": 0.0,
58
  "hidden_size": 1024,
59
  "initializer_range": 0.02,
60
  "intermediate_size": 4096,
61
  "layer_norm_eps": 1e-05,
62
- "layerdrop": 0.0,
 
 
 
 
 
63
  "mask_feature_length": 10,
64
  "mask_feature_min_masks": 0,
65
  "mask_feature_prob": 0.0,
66
  "mask_time_length": 10,
67
  "mask_time_min_masks": 2,
 
 
68
  "mask_time_prob": 0.05,
 
69
  "model_type": "wav2vec2",
70
  "num_adapter_layers": 3,
71
  "num_attention_heads": 16,
@@ -77,8 +85,8 @@
77
  "num_hidden_layers": 24,
78
  "num_negatives": 100,
79
  "output_hidden_size": 1024,
80
- "pad_token_id": 57,
81
- "proj_codevector_dim": 768,
82
  "tdnn_dilation": [
83
  1,
84
  2,
@@ -103,6 +111,6 @@
103
  "torch_dtype": "float32",
104
  "transformers_version": "4.41.2",
105
  "use_weighted_layer_sum": false,
106
- "vocab_size": 60,
107
  "xvector_output_dim": 512
108
  }
 
1
  {
2
+ "_name_or_path": "Amrrs/wav2vec2-large-xlsr-53-tamil",
3
  "activation_dropout": 0.0,
4
  "adapter_attn_dim": null,
5
  "adapter_kernel_size": 3,
 
9
  "architectures": [
10
  "Wav2Vec2ForCTC"
11
  ],
12
+ "attention_dropout": 0.1,
13
  "bos_token_id": 1,
14
  "classifier_proj_size": 256,
15
+ "codevector_dim": 256,
16
  "contrastive_logits_temperature": 0.1,
17
  "conv_bias": true,
18
  "conv_dim": [
 
54
  "feat_quantizer_dropout": 0.0,
55
  "final_dropout": 0.0,
56
  "hidden_act": "gelu",
57
+ "hidden_dropout": 0.1,
58
  "hidden_size": 1024,
59
  "initializer_range": 0.02,
60
  "intermediate_size": 4096,
61
  "layer_norm_eps": 1e-05,
62
+ "layerdrop": 0.1,
63
+ "mask_channel_length": 10,
64
+ "mask_channel_min_space": 1,
65
+ "mask_channel_other": 0.0,
66
+ "mask_channel_prob": 0.0,
67
+ "mask_channel_selection": "static",
68
  "mask_feature_length": 10,
69
  "mask_feature_min_masks": 0,
70
  "mask_feature_prob": 0.0,
71
  "mask_time_length": 10,
72
  "mask_time_min_masks": 2,
73
+ "mask_time_min_space": 1,
74
+ "mask_time_other": 0.0,
75
  "mask_time_prob": 0.05,
76
+ "mask_time_selection": "static",
77
  "model_type": "wav2vec2",
78
  "num_adapter_layers": 3,
79
  "num_attention_heads": 16,
 
85
  "num_hidden_layers": 24,
86
  "num_negatives": 100,
87
  "output_hidden_size": 1024,
88
+ "pad_token_id": 51,
89
+ "proj_codevector_dim": 256,
90
  "tdnn_dilation": [
91
  1,
92
  2,
 
111
  "torch_dtype": "float32",
112
  "transformers_version": "4.41.2",
113
  "use_weighted_layer_sum": false,
114
+ "vocab_size": 52,
115
  "xvector_output_dim": 512
116
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c6a599274c98477615bb1dafaefbbe38828af35311cecfd21151fcd68a2083c6
3
- size 1262053480
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c5a6b1f7a4942d8393e3d748bfee87ee24eb05f9277d78a45186f1775c28e500
3
+ size 1262020680
runs/Jun26_07-17-03_2e29ec53aaf1/events.out.tfevents.1719386266.2e29ec53aaf1.150.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2972e271e218a02425e5bf671ddf22e49a0fd9570845e4e2414e602566923eea
3
+ size 6846
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:32cdd5c4dfc6e62b4b5554d30df58da75c972eafbc01065ea779d9f14d6b65fd
3
  size 5112
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:157b23262ec98eac79179cd488d689286eefe54e696340d9b779cbbf4b03b208
3
  size 5112