Edit model card

# CaptainA_XLS-R_entropy-10_v0

This model is a fine-tuned version of facebook/wav2vec2-xls-r-300m on the Finnish Parliament Corpus with 210h of slow and clean samples. It achieves the following results on the evaluation set (70h of the slow and clean samples from the same corpus):

  • Loss: 196.9006
  • Cer: 0.0178
  • Wer: 0.0592

Model description

This model is used in the CaptainA app.

Intended uses & limitations

The model was fine-tuned with entropy regularization to improve generalization for the Finnish L2 MDD task. Even though the model was trained for the purpose of MDD for L2 Finnish speakers, it was not fine-tuned with any L2 data due to the lack of proper corpus for Finnish L2 MDD.

Therefore, it is important to note that this model is NOT intended for Finnish L2 ASR and will need to be improved further even for the MDD task (hence the model version is 0).

Because the model was fine-tuned with the Finnish Parliament Corpus, it has the same biases from that corpus. The biases are more notable since the model's intended uses are for L2 Finnish speakers. More detail can be read in the Master's thesis: A Mobile App For Practicing Finnish Pronunciation Using Wav2vec 2.0

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 5e-05
  • train_batch_size: 128
  • eval_batch_size: 128
  • seed: 1011
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • lr_scheduler_warmup_steps: 500
  • num_epochs: 10.0
  • mixed_precision_training: Native AMP

Training results

Training Loss Epoch Step Cer Validation Loss Wer
409.2657 0.47 500 1.0 409.9399 1.0
242.0796 0.94 1000 0.0762 222.4043 0.4204
214.4323 1.4 1500 0.0342 205.3006 0.1620
208.1767 1.87 2000 0.0261 201.7304 0.1095
205.3693 2.34 2500 0.0248 200.3012 0.1037
204.3477 2.81 3000 0.0219 199.3383 0.0830
202.9748 3.28 3500 0.0207 199.1589 0.0782
201.9818 3.75 4000 0.0207 198.5560 0.0769
201.8992 4.21 4500 0.0201 198.0990 0.0724
201.6079 4.68 5000 0.0209 197.8516 0.0712
200.6187 5.15 5500 0.0191 197.6185 0.0667
200.5608 5.62 6000 0.0189 197.5194 0.0658
200.1649 6.09 6500 0.0191 197.3655 0.0641
200.1713 6.55 7000 0.0186 197.2977 0.0629
200.1245 7.02 7500 0.0193 197.0914 0.0638
199.5289 7.49 8000 0.0181 197.0704 0.0608
199.4458 7.96 8500 0.0183 196.9986 0.0606
199.1502 8.43 9000 0.0178 197.0260 0.0590
199.4437 8.9 9500 0.0180 196.9412 0.0595
198.8669 9.36 10000 0.0180 196.8834 0.0600
199.1329 9.83 10500 0.0178 196.9176 0.0591

Framework versions

  • Transformers 4.18.0
  • Pytorch 1.12.0.dev20220305
  • Datasets 1.18.4.dev0
  • Tokenizers 0.11.6
Downloads last month
2
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.