|
--- |
|
language: |
|
- multilingual |
|
- ain |
|
license: apache-2.0 |
|
--- |
|
|
|
## Wav2Vec2-Large-XLSR-53 pretrained on Ainu language data |
|
|
|
This is a [wav2vec-large-xlsr-53](https://huggingface.co./facebook/wav2vec2-large-xlsr-53) model adapted for the Ainu language by performing continued pretraining for 100k steps on 234 hours of speech data in Hokkaido Ainu and Sakhalin Ainu. |
|
For details, please refer to the [paper](https://arxiv.org/abs/2301.07295). |
|
|
|
A model fine-tuned for automatic transcription of Sakhalin Ainu is also available: [wav2vec2-large-xlsr-53-ain-sakh](https://huggingface.co./karolnowakowski/wav2vec2-large-xlsr-53-ain-sakh). |
|
|
|
## Citation |
|
When using the model please cite the following paper: |
|
```bibtex |
|
@article{NOWAKOWSKI2023103148, |
|
title = {Adapting multilingual speech representation model for a new, underresourced language through multilingual fine-tuning and continued pretraining}, |
|
journal = {Information Processing & Management}, |
|
volume = {60}, |
|
number = {2}, |
|
pages = {103148}, |
|
year = {2023}, |
|
issn = {0306-4573}, |
|
doi = {https://doi.org/10.1016/j.ipm.2022.103148}, |
|
url = {https://www.sciencedirect.com/science/article/pii/S0306457322002497}, |
|
author = {Karol Nowakowski and Michal Ptaszynski and Kyoko Murasaki and Jagna Nieuważny}, |
|
keywords = {Automatic speech transcription, ASR, Wav2vec 2.0, Pretrained transformer models, Speech representation models, Cross-lingual transfer, Language documentation, Endangered languages, Underresourced languages, Sakhalin Ainu}, |
|
abstract = {In recent years, neural models learned through self-supervised pretraining on large scale multilingual text or speech data have exhibited promising results for underresourced languages, especially when a relatively large amount of data from related language(s) is available. While the technology has a potential for facilitating tasks carried out in language documentation projects, such as speech transcription, pretraining a multilingual model from scratch for every new language would be highly impractical. We investigate the possibility for adapting an existing multilingual wav2vec 2.0 model for a new language, focusing on actual fieldwork data from a critically endangered tongue: Ainu. Specifically, we (i) examine the feasibility of leveraging data from similar languages also in fine-tuning; (ii) verify whether the model’s performance can be improved by further pretraining on target language data. Our results show that continued pretraining is the most effective method to adapt a wav2vec 2.0 model for a new language and leads to considerable reduction in error rates. Furthermore, we find that if a model pretrained on a related speech variety or an unrelated language with similar phonological characteristics is available, multilingual fine-tuning using additional data from that language can have positive impact on speech recognition performance when there is very little labeled data in the target language.} |
|
} |
|
``` |