|
--- |
|
license: apache-2.0 |
|
datasets: |
|
- KBLab/rixvox |
|
language: |
|
- sv |
|
--- |
|
# Whisper Large RixVox Swedish |
|
|
|
This is a [Whisper large](https://huggingface.co./openai/whisper-large-v2) finetuned for Swedish using |
|
the [RixVox](https://huggingface.co./datasets/KBLab/rixvox) dataset. |
|
|
|
## Evaluation |
|
|
|
### [Common Voice 11](#): |
|
* WER: XYZ |
|
* WER (normalized): XYZ |
|
|
|
## Training |
|
|
|
Training was done using Huggingface and Deepspeed with ZeRO stage 2. |
|
|
|
* learning rate: 1e-5 |
|
* optimizer: CPUAdamW (Deepspeed) |
|
* lr scheduler: linear |
|
* warmup steps: 500 |
|
* per device batch size: 20 |
|
* GPUs: 8 x NVIDIA A100 40GB |
|
* total batch size: 160 |
|
* steps: 12000 |
|
* lowercase: no |
|
* fp16 |
|
* entire encoder was frozen |