Finetuned openai/whisper-small on 58000 Swahili training audio samples from mozilla-foundation/common_voice_17_0.

This model was created from the Mozilla.ai Blueprint: speech-to-text-finetune.

Evaluation results on 12253 audio samples of Swahili:

Baseline model (before finetuning) on Swahili

  • Word Error Rate: 133.795
  • Loss: 2.459

Finetuned model (after finetuning) on Swahili

  • Word Error Rate: 43.876
  • Loss: 0.653
Downloads last month
3
Safetensors
Model size
242M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for Mollel/ASR-Swahili-Small

Finetuned
(2324)
this model

Dataset used to train Mollel/ASR-Swahili-Small

Evaluation results