File size: 4,685 Bytes
3b32d4d
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
---
language:
- fa
license: apache-2.0
base_model: openai/whisper-large-v3
tags:
- generated_from_trainer
datasets:
- mozilla-foundation-common-voice-17-0
metrics:
- wer
model-index:
- name: Whisper LargeV3 Persian - Persian ASR
  results:
  - task:
      name: Automatic Speech Recognition
      type: automatic-speech-recognition
    dataset:
      name: common-voice-17-0
      type: mozilla-foundation-common-voice-17-0
      config: default
      split: test[:10%]
      args: 'config: Persian, split: train[:10%]+validation[:10%]'
    metrics:
    - name: Wer
      type: wer
      value: 38.94514767932489
---

<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->

# Whisper LargeV3 Persian - Persian ASR

This model is a fine-tuned version of [openai/whisper-large-v3](https://huggingface.co./openai/whisper-large-v3)on the Common Voice 17.0 dataset in Persian. 
The model has been trained for Automatic Speech Recognition (ASR) and is capable of converting spoken Persian into text. 
The following sections provide more details on its performance, intended uses, training data, and the procedure followed during training.
It achieves the following results on the evaluation set:
- Loss: 0.4072
- Wer: 38.9451

## Model description

This model leverages the Whisper architecture, known for its effectiveness in multilingual ASR tasks. 
Whisper models are trained on a large corpus of multilingual and multitask supervised data, 
enabling them to generalize well across different languages, including low-resource languages like Persian. 
This fine-tuned model is specifically adapted for Persian, improving its accuracy on Persian speech recognition tasks.

## Intended uses & limitations

This model is designed for speech-to-text tasks in the Persian language. It can be used for applications like transcription of audio files, voice-controlled systems,
and any task requiring accurate conversion of spoken Persian into text. However, the model may have limitations when dealing with noisy audio environments, 
diverse accents, or highly technical vocabulary not present in the training data. 
It's recommended to fine-tune the model further if your use case involves specialized language or contexts.

## Training and evaluation data

The model was fine-tuned using the Common Voice 17.0 dataset, which is a crowd-sourced dataset containing diverse voices in Persian. 
The dataset was split into training, validation, and test sets. The training set includes a variety of speakers, ages, and accents, 
making the model robust across different dialects of Persian. The test split used for evaluation represents approximately 10% of the total data, ensuring a reliable assessment of the model's performance.

## Training procedure

### Training hyperparameters

The following hyperparameters were used during training:
- learning_rate: 1e-05
- train_batch_size: 4
- eval_batch_size: 4
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08,which helps in maintaining stability during training.
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 500
- num_epochs: 1 ,meaning the model was trained over the entire dataset once.
- mixed_precision_training: Native AMP, which allows for faster training by using lower precision without significant loss in accuracy.

### Training results

During training, the model achieved the following results:

  -  Training Loss: 0.2083 at the end of 1 epoch.
  -  Validation Loss: 0.4072, showing how well the model generalizes to unseen data.
  -  Word Error Rate (WER): 38.9451, indicating the percentage of words incorrectly predicted during the ASR task on the validation set.

| Training Loss | Epoch | Step | Validation Loss | Wer     |
|:-------------:|:-----:|:----:|:---------------:|:-------:|
| 0.2083        | 1.0   | 987  | 0.4072          | 38.9451 |

These results highlight the model's ability to perform well on the given dataset, though there may be room for further optimization and fine-tuning.

### Framework versions

The model was trained using the following versions of libraries:

  - Transformers: 4.44.0, which provides the necessary tools and APIs to fine-tune transformer models like Whisper.
  - Pytorch: 2.4.0+cu121, the deep learning framework used to build and train the model.
  - Datasets: 2.21.0, which facilitated the loading and preprocessing of the Common Voice dataset.
  - Tokenizers: 0.19, used for efficiently handling text tokenization required by the model.

- Transformers 4.44.0
- Pytorch 2.4.0+cu121
- Datasets 2.21.0
- Tokenizers 0.19.1