CheeLi03 commited on
Commit
9cd3f10
·
verified ·
1 Parent(s): 6427629

Upload tokenizer

Browse files
Files changed (2) hide show
  1. README.md +8 -8
  2. vocab.json +0 -0
README.md CHANGED
@@ -1,21 +1,21 @@
1
  ---
 
 
 
2
  language:
3
  - it
4
  license: apache-2.0
5
- base_model: openai/whisper-tiny
 
6
  tags:
7
  - hf-asr-leaderboard
8
  - generated_from_trainer
9
- datasets:
10
- - fleurs
11
- metrics:
12
- - wer
13
  model-index:
14
  - name: Whisper Tiny Italian 5k - Chee Li
15
  results:
16
  - task:
17
- name: Automatic Speech Recognition
18
  type: automatic-speech-recognition
 
19
  dataset:
20
  name: Google Fleurs
21
  type: fleurs
@@ -23,9 +23,9 @@ model-index:
23
  split: None
24
  args: 'config: it split: test'
25
  metrics:
26
- - name: Wer
27
- type: wer
28
  value: 50.93909245328804
 
29
  ---
30
 
31
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
1
  ---
2
+ base_model: openai/whisper-tiny
3
+ datasets:
4
+ - fleurs
5
  language:
6
  - it
7
  license: apache-2.0
8
+ metrics:
9
+ - wer
10
  tags:
11
  - hf-asr-leaderboard
12
  - generated_from_trainer
 
 
 
 
13
  model-index:
14
  - name: Whisper Tiny Italian 5k - Chee Li
15
  results:
16
  - task:
 
17
  type: automatic-speech-recognition
18
+ name: Automatic Speech Recognition
19
  dataset:
20
  name: Google Fleurs
21
  type: fleurs
 
23
  split: None
24
  args: 'config: it split: test'
25
  metrics:
26
+ - type: wer
 
27
  value: 50.93909245328804
28
+ name: Wer
29
  ---
30
 
31
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
vocab.json CHANGED
The diff for this file is too large to render. See raw diff