asahi417 commited on
Commit
aa89fea
·
verified ·
1 Parent(s): 0137a8d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -5
README.md CHANGED
@@ -20,12 +20,12 @@ Install library and download sample audio.
20
  pip install faster-whisper
21
  wget https://huggingface.co/kotoba-tech/kotoba-whisper-v1.0-ggml/resolve/main/sample_ja_speech.wav
22
  ```
23
- Inference with the kotoba-whisper-v2.0-faster.
24
 
25
  ```python
26
  from faster_whisper import WhisperModel
27
 
28
- model = WhisperModel("kotoba-tech/kotoba-whisper-v2.0-faster")
29
 
30
  segments, info = model.transcribe("sample_ja_speech.wav", language="ja", chunk_length=15, condition_on_previous_text=False)
31
  for segment in segments:
@@ -47,9 +47,10 @@ We measure the inference speed of different kotoba-whisper-v2.0 implementations
47
  |audio 4 | 5.6 | 35 | 126 | 69 |
48
 
49
  Scripts to re-run the experiment can be found bellow:
50
- * [whisper.cpp](https://huggingface.co/kotoba-tech/kotoba-whisper-v2.0-ggml/blob/main/benchmark.sh)
51
- * [faster-whisper](https://huggingface.co/kotoba-tech/kotoba-whisper-v2.0-faster/blob/main/benchmark.sh)
52
- * [hf pipeline](https://huggingface.co/kotoba-tech/kotoba-whisper-v2.0/blob/main/benchmark.sh)
 
53
  Also, currently whisper.cpp and faster-whisper support the [sequential long-form decoding](https://huggingface.co/distil-whisper/distil-large-v3#sequential-long-form),
54
  and only Huggingface pipeline supports the [chunked long-form decoding](https://huggingface.co/distil-whisper/distil-large-v3#chunked-long-form), which we empirically
55
  found better than the sequnential long-form decoding.
 
20
  pip install faster-whisper
21
  wget https://huggingface.co/kotoba-tech/kotoba-whisper-v1.0-ggml/resolve/main/sample_ja_speech.wav
22
  ```
23
+ Inference with the kotoba-whisper-bilingual-v1.0-faster.
24
 
25
  ```python
26
  from faster_whisper import WhisperModel
27
 
28
+ model = WhisperModel("kotoba-tech/kotoba-whisper-bilingual-v1.0-faster")
29
 
30
  segments, info = model.transcribe("sample_ja_speech.wav", language="ja", chunk_length=15, condition_on_previous_text=False)
31
  for segment in segments:
 
47
  |audio 4 | 5.6 | 35 | 126 | 69 |
48
 
49
  Scripts to re-run the experiment can be found bellow:
50
+ * [whisper.cpp](https://huggingface.co/kotoba-tech/kotoba-whisper-v1.0-ggml/blob/main/benchmark.sh)
51
+ * [faster-whisper](https://huggingface.co/kotoba-tech/kotoba-whisper-v1.0-faster/blob/main/benchmark.sh)
52
+ * [hf pipeline](https://huggingface.co/kotoba-tech/kotoba-whisper-v1.0/blob/main/benchmark.sh)
53
+
54
  Also, currently whisper.cpp and faster-whisper support the [sequential long-form decoding](https://huggingface.co/distil-whisper/distil-large-v3#sequential-long-form),
55
  and only Huggingface pipeline supports the [chunked long-form decoding](https://huggingface.co/distil-whisper/distil-large-v3#chunked-long-form), which we empirically
56
  found better than the sequnential long-form decoding.