Update README.md
Browse files
README.md
CHANGED
@@ -20,12 +20,12 @@ Install library and download sample audio.
|
|
20 |
pip install faster-whisper
|
21 |
wget https://huggingface.co/kotoba-tech/kotoba-whisper-v1.0-ggml/resolve/main/sample_ja_speech.wav
|
22 |
```
|
23 |
-
Inference with the kotoba-whisper-
|
24 |
|
25 |
```python
|
26 |
from faster_whisper import WhisperModel
|
27 |
|
28 |
-
model = WhisperModel("kotoba-tech/kotoba-whisper-
|
29 |
|
30 |
segments, info = model.transcribe("sample_ja_speech.wav", language="ja", chunk_length=15, condition_on_previous_text=False)
|
31 |
for segment in segments:
|
@@ -47,9 +47,10 @@ We measure the inference speed of different kotoba-whisper-v2.0 implementations
|
|
47 |
|audio 4 | 5.6 | 35 | 126 | 69 |
|
48 |
|
49 |
Scripts to re-run the experiment can be found bellow:
|
50 |
-
* [whisper.cpp](https://huggingface.co/kotoba-tech/kotoba-whisper-
|
51 |
-
* [faster-whisper](https://huggingface.co/kotoba-tech/kotoba-whisper-
|
52 |
-
* [hf pipeline](https://huggingface.co/kotoba-tech/kotoba-whisper-
|
|
|
53 |
Also, currently whisper.cpp and faster-whisper support the [sequential long-form decoding](https://huggingface.co/distil-whisper/distil-large-v3#sequential-long-form),
|
54 |
and only Huggingface pipeline supports the [chunked long-form decoding](https://huggingface.co/distil-whisper/distil-large-v3#chunked-long-form), which we empirically
|
55 |
found better than the sequnential long-form decoding.
|
|
|
20 |
pip install faster-whisper
|
21 |
wget https://huggingface.co/kotoba-tech/kotoba-whisper-v1.0-ggml/resolve/main/sample_ja_speech.wav
|
22 |
```
|
23 |
+
Inference with the kotoba-whisper-bilingual-v1.0-faster.
|
24 |
|
25 |
```python
|
26 |
from faster_whisper import WhisperModel
|
27 |
|
28 |
+
model = WhisperModel("kotoba-tech/kotoba-whisper-bilingual-v1.0-faster")
|
29 |
|
30 |
segments, info = model.transcribe("sample_ja_speech.wav", language="ja", chunk_length=15, condition_on_previous_text=False)
|
31 |
for segment in segments:
|
|
|
47 |
|audio 4 | 5.6 | 35 | 126 | 69 |
|
48 |
|
49 |
Scripts to re-run the experiment can be found bellow:
|
50 |
+
* [whisper.cpp](https://huggingface.co/kotoba-tech/kotoba-whisper-v1.0-ggml/blob/main/benchmark.sh)
|
51 |
+
* [faster-whisper](https://huggingface.co/kotoba-tech/kotoba-whisper-v1.0-faster/blob/main/benchmark.sh)
|
52 |
+
* [hf pipeline](https://huggingface.co/kotoba-tech/kotoba-whisper-v1.0/blob/main/benchmark.sh)
|
53 |
+
|
54 |
Also, currently whisper.cpp and faster-whisper support the [sequential long-form decoding](https://huggingface.co/distil-whisper/distil-large-v3#sequential-long-form),
|
55 |
and only Huggingface pipeline supports the [chunked long-form decoding](https://huggingface.co/distil-whisper/distil-large-v3#chunked-long-form), which we empirically
|
56 |
found better than the sequnential long-form decoding.
|