update readme to working code
Browse files
README.md
CHANGED
@@ -45,18 +45,19 @@ For more information on how the model was trained, please take a look at the [of
|
|
45 |
To transcribe audio files the model can be used as a standalone acoustic model as follows:
|
46 |
|
47 |
```python
|
48 |
-
|
49 |
import torchaudio
|
|
|
50 |
from transformers import MCTCTForCTC, MCTCTProcessor
|
51 |
|
52 |
-
model = MCTCTForCTC.from_pretrained("
|
53 |
-
processor = MCTCTProcessor.from_pretrained("
|
54 |
|
55 |
# load dummy dataset and read soundfiles
|
56 |
ds = load_dataset("patrickvonplaten/librispeech_asr_dummy", "clean", split="validation")
|
57 |
|
58 |
# tokenize
|
59 |
-
input_features = processor(ds[0]["audio"]["array"], return_tensors="pt"
|
60 |
|
61 |
# retrieve logits
|
62 |
logits = model(input_features).logits
|
|
|
45 |
To transcribe audio files the model can be used as a standalone acoustic model as follows:
|
46 |
|
47 |
```python
|
48 |
+
import torch
|
49 |
import torchaudio
|
50 |
+
from datasets import load_dataset
|
51 |
from transformers import MCTCTForCTC, MCTCTProcessor
|
52 |
|
53 |
+
model = MCTCTForCTC.from_pretrained("cwkeam/mctct-large")
|
54 |
+
processor = MCTCTProcessor.from_pretrained("cwkeam/mctct-large")
|
55 |
|
56 |
# load dummy dataset and read soundfiles
|
57 |
ds = load_dataset("patrickvonplaten/librispeech_asr_dummy", "clean", split="validation")
|
58 |
|
59 |
# tokenize
|
60 |
+
input_features = processor(ds[0]["audio"]["array"], return_tensors="pt").input_features
|
61 |
|
62 |
# retrieve logits
|
63 |
logits = model(input_features).logits
|