Text-to-Speech
PyTorch
ONNX
Catalan
matcha-tts
acoustic modelling
speech
multispeaker
AlexK-PL commited on
Commit
e96083b
1 Parent(s): 0fd399b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +19 -2
README.md CHANGED
@@ -97,10 +97,27 @@ pip install -e .
97
 
98
  #### PyTorch
99
 
100
- Speech inference can be done with **Catalan Matcha-TTS** by loading the model remotely with the HF hub.
 
101
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
102
  ```bash
103
- pip install git+https://github.com/langtech-bsc/Matcha-TTS.git@dev-cat
104
 
105
  ```
106
 
 
97
 
98
  #### PyTorch
99
 
100
+ Speech end-to-end inference can be done together with **Catalan Matcha-TTS**.
101
+ Both models (Catalan Matcha-TTS and Vocos) are loaded remotely from the HF hub.
102
 
103
+ First, export the following environment variables to include the installed espeak-ng version:
104
+
105
+ ```bash
106
+ export PYTHON=/path/to/your/venv/bin/python
107
+ export ESPEAK_DATA_PATH=/path/to/espeak-ng/espeak-ng-data
108
+ export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:/path/to/espeak-ng/lib
109
+ export PATH="/path/to/espeak-ng/bin:$PATH"
110
+
111
+ ```
112
+ Then you can run the inference script:
113
+ ```bash
114
+ cd Matcha-TTS
115
+ python3 matcha_vocos_inference.py --output_path=/output/path --text_input="Bon dia Manel, avui anem a la muntanya."
116
+
117
+ ```
118
+ You can also modify the length scale (speech rate) and the temperature of the generated sample:
119
  ```bash
120
+ python3 matcha_vocos_inference.py --output_path=/output/path --text_input="Bon dia Manel, avui anem a la muntanya." --length_scale=0.8 --temperature=0.7
121
 
122
  ```
123