Update README.md
Browse files
README.md
CHANGED
@@ -6,6 +6,35 @@ requires the tokenizer in tokenizers/
|
|
6 |
|
7 |
Voice latents pre-computed in voices/
|
8 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
9 |
---
|
10 |
license: other
|
11 |
language:
|
|
|
6 |
|
7 |
Voice latents pre-computed in voices/
|
8 |
|
9 |
+
For use in MRQ Voice Cloning WebUI:
|
10 |
+
|
11 |
+
Requires the tokenizer used in training, and code changes to disable text cleaners. At minimum, change english_cleaners to basic_cleaners.
|
12 |
+
|
13 |
+
Code changes:
|
14 |
+
modules\tortoise-tts\tortoise\utils\tokenizer.py
|
15 |
+
Change Line 201: txt = english_cleaners(txt) and replace it
|
16 |
+
with txt = basic_cleaners(txt)
|
17 |
+
|
18 |
+
modules\tortoise-tts\build\lib\tortoise\utils\tokenizer.py
|
19 |
+
Change Line 201: txt = english_cleaners(txt) and replace it
|
20 |
+
with txt = basic_cleaners(txt)
|
21 |
+
|
22 |
+
\modules\dlas\dlas\data\audio\paired_voice_audio_dataset.py
|
23 |
+
Line 133: return text_to_sequence(txt, ['english_cleaners'])
|
24 |
+
and replace it with: return text_to_sequence(txt, ['basic_cleaners'])
|
25 |
+
|
26 |
+
modules\dlas\dlas\data\audio\voice_tokenizer.py
|
27 |
+
Line 14: from dlas.models.audio.tts.tacotron2.text.cleaners import
|
28 |
+
english_cleaners
|
29 |
+
to: from dlas.models.audio.tts.tacotron2.text.cleaners import
|
30 |
+
english_cleaners, basic_cleaners
|
31 |
+
Line 85: txt = english_cleaners(txt) to txt =
|
32 |
+
basic_cleaners(txt)
|
33 |
+
Line 134: word = english_cleaners(word) to basic_cleaners(word)
|
34 |
+
|
35 |
+
Copy and paste German text into the tokenizer tester on the utilities
|
36 |
+
tab, and you should see it tokenized with all of the special
|
37 |
+
characters, and no [UNK].
|
38 |
---
|
39 |
license: other
|
40 |
language:
|