File size: 513 Bytes
e3b20e1 |
1 2 3 4 5 6 7 |
import sentencepiece as spm
spm.SentencePieceTrainer.train(input="/researchdisk/training_dataset_sentences/train.txt", model_prefix='spiece', vocab_size=32000, character_coverage=1.0,
pad_id=0, unk_id=2, eos_id=1, bos_id=-1,
user_defined_symbols=['[NLU]', '[NLG]', '[S2S]'],
train_extremely_large_corpus=True,
num_threads=96, input_sentence_size=50000000, shuffle_input_sentence=True) |