easy-translate / examples /nllb200_3B_fp16.sh
Iker's picture
Add support for SMALL100 and update examples
911ab5a
raw
history blame
435 Bytes
# Run NLLB200-3B model on sample text. We use FP16 precision, which requires a GPU with a lot of VRAM
# For running this model in GPUs with less VRAM, use 8-bit quantization, see examples/nllb200_3B_8bit.sh
python3 translate.py \
--sentences_path sample_text/en.txt \
--output_path sample_text/en2es.translation.nllb-200_3B.txt \
--source_lang eng_Latn \
--target_lang spa_Latn \
--model_name facebook/nllb-200-3.3B \
--precision fp16