Update README.md
Browse files
README.md
CHANGED
@@ -90,7 +90,7 @@ Here is a summary of the model hyper-parameters:
|
|
90 |
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
91 |
model = AutoModelForCausalLM.from_pretrained(model_id)
|
92 |
|
93 |
-
text = '<|im_start|>user\nTranslate the following English source text to Portuguese:\nEnglish: I am a language model for european languages. \nPortuguese: <|im_end|>\n<|im_start|>assistant\n'
|
94 |
|
95 |
inputs = tokenizer(text, return_tensors="pt")
|
96 |
outputs = model.generate(**inputs, max_new_tokens=20)
|
|
|
90 |
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
91 |
model = AutoModelForCausalLM.from_pretrained(model_id)
|
92 |
|
93 |
+
text = '<|im_start|>system\n<|im_end|>\n<|im_start|>user\nTranslate the following English source text to Portuguese:\nEnglish: I am a language model for european languages. \nPortuguese: <|im_end|>\n<|im_start|>assistant\n'
|
94 |
|
95 |
inputs = tokenizer(text, return_tensors="pt")
|
96 |
outputs = model.generate(**inputs, max_new_tokens=20)
|