Update README.md
Browse files
README.md
CHANGED
@@ -135,6 +135,13 @@ model = AutoModelForCausalLM.from_pretrained("norallm/normistral-11b").cuda().ev
|
|
135 |
prompt = """Engelsk: {0}
|
136 |
Bokmål:"""
|
137 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
138 |
# Generation function
|
139 |
@torch.no_grad()
|
140 |
def generate(text):
|
@@ -144,7 +151,7 @@ def generate(text):
|
|
144 |
input_ids,
|
145 |
max_new_tokens=64,
|
146 |
do_sample=False,
|
147 |
-
eos_token_id=
|
148 |
)
|
149 |
return tokenizer.decode(prediction[0, input_ids.size(1):]).strip()
|
150 |
|
|
|
135 |
prompt = """Engelsk: {0}
|
136 |
Bokmål:"""
|
137 |
|
138 |
+
# Define tokens that should end the generation (any token with a newline)
|
139 |
+
eos_token_ids = [
|
140 |
+
token_id
|
141 |
+
for token_id in range(tokenizer.vocab_size)
|
142 |
+
if '\n' in tokenizer.decode([token_id])
|
143 |
+
]
|
144 |
+
|
145 |
# Generation function
|
146 |
@torch.no_grad()
|
147 |
def generate(text):
|
|
|
151 |
input_ids,
|
152 |
max_new_tokens=64,
|
153 |
do_sample=False,
|
154 |
+
eos_token_id=eos_token_ids
|
155 |
)
|
156 |
return tokenizer.decode(prediction[0, input_ids.size(1):]).strip()
|
157 |
|