Update README.md
Browse files
README.md
CHANGED
@@ -37,7 +37,7 @@ More information needed
|
|
37 |
from transformers import AutoModelForCausalLM, AutoTokenizer, TextStreamer
|
38 |
import torch
|
39 |
|
40 |
-
|
41 |
torch.set_float32_matmul_precision('high')
|
42 |
|
43 |
device = "cuda"
|
@@ -49,12 +49,11 @@ model = AutoModelForCausalLM.from_pretrained(model_path,
|
|
49 |
trust_remote_code=True,
|
50 |
).to(device)
|
51 |
|
52 |
-
|
53 |
prompt = "自然言語処理とは、"
|
54 |
-
|
55 |
-
|
|
|
56 |
with torch.no_grad():
|
57 |
-
|
58 |
outputs = model.generate(
|
59 |
inputs["input_ids"],
|
60 |
max_new_tokens=4096,
|
|
|
37 |
from transformers import AutoModelForCausalLM, AutoTokenizer, TextStreamer
|
38 |
import torch
|
39 |
|
40 |
+
model_path = "ce-lery/mistral-2b-base"
|
41 |
torch.set_float32_matmul_precision('high')
|
42 |
|
43 |
device = "cuda"
|
|
|
49 |
trust_remote_code=True,
|
50 |
).to(device)
|
51 |
|
|
|
52 |
prompt = "自然言語処理とは、"
|
53 |
+
inputs = tokenizer(prompt,
|
54 |
+
add_special_tokens=True,
|
55 |
+
return_tensors="pt").to(model.device)
|
56 |
with torch.no_grad():
|
|
|
57 |
outputs = model.generate(
|
58 |
inputs["input_ids"],
|
59 |
max_new_tokens=4096,
|