Update README.md
Browse files
README.md
CHANGED
@@ -45,8 +45,8 @@ For example, to chat with the biomedicine model:
|
|
45 |
```python
|
46 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
47 |
|
48 |
-
model = AutoModelForCausalLM.from_pretrained("AdaptLLM/medicine-
|
49 |
-
tokenizer = AutoTokenizer.from_pretrained("AdaptLLM/medicine-
|
50 |
|
51 |
# Put your input here:
|
52 |
user_input = '''Question: Which of the following is an example of monosomy?
|
@@ -58,11 +58,11 @@ Options:
|
|
58 |
|
59 |
Please provide your choice first and then provide explanations if possible.'''
|
60 |
|
61 |
-
#
|
62 |
-
prompt =
|
63 |
|
64 |
inputs = tokenizer(prompt, return_tensors="pt", add_special_tokens=False).input_ids.to(model.device)
|
65 |
-
outputs = model.generate(input_ids=inputs, max_length=
|
66 |
|
67 |
answer_start = int(inputs.shape[-1])
|
68 |
pred = tokenizer.decode(outputs[answer_start:], skip_special_tokens=True)
|
|
|
45 |
```python
|
46 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
47 |
|
48 |
+
model = AutoModelForCausalLM.from_pretrained("AdaptLLM/medicine-LLM-13B")
|
49 |
+
tokenizer = AutoTokenizer.from_pretrained("AdaptLLM/medicine-LLM-13B", use_fast=False)
|
50 |
|
51 |
# Put your input here:
|
52 |
user_input = '''Question: Which of the following is an example of monosomy?
|
|
|
58 |
|
59 |
Please provide your choice first and then provide explanations if possible.'''
|
60 |
|
61 |
+
# Simply use your input as the prompt
|
62 |
+
prompt = user_input
|
63 |
|
64 |
inputs = tokenizer(prompt, return_tensors="pt", add_special_tokens=False).input_ids.to(model.device)
|
65 |
+
outputs = model.generate(input_ids=inputs, max_length=2048)[0]
|
66 |
|
67 |
answer_start = int(inputs.shape[-1])
|
68 |
pred = tokenizer.decode(outputs[answer_start:], skip_special_tokens=True)
|