Update README.md
Browse files
README.md
CHANGED
@@ -104,3 +104,32 @@ if __name__ == "__main__":
|
|
104 |
|
105 |
trainer.save_model(training_args.output_dir)
|
106 |
```
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
104 |
|
105 |
trainer.save_model(training_args.output_dir)
|
106 |
```
|
107 |
+
|
108 |
+
### Test Script
|
109 |
+
```python
|
110 |
+
from vllm import LLM
|
111 |
+
from datasets import load_dataset
|
112 |
+
from vllm.sampling_params import SamplingParams
|
113 |
+
from transformers import AutoTokenizer
|
114 |
+
|
115 |
+
MODEL_PATH = "autodl-tmp/saves/Qwen2.5-1.5B-ultrachat200k"
|
116 |
+
|
117 |
+
model = LLM(MODEL_PATH,
|
118 |
+
tensor_parallel_size=1,
|
119 |
+
dtype='bfloat16')
|
120 |
+
tokenizer = AutoTokenizer.from_pretrained(MODEL_PATH)
|
121 |
+
|
122 |
+
input = tokenizer.apply_chat_template([{"role": "user", "content": "Which province is Shenyang in?"}],
|
123 |
+
tokenize=False,
|
124 |
+
add_generation_prompt=True)
|
125 |
+
sampling_params = SamplingParams(max_tokens=1024,
|
126 |
+
temperature=0.7,
|
127 |
+
logprobs=1,
|
128 |
+
stop_token_ids=[tokenizer.eos_token_id])
|
129 |
+
|
130 |
+
vllm_generations = model.generate(input,
|
131 |
+
sampling_params)
|
132 |
+
|
133 |
+
print(vllm_generations[0].outputs[0].text)
|
134 |
+
# print result: Shenyang is in Liaoning province, China.
|
135 |
+
```
|