|
--- |
|
license: openrail |
|
--- |
|
|
|
# PMC_LLaMA |
|
|
|
To obtain the foundation model in medical field, we propose [MedLLaMA_13B](https://huggingface.co./chaoyi-wu/MedLLaMA_13B) and PMC_LLaMA_13B. |
|
|
|
MedLLaMA_13B is initialized from LLaMA-13B and further pretrained with medical corpus. Despite the expert knowledge gained, it lacks instruction-following ability. |
|
Hereby we construct a instruction-tuning dataset and evaluate the tuned model. |
|
|
|
As shown in the table, PMC_LLaMA_13B achieves comparable results to ChatGPT on medical QA benchmarks. |
|
|
|
![medical_qa](https://pic4.zhimg.com/80/v2-bf43393cd753018e11fdb1c64a1a87df.png) |
|
|
|
|
|
## Usage |
|
|
|
```python |
|
import transformers |
|
import torch |
|
|
|
tokenizer = transformers.LlamaTokenizer.from_pretrained('axiong/PMC_LLaMA_13B') |
|
model = transformers.LlamaForCausalLM.from_pretrained('axiong/PMC_LLaMA_13B') |
|
|
|
sentence = 'Hello, doctor' |
|
batch = tokenizer( |
|
sentence, |
|
return_tensors="pt", |
|
add_special_tokens=False |
|
) |
|
with torch.no_grad(): |
|
generated = model.generate( |
|
inputs = batch["input_ids"], |
|
max_length=200, |
|
do_sample=True, |
|
top_k=50 |
|
) |
|
print('model predict: ',tokenizer.decode(generated[0])) |
|
``` |
|
|
|
|