shahules786's picture
Update README.md
a33cb2e
## Training details
- Dataset used: Explanation style datasets from psmathur/WizardLM_Orca and Dahoas/cot_gsm8k
- Techniques: fp16 bit precision training + QLoRA + DeepSpeed
- Machine: V100 (16GB) * 2
## Inference
```python
from peft import PeftModel
from huggingface_hub import hf_hub_download
from transformers import LlamaTokenizer, LlamaForCausalLM
import json
model_name = "shahules786/open-llama-7B-orcastyle"
config = hf_hub_download(repo_id=model_name, filename="adapter_config.json", local_dir=".")
config = json.load(open("adapter_config.json"))
base_model = config["base_model_name_or_path"]
tokenizer = LlamaTokenizer.from_pretrained(model_name)
model = LlamaForCausalLM.from_pretrained(base_model)
model.resize_token_embeddings(len(self.tokenizer))
model = PeftModel.from_pretrained(model, model_name).eval()
tokenizer.padding_side = "left"
inputs = tokenizer("This is a sample run", return_tensors="pt")
model.generate(**inputs)
```
Checkout training and inference code [here](https://github.com/explodinggradients/Funtuner/tree/main/funtuner)