shahules786 commited on
Commit
64bbc61
1 Parent(s): 91dbf0a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +13 -1
README.md CHANGED
@@ -1,10 +1,18 @@
 
 
 
 
 
 
 
 
1
 
2
  from peft import PeftModel
3
  from huggingface_hub import hf_hub_download
4
  from transformers import LlamaTokenizer, LlamaForCausalLM
5
  import json
6
 
7
- model_name = "shahules786/open-llama-7B-orcastyle"
8
  config = hf_hub_download(repo_id=model_name, filename="adapter_config.json", local_dir=".")
9
  config = json.load(open("adapter_config.json"))
10
  base_model = config["base_model_name_or_path"]
@@ -16,3 +24,7 @@ tokenizer.padding_side = "left"
16
 
17
  inputs = tokenizer("This is a sample run", return_tensors="pt")
18
  model.generate(**inputs)
 
 
 
 
 
1
+ ## Training details
2
+ - Dataset used: Explanation style datasets from psmathur/WizardLM_Orca and Dahoas/cot_gsm8k
3
+ - Techniques: fp16 bit precision training + LoRA + DeepSpeed
4
+ - Machine: V100 (16GB) * 2
5
+
6
+ ## Inference
7
+
8
+ ```python
9
 
10
  from peft import PeftModel
11
  from huggingface_hub import hf_hub_download
12
  from transformers import LlamaTokenizer, LlamaForCausalLM
13
  import json
14
 
15
+ model_name = "shahules786/open-llama-3B-orcastyle"
16
  config = hf_hub_download(repo_id=model_name, filename="adapter_config.json", local_dir=".")
17
  config = json.load(open("adapter_config.json"))
18
  base_model = config["base_model_name_or_path"]
 
24
 
25
  inputs = tokenizer("This is a sample run", return_tensors="pt")
26
  model.generate(**inputs)
27
+ ```
28
+
29
+ Checkout training and inference code [here](https://github.com/explodinggradients/Funtuner/tree/main/funtuner)
30
+