nmitchko commited on
Commit
bcb8982
1 Parent(s): 302a60d

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -0
README.md CHANGED
@@ -25,6 +25,13 @@ It is based on the Guanaco LORA of LLaMA weighing in at 65B parameters.
25
  The primary goal of this model is to improve question-answering and medical dialogue tasks.
26
  It was trained using [LoRA](https://arxiv.org/abs/2106.09685) and quantized, to reduce memory footprint.
27
 
 
 
 
 
 
 
 
28
 
29
  > The following README is taken from the source page [medalpaca](https://huggingface.co/medalpaca/medalpaca-lora-13b-8bit)
30
 
 
25
  The primary goal of this model is to improve question-answering and medical dialogue tasks.
26
  It was trained using [LoRA](https://arxiv.org/abs/2106.09685) and quantized, to reduce memory footprint.
27
 
28
+ Steps to load this model:
29
+ 1. Load Guanaco-65-GPTQ https://huggingface.co/TheBloke/guanaco-65B-GPTQ
30
+ * I recommend using text-generation-ui to test it out: https://github.com/oobabooga/text-generation-webui/tree/main
31
+ 2. Download this LoRA and apply it to the model
32
+ 3. ensure `--monkey-path` is enabled in the text-generation-ui, 4-bit instructions [here](https://github.com/oobabooga/text-generation-webui/blob/main/docs/GPTQ-models-(4-bit-mode).md)
33
+
34
+ ---
35
 
36
  > The following README is taken from the source page [medalpaca](https://huggingface.co/medalpaca/medalpaca-lora-13b-8bit)
37