TheBloke commited on
Commit
2b4a214
1 Parent(s): 54909dd

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +37 -35
README.md CHANGED
@@ -29,24 +29,54 @@ It is the result of quantising to 4bit using [AutoGPTQ](https://github.com/PanQi
29
  * [3bit GPTQ model for GPU inference](https://huggingface.co/TheBloke/WizardLM-Uncensored-Falcon-40B-3bit-GPTQ).
30
  * [Eric's float16 HF format model for GPU inference and further conversions](https://huggingface.co/ehartford/WizardLM-Uncensored-Falcon-40b).
31
 
 
 
 
 
 
 
 
 
 
32
  ## EXPERIMENTAL
33
 
34
  Please note this is an experimental GPTQ model. Support for it is currently quite limited.
35
 
36
  It is also expected to be **VERY SLOW**. This is unavoidable at the moment, but is being looked at.
37
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
38
  To use it you will require:
39
 
40
- 1. Python 3.10.11
41
- 2. AutoGPTQ v0.2.1 (see below)
42
- 3. Pytorch Stable with CUDA 11.8 (`pip install torch --index-url https://download.pytorch.org/whl/cu118`)
43
  4. einops (`pip install einops`)
44
 
45
- You can then use it immediately from Python code - see example code below - or from text-generation-webui.
46
-
47
  ## AutoGPTQ
48
 
49
- You should install AutoGPTQ of version v0.2.1, thus you can try compiling manually from source:
 
 
50
 
51
  ```
52
  git clone https://github.com/PanQiWei/AutoGPTQ
@@ -85,25 +115,6 @@ output = model.generate(input_ids=tokens, max_new_tokens=100, do_sample=True, te
85
  print(tokenizer.decode(output[0]))
86
  ```
87
 
88
- ## text-generation-webui
89
-
90
- There is also provisional AutoGPTQ support in text-generation-webui.
91
-
92
- This requires a text-generation-webui version of commit `204731952ae59d79ea3805a425c73dd171d943c3` or newer.
93
-
94
- So please first update text-generation-webui to the latest version.
95
-
96
- ### How to download and use this model in text-generation-webui
97
-
98
- 1. Launch text-generation-webui with the following command-line arguments: `--autogptq --trust-remote-code`
99
- 2. Click the **Model tab**.
100
- 3. Under **Download custom model or LoRA**, enter `TheBloke/WizardLM-Uncensored-Falcon-40B-GPTQ`.
101
- 4. Click **Download**.
102
- 5. Wait until it says it's finished downloading.
103
- 6. Click the **Refresh** icon next to **Model** in the top left.
104
- 7. In the **Model drop-down**: choose the model you just downloaded, `WizardLM-Uncensored-Falcon-40B-GPTQ`.
105
- 8. Once it says it's loaded, click the **Text Generation tab** and enter a prompt!
106
-
107
  ## Provided files
108
 
109
  **gptq_model-4bit--1g.safetensors**
@@ -122,22 +133,13 @@ It was created without group_size to reduce VRAM usage, and with `desc_act` (act
122
 
123
  ## FAQ
124
 
125
- ### Prompt template
126
-
127
- Prompt format is WizardLM.
128
-
129
- ```
130
- What is a falcon? Can I keep one as a pet?
131
- ### Response:
132
- ```
133
-
134
  ### About `trust-remote-code`
135
 
136
  Please be aware that this command line argument causes Python code provided by Falcon to be executed on your machine.
137
 
138
  This code is required at the moment because Falcon is too new to be supported by Hugging Face transformers. At some point in the future transformers will support the model natively, and then `trust_remote_code` will no longer be needed.
139
 
140
- In this repo you can see two `.py` files - these are the files that get executed. They are copied from the base repo at [Falcon-7B-Instruct](https://huggingface.co/tiiuae/falcon-7b-instruct).
141
 
142
  <!-- footer start -->
143
  ## Discord
 
29
  * [3bit GPTQ model for GPU inference](https://huggingface.co/TheBloke/WizardLM-Uncensored-Falcon-40B-3bit-GPTQ).
30
  * [Eric's float16 HF format model for GPU inference and further conversions](https://huggingface.co/ehartford/WizardLM-Uncensored-Falcon-40b).
31
 
32
+ ### Prompt template
33
+
34
+ Prompt format is WizardLM.
35
+
36
+ ```
37
+ What is a falcon? Can I keep one as a pet?
38
+ ### Response:
39
+ ```
40
+
41
  ## EXPERIMENTAL
42
 
43
  Please note this is an experimental GPTQ model. Support for it is currently quite limited.
44
 
45
  It is also expected to be **VERY SLOW**. This is unavoidable at the moment, but is being looked at.
46
 
47
+ ## text-generation-webui
48
+
49
+ There is also provisional AutoGPTQ support in text-generation-webui.
50
+
51
+ This requires a text-generation-webui version of commit `204731952ae59d79ea3805a425c73dd171d943c3` or newer.
52
+
53
+ So please first update text-generation-webui to the latest version.
54
+
55
+ ### How to download and use this model in text-generation-webui
56
+
57
+ 1. Launch text-generation-webui
58
+ 2. Click the **Model tab**.
59
+ 3. Under **Download custom model or LoRA**, enter `TheBloke/WizardLM-Uncensored-Falcon-40B-GPTQ`.
60
+ 4. Click **Download**.
61
+ 5. Wait until it says it's finished downloading.
62
+ 6. Tick **Trust Remote Code**
63
+ 7. Click the **Refresh** icon next to **Model** in the top left.
64
+ 8. In the **Model drop-down**: choose the model you just downloaded, `WizardLM-Uncensored-Falcon-40B-GPTQ`.
65
+ 9. Once it says it's loaded, click the **Text Generation tab** and enter a prompt!
66
+
67
+ ## Python inference
68
+
69
  To use it you will require:
70
 
71
+ 1. AutoGPTQ v0.2.1 (see below)
72
+ 2. pytorch 2.0.0 with CUDA 11.7 or 11.8 (eg `pip install torch --index-url https://download.pytorch.org/whl/cu118`)
 
73
  4. einops (`pip install einops`)
74
 
 
 
75
  ## AutoGPTQ
76
 
77
+ You should install AutoGPTQ of version v0.2.1. There are currently problems with automatic installation with `pip install auto-gptq`.
78
+
79
+ Therefore it is recommended to compile manually from source:
80
 
81
  ```
82
  git clone https://github.com/PanQiWei/AutoGPTQ
 
115
  print(tokenizer.decode(output[0]))
116
  ```
117
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
118
  ## Provided files
119
 
120
  **gptq_model-4bit--1g.safetensors**
 
133
 
134
  ## FAQ
135
 
 
 
 
 
 
 
 
 
 
136
  ### About `trust-remote-code`
137
 
138
  Please be aware that this command line argument causes Python code provided by Falcon to be executed on your machine.
139
 
140
  This code is required at the moment because Falcon is too new to be supported by Hugging Face transformers. At some point in the future transformers will support the model natively, and then `trust_remote_code` will no longer be needed.
141
 
142
+ In this repo you can see two `.py` files - these are the files that get executed. They are copied from the base repo at [Falcon-40B-Instruct](https://huggingface.co/tiiuae/falcon-40b-instruct).
143
 
144
  <!-- footer start -->
145
  ## Discord