Spaces:
Runtime error
Runtime error
Commit
Β·
a6c6ee5
1
Parent(s):
4f493f5
updates
Browse files- models/PygmalionAI_pygmalion-350m/huggingface-metadata.txt +1 -1
- models/TheBloke_Pygmalion-13B-SuperHOT-8K-GPTQ/README.md +0 -325
- models/TheBloke_Pygmalion-13B-SuperHOT-8K-GPTQ/config.json +0 -44
- models/TheBloke_Pygmalion-13B-SuperHOT-8K-GPTQ/llama_rope_scaled_monkey_patch.py +0 -65
- models/TheBloke_Pygmalion-13B-SuperHOT-8K-GPTQ/modelling_llama.py +0 -894
- models/TheBloke_Pygmalion-13B-SuperHOT-8K-GPTQ/special_tokens_map.json +0 -23
- models/TheBloke_Pygmalion-2-13B-SuperCOT2-GPTQ/LICENSE.txt +126 -0
- models/TheBloke_Pygmalion-2-13B-SuperCOT2-GPTQ/README.md +308 -0
- models/TheBloke_Pygmalion-2-13B-SuperCOT2-GPTQ/USE_POLICY.md +50 -0
- models/TheBloke_Pygmalion-2-13B-SuperCOT2-GPTQ/config.json +39 -0
- models/{TheBloke_Pygmalion-13B-SuperHOT-8K-GPTQ β TheBloke_Pygmalion-2-13B-SuperCOT2-GPTQ}/generation_config.json +1 -2
- models/{TheBloke_Pygmalion-13B-SuperHOT-8K-GPTQ β TheBloke_Pygmalion-2-13B-SuperCOT2-GPTQ}/huggingface-metadata.txt +3 -3
- models/{TheBloke_Pygmalion-13B-SuperHOT-8K-GPTQ β TheBloke_Pygmalion-2-13B-SuperCOT2-GPTQ}/model.safetensors +2 -2
- models/{TheBloke_Pygmalion-13B-SuperHOT-8K-GPTQ β TheBloke_Pygmalion-2-13B-SuperCOT2-GPTQ}/quantize_config.json +3 -2
- models/TheBloke_Pygmalion-2-13B-SuperCOT2-GPTQ/special_tokens_map.json +6 -0
- models/{TheBloke_Pygmalion-13B-SuperHOT-8K-GPTQ β TheBloke_Pygmalion-2-13B-SuperCOT2-GPTQ}/tokenizer.json +106 -100
- models/{TheBloke_Pygmalion-13B-SuperHOT-8K-GPTQ β TheBloke_Pygmalion-2-13B-SuperCOT2-GPTQ}/tokenizer.model +0 -0
- models/{TheBloke_Pygmalion-13B-SuperHOT-8K-GPTQ β TheBloke_Pygmalion-2-13B-SuperCOT2-GPTQ}/tokenizer_config.json +11 -7
- run.py +2 -2
models/PygmalionAI_pygmalion-350m/huggingface-metadata.txt
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
url: https://huggingface.co/PygmalionAI/pygmalion-350m
|
2 |
branch: main
|
3 |
-
download date: 2023-09-
|
4 |
sha256sum:
|
5 |
356aa4ab61193d13e3e7a097bb5f2c025dc2536d5f127154889202ba3c735ae2 pytorch_model.bin
|
|
|
1 |
url: https://huggingface.co/PygmalionAI/pygmalion-350m
|
2 |
branch: main
|
3 |
+
download date: 2023-09-20 03:12:34
|
4 |
sha256sum:
|
5 |
356aa4ab61193d13e3e7a097bb5f2c025dc2536d5f127154889202ba3c735ae2 pytorch_model.bin
|
models/TheBloke_Pygmalion-13B-SuperHOT-8K-GPTQ/README.md
DELETED
@@ -1,325 +0,0 @@
|
|
1 |
-
---
|
2 |
-
inference: false
|
3 |
-
license: other
|
4 |
-
---
|
5 |
-
|
6 |
-
<!-- header start -->
|
7 |
-
<!-- 200823 -->
|
8 |
-
<div style="width: auto; margin-left: auto; margin-right: auto">
|
9 |
-
<img src="https://i.imgur.com/EBdldam.jpg" alt="TheBlokeAI" style="width: 100%; min-width: 400px; display: block; margin: auto;">
|
10 |
-
</div>
|
11 |
-
<div style="display: flex; justify-content: space-between; width: 100%;">
|
12 |
-
<div style="display: flex; flex-direction: column; align-items: flex-start;">
|
13 |
-
<p style="margin-top: 0.5em; margin-bottom: 0em;"><a href="https://discord.gg/theblokeai">Chat & support: TheBloke's Discord server</a></p>
|
14 |
-
</div>
|
15 |
-
<div style="display: flex; flex-direction: column; align-items: flex-end;">
|
16 |
-
<p style="margin-top: 0.5em; margin-bottom: 0em;"><a href="https://www.patreon.com/TheBlokeAI">Want to contribute? TheBloke's Patreon page</a></p>
|
17 |
-
</div>
|
18 |
-
</div>
|
19 |
-
<div style="text-align:center; margin-top: 0em; margin-bottom: 0em"><p style="margin-top: 0.25em; margin-bottom: 0em;">TheBloke's LLM work is generously supported by a grant from <a href="https://a16z.com">andreessen horowitz (a16z)</a></p></div>
|
20 |
-
<hr style="margin-top: 1.0em; margin-bottom: 1.0em;">
|
21 |
-
<!-- header end -->
|
22 |
-
|
23 |
-
# TehVenom's merge of PygmalionAI's Pygmalion 13B GPTQ
|
24 |
-
|
25 |
-
These files are GPTQ 4bit model files for [TehVenom's merge of PygmalionAI's Pygmalion 13B](https://huggingface.co/TehVenom/Pygmalion-13b-Merged) merged with [Kaio Ken's SuperHOT 8K](https://huggingface.co/kaiokendev/superhot-13b-8k-no-rlhf-test).
|
26 |
-
|
27 |
-
It is the result of quantising to 4bit using [GPTQ-for-LLaMa](https://github.com/qwopqwop200/GPTQ-for-LLaMa).
|
28 |
-
|
29 |
-
**This is an experimental new GPTQ which offers up to 8K context size**
|
30 |
-
|
31 |
-
The increased context is tested to work with [ExLlama](https://github.com/turboderp/exllama), via the latest release of [text-generation-webui](https://github.com/oobabooga/text-generation-webui).
|
32 |
-
|
33 |
-
It has also been tested from Python code using AutoGPTQ, and `trust_remote_code=True`.
|
34 |
-
|
35 |
-
Code credits:
|
36 |
-
- Original concept and code for increasing context length: [kaiokendev](https://huggingface.co/kaiokendev)
|
37 |
-
- Updated Llama modelling code that includes this automatically via trust_remote_code: [emozilla](https://huggingface.co/emozilla).
|
38 |
-
|
39 |
-
Please read carefully below to see how to use it.
|
40 |
-
|
41 |
-
GGML versions are not yet provided, as there is not yet support for SuperHOT in llama.cpp. This is being investigated and will hopefully come soon.
|
42 |
-
|
43 |
-
## Repositories available
|
44 |
-
|
45 |
-
* [4-bit GPTQ models for GPU inference](https://huggingface.co/TheBloke/Pygmalion-13B-SuperHOT-8K-GPTQ)
|
46 |
-
* [2, 3, 4, 5, 6 and 8-bit GGML models for CPU inference](https://huggingface.co/TheBloke/Pygmalion-13B-SuperHOT-8K-GGML)
|
47 |
-
* [Unquantised SuperHOT fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/TheBloke/Pygmalion-13B-SuperHOT-8K-fp16)
|
48 |
-
* [Unquantised base fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/PygmalionAI/pygmalion-13b)
|
49 |
-
|
50 |
-
## How to easily download and use this model in text-generation-webui with ExLlama
|
51 |
-
|
52 |
-
Please make sure you're using the latest version of text-generation-webui
|
53 |
-
|
54 |
-
1. Click the **Model tab**.
|
55 |
-
2. Under **Download custom model or LoRA**, enter `TheBloke/Pygmalion-13B-SuperHOT-8K-GPTQ`.
|
56 |
-
3. Click **Download**.
|
57 |
-
4. The model will start downloading. Once it's finished it will say "Done"
|
58 |
-
5. Untick **Autoload the model**
|
59 |
-
6. In the top left, click the refresh icon next to **Model**.
|
60 |
-
7. In the **Model** dropdown, choose the model you just downloaded: `Pygmalion-13B-SuperHOT-8K-GPTQ`
|
61 |
-
8. To use the increased context, set the **Loader** to **ExLlama**, set **max_seq_len** to 8192 or 4096, and set **compress_pos_emb** to **4** for 8192 context, or to **2** for 4096 context.
|
62 |
-
9. Now click **Save Settings** followed by **Reload**
|
63 |
-
10. The model will automatically load, and is now ready for use!
|
64 |
-
11. Once you're ready, click the **Text Generation tab** and enter a prompt to get started!
|
65 |
-
|
66 |
-
## How to use this GPTQ model from Python code with AutoGPTQ
|
67 |
-
|
68 |
-
First make sure you have AutoGPTQ and Einops installed:
|
69 |
-
|
70 |
-
```
|
71 |
-
pip3 install einops auto-gptq
|
72 |
-
```
|
73 |
-
|
74 |
-
Then run the following code. Note that in order to get this to work, `config.json` has been hardcoded to a sequence length of 8192.
|
75 |
-
|
76 |
-
If you want to try 4096 instead to reduce VRAM usage, please manually edit `config.json` to set `max_position_embeddings` to the value you want.
|
77 |
-
|
78 |
-
```python
|
79 |
-
from transformers import AutoTokenizer, pipeline, logging
|
80 |
-
from auto_gptq import AutoGPTQForCausalLM, BaseQuantizeConfig
|
81 |
-
import argparse
|
82 |
-
|
83 |
-
model_name_or_path = "TheBloke/Pygmalion-13B-SuperHOT-8K-GPTQ"
|
84 |
-
model_basename = "pygmalion-13b-superhot-8k-GPTQ-4bit-128g.no-act.order"
|
85 |
-
|
86 |
-
use_triton = False
|
87 |
-
|
88 |
-
tokenizer = AutoTokenizer.from_pretrained(model_name_or_path, use_fast=True)
|
89 |
-
|
90 |
-
model = AutoGPTQForCausalLM.from_quantized(model_name_or_path,
|
91 |
-
model_basename=model_basename,
|
92 |
-
use_safetensors=True,
|
93 |
-
trust_remote_code=True,
|
94 |
-
device_map='auto',
|
95 |
-
use_triton=use_triton,
|
96 |
-
quantize_config=None)
|
97 |
-
|
98 |
-
model.seqlen = 8192
|
99 |
-
|
100 |
-
# Note: check the prompt template is correct for this model.
|
101 |
-
prompt = "Tell me about AI"
|
102 |
-
prompt_template=f'''USER: {prompt}
|
103 |
-
ASSISTANT:'''
|
104 |
-
|
105 |
-
print("\n\n*** Generate:")
|
106 |
-
|
107 |
-
input_ids = tokenizer(prompt_template, return_tensors='pt').input_ids.cuda()
|
108 |
-
output = model.generate(inputs=input_ids, temperature=0.7, max_new_tokens=512)
|
109 |
-
print(tokenizer.decode(output[0]))
|
110 |
-
|
111 |
-
# Inference can also be done using transformers' pipeline
|
112 |
-
|
113 |
-
# Prevent printing spurious transformers error when using pipeline with AutoGPTQ
|
114 |
-
logging.set_verbosity(logging.CRITICAL)
|
115 |
-
|
116 |
-
print("*** Pipeline:")
|
117 |
-
pipe = pipeline(
|
118 |
-
"text-generation",
|
119 |
-
model=model,
|
120 |
-
tokenizer=tokenizer,
|
121 |
-
max_new_tokens=512,
|
122 |
-
temperature=0.7,
|
123 |
-
top_p=0.95,
|
124 |
-
repetition_penalty=1.15
|
125 |
-
)
|
126 |
-
|
127 |
-
print(pipe(prompt_template)[0]['generated_text'])
|
128 |
-
```
|
129 |
-
|
130 |
-
## Using other UIs: monkey patch
|
131 |
-
|
132 |
-
Provided in the repo is `llama_rope_scaled_monkey_patch.py`, written by @kaiokendev.
|
133 |
-
|
134 |
-
It can be theoretically be added to any Python UI or custom code to enable the same result as `trust_remote_code=True`. I have not tested this, and it should be superseded by using `trust_remote_code=True`, but I include it for completeness and for interest.
|
135 |
-
|
136 |
-
## Provided files
|
137 |
-
|
138 |
-
**pygmalion-13b-superhot-8k-GPTQ-4bit-128g.no-act.order.safetensors**
|
139 |
-
|
140 |
-
This will work with AutoGPTQ, ExLlama, and CUDA versions of GPTQ-for-LLaMa. There are reports of issues with Triton mode of recent GPTQ-for-LLaMa. If you have issues, please use AutoGPTQ instead.
|
141 |
-
|
142 |
-
It was created with group_size 128 to increase inference accuracy, but without --act-order (desc_act) to increase compatibility and improve inference speed.
|
143 |
-
|
144 |
-
* `pygmalion-13b-superhot-8k-GPTQ-4bit-128g.no-act.order.safetensors`
|
145 |
-
* Works for use with ExLlama with increased context (4096 or 8192)
|
146 |
-
* Works with AutoGPTQ in Python code, including with increased context, if `trust_remote_code=True` is set.
|
147 |
-
* Should work with GPTQ-for-LLaMa in CUDA mode, but unknown if increased context works - TBC. May have issues with GPTQ-for-LLaMa Triton mode.
|
148 |
-
* Works with text-generation-webui, including one-click-installers.
|
149 |
-
* Parameters: Groupsize = 128. Act Order / desc_act = False.
|
150 |
-
|
151 |
-
<!-- footer start -->
|
152 |
-
<!-- 200823 -->
|
153 |
-
## Discord
|
154 |
-
|
155 |
-
For further support, and discussions on these models and AI in general, join us at:
|
156 |
-
|
157 |
-
[TheBloke AI's Discord server](https://discord.gg/theblokeai)
|
158 |
-
|
159 |
-
## Thanks, and how to contribute.
|
160 |
-
|
161 |
-
Thanks to the [chirper.ai](https://chirper.ai) team!
|
162 |
-
|
163 |
-
I've had a lot of people ask if they can contribute. I enjoy providing models and helping people, and would love to be able to spend even more time doing it, as well as expanding into new projects like fine tuning/training.
|
164 |
-
|
165 |
-
If you're able and willing to contribute it will be most gratefully received and will help me to keep providing more models, and to start work on new AI projects.
|
166 |
-
|
167 |
-
Donaters will get priority support on any and all AI/LLM/model questions and requests, access to a private Discord room, plus other benefits.
|
168 |
-
|
169 |
-
* Patreon: https://patreon.com/TheBlokeAI
|
170 |
-
* Ko-Fi: https://ko-fi.com/TheBlokeAI
|
171 |
-
|
172 |
-
**Special thanks to**: Aemon Algiz.
|
173 |
-
|
174 |
-
**Patreon special mentions**: Sam, theTransient, Jonathan Leane, Steven Wood, webtim, Johann-Peter Hartmann, Geoffrey Montalvo, Gabriel Tamborski, Willem Michiel, John Villwock, Derek Yates, Mesiah Bishop, Eugene Pentland, Pieter, Chadd, Stephen Murray, Daniel P. Andersen, terasurfer, Brandon Frisco, Thomas Belote, Sid, Nathan LeClaire, Magnesian, Alps Aficionado, Stanislav Ovsiannikov, Alex, Joseph William Delisle, Nikolai Manek, Michael Davis, Junyu Yang, K, J, Spencer Kim, Stefan Sabev, Olusegun Samson, transmissions 11, Michael Levine, Cory Kujawski, Rainer Wilmers, zynix, Kalila, Luke @flexchar, Ajan Kanaga, Mandus, vamX, Ai Maven, Mano Prime, Matthew Berman, subjectnull, Vitor Caleffi, Clay Pascal, biorpg, alfie_i, ιΏζ, Jeffrey Morgan, ya boyyy, Raymond Fosdick, knownsqashed, Olakabola, Leonard Tan, ReadyPlayerEmma, Enrico Ros, Dave, Talal Aujan, Illia Dulskyi, Sean Connelly, senxiiz, Artur Olbinski, Elle, Raven Klaugh, Fen Risland, Deep Realms, Imad Khwaja, Fred von Graf, Will Dee, usrbinkat, SuperWojo, Alexandros Triantafyllidis, Swaroop Kallakuri, Dan Guido, John Detwiler, Pedro Madruga, Iucharbius, Viktor Bowallius, Asp the Wyvern, Edmond Seymore, Trenton Dambrowitz, Space Cruiser, Spiking Neurons AB, Pyrater, LangChain4j, Tony Hughes, Kacper WikieΕ, Rishabh Srivastava, David Ziegler, Luke Pendergrass, Andrey, Gabriel Puliatti, Lone Striker, Sebastain Graf, Pierre Kircher, Randy H, NimbleBox.ai, Vadim, danny, Deo Leter
|
175 |
-
|
176 |
-
|
177 |
-
Thank you to all my generous patrons and donaters!
|
178 |
-
|
179 |
-
And thank you again to a16z for their generous grant.
|
180 |
-
|
181 |
-
<!-- footer end -->
|
182 |
-
|
183 |
-
# Original model card: Kaio Ken's SuperHOT 8K
|
184 |
-
|
185 |
-
### SuperHOT Prototype 2 w/ 8K Context
|
186 |
-
|
187 |
-
This is a second prototype of SuperHOT, this time 30B with 8K context and no RLHF, using the same technique described in [the github blog](https://kaiokendev.github.io/til#extending-context-to-8k).
|
188 |
-
Tests have shown that the model does indeed leverage the extended context at 8K.
|
189 |
-
|
190 |
-
You will need to **use either the monkeypatch** or, if you are already using the monkeypatch, **change the scaling factor to 0.25 and the maximum sequence length to 8192**
|
191 |
-
|
192 |
-
#### Looking for Merged & Quantized Models?
|
193 |
-
- 30B 4-bit CUDA: [tmpupload/superhot-30b-8k-4bit-safetensors](https://huggingface.co/tmpupload/superhot-30b-8k-4bit-safetensors)
|
194 |
-
- 30B 4-bit CUDA 128g: [tmpupload/superhot-30b-8k-4bit-128g-safetensors](https://huggingface.co/tmpupload/superhot-30b-8k-4bit-128g-safetensors)
|
195 |
-
|
196 |
-
|
197 |
-
#### Training Details
|
198 |
-
I trained the LoRA with the following configuration:
|
199 |
-
- 1200 samples (~400 samples over 2048 sequence length)
|
200 |
-
- learning rate of 3e-4
|
201 |
-
- 3 epochs
|
202 |
-
- The exported modules are:
|
203 |
-
- q_proj
|
204 |
-
- k_proj
|
205 |
-
- v_proj
|
206 |
-
- o_proj
|
207 |
-
- no bias
|
208 |
-
- Rank = 4
|
209 |
-
- Alpha = 8
|
210 |
-
- no dropout
|
211 |
-
- weight decay of 0.1
|
212 |
-
- AdamW beta1 of 0.9 and beta2 0.99, epsilon of 1e-5
|
213 |
-
- Trained on 4-bit base model
|
214 |
-
|
215 |
-
# Original model card: TehVenom's merge of PygmalionAI's Pygmalion 13B
|
216 |
-
|
217 |
-
<h1 style="text-align: center">Pygmalion 13b</h1>
|
218 |
-
<h2 style="text-align: center">A conversational LLaMA fine-tune.</h2>
|
219 |
-
|
220 |
-
## Model Details:
|
221 |
-
|
222 |
-
Pygmalion 13b is a dialogue model based on Meta's LLaMA-13b.
|
223 |
-
|
224 |
-
This is version 1. It has been fine-tuned using a subset of the data from Pygmalion-6B-v8-pt4,
|
225 |
-
for those of you familiar with the project.
|
226 |
-
|
227 |
-
The current Pygmalion-13b has been trained as a LoRA, then merged down to the base model for distribuition.
|
228 |
-
|
229 |
-
## Applying the XORs
|
230 |
-
|
231 |
-
This models has the XOR files pre-applied out of the box.
|
232 |
-
Converted from the XORs weights from PygmalionAI's release https://huggingface.co/PygmalionAI/pygmalion-13b
|
233 |
-
|
234 |
-
## Prompting
|
235 |
-
|
236 |
-
The model was trained on the usual Pygmalion persona + chat format, so any of the usual UIs should already handle everything correctly. If you're using the model directly, this is the expected formatting:
|
237 |
-
|
238 |
-
```
|
239 |
-
[CHARACTER]'s Persona: [A few sentences about the character you want the model to play]
|
240 |
-
<START>
|
241 |
-
[DIALOGUE HISTORY]
|
242 |
-
You: [User's input message here]
|
243 |
-
[CHARACTER]:
|
244 |
-
```
|
245 |
-
|
246 |
-
Where `[CHARACTER]` is, as you can probably guess, the name of the character you want the model to portray, `<START>` should be used verbatim as a delimiter token to separate persona and scenario data from the dialogue, and `[DIALOGUE HISTORY]` is a sliding window of chat history so the model can have conversational context to draw from. Here's a concrete example:
|
247 |
-
|
248 |
-
```
|
249 |
-
Assistant's Persona: Assistant is a highly intelligent language model trained to comply with user requests.
|
250 |
-
<START>
|
251 |
-
Assistant: Hello! How may I help you today?
|
252 |
-
You: What is Zork?
|
253 |
-
Assistant:
|
254 |
-
```
|
255 |
-
|
256 |
-
Which will generate something like:
|
257 |
-
|
258 |
-
```
|
259 |
-
Zork is an interactive fiction computer game created in the 1970s by Infocom, Inc., which was later acquired by Activision Blizzard. It is widely considered one of the most influential games ever made and has been credited with popularizing text-based adventure games. The original version of Zork was written in the programming language MACRO-10, but it was ported to many other platforms over the years."
|
260 |
-
```
|
261 |
-
|
262 |
-
The model will automatically emit an end-of-text token (`</s>`) when it judges that the response is complete.
|
263 |
-
|
264 |
-
## Eval / Benchmark scores
|
265 |
-
|
266 |
-
|
267 |
-
Current evals out of the Pygmalion-13b model: <br>
|
268 |
-
<html>
|
269 |
-
<head>
|
270 |
-
<style>
|
271 |
-
table {
|
272 |
-
border:1px solid #b3adad;
|
273 |
-
border-collapse:collapse;
|
274 |
-
padding:5px;
|
275 |
-
}
|
276 |
-
table th {
|
277 |
-
border:1px solid #b3adad;
|
278 |
-
padding:5px;
|
279 |
-
background: #f0f0f0;
|
280 |
-
color: #313030;
|
281 |
-
}
|
282 |
-
table td {
|
283 |
-
border:1px solid #b3adad;
|
284 |
-
text-align:center;
|
285 |
-
padding:5px;
|
286 |
-
background: #ffffff;
|
287 |
-
color: #313030;
|
288 |
-
}
|
289 |
-
</style>
|
290 |
-
</head>
|
291 |
-
<body>
|
292 |
-
<table>
|
293 |
-
<thead>
|
294 |
-
<tr>
|
295 |
-
<th>Model:</th>
|
296 |
-
<th>Wikitext2</th>
|
297 |
-
<th>Ptb-New</th>
|
298 |
-
<th>C4-New</th>
|
299 |
-
</tr>
|
300 |
-
</thead>
|
301 |
-
<tbody>
|
302 |
-
<tr>
|
303 |
-
<td>Pygmalion 13b - 16bit</td>
|
304 |
-
<td>5.710726737976074</td>
|
305 |
-
<td>23.633684158325195</td>
|
306 |
-
<td>7.6324849128723145</td>
|
307 |
-
</tr>
|
308 |
-
</tbody>
|
309 |
-
</table>
|
310 |
-
</body>
|
311 |
-
</html>
|
312 |
-
<br>Thanks to YellowRose#1776 for the numbers.
|
313 |
-
<hr>
|
314 |
-
|
315 |
-
## Other notes
|
316 |
-
|
317 |
-
- When prompted correctly, the model will always start by generating a BOS token. This behavior is an accidental side-effect which we plan to address in future model versions and should not be relied upon.
|
318 |
-
- The model was trained as a LoRA with a somewhat unorthodox configuration which causes errors when used with the current version of `peft`, hence we release it as a full model instead.
|
319 |
-
|
320 |
-
|
321 |
-
## Limitations and biases
|
322 |
-
|
323 |
-
The intended use-case for this model is fictional conversation for entertainment purposes. Any other sort of usage is out of scope.
|
324 |
-
|
325 |
-
As such, it was **not** fine-tuned to be safe and harmless: the base model _and_ this fine-tune have been trained on data known to contain profanity and texts that are lewd or otherwise offensive. It may produce socially unacceptable or undesirable text, even if the prompt itself does not include anything explicitly offensive. Outputs might often be factually wrong or misleading.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
models/TheBloke_Pygmalion-13B-SuperHOT-8K-GPTQ/config.json
DELETED
@@ -1,44 +0,0 @@
|
|
1 |
-
{
|
2 |
-
"_name_or_path": "/workspace/superhot_process/pygmalion-13b/source",
|
3 |
-
"architectures": [
|
4 |
-
"LlamaForCausalLM"
|
5 |
-
],
|
6 |
-
"badwordsids": [
|
7 |
-
[
|
8 |
-
0
|
9 |
-
]
|
10 |
-
],
|
11 |
-
"bos_token_id": 1,
|
12 |
-
"eos_token_id": 2,
|
13 |
-
"hidden_act": "silu",
|
14 |
-
"hidden_size": 5120,
|
15 |
-
"initializer_range": 0.02,
|
16 |
-
"intermediate_size": 13824,
|
17 |
-
"max_position_embeddings": 8192,
|
18 |
-
"max_sequence_length": 2048,
|
19 |
-
"model_type": "llama",
|
20 |
-
"num_attention_heads": 40,
|
21 |
-
"num_hidden_layers": 40,
|
22 |
-
"pad_token_id": 0,
|
23 |
-
"rms_norm_eps": 1e-06,
|
24 |
-
"tie_word_embeddings": false,
|
25 |
-
"torch_dtype": "float16",
|
26 |
-
"transformers_version": "4.30.0.dev0",
|
27 |
-
"use_cache": true,
|
28 |
-
"vocab_size": 32000,
|
29 |
-
"auto_map": {
|
30 |
-
"AutoModel": "modelling_llama.LlamaModel",
|
31 |
-
"AutoModelForCausalLM": "modelling_llama.LlamaForCausalLM",
|
32 |
-
"AutoModelForSequenceClassification": "modelling_llama.LlamaForSequenceClassification"
|
33 |
-
},
|
34 |
-
"quantization_config": {
|
35 |
-
"bits": 4,
|
36 |
-
"group_size": 128,
|
37 |
-
"damp_percent": 0.01,
|
38 |
-
"desc_act": false,
|
39 |
-
"sym": true,
|
40 |
-
"true_sequential": true,
|
41 |
-
"model_file_base_name": "model",
|
42 |
-
"quant_method": "gptq"
|
43 |
-
}
|
44 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
models/TheBloke_Pygmalion-13B-SuperHOT-8K-GPTQ/llama_rope_scaled_monkey_patch.py
DELETED
@@ -1,65 +0,0 @@
|
|
1 |
-
import torch
|
2 |
-
import transformers
|
3 |
-
import transformers.models.llama.modeling_llama
|
4 |
-
from einops import rearrange
|
5 |
-
import random
|
6 |
-
|
7 |
-
# This monkey patch file is not needed if using ExLlama, or if using `trust_remote_code=True``
|
8 |
-
|
9 |
-
class ScaledRotaryEmbedding(torch.nn.Module):
|
10 |
-
def __init__(self, dim, max_position_embeddings=2048, base=10000, device=None):
|
11 |
-
super().__init__()
|
12 |
-
inv_freq = 1.0 / (base ** (torch.arange(0, dim, 2).float().to(device) / dim))
|
13 |
-
self.register_buffer("inv_freq", inv_freq)
|
14 |
-
|
15 |
-
max_position_embeddings = 8192
|
16 |
-
|
17 |
-
# Build here to make `torch.jit.trace` work.
|
18 |
-
self.max_seq_len_cached = max_position_embeddings
|
19 |
-
t = torch.arange(
|
20 |
-
self.max_seq_len_cached,
|
21 |
-
device=self.inv_freq.device,
|
22 |
-
dtype=self.inv_freq.dtype,
|
23 |
-
)
|
24 |
-
|
25 |
-
self.scale = 1 / 4
|
26 |
-
t *= self.scale
|
27 |
-
|
28 |
-
freqs = torch.einsum("i,j->ij", t, self.inv_freq)
|
29 |
-
# Different from paper, but it uses a different permutation in order to obtain the same calculation
|
30 |
-
emb = torch.cat((freqs, freqs), dim=-1)
|
31 |
-
self.register_buffer(
|
32 |
-
"cos_cached", emb.cos()[None, None, :, :], persistent=False
|
33 |
-
)
|
34 |
-
self.register_buffer(
|
35 |
-
"sin_cached", emb.sin()[None, None, :, :], persistent=False
|
36 |
-
)
|
37 |
-
|
38 |
-
def forward(self, x, seq_len=None):
|
39 |
-
# x: [bs, num_attention_heads, seq_len, head_size]
|
40 |
-
# This `if` block is unlikely to be run after we build sin/cos in `__init__`. Keep the logic here just in case.
|
41 |
-
if seq_len > self.max_seq_len_cached:
|
42 |
-
self.max_seq_len_cached = seq_len
|
43 |
-
t = torch.arange(
|
44 |
-
self.max_seq_len_cached, device=x.device, dtype=self.inv_freq.dtype
|
45 |
-
)
|
46 |
-
t *= self.scale
|
47 |
-
freqs = torch.einsum("i,j->ij", t, self.inv_freq)
|
48 |
-
# Different from paper, but it uses a different permutation in order to obtain the same calculation
|
49 |
-
emb = torch.cat((freqs, freqs), dim=-1).to(x.device)
|
50 |
-
self.register_buffer(
|
51 |
-
"cos_cached", emb.cos()[None, None, :, :], persistent=False
|
52 |
-
)
|
53 |
-
self.register_buffer(
|
54 |
-
"sin_cached", emb.sin()[None, None, :, :], persistent=False
|
55 |
-
)
|
56 |
-
return (
|
57 |
-
self.cos_cached[:, :, :seq_len, ...].to(dtype=x.dtype),
|
58 |
-
self.sin_cached[:, :, :seq_len, ...].to(dtype=x.dtype),
|
59 |
-
)
|
60 |
-
|
61 |
-
|
62 |
-
def replace_llama_rope_with_scaled_rope():
|
63 |
-
transformers.models.llama.modeling_llama.LlamaRotaryEmbedding = (
|
64 |
-
ScaledRotaryEmbedding
|
65 |
-
)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
models/TheBloke_Pygmalion-13B-SuperHOT-8K-GPTQ/modelling_llama.py
DELETED
@@ -1,894 +0,0 @@
|
|
1 |
-
# coding=utf-8
|
2 |
-
# Copyright 2022 EleutherAI and the HuggingFace Inc. team. All rights reserved.
|
3 |
-
#
|
4 |
-
# This code is based on EleutherAI's GPT-NeoX library and the GPT-NeoX
|
5 |
-
# and OPT implementations in this library. It has been modified from its
|
6 |
-
# original forms to accommodate minor architectural differences compared
|
7 |
-
# to GPT-NeoX and OPT used by the Meta AI team that trained the model.
|
8 |
-
#
|
9 |
-
# Licensed under the Apache License, Version 2.0 (the "License");
|
10 |
-
# you may not use this file except in compliance with the License.
|
11 |
-
# You may obtain a copy of the License at
|
12 |
-
#
|
13 |
-
# http://www.apache.org/licenses/LICENSE-2.0
|
14 |
-
#
|
15 |
-
# Unless required by applicable law or agreed to in writing, software
|
16 |
-
# distributed under the License is distributed on an "AS IS" BASIS,
|
17 |
-
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
18 |
-
# See the License for the specific language governing permissions and
|
19 |
-
# limitations under the License.
|
20 |
-
""" PyTorch LLaMA model."""
|
21 |
-
import math
|
22 |
-
from typing import List, Optional, Tuple, Union
|
23 |
-
|
24 |
-
import torch
|
25 |
-
import torch.utils.checkpoint
|
26 |
-
from torch import nn
|
27 |
-
from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss
|
28 |
-
|
29 |
-
from transformers.activations import ACT2FN
|
30 |
-
from transformers.modeling_outputs import BaseModelOutputWithPast, CausalLMOutputWithPast, SequenceClassifierOutputWithPast
|
31 |
-
from transformers.modeling_utils import PreTrainedModel
|
32 |
-
from transformers.utils import add_start_docstrings, add_start_docstrings_to_model_forward, logging, replace_return_docstrings
|
33 |
-
from transformers.models.llama.modeling_llama import LlamaConfig
|
34 |
-
|
35 |
-
logger = logging.get_logger(__name__)
|
36 |
-
|
37 |
-
_CONFIG_FOR_DOC = "LlamaConfig"
|
38 |
-
|
39 |
-
|
40 |
-
# Copied from transformers.models.bart.modeling_bart._make_causal_mask
|
41 |
-
def _make_causal_mask(
|
42 |
-
input_ids_shape: torch.Size, dtype: torch.dtype, device: torch.device, past_key_values_length: int = 0
|
43 |
-
):
|
44 |
-
"""
|
45 |
-
Make causal mask used for bi-directional self-attention.
|
46 |
-
"""
|
47 |
-
bsz, tgt_len = input_ids_shape
|
48 |
-
mask = torch.full((tgt_len, tgt_len), torch.tensor(torch.finfo(dtype).min, device=device), device=device)
|
49 |
-
mask_cond = torch.arange(mask.size(-1), device=device)
|
50 |
-
mask.masked_fill_(mask_cond < (mask_cond + 1).view(mask.size(-1), 1), 0)
|
51 |
-
mask = mask.to(dtype)
|
52 |
-
|
53 |
-
if past_key_values_length > 0:
|
54 |
-
mask = torch.cat([torch.zeros(tgt_len, past_key_values_length, dtype=dtype, device=device), mask], dim=-1)
|
55 |
-
return mask[None, None, :, :].expand(bsz, 1, tgt_len, tgt_len + past_key_values_length)
|
56 |
-
|
57 |
-
|
58 |
-
# Copied from transformers.models.bart.modeling_bart._expand_mask
|
59 |
-
def _expand_mask(mask: torch.Tensor, dtype: torch.dtype, tgt_len: Optional[int] = None):
|
60 |
-
"""
|
61 |
-
Expands attention_mask from `[bsz, seq_len]` to `[bsz, 1, tgt_seq_len, src_seq_len]`.
|
62 |
-
"""
|
63 |
-
bsz, src_len = mask.size()
|
64 |
-
tgt_len = tgt_len if tgt_len is not None else src_len
|
65 |
-
|
66 |
-
expanded_mask = mask[:, None, None, :].expand(bsz, 1, tgt_len, src_len).to(dtype)
|
67 |
-
|
68 |
-
inverted_mask = 1.0 - expanded_mask
|
69 |
-
|
70 |
-
return inverted_mask.masked_fill(inverted_mask.to(torch.bool), torch.finfo(dtype).min)
|
71 |
-
|
72 |
-
|
73 |
-
class LlamaRMSNorm(nn.Module):
|
74 |
-
def __init__(self, hidden_size, eps=1e-6):
|
75 |
-
"""
|
76 |
-
LlamaRMSNorm is equivalent to T5LayerNorm
|
77 |
-
"""
|
78 |
-
super().__init__()
|
79 |
-
self.weight = nn.Parameter(torch.ones(hidden_size))
|
80 |
-
self.variance_epsilon = eps
|
81 |
-
|
82 |
-
def forward(self, hidden_states):
|
83 |
-
input_dtype = hidden_states.dtype
|
84 |
-
variance = hidden_states.to(torch.float32).pow(2).mean(-1, keepdim=True)
|
85 |
-
hidden_states = hidden_states * torch.rsqrt(variance + self.variance_epsilon)
|
86 |
-
|
87 |
-
return (self.weight * hidden_states).to(input_dtype)
|
88 |
-
|
89 |
-
|
90 |
-
class LlamaRotaryEmbedding(torch.nn.Module):
|
91 |
-
def __init__(self, dim, max_position_embeddings=2048, base=10000, scale=1, device=None):
|
92 |
-
super().__init__()
|
93 |
-
inv_freq = 1.0 / (base ** (torch.arange(0, dim, 2).float().to(device) / dim))
|
94 |
-
self.register_buffer("inv_freq", inv_freq)
|
95 |
-
|
96 |
-
# Build here to make `torch.jit.trace` work.
|
97 |
-
self.max_seq_len_cached = max_position_embeddings
|
98 |
-
t = torch.arange(self.max_seq_len_cached, device=self.inv_freq.device, dtype=self.inv_freq.dtype)
|
99 |
-
|
100 |
-
self.scale = scale
|
101 |
-
t *= self.scale
|
102 |
-
|
103 |
-
freqs = torch.einsum("i,j->ij", t, self.inv_freq)
|
104 |
-
# Different from paper, but it uses a different permutation in order to obtain the same calculation
|
105 |
-
emb = torch.cat((freqs, freqs), dim=-1)
|
106 |
-
dtype = torch.get_default_dtype()
|
107 |
-
self.register_buffer("cos_cached", emb.cos()[None, None, :, :].to(dtype), persistent=False)
|
108 |
-
self.register_buffer("sin_cached", emb.sin()[None, None, :, :].to(dtype), persistent=False)
|
109 |
-
|
110 |
-
def forward(self, x, seq_len=None):
|
111 |
-
# x: [bs, num_attention_heads, seq_len, head_size]
|
112 |
-
# This `if` block is unlikely to be run after we build sin/cos in `__init__`. Keep the logic here just in case.
|
113 |
-
if seq_len > self.max_seq_len_cached:
|
114 |
-
self.max_seq_len_cached = seq_len
|
115 |
-
t = torch.arange(self.max_seq_len_cached, device=x.device, dtype=self.inv_freq.dtype)
|
116 |
-
freqs = torch.einsum("i,j->ij", t, self.inv_freq)
|
117 |
-
# Different from paper, but it uses a different permutation in order to obtain the same calculation
|
118 |
-
emb = torch.cat((freqs, freqs), dim=-1).to(x.device)
|
119 |
-
self.register_buffer("cos_cached", emb.cos()[None, None, :, :].to(x.dtype), persistent=False)
|
120 |
-
self.register_buffer("sin_cached", emb.sin()[None, None, :, :].to(x.dtype), persistent=False)
|
121 |
-
return (
|
122 |
-
self.cos_cached[:, :, :seq_len, ...].to(dtype=x.dtype),
|
123 |
-
self.sin_cached[:, :, :seq_len, ...].to(dtype=x.dtype),
|
124 |
-
)
|
125 |
-
|
126 |
-
|
127 |
-
def rotate_half(x):
|
128 |
-
"""Rotates half the hidden dims of the input."""
|
129 |
-
x1 = x[..., : x.shape[-1] // 2]
|
130 |
-
x2 = x[..., x.shape[-1] // 2 :]
|
131 |
-
return torch.cat((-x2, x1), dim=-1)
|
132 |
-
|
133 |
-
|
134 |
-
def apply_rotary_pos_emb(q, k, cos, sin, position_ids):
|
135 |
-
# The first two dimensions of cos and sin are always 1, so we can `squeeze` them.
|
136 |
-
cos = cos.squeeze(1).squeeze(0) # [seq_len, dim]
|
137 |
-
sin = sin.squeeze(1).squeeze(0) # [seq_len, dim]
|
138 |
-
cos = cos[position_ids].unsqueeze(1) # [bs, 1, seq_len, dim]
|
139 |
-
sin = sin[position_ids].unsqueeze(1) # [bs, 1, seq_len, dim]
|
140 |
-
q_embed = (q * cos) + (rotate_half(q) * sin)
|
141 |
-
k_embed = (k * cos) + (rotate_half(k) * sin)
|
142 |
-
return q_embed, k_embed
|
143 |
-
|
144 |
-
|
145 |
-
class LlamaMLP(nn.Module):
|
146 |
-
def __init__(
|
147 |
-
self,
|
148 |
-
hidden_size: int,
|
149 |
-
intermediate_size: int,
|
150 |
-
hidden_act: str,
|
151 |
-
):
|
152 |
-
super().__init__()
|
153 |
-
self.gate_proj = nn.Linear(hidden_size, intermediate_size, bias=False)
|
154 |
-
self.down_proj = nn.Linear(intermediate_size, hidden_size, bias=False)
|
155 |
-
self.up_proj = nn.Linear(hidden_size, intermediate_size, bias=False)
|
156 |
-
self.act_fn = ACT2FN[hidden_act]
|
157 |
-
|
158 |
-
def forward(self, x):
|
159 |
-
return self.down_proj(self.act_fn(self.gate_proj(x)) * self.up_proj(x))
|
160 |
-
|
161 |
-
|
162 |
-
class LlamaAttention(nn.Module):
|
163 |
-
"""Multi-headed attention from 'Attention Is All You Need' paper"""
|
164 |
-
|
165 |
-
def __init__(self, config: LlamaConfig):
|
166 |
-
super().__init__()
|
167 |
-
self.config = config
|
168 |
-
self.hidden_size = config.hidden_size
|
169 |
-
self.num_heads = config.num_attention_heads
|
170 |
-
self.head_dim = self.hidden_size // self.num_heads
|
171 |
-
self.max_position_embeddings = config.max_position_embeddings
|
172 |
-
self.position_embeddings_scale = 2048 / self.max_position_embeddings
|
173 |
-
|
174 |
-
if (self.head_dim * self.num_heads) != self.hidden_size:
|
175 |
-
raise ValueError(
|
176 |
-
f"hidden_size must be divisible by num_heads (got `hidden_size`: {self.hidden_size}"
|
177 |
-
f" and `num_heads`: {self.num_heads})."
|
178 |
-
)
|
179 |
-
self.q_proj = nn.Linear(self.hidden_size, self.num_heads * self.head_dim, bias=False)
|
180 |
-
self.k_proj = nn.Linear(self.hidden_size, self.num_heads * self.head_dim, bias=False)
|
181 |
-
self.v_proj = nn.Linear(self.hidden_size, self.num_heads * self.head_dim, bias=False)
|
182 |
-
self.o_proj = nn.Linear(self.num_heads * self.head_dim, self.hidden_size, bias=False)
|
183 |
-
self.rotary_emb = LlamaRotaryEmbedding(self.head_dim, max_position_embeddings=self.max_position_embeddings, scale=self.position_embeddings_scale)
|
184 |
-
|
185 |
-
def _shape(self, tensor: torch.Tensor, seq_len: int, bsz: int):
|
186 |
-
return tensor.view(bsz, seq_len, self.num_heads, self.head_dim).transpose(1, 2).contiguous()
|
187 |
-
|
188 |
-
def forward(
|
189 |
-
self,
|
190 |
-
hidden_states: torch.Tensor,
|
191 |
-
attention_mask: Optional[torch.Tensor] = None,
|
192 |
-
position_ids: Optional[torch.LongTensor] = None,
|
193 |
-
past_key_value: Optional[Tuple[torch.Tensor]] = None,
|
194 |
-
output_attentions: bool = False,
|
195 |
-
use_cache: bool = False,
|
196 |
-
) -> Tuple[torch.Tensor, Optional[torch.Tensor], Optional[Tuple[torch.Tensor]]]:
|
197 |
-
bsz, q_len, _ = hidden_states.size()
|
198 |
-
|
199 |
-
query_states = self.q_proj(hidden_states).view(bsz, q_len, self.num_heads, self.head_dim).transpose(1, 2)
|
200 |
-
key_states = self.k_proj(hidden_states).view(bsz, q_len, self.num_heads, self.head_dim).transpose(1, 2)
|
201 |
-
value_states = self.v_proj(hidden_states).view(bsz, q_len, self.num_heads, self.head_dim).transpose(1, 2)
|
202 |
-
|
203 |
-
kv_seq_len = key_states.shape[-2]
|
204 |
-
if past_key_value is not None:
|
205 |
-
kv_seq_len += past_key_value[0].shape[-2]
|
206 |
-
cos, sin = self.rotary_emb(value_states, seq_len=kv_seq_len)
|
207 |
-
query_states, key_states = apply_rotary_pos_emb(query_states, key_states, cos, sin, position_ids)
|
208 |
-
# [bsz, nh, t, hd]
|
209 |
-
|
210 |
-
if past_key_value is not None:
|
211 |
-
# reuse k, v, self_attention
|
212 |
-
key_states = torch.cat([past_key_value[0], key_states], dim=2)
|
213 |
-
value_states = torch.cat([past_key_value[1], value_states], dim=2)
|
214 |
-
|
215 |
-
past_key_value = (key_states, value_states) if use_cache else None
|
216 |
-
|
217 |
-
attn_weights = torch.matmul(query_states, key_states.transpose(2, 3)) / math.sqrt(self.head_dim)
|
218 |
-
|
219 |
-
if attn_weights.size() != (bsz, self.num_heads, q_len, kv_seq_len):
|
220 |
-
raise ValueError(
|
221 |
-
f"Attention weights should be of size {(bsz, self.num_heads, q_len, kv_seq_len)}, but is"
|
222 |
-
f" {attn_weights.size()}"
|
223 |
-
)
|
224 |
-
|
225 |
-
if attention_mask is not None:
|
226 |
-
if attention_mask.size() != (bsz, 1, q_len, kv_seq_len):
|
227 |
-
raise ValueError(
|
228 |
-
f"Attention mask should be of size {(bsz, 1, q_len, kv_seq_len)}, but is {attention_mask.size()}"
|
229 |
-
)
|
230 |
-
attn_weights = attn_weights + attention_mask
|
231 |
-
attn_weights = torch.max(
|
232 |
-
attn_weights, torch.tensor(torch.finfo(attn_weights.dtype).min, device=attn_weights.device)
|
233 |
-
)
|
234 |
-
|
235 |
-
# upcast attention to fp32
|
236 |
-
attn_weights = nn.functional.softmax(attn_weights, dim=-1, dtype=torch.float32).to(query_states.dtype)
|
237 |
-
attn_output = torch.matmul(attn_weights, value_states)
|
238 |
-
|
239 |
-
if attn_output.size() != (bsz, self.num_heads, q_len, self.head_dim):
|
240 |
-
raise ValueError(
|
241 |
-
f"`attn_output` should be of size {(bsz, self.num_heads, q_len, self.head_dim)}, but is"
|
242 |
-
f" {attn_output.size()}"
|
243 |
-
)
|
244 |
-
|
245 |
-
attn_output = attn_output.transpose(1, 2)
|
246 |
-
attn_output = attn_output.reshape(bsz, q_len, self.hidden_size)
|
247 |
-
|
248 |
-
attn_output = self.o_proj(attn_output)
|
249 |
-
|
250 |
-
if not output_attentions:
|
251 |
-
attn_weights = None
|
252 |
-
|
253 |
-
return attn_output, attn_weights, past_key_value
|
254 |
-
|
255 |
-
|
256 |
-
class LlamaDecoderLayer(nn.Module):
|
257 |
-
def __init__(self, config: LlamaConfig):
|
258 |
-
super().__init__()
|
259 |
-
self.hidden_size = config.hidden_size
|
260 |
-
self.self_attn = LlamaAttention(config=config)
|
261 |
-
self.mlp = LlamaMLP(
|
262 |
-
hidden_size=self.hidden_size,
|
263 |
-
intermediate_size=config.intermediate_size,
|
264 |
-
hidden_act=config.hidden_act,
|
265 |
-
)
|
266 |
-
self.input_layernorm = LlamaRMSNorm(config.hidden_size, eps=config.rms_norm_eps)
|
267 |
-
self.post_attention_layernorm = LlamaRMSNorm(config.hidden_size, eps=config.rms_norm_eps)
|
268 |
-
|
269 |
-
def forward(
|
270 |
-
self,
|
271 |
-
hidden_states: torch.Tensor,
|
272 |
-
attention_mask: Optional[torch.Tensor] = None,
|
273 |
-
position_ids: Optional[torch.LongTensor] = None,
|
274 |
-
past_key_value: Optional[Tuple[torch.Tensor]] = None,
|
275 |
-
output_attentions: Optional[bool] = False,
|
276 |
-
use_cache: Optional[bool] = False,
|
277 |
-
) -> Tuple[torch.FloatTensor, Optional[Tuple[torch.FloatTensor, torch.FloatTensor]]]:
|
278 |
-
"""
|
279 |
-
Args:
|
280 |
-
hidden_states (`torch.FloatTensor`): input to the layer of shape `(batch, seq_len, embed_dim)`
|
281 |
-
attention_mask (`torch.FloatTensor`, *optional*): attention mask of size
|
282 |
-
`(batch, 1, tgt_len, src_len)` where padding elements are indicated by very large negative values.
|
283 |
-
output_attentions (`bool`, *optional*):
|
284 |
-
Whether or not to return the attentions tensors of all attention layers. See `attentions` under
|
285 |
-
returned tensors for more detail.
|
286 |
-
use_cache (`bool`, *optional*):
|
287 |
-
If set to `True`, `past_key_values` key value states are returned and can be used to speed up decoding
|
288 |
-
(see `past_key_values`).
|
289 |
-
past_key_value (`Tuple(torch.FloatTensor)`, *optional*): cached past key and value projection states
|
290 |
-
"""
|
291 |
-
|
292 |
-
residual = hidden_states
|
293 |
-
|
294 |
-
hidden_states = self.input_layernorm(hidden_states)
|
295 |
-
|
296 |
-
# Self Attention
|
297 |
-
hidden_states, self_attn_weights, present_key_value = self.self_attn(
|
298 |
-
hidden_states=hidden_states,
|
299 |
-
attention_mask=attention_mask,
|
300 |
-
position_ids=position_ids,
|
301 |
-
past_key_value=past_key_value,
|
302 |
-
output_attentions=output_attentions,
|
303 |
-
use_cache=use_cache,
|
304 |
-
)
|
305 |
-
hidden_states = residual + hidden_states
|
306 |
-
|
307 |
-
# Fully Connected
|
308 |
-
residual = hidden_states
|
309 |
-
hidden_states = self.post_attention_layernorm(hidden_states)
|
310 |
-
hidden_states = self.mlp(hidden_states)
|
311 |
-
hidden_states = residual + hidden_states
|
312 |
-
|
313 |
-
outputs = (hidden_states,)
|
314 |
-
|
315 |
-
if output_attentions:
|
316 |
-
outputs += (self_attn_weights,)
|
317 |
-
|
318 |
-
if use_cache:
|
319 |
-
outputs += (present_key_value,)
|
320 |
-
|
321 |
-
return outputs
|
322 |
-
|
323 |
-
|
324 |
-
LLAMA_START_DOCSTRING = r"""
|
325 |
-
This model inherits from [`PreTrainedModel`]. Check the superclass documentation for the generic methods the
|
326 |
-
library implements for all its model (such as downloading or saving, resizing the input embeddings, pruning heads
|
327 |
-
etc.)
|
328 |
-
|
329 |
-
This model is also a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) subclass.
|
330 |
-
Use it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage
|
331 |
-
and behavior.
|
332 |
-
|
333 |
-
Parameters:
|
334 |
-
config ([`LlamaConfig`]):
|
335 |
-
Model configuration class with all the parameters of the model. Initializing with a config file does not
|
336 |
-
load the weights associated with the model, only the configuration. Check out the
|
337 |
-
[`~PreTrainedModel.from_pretrained`] method to load the model weights.
|
338 |
-
"""
|
339 |
-
|
340 |
-
|
341 |
-
@add_start_docstrings(
|
342 |
-
"The bare LLaMA Model outputting raw hidden-states without any specific head on top.",
|
343 |
-
LLAMA_START_DOCSTRING,
|
344 |
-
)
|
345 |
-
class LlamaPreTrainedModel(PreTrainedModel):
|
346 |
-
config_class = LlamaConfig
|
347 |
-
base_model_prefix = "model"
|
348 |
-
supports_gradient_checkpointing = True
|
349 |
-
_no_split_modules = ["LlamaDecoderLayer"]
|
350 |
-
_skip_keys_device_placement = "past_key_values"
|
351 |
-
_keys_to_ignore_on_load_unexpected = [r"decoder\.version"]
|
352 |
-
|
353 |
-
def _init_weights(self, module):
|
354 |
-
std = self.config.initializer_range
|
355 |
-
if isinstance(module, nn.Linear):
|
356 |
-
module.weight.data.normal_(mean=0.0, std=std)
|
357 |
-
if module.bias is not None:
|
358 |
-
module.bias.data.zero_()
|
359 |
-
elif isinstance(module, nn.Embedding):
|
360 |
-
module.weight.data.normal_(mean=0.0, std=std)
|
361 |
-
if module.padding_idx is not None:
|
362 |
-
module.weight.data[module.padding_idx].zero_()
|
363 |
-
|
364 |
-
def _set_gradient_checkpointing(self, module, value=False):
|
365 |
-
if isinstance(module, LlamaModel):
|
366 |
-
module.gradient_checkpointing = value
|
367 |
-
|
368 |
-
|
369 |
-
LLAMA_INPUTS_DOCSTRING = r"""
|
370 |
-
Args:
|
371 |
-
input_ids (`torch.LongTensor` of shape `(batch_size, sequence_length)`):
|
372 |
-
Indices of input sequence tokens in the vocabulary. Padding will be ignored by default should you provide
|
373 |
-
it.
|
374 |
-
|
375 |
-
Indices can be obtained using [`AutoTokenizer`]. See [`PreTrainedTokenizer.encode`] and
|
376 |
-
[`PreTrainedTokenizer.__call__`] for details.
|
377 |
-
|
378 |
-
[What are input IDs?](../glossary#input-ids)
|
379 |
-
attention_mask (`torch.Tensor` of shape `(batch_size, sequence_length)`, *optional*):
|
380 |
-
Mask to avoid performing attention on padding token indices. Mask values selected in `[0, 1]`:
|
381 |
-
|
382 |
-
- 1 for tokens that are **not masked**,
|
383 |
-
- 0 for tokens that are **masked**.
|
384 |
-
|
385 |
-
[What are attention masks?](../glossary#attention-mask)
|
386 |
-
|
387 |
-
Indices can be obtained using [`AutoTokenizer`]. See [`PreTrainedTokenizer.encode`] and
|
388 |
-
[`PreTrainedTokenizer.__call__`] for details.
|
389 |
-
|
390 |
-
If `past_key_values` is used, optionally only the last `decoder_input_ids` have to be input (see
|
391 |
-
`past_key_values`).
|
392 |
-
|
393 |
-
If you want to change padding behavior, you should read [`modeling_opt._prepare_decoder_attention_mask`]
|
394 |
-
and modify to your needs. See diagram 1 in [the paper](https://arxiv.org/abs/1910.13461) for more
|
395 |
-
information on the default strategy.
|
396 |
-
|
397 |
-
- 1 indicates the head is **not masked**,
|
398 |
-
- 0 indicates the head is **masked**.
|
399 |
-
position_ids (`torch.LongTensor` of shape `(batch_size, sequence_length)`, *optional*):
|
400 |
-
Indices of positions of each input sequence tokens in the position embeddings. Selected in the range `[0,
|
401 |
-
config.n_positions - 1]`.
|
402 |
-
|
403 |
-
[What are position IDs?](../glossary#position-ids)
|
404 |
-
past_key_values (`tuple(tuple(torch.FloatTensor))`, *optional*, returned when `use_cache=True` is passed or when `config.use_cache=True`):
|
405 |
-
Tuple of `tuple(torch.FloatTensor)` of length `config.n_layers`, with each tuple having 2 tensors of shape
|
406 |
-
`(batch_size, num_heads, sequence_length, embed_size_per_head)`) and 2 additional tensors of shape
|
407 |
-
`(batch_size, num_heads, encoder_sequence_length, embed_size_per_head)`.
|
408 |
-
|
409 |
-
Contains pre-computed hidden-states (key and values in the self-attention blocks and in the cross-attention
|
410 |
-
blocks) that can be used (see `past_key_values` input) to speed up sequential decoding.
|
411 |
-
|
412 |
-
If `past_key_values` are used, the user can optionally input only the last `decoder_input_ids` (those that
|
413 |
-
don't have their past key value states given to this model) of shape `(batch_size, 1)` instead of all
|
414 |
-
`decoder_input_ids` of shape `(batch_size, sequence_length)`.
|
415 |
-
inputs_embeds (`torch.FloatTensor` of shape `(batch_size, sequence_length, hidden_size)`, *optional*):
|
416 |
-
Optionally, instead of passing `input_ids` you can choose to directly pass an embedded representation. This
|
417 |
-
is useful if you want more control over how to convert `input_ids` indices into associated vectors than the
|
418 |
-
model's internal embedding lookup matrix.
|
419 |
-
use_cache (`bool`, *optional*):
|
420 |
-
If set to `True`, `past_key_values` key value states are returned and can be used to speed up decoding (see
|
421 |
-
`past_key_values`).
|
422 |
-
output_attentions (`bool`, *optional*):
|
423 |
-
Whether or not to return the attentions tensors of all attention layers. See `attentions` under returned
|
424 |
-
tensors for more detail.
|
425 |
-
output_hidden_states (`bool`, *optional*):
|
426 |
-
Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for
|
427 |
-
more detail.
|
428 |
-
return_dict (`bool`, *optional*):
|
429 |
-
Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple.
|
430 |
-
"""
|
431 |
-
|
432 |
-
|
433 |
-
@add_start_docstrings(
|
434 |
-
"The bare LLaMA Model outputting raw hidden-states without any specific head on top.",
|
435 |
-
LLAMA_START_DOCSTRING,
|
436 |
-
)
|
437 |
-
class LlamaModel(LlamaPreTrainedModel):
|
438 |
-
"""
|
439 |
-
Transformer decoder consisting of *config.num_hidden_layers* layers. Each layer is a [`LlamaDecoderLayer`]
|
440 |
-
|
441 |
-
Args:
|
442 |
-
config: LlamaConfig
|
443 |
-
"""
|
444 |
-
|
445 |
-
def __init__(self, config: LlamaConfig):
|
446 |
-
super().__init__(config)
|
447 |
-
self.padding_idx = config.pad_token_id
|
448 |
-
self.vocab_size = config.vocab_size
|
449 |
-
|
450 |
-
self.embed_tokens = nn.Embedding(config.vocab_size, config.hidden_size, self.padding_idx)
|
451 |
-
self.layers = nn.ModuleList([LlamaDecoderLayer(config) for _ in range(config.num_hidden_layers)])
|
452 |
-
self.norm = LlamaRMSNorm(config.hidden_size, eps=config.rms_norm_eps)
|
453 |
-
|
454 |
-
self.gradient_checkpointing = False
|
455 |
-
# Initialize weights and apply final processing
|
456 |
-
self.post_init()
|
457 |
-
|
458 |
-
def get_input_embeddings(self):
|
459 |
-
return self.embed_tokens
|
460 |
-
|
461 |
-
def set_input_embeddings(self, value):
|
462 |
-
self.embed_tokens = value
|
463 |
-
|
464 |
-
# Copied from transformers.models.bart.modeling_bart.BartDecoder._prepare_decoder_attention_mask
|
465 |
-
def _prepare_decoder_attention_mask(self, attention_mask, input_shape, inputs_embeds, past_key_values_length):
|
466 |
-
# create causal mask
|
467 |
-
# [bsz, seq_len] -> [bsz, 1, tgt_seq_len, src_seq_len]
|
468 |
-
combined_attention_mask = None
|
469 |
-
if input_shape[-1] > 1:
|
470 |
-
combined_attention_mask = _make_causal_mask(
|
471 |
-
input_shape,
|
472 |
-
inputs_embeds.dtype,
|
473 |
-
device=inputs_embeds.device,
|
474 |
-
past_key_values_length=past_key_values_length,
|
475 |
-
)
|
476 |
-
|
477 |
-
if attention_mask is not None:
|
478 |
-
# [bsz, seq_len] -> [bsz, 1, tgt_seq_len, src_seq_len]
|
479 |
-
expanded_attn_mask = _expand_mask(attention_mask, inputs_embeds.dtype, tgt_len=input_shape[-1]).to(
|
480 |
-
inputs_embeds.device
|
481 |
-
)
|
482 |
-
combined_attention_mask = (
|
483 |
-
expanded_attn_mask if combined_attention_mask is None else expanded_attn_mask + combined_attention_mask
|
484 |
-
)
|
485 |
-
|
486 |
-
return combined_attention_mask
|
487 |
-
|
488 |
-
@add_start_docstrings_to_model_forward(LLAMA_INPUTS_DOCSTRING)
|
489 |
-
def forward(
|
490 |
-
self,
|
491 |
-
input_ids: torch.LongTensor = None,
|
492 |
-
attention_mask: Optional[torch.Tensor] = None,
|
493 |
-
position_ids: Optional[torch.LongTensor] = None,
|
494 |
-
past_key_values: Optional[List[torch.FloatTensor]] = None,
|
495 |
-
inputs_embeds: Optional[torch.FloatTensor] = None,
|
496 |
-
use_cache: Optional[bool] = None,
|
497 |
-
output_attentions: Optional[bool] = None,
|
498 |
-
output_hidden_states: Optional[bool] = None,
|
499 |
-
return_dict: Optional[bool] = None,
|
500 |
-
) -> Union[Tuple, BaseModelOutputWithPast]:
|
501 |
-
output_attentions = output_attentions if output_attentions is not None else self.config.output_attentions
|
502 |
-
output_hidden_states = (
|
503 |
-
output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states
|
504 |
-
)
|
505 |
-
use_cache = use_cache if use_cache is not None else self.config.use_cache
|
506 |
-
|
507 |
-
return_dict = return_dict if return_dict is not None else self.config.use_return_dict
|
508 |
-
|
509 |
-
# retrieve input_ids and inputs_embeds
|
510 |
-
if input_ids is not None and inputs_embeds is not None:
|
511 |
-
raise ValueError("You cannot specify both decoder_input_ids and decoder_inputs_embeds at the same time")
|
512 |
-
elif input_ids is not None:
|
513 |
-
batch_size, seq_length = input_ids.shape
|
514 |
-
elif inputs_embeds is not None:
|
515 |
-
batch_size, seq_length, _ = inputs_embeds.shape
|
516 |
-
else:
|
517 |
-
raise ValueError("You have to specify either decoder_input_ids or decoder_inputs_embeds")
|
518 |
-
|
519 |
-
seq_length_with_past = seq_length
|
520 |
-
past_key_values_length = 0
|
521 |
-
|
522 |
-
if past_key_values is not None:
|
523 |
-
past_key_values_length = past_key_values[0][0].shape[2]
|
524 |
-
seq_length_with_past = seq_length_with_past + past_key_values_length
|
525 |
-
|
526 |
-
if position_ids is None:
|
527 |
-
device = input_ids.device if input_ids is not None else inputs_embeds.device
|
528 |
-
position_ids = torch.arange(
|
529 |
-
past_key_values_length, seq_length + past_key_values_length, dtype=torch.long, device=device
|
530 |
-
)
|
531 |
-
position_ids = position_ids.unsqueeze(0).view(-1, seq_length)
|
532 |
-
else:
|
533 |
-
position_ids = position_ids.view(-1, seq_length).long()
|
534 |
-
|
535 |
-
if inputs_embeds is None:
|
536 |
-
inputs_embeds = self.embed_tokens(input_ids)
|
537 |
-
# embed positions
|
538 |
-
if attention_mask is None:
|
539 |
-
attention_mask = torch.ones(
|
540 |
-
(batch_size, seq_length_with_past), dtype=torch.bool, device=inputs_embeds.device
|
541 |
-
)
|
542 |
-
attention_mask = self._prepare_decoder_attention_mask(
|
543 |
-
attention_mask, (batch_size, seq_length), inputs_embeds, past_key_values_length
|
544 |
-
)
|
545 |
-
|
546 |
-
hidden_states = inputs_embeds
|
547 |
-
|
548 |
-
if self.gradient_checkpointing and self.training:
|
549 |
-
if use_cache:
|
550 |
-
logger.warning_once(
|
551 |
-
"`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..."
|
552 |
-
)
|
553 |
-
use_cache = False
|
554 |
-
|
555 |
-
# decoder layers
|
556 |
-
all_hidden_states = () if output_hidden_states else None
|
557 |
-
all_self_attns = () if output_attentions else None
|
558 |
-
next_decoder_cache = () if use_cache else None
|
559 |
-
|
560 |
-
for idx, decoder_layer in enumerate(self.layers):
|
561 |
-
if output_hidden_states:
|
562 |
-
all_hidden_states += (hidden_states,)
|
563 |
-
|
564 |
-
past_key_value = past_key_values[idx] if past_key_values is not None else None
|
565 |
-
|
566 |
-
if self.gradient_checkpointing and self.training:
|
567 |
-
|
568 |
-
def create_custom_forward(module):
|
569 |
-
def custom_forward(*inputs):
|
570 |
-
# None for past_key_value
|
571 |
-
return module(*inputs, output_attentions, None)
|
572 |
-
|
573 |
-
return custom_forward
|
574 |
-
|
575 |
-
layer_outputs = torch.utils.checkpoint.checkpoint(
|
576 |
-
create_custom_forward(decoder_layer),
|
577 |
-
hidden_states,
|
578 |
-
attention_mask,
|
579 |
-
position_ids,
|
580 |
-
None,
|
581 |
-
)
|
582 |
-
else:
|
583 |
-
layer_outputs = decoder_layer(
|
584 |
-
hidden_states,
|
585 |
-
attention_mask=attention_mask,
|
586 |
-
position_ids=position_ids,
|
587 |
-
past_key_value=past_key_value,
|
588 |
-
output_attentions=output_attentions,
|
589 |
-
use_cache=use_cache,
|
590 |
-
)
|
591 |
-
|
592 |
-
hidden_states = layer_outputs[0]
|
593 |
-
|
594 |
-
if use_cache:
|
595 |
-
next_decoder_cache += (layer_outputs[2 if output_attentions else 1],)
|
596 |
-
|
597 |
-
if output_attentions:
|
598 |
-
all_self_attns += (layer_outputs[1],)
|
599 |
-
|
600 |
-
hidden_states = self.norm(hidden_states)
|
601 |
-
|
602 |
-
# add hidden states from the last decoder layer
|
603 |
-
if output_hidden_states:
|
604 |
-
all_hidden_states += (hidden_states,)
|
605 |
-
|
606 |
-
next_cache = next_decoder_cache if use_cache else None
|
607 |
-
if not return_dict:
|
608 |
-
return tuple(v for v in [hidden_states, next_cache, all_hidden_states, all_self_attns] if v is not None)
|
609 |
-
return BaseModelOutputWithPast(
|
610 |
-
last_hidden_state=hidden_states,
|
611 |
-
past_key_values=next_cache,
|
612 |
-
hidden_states=all_hidden_states,
|
613 |
-
attentions=all_self_attns,
|
614 |
-
)
|
615 |
-
|
616 |
-
|
617 |
-
class LlamaForCausalLM(LlamaPreTrainedModel):
|
618 |
-
_tied_weights_keys = ["lm_head.weight"]
|
619 |
-
|
620 |
-
def __init__(self, config):
|
621 |
-
super().__init__(config)
|
622 |
-
self.model = LlamaModel(config)
|
623 |
-
|
624 |
-
self.lm_head = nn.Linear(config.hidden_size, config.vocab_size, bias=False)
|
625 |
-
|
626 |
-
# Initialize weights and apply final processing
|
627 |
-
self.post_init()
|
628 |
-
|
629 |
-
def get_input_embeddings(self):
|
630 |
-
return self.model.embed_tokens
|
631 |
-
|
632 |
-
def set_input_embeddings(self, value):
|
633 |
-
self.model.embed_tokens = value
|
634 |
-
|
635 |
-
def get_output_embeddings(self):
|
636 |
-
return self.lm_head
|
637 |
-
|
638 |
-
def set_output_embeddings(self, new_embeddings):
|
639 |
-
self.lm_head = new_embeddings
|
640 |
-
|
641 |
-
def set_decoder(self, decoder):
|
642 |
-
self.model = decoder
|
643 |
-
|
644 |
-
def get_decoder(self):
|
645 |
-
return self.model
|
646 |
-
|
647 |
-
@add_start_docstrings_to_model_forward(LLAMA_INPUTS_DOCSTRING)
|
648 |
-
@replace_return_docstrings(output_type=CausalLMOutputWithPast, config_class=_CONFIG_FOR_DOC)
|
649 |
-
def forward(
|
650 |
-
self,
|
651 |
-
input_ids: torch.LongTensor = None,
|
652 |
-
attention_mask: Optional[torch.Tensor] = None,
|
653 |
-
position_ids: Optional[torch.LongTensor] = None,
|
654 |
-
past_key_values: Optional[List[torch.FloatTensor]] = None,
|
655 |
-
inputs_embeds: Optional[torch.FloatTensor] = None,
|
656 |
-
labels: Optional[torch.LongTensor] = None,
|
657 |
-
use_cache: Optional[bool] = None,
|
658 |
-
output_attentions: Optional[bool] = None,
|
659 |
-
output_hidden_states: Optional[bool] = None,
|
660 |
-
return_dict: Optional[bool] = None,
|
661 |
-
) -> Union[Tuple, CausalLMOutputWithPast]:
|
662 |
-
r"""
|
663 |
-
Args:
|
664 |
-
labels (`torch.LongTensor` of shape `(batch_size, sequence_length)`, *optional*):
|
665 |
-
Labels for computing the masked language modeling loss. Indices should either be in `[0, ...,
|
666 |
-
config.vocab_size]` or -100 (see `input_ids` docstring). Tokens with indices set to `-100` are ignored
|
667 |
-
(masked), the loss is only computed for the tokens with labels in `[0, ..., config.vocab_size]`.
|
668 |
-
|
669 |
-
Returns:
|
670 |
-
|
671 |
-
Example:
|
672 |
-
|
673 |
-
```python
|
674 |
-
>>> from transformers import AutoTokenizer, LlamaForCausalLM
|
675 |
-
|
676 |
-
>>> model = LlamaForCausalLM.from_pretrained(PATH_TO_CONVERTED_WEIGHTS)
|
677 |
-
>>> tokenizer = AutoTokenizer.from_pretrained(PATH_TO_CONVERTED_TOKENIZER)
|
678 |
-
|
679 |
-
>>> prompt = "Hey, are you conscious? Can you talk to me?"
|
680 |
-
>>> inputs = tokenizer(prompt, return_tensors="pt")
|
681 |
-
|
682 |
-
>>> # Generate
|
683 |
-
>>> generate_ids = model.generate(inputs.input_ids, max_length=30)
|
684 |
-
>>> tokenizer.batch_decode(generate_ids, skip_special_tokens=True, clean_up_tokenization_spaces=False)[0]
|
685 |
-
"Hey, are you conscious? Can you talk to me?\nI'm not conscious, but I can talk to you."
|
686 |
-
```"""
|
687 |
-
|
688 |
-
output_attentions = output_attentions if output_attentions is not None else self.config.output_attentions
|
689 |
-
output_hidden_states = (
|
690 |
-
output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states
|
691 |
-
)
|
692 |
-
return_dict = return_dict if return_dict is not None else self.config.use_return_dict
|
693 |
-
|
694 |
-
# decoder outputs consists of (dec_features, layer_state, dec_hidden, dec_attn)
|
695 |
-
outputs = self.model(
|
696 |
-
input_ids=input_ids,
|
697 |
-
attention_mask=attention_mask,
|
698 |
-
position_ids=position_ids,
|
699 |
-
past_key_values=past_key_values,
|
700 |
-
inputs_embeds=inputs_embeds,
|
701 |
-
use_cache=use_cache,
|
702 |
-
output_attentions=output_attentions,
|
703 |
-
output_hidden_states=output_hidden_states,
|
704 |
-
return_dict=return_dict,
|
705 |
-
)
|
706 |
-
|
707 |
-
hidden_states = outputs[0]
|
708 |
-
logits = self.lm_head(hidden_states)
|
709 |
-
|
710 |
-
loss = None
|
711 |
-
if labels is not None:
|
712 |
-
# Shift so that tokens < n predict n
|
713 |
-
shift_logits = logits[..., :-1, :].contiguous()
|
714 |
-
shift_labels = labels[..., 1:].contiguous()
|
715 |
-
# Flatten the tokens
|
716 |
-
loss_fct = CrossEntropyLoss()
|
717 |
-
shift_logits = shift_logits.view(-1, self.config.vocab_size)
|
718 |
-
shift_labels = shift_labels.view(-1)
|
719 |
-
# Enable model parallelism
|
720 |
-
shift_labels = shift_labels.to(shift_logits.device)
|
721 |
-
loss = loss_fct(shift_logits, shift_labels)
|
722 |
-
|
723 |
-
if not return_dict:
|
724 |
-
output = (logits,) + outputs[1:]
|
725 |
-
return (loss,) + output if loss is not None else output
|
726 |
-
|
727 |
-
return CausalLMOutputWithPast(
|
728 |
-
loss=loss,
|
729 |
-
logits=logits,
|
730 |
-
past_key_values=outputs.past_key_values,
|
731 |
-
hidden_states=outputs.hidden_states,
|
732 |
-
attentions=outputs.attentions,
|
733 |
-
)
|
734 |
-
|
735 |
-
def prepare_inputs_for_generation(
|
736 |
-
self, input_ids, past_key_values=None, attention_mask=None, inputs_embeds=None, **kwargs
|
737 |
-
):
|
738 |
-
if past_key_values:
|
739 |
-
input_ids = input_ids[:, -1:]
|
740 |
-
|
741 |
-
position_ids = kwargs.get("position_ids", None)
|
742 |
-
if attention_mask is not None and position_ids is None:
|
743 |
-
# create position_ids on the fly for batch generation
|
744 |
-
position_ids = attention_mask.long().cumsum(-1) - 1
|
745 |
-
position_ids.masked_fill_(attention_mask == 0, 1)
|
746 |
-
if past_key_values:
|
747 |
-
position_ids = position_ids[:, -1].unsqueeze(-1)
|
748 |
-
|
749 |
-
# if `inputs_embeds` are passed, we only want to use them in the 1st generation step
|
750 |
-
if inputs_embeds is not None and past_key_values is None:
|
751 |
-
model_inputs = {"inputs_embeds": inputs_embeds}
|
752 |
-
else:
|
753 |
-
model_inputs = {"input_ids": input_ids}
|
754 |
-
|
755 |
-
model_inputs.update(
|
756 |
-
{
|
757 |
-
"position_ids": position_ids,
|
758 |
-
"past_key_values": past_key_values,
|
759 |
-
"use_cache": kwargs.get("use_cache"),
|
760 |
-
"attention_mask": attention_mask,
|
761 |
-
}
|
762 |
-
)
|
763 |
-
return model_inputs
|
764 |
-
|
765 |
-
@staticmethod
|
766 |
-
def _reorder_cache(past_key_values, beam_idx):
|
767 |
-
reordered_past = ()
|
768 |
-
for layer_past in past_key_values:
|
769 |
-
reordered_past += (
|
770 |
-
tuple(past_state.index_select(0, beam_idx.to(past_state.device)) for past_state in layer_past),
|
771 |
-
)
|
772 |
-
return reordered_past
|
773 |
-
|
774 |
-
|
775 |
-
@add_start_docstrings(
|
776 |
-
"""
|
777 |
-
The LLaMa Model transformer with a sequence classification head on top (linear layer).
|
778 |
-
|
779 |
-
[`LlamaForSequenceClassification`] uses the last token in order to do the classification, as other causal models
|
780 |
-
(e.g. GPT-2) do.
|
781 |
-
|
782 |
-
Since it does classification on the last token, it requires to know the position of the last token. If a
|
783 |
-
`pad_token_id` is defined in the configuration, it finds the last token that is not a padding token in each row. If
|
784 |
-
no `pad_token_id` is defined, it simply takes the last value in each row of the batch. Since it cannot guess the
|
785 |
-
padding tokens when `inputs_embeds` are passed instead of `input_ids`, it does the same (take the last value in
|
786 |
-
each row of the batch).
|
787 |
-
""",
|
788 |
-
LLAMA_START_DOCSTRING,
|
789 |
-
)
|
790 |
-
class LlamaForSequenceClassification(LlamaPreTrainedModel):
|
791 |
-
_keys_to_ignore_on_load_missing = [r"lm_head.weight"]
|
792 |
-
|
793 |
-
def __init__(self, config):
|
794 |
-
super().__init__(config)
|
795 |
-
self.num_labels = config.num_labels
|
796 |
-
self.model = LlamaModel(config)
|
797 |
-
self.score = nn.Linear(config.hidden_size, self.num_labels, bias=False)
|
798 |
-
|
799 |
-
# Initialize weights and apply final processing
|
800 |
-
self.post_init()
|
801 |
-
|
802 |
-
def get_input_embeddings(self):
|
803 |
-
return self.model.embed_tokens
|
804 |
-
|
805 |
-
def set_input_embeddings(self, value):
|
806 |
-
self.model.embed_tokens = value
|
807 |
-
|
808 |
-
@add_start_docstrings_to_model_forward(LLAMA_INPUTS_DOCSTRING)
|
809 |
-
def forward(
|
810 |
-
self,
|
811 |
-
input_ids: torch.LongTensor = None,
|
812 |
-
attention_mask: Optional[torch.Tensor] = None,
|
813 |
-
position_ids: Optional[torch.LongTensor] = None,
|
814 |
-
past_key_values: Optional[List[torch.FloatTensor]] = None,
|
815 |
-
inputs_embeds: Optional[torch.FloatTensor] = None,
|
816 |
-
labels: Optional[torch.LongTensor] = None,
|
817 |
-
use_cache: Optional[bool] = None,
|
818 |
-
output_attentions: Optional[bool] = None,
|
819 |
-
output_hidden_states: Optional[bool] = None,
|
820 |
-
return_dict: Optional[bool] = None,
|
821 |
-
) -> Union[Tuple, SequenceClassifierOutputWithPast]:
|
822 |
-
r"""
|
823 |
-
labels (`torch.LongTensor` of shape `(batch_size,)`, *optional*):
|
824 |
-
Labels for computing the sequence classification/regression loss. Indices should be in `[0, ...,
|
825 |
-
config.num_labels - 1]`. If `config.num_labels == 1` a regression loss is computed (Mean-Square loss), If
|
826 |
-
`config.num_labels > 1` a classification loss is computed (Cross-Entropy).
|
827 |
-
"""
|
828 |
-
return_dict = return_dict if return_dict is not None else self.config.use_return_dict
|
829 |
-
|
830 |
-
transformer_outputs = self.model(
|
831 |
-
input_ids,
|
832 |
-
attention_mask=attention_mask,
|
833 |
-
position_ids=position_ids,
|
834 |
-
past_key_values=past_key_values,
|
835 |
-
inputs_embeds=inputs_embeds,
|
836 |
-
use_cache=use_cache,
|
837 |
-
output_attentions=output_attentions,
|
838 |
-
output_hidden_states=output_hidden_states,
|
839 |
-
return_dict=return_dict,
|
840 |
-
)
|
841 |
-
hidden_states = transformer_outputs[0]
|
842 |
-
logits = self.score(hidden_states)
|
843 |
-
|
844 |
-
if input_ids is not None:
|
845 |
-
batch_size = input_ids.shape[0]
|
846 |
-
else:
|
847 |
-
batch_size = inputs_embeds.shape[0]
|
848 |
-
|
849 |
-
if self.config.pad_token_id is None and batch_size != 1:
|
850 |
-
raise ValueError("Cannot handle batch sizes > 1 if no padding token is defined.")
|
851 |
-
if self.config.pad_token_id is None:
|
852 |
-
sequence_lengths = -1
|
853 |
-
else:
|
854 |
-
if input_ids is not None:
|
855 |
-
sequence_lengths = (torch.ne(input_ids, self.config.pad_token_id).sum(-1) - 1).to(logits.device)
|
856 |
-
else:
|
857 |
-
sequence_lengths = -1
|
858 |
-
|
859 |
-
pooled_logits = logits[torch.arange(batch_size, device=logits.device), sequence_lengths]
|
860 |
-
|
861 |
-
loss = None
|
862 |
-
if labels is not None:
|
863 |
-
labels = labels.to(logits.device)
|
864 |
-
if self.config.problem_type is None:
|
865 |
-
if self.num_labels == 1:
|
866 |
-
self.config.problem_type = "regression"
|
867 |
-
elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int):
|
868 |
-
self.config.problem_type = "single_label_classification"
|
869 |
-
else:
|
870 |
-
self.config.problem_type = "multi_label_classification"
|
871 |
-
|
872 |
-
if self.config.problem_type == "regression":
|
873 |
-
loss_fct = MSELoss()
|
874 |
-
if self.num_labels == 1:
|
875 |
-
loss = loss_fct(pooled_logits.squeeze(), labels.squeeze())
|
876 |
-
else:
|
877 |
-
loss = loss_fct(pooled_logits, labels)
|
878 |
-
elif self.config.problem_type == "single_label_classification":
|
879 |
-
loss_fct = CrossEntropyLoss()
|
880 |
-
loss = loss_fct(pooled_logits.view(-1, self.num_labels), labels.view(-1))
|
881 |
-
elif self.config.problem_type == "multi_label_classification":
|
882 |
-
loss_fct = BCEWithLogitsLoss()
|
883 |
-
loss = loss_fct(pooled_logits, labels)
|
884 |
-
if not return_dict:
|
885 |
-
output = (pooled_logits,) + transformer_outputs[1:]
|
886 |
-
return ((loss,) + output) if loss is not None else output
|
887 |
-
|
888 |
-
return SequenceClassifierOutputWithPast(
|
889 |
-
loss=loss,
|
890 |
-
logits=pooled_logits,
|
891 |
-
past_key_values=transformer_outputs.past_key_values,
|
892 |
-
hidden_states=transformer_outputs.hidden_states,
|
893 |
-
attentions=transformer_outputs.attentions,
|
894 |
-
)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
models/TheBloke_Pygmalion-13B-SuperHOT-8K-GPTQ/special_tokens_map.json
DELETED
@@ -1,23 +0,0 @@
|
|
1 |
-
{
|
2 |
-
"bos_token": {
|
3 |
-
"content": "<s>",
|
4 |
-
"lstrip": false,
|
5 |
-
"normalized": true,
|
6 |
-
"rstrip": false,
|
7 |
-
"single_word": false
|
8 |
-
},
|
9 |
-
"eos_token": {
|
10 |
-
"content": "</s>",
|
11 |
-
"lstrip": false,
|
12 |
-
"normalized": true,
|
13 |
-
"rstrip": false,
|
14 |
-
"single_word": false
|
15 |
-
},
|
16 |
-
"unk_token": {
|
17 |
-
"content": "<unk>",
|
18 |
-
"lstrip": false,
|
19 |
-
"normalized": true,
|
20 |
-
"rstrip": false,
|
21 |
-
"single_word": false
|
22 |
-
}
|
23 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
models/TheBloke_Pygmalion-2-13B-SuperCOT2-GPTQ/LICENSE.txt
ADDED
@@ -0,0 +1,126 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
LLAMA 2 COMMUNITY LICENSE AGREEMENT
|
2 |
+
Llama 2 Version Release Date: July 18, 2023
|
3 |
+
|
4 |
+
"Agreement" means the terms and conditions for use, reproduction, distribution and
|
5 |
+
modification of the Llama Materials set forth herein.
|
6 |
+
|
7 |
+
"Documentation" means the specifications, manuals and documentation
|
8 |
+
accompanying Llama 2 distributed by Meta at ai.meta.com/resources/models-and-
|
9 |
+
libraries/llama-downloads/.
|
10 |
+
|
11 |
+
"Licensee" or "you" means you, or your employer or any other person or entity (if
|
12 |
+
you are entering into this Agreement on such person or entity's behalf), of the age
|
13 |
+
required under applicable laws, rules or regulations to provide legal consent and that
|
14 |
+
has legal authority to bind your employer or such other person or entity if you are
|
15 |
+
entering in this Agreement on their behalf.
|
16 |
+
|
17 |
+
"Llama 2" means the foundational large language models and software and
|
18 |
+
algorithms, including machine-learning model code, trained model weights,
|
19 |
+
inference-enabling code, training-enabling code, fine-tuning enabling code and other
|
20 |
+
elements of the foregoing distributed by Meta at ai.meta.com/resources/models-and-
|
21 |
+
libraries/llama-downloads/.
|
22 |
+
|
23 |
+
"Llama Materials" means, collectively, Meta's proprietary Llama 2 and
|
24 |
+
Documentation (and any portion thereof) made available under this Agreement.
|
25 |
+
|
26 |
+
"Meta" or "we" means Meta Platforms Ireland Limited (if you are located in or, if you
|
27 |
+
are an entity, your principal place of business is in the EEA or Switzerland) and Meta
|
28 |
+
Platforms, Inc. (if you are located outside of the EEA or Switzerland).
|
29 |
+
|
30 |
+
By clicking "I Accept" below or by using or distributing any portion or element of the
|
31 |
+
Llama Materials, you agree to be bound by this Agreement.
|
32 |
+
|
33 |
+
1. License Rights and Redistribution.
|
34 |
+
|
35 |
+
a. Grant of Rights. You are granted a non-exclusive, worldwide, non-
|
36 |
+
transferable and royalty-free limited license under Meta's intellectual property or
|
37 |
+
other rights owned by Meta embodied in the Llama Materials to use, reproduce,
|
38 |
+
distribute, copy, create derivative works of, and make modifications to the Llama
|
39 |
+
Materials.
|
40 |
+
|
41 |
+
b. Redistribution and Use.
|
42 |
+
|
43 |
+
i. If you distribute or make the Llama Materials, or any derivative works
|
44 |
+
thereof, available to a third party, you shall provide a copy of this Agreement to such
|
45 |
+
third party.
|
46 |
+
ii. If you receive Llama Materials, or any derivative works thereof, from
|
47 |
+
a Licensee as part of an integrated end user product, then Section 2 of this
|
48 |
+
Agreement will not apply to you.
|
49 |
+
|
50 |
+
iii. You must retain in all copies of the Llama Materials that you
|
51 |
+
distribute the following attribution notice within a "Notice" text file distributed as a
|
52 |
+
part of such copies: "Llama 2 is licensed under the LLAMA 2 Community License,
|
53 |
+
Copyright (c) Meta Platforms, Inc. All Rights Reserved."
|
54 |
+
|
55 |
+
iv. Your use of the Llama Materials must comply with applicable laws
|
56 |
+
and regulations (including trade compliance laws and regulations) and adhere to the
|
57 |
+
Acceptable Use Policy for the Llama Materials (available at
|
58 |
+
https://ai.meta.com/llama/use-policy), which is hereby incorporated by reference into
|
59 |
+
this Agreement.
|
60 |
+
|
61 |
+
v. You will not use the Llama Materials or any output or results of the
|
62 |
+
Llama Materials to improve any other large language model (excluding Llama 2 or
|
63 |
+
derivative works thereof).
|
64 |
+
|
65 |
+
2. Additional Commercial Terms. If, on the Llama 2 version release date, the
|
66 |
+
monthly active users of the products or services made available by or for Licensee,
|
67 |
+
or Licensee's affiliates, is greater than 700 million monthly active users in the
|
68 |
+
preceding calendar month, you must request a license from Meta, which Meta may
|
69 |
+
grant to you in its sole discretion, and you are not authorized to exercise any of the
|
70 |
+
rights under this Agreement unless or until Meta otherwise expressly grants you
|
71 |
+
such rights.
|
72 |
+
|
73 |
+
3. Disclaimer of Warranty. UNLESS REQUIRED BY APPLICABLE LAW, THE
|
74 |
+
LLAMA MATERIALS AND ANY OUTPUT AND RESULTS THEREFROM ARE
|
75 |
+
PROVIDED ON AN "AS IS" BASIS, WITHOUT WARRANTIES OF ANY KIND,
|
76 |
+
EITHER EXPRESS OR IMPLIED, INCLUDING, WITHOUT LIMITATION, ANY
|
77 |
+
WARRANTIES OF TITLE, NON-INFRINGEMENT, MERCHANTABILITY, OR
|
78 |
+
FITNESS FOR A PARTICULAR PURPOSE. YOU ARE SOLELY RESPONSIBLE
|
79 |
+
FOR DETERMINING THE APPROPRIATENESS OF USING OR REDISTRIBUTING
|
80 |
+
THE LLAMA MATERIALS AND ASSUME ANY RISKS ASSOCIATED WITH YOUR
|
81 |
+
USE OF THE LLAMA MATERIALS AND ANY OUTPUT AND RESULTS.
|
82 |
+
|
83 |
+
4. Limitation of Liability. IN NO EVENT WILL META OR ITS AFFILIATES BE
|
84 |
+
LIABLE UNDER ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, TORT,
|
85 |
+
NEGLIGENCE, PRODUCTS LIABILITY, OR OTHERWISE, ARISING OUT OF THIS
|
86 |
+
AGREEMENT, FOR ANY LOST PROFITS OR ANY INDIRECT, SPECIAL,
|
87 |
+
CONSEQUENTIAL, INCIDENTAL, EXEMPLARY OR PUNITIVE DAMAGES, EVEN
|
88 |
+
IF META OR ITS AFFILIATES HAVE BEEN ADVISED OF THE POSSIBILITY OF
|
89 |
+
ANY OF THE FOREGOING.
|
90 |
+
|
91 |
+
5. Intellectual Property.
|
92 |
+
|
93 |
+
a. No trademark licenses are granted under this Agreement, and in
|
94 |
+
connection with the Llama Materials, neither Meta nor Licensee may use any name
|
95 |
+
or mark owned by or associated with the other or any of its affiliates, except as
|
96 |
+
required for reasonable and customary use in describing and redistributing the
|
97 |
+
Llama Materials.
|
98 |
+
|
99 |
+
b. Subject to Meta's ownership of Llama Materials and derivatives made by or
|
100 |
+
for Meta, with respect to any derivative works and modifications of the Llama
|
101 |
+
Materials that are made by you, as between you and Meta, you are and will be the
|
102 |
+
owner of such derivative works and modifications.
|
103 |
+
|
104 |
+
c. If you institute litigation or other proceedings against Meta or any entity
|
105 |
+
(including a cross-claim or counterclaim in a lawsuit) alleging that the Llama
|
106 |
+
Materials or Llama 2 outputs or results, or any portion of any of the foregoing,
|
107 |
+
constitutes infringement of intellectual property or other rights owned or licensable
|
108 |
+
by you, then any licenses granted to you under this Agreement shall terminate as of
|
109 |
+
the date such litigation or claim is filed or instituted. You will indemnify and hold
|
110 |
+
harmless Meta from and against any claim by any third party arising out of or related
|
111 |
+
to your use or distribution of the Llama Materials.
|
112 |
+
|
113 |
+
6. Term and Termination. The term of this Agreement will commence upon your
|
114 |
+
acceptance of this Agreement or access to the Llama Materials and will continue in
|
115 |
+
full force and effect until terminated in accordance with the terms and conditions
|
116 |
+
herein. Meta may terminate this Agreement if you are in breach of any term or
|
117 |
+
condition of this Agreement. Upon termination of this Agreement, you shall delete
|
118 |
+
and cease use of the Llama Materials. Sections 3, 4 and 7 shall survive the
|
119 |
+
termination of this Agreement.
|
120 |
+
|
121 |
+
7. Governing Law and Jurisdiction. This Agreement will be governed and
|
122 |
+
construed under the laws of the State of California without regard to choice of law
|
123 |
+
principles, and the UN Convention on Contracts for the International Sale of Goods
|
124 |
+
does not apply to this Agreement. The courts of California shall have exclusive
|
125 |
+
jurisdiction of any dispute arising out of this Agreement.
|
126 |
+
|
models/TheBloke_Pygmalion-2-13B-SuperCOT2-GPTQ/README.md
ADDED
@@ -0,0 +1,308 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
base_model: https://huggingface.co/royallab/Pygmalion-2-13b-SuperCOT2
|
3 |
+
inference: false
|
4 |
+
language:
|
5 |
+
- en
|
6 |
+
library_name: transformers
|
7 |
+
license: llama2
|
8 |
+
model_creator: royallab
|
9 |
+
model_name: Pygmalion 2 13B SuperCOT2
|
10 |
+
model_type: llama
|
11 |
+
pipeline_tag: text-generation
|
12 |
+
prompt_template: 'Below is an instruction that describes a task. Write a response
|
13 |
+
that appropriately completes the request.
|
14 |
+
|
15 |
+
|
16 |
+
### Instruction:
|
17 |
+
|
18 |
+
{prompt}
|
19 |
+
|
20 |
+
|
21 |
+
### Response:
|
22 |
+
|
23 |
+
'
|
24 |
+
quantized_by: TheBloke
|
25 |
+
tags:
|
26 |
+
- llama
|
27 |
+
- llama-2
|
28 |
+
---
|
29 |
+
|
30 |
+
<!-- header start -->
|
31 |
+
<!-- 200823 -->
|
32 |
+
<div style="width: auto; margin-left: auto; margin-right: auto">
|
33 |
+
<img src="https://i.imgur.com/EBdldam.jpg" alt="TheBlokeAI" style="width: 100%; min-width: 400px; display: block; margin: auto;">
|
34 |
+
</div>
|
35 |
+
<div style="display: flex; justify-content: space-between; width: 100%;">
|
36 |
+
<div style="display: flex; flex-direction: column; align-items: flex-start;">
|
37 |
+
<p style="margin-top: 0.5em; margin-bottom: 0em;"><a href="https://discord.gg/theblokeai">Chat & support: TheBloke's Discord server</a></p>
|
38 |
+
</div>
|
39 |
+
<div style="display: flex; flex-direction: column; align-items: flex-end;">
|
40 |
+
<p style="margin-top: 0.5em; margin-bottom: 0em;"><a href="https://www.patreon.com/TheBlokeAI">Want to contribute? TheBloke's Patreon page</a></p>
|
41 |
+
</div>
|
42 |
+
</div>
|
43 |
+
<div style="text-align:center; margin-top: 0em; margin-bottom: 0em"><p style="margin-top: 0.25em; margin-bottom: 0em;">TheBloke's LLM work is generously supported by a grant from <a href="https://a16z.com">andreessen horowitz (a16z)</a></p></div>
|
44 |
+
<hr style="margin-top: 1.0em; margin-bottom: 1.0em;">
|
45 |
+
<!-- header end -->
|
46 |
+
|
47 |
+
# Pygmalion 2 13B SuperCOT2 - GPTQ
|
48 |
+
- Model creator: [royallab](https://huggingface.co/royallab)
|
49 |
+
- Original model: [Pygmalion 2 13B SuperCOT2](https://huggingface.co/royallab/Pygmalion-2-13b-SuperCOT2)
|
50 |
+
|
51 |
+
<!-- description start -->
|
52 |
+
## Description
|
53 |
+
|
54 |
+
This repo contains GPTQ model files for [royallab's Pygmalion 2 13B SuperCOT2](https://huggingface.co/royallab/Pygmalion-2-13b-SuperCOT2).
|
55 |
+
|
56 |
+
Multiple GPTQ parameter permutations are provided; see Provided Files below for details of the options provided, their parameters, and the software used to create them.
|
57 |
+
|
58 |
+
<!-- description end -->
|
59 |
+
<!-- repositories-available start -->
|
60 |
+
## Repositories available
|
61 |
+
|
62 |
+
* [AWQ model(s) for GPU inference.](https://huggingface.co/TheBloke/Pygmalion-2-13B-SuperCOT2-AWQ)
|
63 |
+
* [GPTQ models for GPU inference, with multiple quantisation parameter options.](https://huggingface.co/TheBloke/Pygmalion-2-13B-SuperCOT2-GPTQ)
|
64 |
+
* [2, 3, 4, 5, 6 and 8-bit GGUF models for CPU+GPU inference](https://huggingface.co/TheBloke/Pygmalion-2-13B-SuperCOT2-GGUF)
|
65 |
+
* [royallab's original unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/royallab/Pygmalion-2-13b-SuperCOT2)
|
66 |
+
<!-- repositories-available end -->
|
67 |
+
|
68 |
+
<!-- prompt-template start -->
|
69 |
+
## Prompt template: Alpaca
|
70 |
+
|
71 |
+
```
|
72 |
+
Below is an instruction that describes a task. Write a response that appropriately completes the request.
|
73 |
+
|
74 |
+
### Instruction:
|
75 |
+
{prompt}
|
76 |
+
|
77 |
+
### Response:
|
78 |
+
|
79 |
+
```
|
80 |
+
|
81 |
+
<!-- prompt-template end -->
|
82 |
+
|
83 |
+
|
84 |
+
<!-- README_GPTQ.md-provided-files start -->
|
85 |
+
## Provided files and GPTQ parameters
|
86 |
+
|
87 |
+
Multiple quantisation parameters are provided, to allow you to choose the best one for your hardware and requirements.
|
88 |
+
|
89 |
+
Each separate quant is in a different branch. See below for instructions on fetching from different branches.
|
90 |
+
|
91 |
+
All recent GPTQ files are made with AutoGPTQ, and all files in non-main branches are made with AutoGPTQ. Files in the `main` branch which were uploaded before August 2023 were made with GPTQ-for-LLaMa.
|
92 |
+
|
93 |
+
<details>
|
94 |
+
<summary>Explanation of GPTQ parameters</summary>
|
95 |
+
|
96 |
+
- Bits: The bit size of the quantised model.
|
97 |
+
- GS: GPTQ group size. Higher numbers use less VRAM, but have lower quantisation accuracy. "None" is the lowest possible value.
|
98 |
+
- Act Order: True or False. Also known as `desc_act`. True results in better quantisation accuracy. Some GPTQ clients have had issues with models that use Act Order plus Group Size, but this is generally resolved now.
|
99 |
+
- Damp %: A GPTQ parameter that affects how samples are processed for quantisation. 0.01 is default, but 0.1 results in slightly better accuracy.
|
100 |
+
- GPTQ dataset: The dataset used for quantisation. Using a dataset more appropriate to the model's training can improve quantisation accuracy. Note that the GPTQ dataset is not the same as the dataset used to train the model - please refer to the original model repo for details of the training dataset(s).
|
101 |
+
- Sequence Length: The length of the dataset sequences used for quantisation. Ideally this is the same as the model sequence length. For some very long sequence models (16+K), a lower sequence length may have to be used. Note that a lower sequence length does not limit the sequence length of the quantised model. It only impacts the quantisation accuracy on longer inference sequences.
|
102 |
+
- ExLlama Compatibility: Whether this file can be loaded with ExLlama, which currently only supports Llama models in 4-bit.
|
103 |
+
|
104 |
+
</details>
|
105 |
+
|
106 |
+
| Branch | Bits | GS | Act Order | Damp % | GPTQ Dataset | Seq Len | Size | ExLlama | Desc |
|
107 |
+
| ------ | ---- | -- | --------- | ------ | ------------ | ------- | ---- | ------- | ---- |
|
108 |
+
| [main](https://huggingface.co/TheBloke/Pygmalion-2-13B-SuperCOT2-GPTQ/tree/main) | 4 | 128 | Yes | 0.1 | [wikitext](https://huggingface.co/datasets/wikitext/viewer/wikitext-2-v1/test) | 4096 | 7.26 GB | Yes | 4-bit, with Act Order and group size 128g. Uses even less VRAM than 64g, but with slightly lower accuracy. |
|
109 |
+
| [gptq-4bit-32g-actorder_True](https://huggingface.co/TheBloke/Pygmalion-2-13B-SuperCOT2-GPTQ/tree/gptq-4bit-32g-actorder_True) | 4 | 32 | Yes | 0.1 | [wikitext](https://huggingface.co/datasets/wikitext/viewer/wikitext-2-v1/test) | 4096 | 8.00 GB | Yes | 4-bit, with Act Order and group size 32g. Gives highest possible inference quality, with maximum VRAM usage. |
|
110 |
+
| [gptq-8bit--1g-actorder_True](https://huggingface.co/TheBloke/Pygmalion-2-13B-SuperCOT2-GPTQ/tree/gptq-8bit--1g-actorder_True) | 8 | None | Yes | 0.1 | [wikitext](https://huggingface.co/datasets/wikitext/viewer/wikitext-2-v1/test) | 4096 | 13.36 GB | No | 8-bit, with Act Order. No group size, to lower VRAM requirements. |
|
111 |
+
| [gptq-8bit-128g-actorder_True](https://huggingface.co/TheBloke/Pygmalion-2-13B-SuperCOT2-GPTQ/tree/gptq-8bit-128g-actorder_True) | 8 | 128 | Yes | 0.1 | [wikitext](https://huggingface.co/datasets/wikitext/viewer/wikitext-2-v1/test) | 4096 | 13.65 GB | No | 8-bit, with group size 128g for higher inference quality and with Act Order for even higher accuracy. |
|
112 |
+
|
113 |
+
<!-- README_GPTQ.md-provided-files end -->
|
114 |
+
|
115 |
+
<!-- README_GPTQ.md-download-from-branches start -->
|
116 |
+
## How to download from branches
|
117 |
+
|
118 |
+
- In text-generation-webui, you can add `:branch` to the end of the download name, eg `TheBloke/Pygmalion-2-13B-SuperCOT2-GPTQ:main`
|
119 |
+
- With Git, you can clone a branch with:
|
120 |
+
```
|
121 |
+
git clone --single-branch --branch main https://huggingface.co/TheBloke/Pygmalion-2-13B-SuperCOT2-GPTQ
|
122 |
+
```
|
123 |
+
- In Python Transformers code, the branch is the `revision` parameter; see below.
|
124 |
+
<!-- README_GPTQ.md-download-from-branches end -->
|
125 |
+
<!-- README_GPTQ.md-text-generation-webui start -->
|
126 |
+
## How to easily download and use this model in [text-generation-webui](https://github.com/oobabooga/text-generation-webui).
|
127 |
+
|
128 |
+
Please make sure you're using the latest version of [text-generation-webui](https://github.com/oobabooga/text-generation-webui).
|
129 |
+
|
130 |
+
It is strongly recommended to use the text-generation-webui one-click-installers unless you're sure you know how to make a manual install.
|
131 |
+
|
132 |
+
1. Click the **Model tab**.
|
133 |
+
2. Under **Download custom model or LoRA**, enter `TheBloke/Pygmalion-2-13B-SuperCOT2-GPTQ`.
|
134 |
+
- To download from a specific branch, enter for example `TheBloke/Pygmalion-2-13B-SuperCOT2-GPTQ:main`
|
135 |
+
- see Provided Files above for the list of branches for each option.
|
136 |
+
3. Click **Download**.
|
137 |
+
4. The model will start downloading. Once it's finished it will say "Done".
|
138 |
+
5. In the top left, click the refresh icon next to **Model**.
|
139 |
+
6. In the **Model** dropdown, choose the model you just downloaded: `Pygmalion-2-13B-SuperCOT2-GPTQ`
|
140 |
+
7. The model will automatically load, and is now ready for use!
|
141 |
+
8. If you want any custom settings, set them and then click **Save settings for this model** followed by **Reload the Model** in the top right.
|
142 |
+
* Note that you do not need to and should not set manual GPTQ parameters any more. These are set automatically from the file `quantize_config.json`.
|
143 |
+
9. Once you're ready, click the **Text Generation tab** and enter a prompt to get started!
|
144 |
+
<!-- README_GPTQ.md-text-generation-webui end -->
|
145 |
+
|
146 |
+
<!-- README_GPTQ.md-use-from-python start -->
|
147 |
+
## How to use this GPTQ model from Python code
|
148 |
+
|
149 |
+
### Install the necessary packages
|
150 |
+
|
151 |
+
Requires: Transformers 4.32.0 or later, Optimum 1.12.0 or later, and AutoGPTQ 0.4.2 or later.
|
152 |
+
|
153 |
+
```shell
|
154 |
+
pip3 install transformers>=4.32.0 optimum>=1.12.0
|
155 |
+
pip3 install auto-gptq --extra-index-url https://huggingface.github.io/autogptq-index/whl/cu118/ # Use cu117 if on CUDA 11.7
|
156 |
+
```
|
157 |
+
|
158 |
+
If you have problems installing AutoGPTQ using the pre-built wheels, install it from source instead:
|
159 |
+
|
160 |
+
```shell
|
161 |
+
pip3 uninstall -y auto-gptq
|
162 |
+
git clone https://github.com/PanQiWei/AutoGPTQ
|
163 |
+
cd AutoGPTQ
|
164 |
+
pip3 install .
|
165 |
+
```
|
166 |
+
|
167 |
+
### For CodeLlama models only: you must use Transformers 4.33.0 or later.
|
168 |
+
|
169 |
+
If 4.33.0 is not yet released when you read this, you will need to install Transformers from source:
|
170 |
+
```shell
|
171 |
+
pip3 uninstall -y transformers
|
172 |
+
pip3 install git+https://github.com/huggingface/transformers.git
|
173 |
+
```
|
174 |
+
|
175 |
+
### You can then use the following code
|
176 |
+
|
177 |
+
```python
|
178 |
+
from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
|
179 |
+
|
180 |
+
model_name_or_path = "TheBloke/Pygmalion-2-13B-SuperCOT2-GPTQ"
|
181 |
+
# To use a different branch, change revision
|
182 |
+
# For example: revision="main"
|
183 |
+
model = AutoModelForCausalLM.from_pretrained(model_name_or_path,
|
184 |
+
device_map="auto",
|
185 |
+
trust_remote_code=False,
|
186 |
+
revision="main")
|
187 |
+
|
188 |
+
tokenizer = AutoTokenizer.from_pretrained(model_name_or_path, use_fast=True)
|
189 |
+
|
190 |
+
prompt = "Tell me about AI"
|
191 |
+
prompt_template=f'''Below is an instruction that describes a task. Write a response that appropriately completes the request.
|
192 |
+
|
193 |
+
### Instruction:
|
194 |
+
{prompt}
|
195 |
+
|
196 |
+
### Response:
|
197 |
+
|
198 |
+
'''
|
199 |
+
|
200 |
+
print("\n\n*** Generate:")
|
201 |
+
|
202 |
+
input_ids = tokenizer(prompt_template, return_tensors='pt').input_ids.cuda()
|
203 |
+
output = model.generate(inputs=input_ids, temperature=0.7, do_sample=True, top_p=0.95, top_k=40, max_new_tokens=512)
|
204 |
+
print(tokenizer.decode(output[0]))
|
205 |
+
|
206 |
+
# Inference can also be done using transformers' pipeline
|
207 |
+
|
208 |
+
print("*** Pipeline:")
|
209 |
+
pipe = pipeline(
|
210 |
+
"text-generation",
|
211 |
+
model=model,
|
212 |
+
tokenizer=tokenizer,
|
213 |
+
max_new_tokens=512,
|
214 |
+
do_sample=True,
|
215 |
+
temperature=0.7,
|
216 |
+
top_p=0.95,
|
217 |
+
top_k=40,
|
218 |
+
repetition_penalty=1.1
|
219 |
+
)
|
220 |
+
|
221 |
+
print(pipe(prompt_template)[0]['generated_text'])
|
222 |
+
```
|
223 |
+
<!-- README_GPTQ.md-use-from-python end -->
|
224 |
+
|
225 |
+
<!-- README_GPTQ.md-compatibility start -->
|
226 |
+
## Compatibility
|
227 |
+
|
228 |
+
The files provided are tested to work with AutoGPTQ, both via Transformers and using AutoGPTQ directly. They should also work with [Occ4m's GPTQ-for-LLaMa fork](https://github.com/0cc4m/KoboldAI).
|
229 |
+
|
230 |
+
[ExLlama](https://github.com/turboderp/exllama) is compatible with Llama models in 4-bit. Please see the Provided Files table above for per-file compatibility.
|
231 |
+
|
232 |
+
[Huggingface Text Generation Inference (TGI)](https://github.com/huggingface/text-generation-inference) is compatible with all GPTQ models.
|
233 |
+
<!-- README_GPTQ.md-compatibility end -->
|
234 |
+
|
235 |
+
<!-- footer start -->
|
236 |
+
<!-- 200823 -->
|
237 |
+
## Discord
|
238 |
+
|
239 |
+
For further support, and discussions on these models and AI in general, join us at:
|
240 |
+
|
241 |
+
[TheBloke AI's Discord server](https://discord.gg/theblokeai)
|
242 |
+
|
243 |
+
## Thanks, and how to contribute
|
244 |
+
|
245 |
+
Thanks to the [chirper.ai](https://chirper.ai) team!
|
246 |
+
|
247 |
+
Thanks to Clay from [gpus.llm-utils.org](llm-utils)!
|
248 |
+
|
249 |
+
I've had a lot of people ask if they can contribute. I enjoy providing models and helping people, and would love to be able to spend even more time doing it, as well as expanding into new projects like fine tuning/training.
|
250 |
+
|
251 |
+
If you're able and willing to contribute it will be most gratefully received and will help me to keep providing more models, and to start work on new AI projects.
|
252 |
+
|
253 |
+
Donaters will get priority support on any and all AI/LLM/model questions and requests, access to a private Discord room, plus other benefits.
|
254 |
+
|
255 |
+
* Patreon: https://patreon.com/TheBlokeAI
|
256 |
+
* Ko-Fi: https://ko-fi.com/TheBlokeAI
|
257 |
+
|
258 |
+
**Special thanks to**: Aemon Algiz.
|
259 |
+
|
260 |
+
**Patreon special mentions**: Alicia Loh, Stephen Murray, K, Ajan Kanaga, RoA, Magnesian, Deo Leter, Olakabola, Eugene Pentland, zynix, Deep Realms, Raymond Fosdick, Elijah Stavena, Iucharbius, Erik BjΓ€reholt, Luis Javier Navarrete Lozano, Nicholas, theTransient, John Detwiler, alfie_i, knownsqashed, Mano Prime, Willem Michiel, Enrico Ros, LangChain4j, OG, Michael Dempsey, Pierre Kircher, Pedro Madruga, James Bentley, Thomas Belote, Luke @flexchar, Leonard Tan, Johann-Peter Hartmann, Illia Dulskyi, Fen Risland, Chadd, S_X, Jeff Scroggin, Ken Nordquist, Sean Connelly, Artur Olbinski, Swaroop Kallakuri, Jack West, Ai Maven, David Ziegler, Russ Johnson, transmissions 11, John Villwock, Alps Aficionado, Clay Pascal, Viktor Bowallius, Subspace Studios, Rainer Wilmers, Trenton Dambrowitz, vamX, Michael Levine, μ€κ΅ κΉ, Brandon Frisco, Kalila, Trailburnt, Randy H, Talal Aujan, Nathan Dryer, Vadim, ιΏζ, ReadyPlayerEmma, Tiffany J. Kim, George Stoitzev, Spencer Kim, Jerry Meng, Gabriel Tamborski, Cory Kujawski, Jeffrey Morgan, Spiking Neurons AB, Edmond Seymore, Alexandros Triantafyllidis, Lone Striker, Cap'n Zoog, Nikolai Manek, danny, ya boyyy, Derek Yates, usrbinkat, Mandus, TL, Nathan LeClaire, subjectnull, Imad Khwaja, webtim, Raven Klaugh, Asp the Wyvern, Gabriel Puliatti, Caitlyn Gatomon, Joseph William Delisle, Jonathan Leane, Luke Pendergrass, SuperWojo, Sebastain Graf, Will Dee, Fred von Graf, Andrey, Dan Guido, Daniel P. Andersen, Nitin Borwankar, Elle, Vitor Caleffi, biorpg, jjj, NimbleBox.ai, Pieter, Matthew Berman, terasurfer, Michael Davis, Alex, Stanislav Ovsiannikov
|
261 |
+
|
262 |
+
|
263 |
+
Thank you to all my generous patrons and donaters!
|
264 |
+
|
265 |
+
And thank you again to a16z for their generous grant.
|
266 |
+
|
267 |
+
<!-- footer end -->
|
268 |
+
|
269 |
+
# Original model card: royallab's Pygmalion 2 13B SuperCOT2
|
270 |
+
|
271 |
+
|
272 |
+
# Model Card: Pygmalion-2-13b-SuperCOT2
|
273 |
+
|
274 |
+
This is a merge between:
|
275 |
+
- [Pygmalion 2 13b](https://huggingface.co/PygmalionAI/pygmalion-2-13b)
|
276 |
+
- [Ausboss's Llama2 SuperCOT2 loras](https://huggingface.co/ausboss/llama2-13b-supercot-loras2) at a weight of 1.00.
|
277 |
+
|
278 |
+
The merge was performed by a commandline version of [EzTrainer](https://github.com/CoffeeVampir3/ez-trainer) by CoffeeVampire/Blackroot via [zaraki-tools](https://github.com/zarakiquemparte/zaraki-tools) by Zaraki.
|
279 |
+
|
280 |
+
This merge differs from the previous Pyg-2-SuperCOT merge. According to AusBoss, this version was trained closer to SuperCOT llama1. The intended objective is the same, which is to make Pygmalion smarter.
|
281 |
+
|
282 |
+
The SuperCOT2 lora was merged at a weight of 1.
|
283 |
+
|
284 |
+
## Usage:
|
285 |
+
|
286 |
+
Since this is a merge between Pygmalion-2 and SuperCOT2, the following instruction formats should work:
|
287 |
+
|
288 |
+
Metharme:
|
289 |
+
|
290 |
+
```
|
291 |
+
<|system|>This is a text adventure game. Describe the scenario to the user and give him three options to pick from on each turn.<|user|>Start!<|model|>
|
292 |
+
```
|
293 |
+
|
294 |
+
Alpaca:
|
295 |
+
|
296 |
+
```
|
297 |
+
### Instruction:
|
298 |
+
Your instruction or question here.
|
299 |
+
### Response:
|
300 |
+
```
|
301 |
+
|
302 |
+
## Bias, Risks, and Limitations
|
303 |
+
|
304 |
+
The model will show biases similar to those observed in niche roleplaying forums on the Internet, besides those exhibited by the base model. It is not intended for supplying factual information or advice in any form.
|
305 |
+
|
306 |
+
## Training Details
|
307 |
+
|
308 |
+
This model is merged and can be reproduced using the tools mentioned above. Please refer to all provided links for extra model-specific details.
|
models/TheBloke_Pygmalion-2-13B-SuperCOT2-GPTQ/USE_POLICY.md
ADDED
@@ -0,0 +1,50 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Llama 2 Acceptable Use Policy
|
2 |
+
|
3 |
+
Meta is committed to promoting safe and fair use of its tools and features, including Llama 2. If you access or use Llama 2, you agree to this Acceptable Use Policy (βPolicyβ). The most recent copy of this policy can be found at [ai.meta.com/llama/use-policy](http://ai.meta.com/llama/use-policy).
|
4 |
+
|
5 |
+
## Prohibited Uses
|
6 |
+
We want everyone to use Llama 2 safely and responsibly. You agree you will not use, or allow others to use, Llama 2 to:
|
7 |
+
|
8 |
+
1. Violate the law or othersβ rights, including to:
|
9 |
+
1. Engage in, promote, generate, contribute to, encourage, plan, incite, or further illegal or unlawful activity or content, such as:
|
10 |
+
1. Violence or terrorism
|
11 |
+
2. Exploitation or harm to children, including the solicitation, creation, acquisition, or dissemination of child exploitative content or failure to report Child Sexual Abuse Material
|
12 |
+
3. Human trafficking, exploitation, and sexual violence
|
13 |
+
4. The illegal distribution of information or materials to minors, including obscene materials, or failure to employ legally required age-gating in connection with such information or materials.
|
14 |
+
5. Sexual solicitation
|
15 |
+
6. Any other criminal activity
|
16 |
+
2. Engage in, promote, incite, or facilitate the harassment, abuse, threatening, or bullying of individuals or groups of individuals
|
17 |
+
3. Engage in, promote, incite, or facilitate discrimination or other unlawful or harmful conduct in the provision of employment, employment benefits, credit, housing, other economic benefits, or other essential goods and services
|
18 |
+
4. Engage in the unauthorized or unlicensed practice of any profession including, but not limited to, financial, legal, medical/health, or related professional practices
|
19 |
+
5. Collect, process, disclose, generate, or infer health, demographic, or other sensitive personal or private information about individuals without rights and consents required by applicable laws
|
20 |
+
6. Engage in or facilitate any action or generate any content that infringes, misappropriates, or otherwise violates any third-party rights, including the outputs or results of any products or services using the Llama 2 Materials
|
21 |
+
7. Create, generate, or facilitate the creation of malicious code, malware, computer viruses or do anything else that could disable, overburden, interfere with or impair the proper working, integrity, operation or appearance of a website or computer system
|
22 |
+
|
23 |
+
|
24 |
+
|
25 |
+
2. Engage in, promote, incite, facilitate, or assist in the planning or development of activities that present a risk of death or bodily harm to individuals, including use of Llama 2 related to the following:
|
26 |
+
1. Military, warfare, nuclear industries or applications, espionage, use for materials or activities that are subject to the International Traffic Arms Regulations (ITAR) maintained by the United States Department of State
|
27 |
+
2. Guns and illegal weapons (including weapon development)
|
28 |
+
3. Illegal drugs and regulated/controlled substances
|
29 |
+
4. Operation of critical infrastructure, transportation technologies, or heavy machinery
|
30 |
+
5. Self-harm or harm to others, including suicide, cutting, and eating disorders
|
31 |
+
6. Any content intended to incite or promote violence, abuse, or any infliction of bodily harm to an individual
|
32 |
+
|
33 |
+
|
34 |
+
|
35 |
+
3. Intentionally deceive or mislead others, including use of Llama 2 related to the following:
|
36 |
+
1. Generating, promoting, or furthering fraud or the creation or promotion of disinformation
|
37 |
+
2. Generating, promoting, or furthering defamatory content, including the creation of defamatory statements, images, or other content
|
38 |
+
3. Generating, promoting, or further distributing spam
|
39 |
+
4. Impersonating another individual without consent, authorization, or legal right
|
40 |
+
5. Representing that the use of Llama 2 or outputs are human-generated
|
41 |
+
6. Generating or facilitating false online engagement, including fake reviews and other means of fake online engagement
|
42 |
+
4. Fail to appropriately disclose to end users any known dangers of your AI system
|
43 |
+
|
44 |
+
Please report any violation of this Policy, software βbug,β or other problems that could lead to a violation of this Policy through one of the following means:
|
45 |
+
|
46 |
+
* Reporting issues with the model: [github.com/facebookresearch/llama](http://github.com/facebookresearch/llama)
|
47 |
+
* Reporting risky content generated by the model: [developers.facebook.com/llama_output_feedback](http://developers.facebook.com/llama_output_feedback)
|
48 |
+
* Reporting bugs and security concerns: [facebook.com/whitehat/info](http://facebook.com/whitehat/info)
|
49 |
+
* Reporting violations of the Acceptable Use Policy or unlicensed uses of Llama: [[email protected]](mailto:[email protected])
|
50 |
+
|
models/TheBloke_Pygmalion-2-13B-SuperCOT2-GPTQ/config.json
ADDED
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "royallab/Pygmalion-2-13b-SuperCOT2",
|
3 |
+
"architectures": [
|
4 |
+
"LlamaForCausalLM"
|
5 |
+
],
|
6 |
+
"badwordsids": "[[29961], [14352], [24630], [29962], [11759], [15974], [5519], [25473], [18899], [25901], [7110], [9341], [13531], [518], [9310], [2636], [3366], [21069], [11970], [23098], [16733], [21298], [18173], [10846], [3816], [28513], [15625], [23192], [28166], [10062], [1385], [11724], [3108], [15555], [10834], [10370], [14330], [1822], [12436], [5262], [17094], [10725], [17077], [11424], [4197], [24406], [13359], [17531], [24566], [23076], [4514], [13192], [19942], [16261], [7072], [6024], [1402], [1839], [2033], [13970], [850], [5913], [28895], [5387], [8308], [24927], [5691], [12940], [19997], [18959], [11287], [16862], [4638], [22322], [29861], [21251], [14704], [17548], [12452], [17288], [23160], [24960], [8219], [18024], [5539], [7464], [27865], [29588], [20068], [19660], [27706], [22896], [24264], [12258], [2314], [4400], [5586], [12622], [6796], [7226], [21939], [18456], [14178], [21540], [21945], [14664], [16215], [10338], [17361], [7503], [13769], [26073], [9601], [26909], [7961], [8999], [20840], [16272], [21545], [3199], [10514], [5159], [22689], [6525], [20526], [27077], [18017]]",
|
7 |
+
"bos_token_id": 1,
|
8 |
+
"eos_token_id": 2,
|
9 |
+
"hidden_act": "silu",
|
10 |
+
"hidden_size": 5120,
|
11 |
+
"initializer_range": 0.02,
|
12 |
+
"intermediate_size": 13824,
|
13 |
+
"max_position_embeddings": 4096,
|
14 |
+
"model_type": "llama",
|
15 |
+
"num_attention_heads": 40,
|
16 |
+
"num_hidden_layers": 40,
|
17 |
+
"num_key_value_heads": 40,
|
18 |
+
"pretraining_tp": 1,
|
19 |
+
"rms_norm_eps": 1e-05,
|
20 |
+
"rope_scaling": null,
|
21 |
+
"rope_theta": 10000.0,
|
22 |
+
"tie_word_embeddings": false,
|
23 |
+
"torch_dtype": "float16",
|
24 |
+
"transformers_version": "4.34.0.dev0",
|
25 |
+
"use_cache": true,
|
26 |
+
"vocab_size": 32000,
|
27 |
+
"pad_token_id": 0,
|
28 |
+
"quantization_config": {
|
29 |
+
"bits": 4,
|
30 |
+
"group_size": 128,
|
31 |
+
"damp_percent": 0.1,
|
32 |
+
"desc_act": true,
|
33 |
+
"sym": true,
|
34 |
+
"true_sequential": true,
|
35 |
+
"model_name_or_path": null,
|
36 |
+
"model_file_base_name": "model",
|
37 |
+
"quant_method": "gptq"
|
38 |
+
}
|
39 |
+
}
|
models/{TheBloke_Pygmalion-13B-SuperHOT-8K-GPTQ β TheBloke_Pygmalion-2-13B-SuperCOT2-GPTQ}/generation_config.json
RENAMED
@@ -2,6 +2,5 @@
|
|
2 |
"_from_model_config": true,
|
3 |
"bos_token_id": 1,
|
4 |
"eos_token_id": 2,
|
5 |
-
"
|
6 |
-
"transformers_version": "4.30.0.dev0"
|
7 |
}
|
|
|
2 |
"_from_model_config": true,
|
3 |
"bos_token_id": 1,
|
4 |
"eos_token_id": 2,
|
5 |
+
"transformers_version": "4.34.0.dev0"
|
|
|
6 |
}
|
models/{TheBloke_Pygmalion-13B-SuperHOT-8K-GPTQ β TheBloke_Pygmalion-2-13B-SuperCOT2-GPTQ}/huggingface-metadata.txt
RENAMED
@@ -1,6 +1,6 @@
|
|
1 |
-
url: https://huggingface.co/TheBloke/Pygmalion-13B-
|
2 |
branch: main
|
3 |
-
download date: 2023-09-
|
4 |
sha256sum:
|
5 |
-
|
6 |
9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347 tokenizer.model
|
|
|
1 |
+
url: https://huggingface.co/TheBloke/Pygmalion-2-13B-SuperCOT2-GPTQ
|
2 |
branch: main
|
3 |
+
download date: 2023-09-20 03:13:55
|
4 |
sha256sum:
|
5 |
+
457189eea624e2822724055c5bb4d2c5fa9da26adfca7ca1b3900301f0b86a57 model.safetensors
|
6 |
9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347 tokenizer.model
|
models/{TheBloke_Pygmalion-13B-SuperHOT-8K-GPTQ β TheBloke_Pygmalion-2-13B-SuperCOT2-GPTQ}/model.safetensors
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:457189eea624e2822724055c5bb4d2c5fa9da26adfca7ca1b3900301f0b86a57
|
3 |
+
size 7259435192
|
models/{TheBloke_Pygmalion-13B-SuperHOT-8K-GPTQ β TheBloke_Pygmalion-2-13B-SuperCOT2-GPTQ}/quantize_config.json
RENAMED
@@ -1,9 +1,10 @@
|
|
1 |
{
|
2 |
"bits": 4,
|
3 |
"group_size": 128,
|
4 |
-
"damp_percent": 0.
|
5 |
-
"desc_act":
|
6 |
"sym": true,
|
7 |
"true_sequential": true,
|
|
|
8 |
"model_file_base_name": "model"
|
9 |
}
|
|
|
1 |
{
|
2 |
"bits": 4,
|
3 |
"group_size": 128,
|
4 |
+
"damp_percent": 0.1,
|
5 |
+
"desc_act": true,
|
6 |
"sym": true,
|
7 |
"true_sequential": true,
|
8 |
+
"model_name_or_path": null,
|
9 |
"model_file_base_name": "model"
|
10 |
}
|
models/TheBloke_Pygmalion-2-13B-SuperCOT2-GPTQ/special_tokens_map.json
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bos_token": "<s>",
|
3 |
+
"eos_token": "</s>",
|
4 |
+
"pad_token": "</s>",
|
5 |
+
"unk_token": "<unk>"
|
6 |
+
}
|
models/{TheBloke_Pygmalion-13B-SuperHOT-8K-GPTQ β TheBloke_Pygmalion-2-13B-SuperCOT2-GPTQ}/tokenizer.json
RENAMED
@@ -9,7 +9,7 @@
|
|
9 |
"single_word": false,
|
10 |
"lstrip": false,
|
11 |
"rstrip": false,
|
12 |
-
"normalized":
|
13 |
"special": true
|
14 |
},
|
15 |
{
|
@@ -18,7 +18,7 @@
|
|
18 |
"single_word": false,
|
19 |
"lstrip": false,
|
20 |
"rstrip": false,
|
21 |
-
"normalized":
|
22 |
"special": true
|
23 |
},
|
24 |
{
|
@@ -27,7 +27,7 @@
|
|
27 |
"single_word": false,
|
28 |
"lstrip": false,
|
29 |
"rstrip": false,
|
30 |
-
"normalized":
|
31 |
"special": true
|
32 |
}
|
33 |
],
|
@@ -77,10 +77,16 @@
|
|
77 |
"type_id": 0
|
78 |
}
|
79 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
80 |
{
|
81 |
"Sequence": {
|
82 |
"id": "B",
|
83 |
-
"type_id":
|
84 |
}
|
85 |
}
|
86 |
],
|
@@ -93260,126 +93266,126 @@
|
|
93260 |
"βlivre s",
|
93261 |
"lu b",
|
93262 |
"l ub",
|
|
|
93263 |
"ββ ββ",
|
93264 |
-
"
|
93265 |
-
"
|
93266 |
-
"ββ βββββ",
|
93267 |
"ββ ββββββ",
|
93268 |
-
"ββ ββββββββββββ",
|
93269 |
-
"ββ βββββββββββββ",
|
93270 |
-
"ββ ββββββββββ",
|
93271 |
-
"ββ ββββββββββββββ",
|
93272 |
-
"ββ βββ",
|
93273 |
-
"ββ βββββββββ",
|
93274 |
-
"ββ βββββββ",
|
93275 |
-
"ββ βββββββββββ",
|
93276 |
-
"ββ β",
|
93277 |
-
"ββββ ββ",
|
93278 |
"ββββ ββββ",
|
93279 |
-
"
|
93280 |
-
"
|
93281 |
-
"
|
93282 |
-
"
|
93283 |
-
"
|
93284 |
-
"
|
93285 |
-
"ββββ βββββββββ",
|
93286 |
-
"ββββ βββββββ",
|
93287 |
-
"ββββ βββββββββββ",
|
93288 |
"ββββ β",
|
93289 |
-
"
|
93290 |
-
"
|
|
|
|
|
93291 |
"ββββββββ ββββββββ",
|
93292 |
-
"ββββββββ βββββ",
|
93293 |
-
"ββββββββ ββββββ",
|
93294 |
-
"ββββββββ βββ",
|
93295 |
-
"ββββββββ βββββββ",
|
93296 |
-
"ββββββββ β",
|
93297 |
-
"βββββ ββ",
|
93298 |
-
"βββββ ββββ",
|
93299 |
-
"βββββ ββββββββ",
|
93300 |
-
"βββββ βββββ",
|
93301 |
-
"βββββ ββββββ",
|
93302 |
-
"βββββ ββββββββββ",
|
93303 |
-
"βββββ βββ",
|
93304 |
-
"βββββ βββββββββ",
|
93305 |
-
"βββββ βββββββ",
|
93306 |
"βββββ βββββββββββ",
|
93307 |
-
"βββββ β",
|
93308 |
-
"ββββββ ββ",
|
93309 |
-
"ββββββ ββββ",
|
93310 |
-
"ββββββ ββββββββ",
|
93311 |
-
"ββββββ βββββ",
|
93312 |
-
"ββββββ ββββββ",
|
93313 |
"ββββββ ββββββββββ",
|
93314 |
-
"ββββββ βββ",
|
93315 |
-
"ββββββ βββββββββ",
|
93316 |
-
"ββββββ βββββββ",
|
93317 |
-
"ββββββ β",
|
93318 |
-
"ββββββββββββ ββ",
|
93319 |
"ββββββββββββ ββββ",
|
93320 |
-
"ββββββββββββ βββ",
|
93321 |
-
"ββββββββββββ β",
|
93322 |
-
"βββββββββββββ ββ",
|
93323 |
"βββββββββββββ βββ",
|
93324 |
-
"βββββββββββββ β",
|
93325 |
-
"ββββββββββ ββ",
|
93326 |
-
"ββββββββββ ββββ",
|
93327 |
-
"ββββββββββ βββββ",
|
93328 |
"ββββββββββ ββββββ",
|
93329 |
-
"ββββββββββ βββ",
|
93330 |
-
"ββββββββββ β",
|
93331 |
"ββββββββββββββ ββ",
|
93332 |
-
"ββββββββββββββ β",
|
93333 |
-
"βββ ββ",
|
93334 |
-
"βββ ββββ",
|
93335 |
-
"βββ ββββββββ",
|
93336 |
-
"βββ βββββ",
|
93337 |
-
"βββ ββββββ",
|
93338 |
-
"βββ ββββββββββββ",
|
93339 |
"βββ βββββββββββββ",
|
93340 |
-
"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
93341 |
"βββ βββ",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
93342 |
"βββ βββββββββ",
|
93343 |
-
"βββ βββββββ",
|
93344 |
-
"βββ βββββββββββ",
|
93345 |
-
"βββ β",
|
93346 |
-
"βββββββββ ββ",
|
93347 |
-
"βββββββββ ββββ",
|
93348 |
-
"βββββββββ βββββ",
|
93349 |
-
"βββββββββ ββββββ",
|
93350 |
"βββββββββ βββ",
|
93351 |
-
"βββββββββ βββββββ",
|
93352 |
-
"βββββββββ β",
|
93353 |
-
"βββββββ ββ",
|
93354 |
-
"βββββββ ββββ",
|
93355 |
-
"βββββββ ββββββββ",
|
93356 |
"βββββββ βββββ",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
93357 |
"βββββββ ββββββ",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
93358 |
"βββββββ βββ",
|
93359 |
-
"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
93360 |
"βββββββ βββββββ",
|
93361 |
-
"βββββββ β",
|
93362 |
-
"βββββββββββ ββ",
|
93363 |
-
"βββββββββββ ββββ",
|
93364 |
-
"βββββββββββ βββββ",
|
93365 |
"βββββββββββ βββ",
|
93366 |
-
"
|
93367 |
-
"
|
93368 |
"β ββ",
|
93369 |
-
"
|
|
|
|
|
|
|
|
|
|
|
|
|
93370 |
"β ββββββββ",
|
93371 |
-
"
|
|
|
|
|
|
|
|
|
93372 |
"β ββββββ",
|
93373 |
-
"
|
93374 |
-
"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
93375 |
"β ββββββββββ",
|
93376 |
-
"
|
93377 |
-
"
|
93378 |
-
"
|
93379 |
-
"
|
93380 |
-
"
|
93381 |
-
"
|
93382 |
-
"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
93383 |
]
|
93384 |
}
|
93385 |
}
|
|
|
9 |
"single_word": false,
|
10 |
"lstrip": false,
|
11 |
"rstrip": false,
|
12 |
+
"normalized": false,
|
13 |
"special": true
|
14 |
},
|
15 |
{
|
|
|
18 |
"single_word": false,
|
19 |
"lstrip": false,
|
20 |
"rstrip": false,
|
21 |
+
"normalized": false,
|
22 |
"special": true
|
23 |
},
|
24 |
{
|
|
|
27 |
"single_word": false,
|
28 |
"lstrip": false,
|
29 |
"rstrip": false,
|
30 |
+
"normalized": false,
|
31 |
"special": true
|
32 |
}
|
33 |
],
|
|
|
77 |
"type_id": 0
|
78 |
}
|
79 |
},
|
80 |
+
{
|
81 |
+
"SpecialToken": {
|
82 |
+
"id": "<s>",
|
83 |
+
"type_id": 1
|
84 |
+
}
|
85 |
+
},
|
86 |
{
|
87 |
"Sequence": {
|
88 |
"id": "B",
|
89 |
+
"type_id": 1
|
90 |
}
|
91 |
}
|
92 |
],
|
|
|
93266 |
"βlivre s",
|
93267 |
"lu b",
|
93268 |
"l ub",
|
93269 |
+
"β β",
|
93270 |
"ββ ββ",
|
93271 |
+
"βββ β",
|
93272 |
+
"β βββ",
|
|
|
93273 |
"ββ ββββββ",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
93274 |
"ββββ ββββ",
|
93275 |
+
"βββββ βββ",
|
93276 |
+
"ββββββ ββ",
|
93277 |
+
"βββ βββββ",
|
93278 |
+
"βββββββ β",
|
93279 |
+
"β βββββββ",
|
93280 |
+
"ββ βββ",
|
|
|
|
|
|
|
93281 |
"ββββ β",
|
93282 |
+
"βββ ββ",
|
93283 |
+
"β ββββ",
|
93284 |
+
"ββ ββββββββββββββ",
|
93285 |
+
"ββββ ββββββββββββ",
|
93286 |
"ββββββββ ββββββββ",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
93287 |
"βββββ βββββββββββ",
|
|
|
|
|
|
|
|
|
|
|
|
|
93288 |
"ββββββ ββββββββββ",
|
|
|
|
|
|
|
|
|
|
|
93289 |
"ββββββββββββ ββββ",
|
|
|
|
|
|
|
93290 |
"βββββββββββββ βββ",
|
|
|
|
|
|
|
|
|
93291 |
"ββββββββββ ββββββ",
|
|
|
|
|
93292 |
"ββββββββββββββ ββ",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
93293 |
"βββ βββββββββββββ",
|
93294 |
+
"βββββββββ βββββββ",
|
93295 |
+
"βββββββ ββββββββοΏ½οΏ½",
|
93296 |
+
"βββββββββββ βββββ",
|
93297 |
+
"βββββββββββββββ β",
|
93298 |
+
"β βββββββββββββββ",
|
93299 |
+
"ββ ββββ",
|
93300 |
+
"ββββ ββ",
|
93301 |
+
"βββββ β",
|
93302 |
"βββ βββ",
|
93303 |
+
"β βββββ",
|
93304 |
+
"ββ ββββββββββ",
|
93305 |
+
"ββββ ββββββββ",
|
93306 |
+
"ββββββββ ββββ",
|
93307 |
+
"βββββ βββββββ",
|
93308 |
+
"ββββββ ββββββ",
|
93309 |
+
"ββββββββββ ββ",
|
93310 |
"βββ βββββββββ",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
93311 |
"βββββββββ βββ",
|
|
|
|
|
|
|
|
|
|
|
93312 |
"βββββββ βββββ",
|
93313 |
+
"βββββββββββ β",
|
93314 |
+
"β βββββββββββ",
|
93315 |
+
"ββ βββββββββββ",
|
93316 |
+
"ββββ βββββββββ",
|
93317 |
+
"ββββββββ βββββ",
|
93318 |
+
"βββββ ββββββββ",
|
93319 |
+
"ββββββ βββββββ",
|
93320 |
+
"ββββββββββββ β",
|
93321 |
+
"ββββββββββ βββ",
|
93322 |
+
"βββ ββββββββββ",
|
93323 |
+
"βββββββββ ββββ",
|
93324 |
"βββββββ ββββββ",
|
93325 |
+
"βββββββββββ ββ",
|
93326 |
+
"β ββββββββββββ",
|
93327 |
+
"ββ ββββββββ",
|
93328 |
+
"ββββ ββββββ",
|
93329 |
+
"ββββββββ ββ",
|
93330 |
+
"βββββ βββββ",
|
93331 |
+
"ββββββ ββββ",
|
93332 |
+
"βββ βββββββ",
|
93333 |
+
"βββββββββ β",
|
93334 |
"βββββββ βββ",
|
93335 |
+
"β βββββββββ",
|
93336 |
+
"ββ ββββββββββββ",
|
93337 |
+
"ββββ ββββββββββ",
|
93338 |
+
"ββββββββ ββββββ",
|
93339 |
+
"βββββ βββββββββ",
|
93340 |
+
"ββββββ ββββββββ",
|
93341 |
+
"ββββββββββββ ββ",
|
93342 |
+
"βββββββββββββ β",
|
93343 |
+
"ββββββββββ ββββ",
|
93344 |
+
"βββ βββββββββββ",
|
93345 |
+
"βββββββββ βββββ",
|
93346 |
"βββββββ βββββββ",
|
|
|
|
|
|
|
|
|
93347 |
"βββββββββββ βββ",
|
93348 |
+
"β βββββββββββββ",
|
93349 |
+
"ββ β",
|
93350 |
"β ββ",
|
93351 |
+
"ββ βββββββ",
|
93352 |
+
"ββββ βββββ",
|
93353 |
+
"ββββββββ β",
|
93354 |
+
"βββββ ββββ",
|
93355 |
+
"ββββββ βββ",
|
93356 |
+
"βββ ββββββ",
|
93357 |
+
"βββββββ ββ",
|
93358 |
"β ββββββββ",
|
93359 |
+
"ββ βββββ",
|
93360 |
+
"ββββ βββ",
|
93361 |
+
"βββββ ββ",
|
93362 |
+
"ββββββ β",
|
93363 |
+
"βββ ββββ",
|
93364 |
"β ββββββ",
|
93365 |
+
"ββ βββββββββ",
|
93366 |
+
"ββββ βββββββ",
|
93367 |
+
"ββββββββ βββ",
|
93368 |
+
"βββββ ββββββ",
|
93369 |
+
"ββββββ βββββ",
|
93370 |
+
"ββββββββββ β",
|
93371 |
+
"βββ ββββββββ",
|
93372 |
+
"βββββββββ ββ",
|
93373 |
+
"βββββββ ββββ",
|
93374 |
"β ββββββββββ",
|
93375 |
+
"ββ βββββββββββββ",
|
93376 |
+
"ββββ βββββββββββ",
|
93377 |
+
"ββββββββ βββββββ",
|
93378 |
+
"βββββ ββββββββββ",
|
93379 |
+
"ββββββ βββββββββ",
|
93380 |
+
"ββββββββββββ βββ",
|
93381 |
+
"βββββββββββββ ββ",
|
93382 |
+
"ββββββββββ βββββ",
|
93383 |
+
"ββββββββββββββ β",
|
93384 |
+
"βββ ββββββββββββ",
|
93385 |
+
"βββββββββ ββββββ",
|
93386 |
+
"βββββββ ββββββββ",
|
93387 |
+
"βββββββββββ ββββ",
|
93388 |
+
"β ββββββββββββββ"
|
93389 |
]
|
93390 |
}
|
93391 |
}
|
models/{TheBloke_Pygmalion-13B-SuperHOT-8K-GPTQ β TheBloke_Pygmalion-2-13B-SuperCOT2-GPTQ}/tokenizer.model
RENAMED
File without changes
|
models/{TheBloke_Pygmalion-13B-SuperHOT-8K-GPTQ β TheBloke_Pygmalion-2-13B-SuperCOT2-GPTQ}/tokenizer_config.json
RENAMED
@@ -1,11 +1,9 @@
|
|
1 |
{
|
2 |
-
"add_bos_token": true,
|
3 |
-
"add_eos_token": false,
|
4 |
"bos_token": {
|
5 |
"__type": "AddedToken",
|
6 |
"content": "<s>",
|
7 |
"lstrip": false,
|
8 |
-
"normalized":
|
9 |
"rstrip": false,
|
10 |
"single_word": false
|
11 |
},
|
@@ -14,20 +12,26 @@
|
|
14 |
"__type": "AddedToken",
|
15 |
"content": "</s>",
|
16 |
"lstrip": false,
|
17 |
-
"normalized":
|
18 |
"rstrip": false,
|
19 |
"single_word": false
|
20 |
},
|
21 |
-
"
|
|
|
22 |
"pad_token": null,
|
|
|
23 |
"sp_model_kwargs": {},
|
|
|
24 |
"tokenizer_class": "LlamaTokenizer",
|
|
|
25 |
"unk_token": {
|
26 |
"__type": "AddedToken",
|
27 |
"content": "<unk>",
|
28 |
"lstrip": false,
|
29 |
-
"normalized":
|
30 |
"rstrip": false,
|
31 |
"single_word": false
|
32 |
-
}
|
|
|
|
|
33 |
}
|
|
|
1 |
{
|
|
|
|
|
2 |
"bos_token": {
|
3 |
"__type": "AddedToken",
|
4 |
"content": "<s>",
|
5 |
"lstrip": false,
|
6 |
+
"normalized": false,
|
7 |
"rstrip": false,
|
8 |
"single_word": false
|
9 |
},
|
|
|
12 |
"__type": "AddedToken",
|
13 |
"content": "</s>",
|
14 |
"lstrip": false,
|
15 |
+
"normalized": false,
|
16 |
"rstrip": false,
|
17 |
"single_word": false
|
18 |
},
|
19 |
+
"legacy": true,
|
20 |
+
"model_max_length": 1000000000000000019884624838656,
|
21 |
"pad_token": null,
|
22 |
+
"padding_side": "right",
|
23 |
"sp_model_kwargs": {},
|
24 |
+
"spaces_between_special_tokens": false,
|
25 |
"tokenizer_class": "LlamaTokenizer",
|
26 |
+
"trust_remote_code": false,
|
27 |
"unk_token": {
|
28 |
"__type": "AddedToken",
|
29 |
"content": "<unk>",
|
30 |
"lstrip": false,
|
31 |
+
"normalized": false,
|
32 |
"rstrip": false,
|
33 |
"single_word": false
|
34 |
+
},
|
35 |
+
"use_default_system_prompt": true,
|
36 |
+
"use_fast": true
|
37 |
}
|
run.py
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
import os
|
2 |
-
os.system('
|
3 |
-
os.system('
|
|
|
1 |
import os
|
2 |
+
os.system('python3 download-model.py TheBloke/Pygmalion-2-13B-SuperCOT2-GPTQ --branch main')
|
3 |
+
os.system('python3 server.py --auto-devices --gpu-memory 24 --public-api')
|