danielhanchen commited on
Commit
8c3465c
·
verified ·
1 Parent(s): e964007

Upload Qwen2ForCausalLM

Browse files
config.json CHANGED
@@ -39,9 +39,7 @@
39
  "model.layers.2.mlp",
40
  "model.layers.3.self_attn",
41
  "model.layers.3.mlp",
42
- "model.layers.4.self_attn",
43
  "model.layers.4.mlp",
44
- "model.layers.6.mlp",
45
  "model.layers.11.mlp",
46
  "model.layers.24.mlp",
47
  "model.layers.25.mlp",
 
39
  "model.layers.2.mlp",
40
  "model.layers.3.self_attn",
41
  "model.layers.3.mlp",
 
42
  "model.layers.4.mlp",
 
43
  "model.layers.11.mlp",
44
  "model.layers.24.mlp",
45
  "model.layers.25.mlp",
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e8231319d72b4fc1c981edaa8cdc5186c7d453dfeba29e3d2570b6af7d872796
3
- size 4995769972
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d89951edcbded32eea6718bf5957114c7151be76c8be277a381fa98f7105209b
3
+ size 4967113023
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f5beca751f42e159960c6753202b9f1e4a50e55e45bd32f86400de343d17ac01
3
- size 3835794652
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:32152fcce250c61031324ce54a600de4e483becc307df492b0c910314b5cf2cd
3
+ size 3518593782
model.safetensors.index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "metadata": {
3
- "total_size": 8831440744
4
  },
5
  "weight_map": {
6
  "lm_head.weight": "model-00002-of-00002.safetensors",
@@ -155,26 +155,26 @@
155
  "model.layers.12.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
156
  "model.layers.12.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
157
  "model.layers.12.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
158
- "model.layers.13.input_layernorm.weight": "model-00002-of-00002.safetensors",
159
- "model.layers.13.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
160
- "model.layers.13.mlp.down_proj.weight.absmax": "model-00002-of-00002.safetensors",
161
- "model.layers.13.mlp.down_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
162
- "model.layers.13.mlp.down_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
163
- "model.layers.13.mlp.down_proj.weight.quant_map": "model-00002-of-00002.safetensors",
164
- "model.layers.13.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
165
- "model.layers.13.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
166
- "model.layers.13.mlp.gate_proj.weight.absmax": "model-00002-of-00002.safetensors",
167
- "model.layers.13.mlp.gate_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
168
- "model.layers.13.mlp.gate_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
169
- "model.layers.13.mlp.gate_proj.weight.quant_map": "model-00002-of-00002.safetensors",
170
- "model.layers.13.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
171
- "model.layers.13.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
172
- "model.layers.13.mlp.up_proj.weight.absmax": "model-00002-of-00002.safetensors",
173
- "model.layers.13.mlp.up_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
174
- "model.layers.13.mlp.up_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
175
- "model.layers.13.mlp.up_proj.weight.quant_map": "model-00002-of-00002.safetensors",
176
- "model.layers.13.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
177
- "model.layers.13.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
178
  "model.layers.13.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
179
  "model.layers.13.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
180
  "model.layers.13.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
@@ -182,12 +182,12 @@
182
  "model.layers.13.self_attn.k_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
183
  "model.layers.13.self_attn.k_proj.weight.quant_map": "model-00001-of-00002.safetensors",
184
  "model.layers.13.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
185
- "model.layers.13.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
186
- "model.layers.13.self_attn.o_proj.weight.absmax": "model-00002-of-00002.safetensors",
187
- "model.layers.13.self_attn.o_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
188
- "model.layers.13.self_attn.o_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
189
- "model.layers.13.self_attn.o_proj.weight.quant_map": "model-00002-of-00002.safetensors",
190
- "model.layers.13.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
191
  "model.layers.13.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
192
  "model.layers.13.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
193
  "model.layers.13.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
@@ -202,53 +202,53 @@
202
  "model.layers.13.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
203
  "model.layers.13.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
204
  "model.layers.13.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
205
- "model.layers.14.input_layernorm.weight": "model-00002-of-00002.safetensors",
206
- "model.layers.14.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
207
- "model.layers.14.mlp.down_proj.weight.absmax": "model-00002-of-00002.safetensors",
208
- "model.layers.14.mlp.down_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
209
- "model.layers.14.mlp.down_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
210
- "model.layers.14.mlp.down_proj.weight.quant_map": "model-00002-of-00002.safetensors",
211
- "model.layers.14.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
212
- "model.layers.14.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
213
- "model.layers.14.mlp.gate_proj.weight.absmax": "model-00002-of-00002.safetensors",
214
- "model.layers.14.mlp.gate_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
215
- "model.layers.14.mlp.gate_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
216
- "model.layers.14.mlp.gate_proj.weight.quant_map": "model-00002-of-00002.safetensors",
217
- "model.layers.14.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
218
- "model.layers.14.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
219
- "model.layers.14.mlp.up_proj.weight.absmax": "model-00002-of-00002.safetensors",
220
- "model.layers.14.mlp.up_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
221
- "model.layers.14.mlp.up_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
222
- "model.layers.14.mlp.up_proj.weight.quant_map": "model-00002-of-00002.safetensors",
223
- "model.layers.14.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
224
- "model.layers.14.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
225
- "model.layers.14.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
226
- "model.layers.14.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
227
- "model.layers.14.self_attn.k_proj.weight.absmax": "model-00002-of-00002.safetensors",
228
- "model.layers.14.self_attn.k_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
229
- "model.layers.14.self_attn.k_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
230
- "model.layers.14.self_attn.k_proj.weight.quant_map": "model-00002-of-00002.safetensors",
231
- "model.layers.14.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
232
- "model.layers.14.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
233
- "model.layers.14.self_attn.o_proj.weight.absmax": "model-00002-of-00002.safetensors",
234
- "model.layers.14.self_attn.o_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
235
- "model.layers.14.self_attn.o_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
236
- "model.layers.14.self_attn.o_proj.weight.quant_map": "model-00002-of-00002.safetensors",
237
- "model.layers.14.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
238
- "model.layers.14.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
239
- "model.layers.14.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
240
- "model.layers.14.self_attn.q_proj.weight.absmax": "model-00002-of-00002.safetensors",
241
- "model.layers.14.self_attn.q_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
242
- "model.layers.14.self_attn.q_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
243
- "model.layers.14.self_attn.q_proj.weight.quant_map": "model-00002-of-00002.safetensors",
244
- "model.layers.14.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
245
- "model.layers.14.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
246
- "model.layers.14.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
247
- "model.layers.14.self_attn.v_proj.weight.absmax": "model-00002-of-00002.safetensors",
248
- "model.layers.14.self_attn.v_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
249
- "model.layers.14.self_attn.v_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
250
- "model.layers.14.self_attn.v_proj.weight.quant_map": "model-00002-of-00002.safetensors",
251
- "model.layers.14.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
252
  "model.layers.15.input_layernorm.weight": "model-00002-of-00002.safetensors",
253
  "model.layers.15.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
254
  "model.layers.15.mlp.down_proj.weight.absmax": "model-00002-of-00002.safetensors",
@@ -256,46 +256,46 @@
256
  "model.layers.15.mlp.down_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
257
  "model.layers.15.mlp.down_proj.weight.quant_map": "model-00002-of-00002.safetensors",
258
  "model.layers.15.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
259
- "model.layers.15.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
260
- "model.layers.15.mlp.gate_proj.weight.absmax": "model-00002-of-00002.safetensors",
261
- "model.layers.15.mlp.gate_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
262
- "model.layers.15.mlp.gate_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
263
- "model.layers.15.mlp.gate_proj.weight.quant_map": "model-00002-of-00002.safetensors",
264
- "model.layers.15.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
265
- "model.layers.15.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
266
- "model.layers.15.mlp.up_proj.weight.absmax": "model-00002-of-00002.safetensors",
267
- "model.layers.15.mlp.up_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
268
- "model.layers.15.mlp.up_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
269
- "model.layers.15.mlp.up_proj.weight.quant_map": "model-00002-of-00002.safetensors",
270
- "model.layers.15.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
271
  "model.layers.15.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
272
- "model.layers.15.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
273
- "model.layers.15.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
274
- "model.layers.15.self_attn.k_proj.weight.absmax": "model-00002-of-00002.safetensors",
275
- "model.layers.15.self_attn.k_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
276
- "model.layers.15.self_attn.k_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
277
- "model.layers.15.self_attn.k_proj.weight.quant_map": "model-00002-of-00002.safetensors",
278
- "model.layers.15.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
279
- "model.layers.15.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
280
- "model.layers.15.self_attn.o_proj.weight.absmax": "model-00002-of-00002.safetensors",
281
- "model.layers.15.self_attn.o_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
282
- "model.layers.15.self_attn.o_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
283
- "model.layers.15.self_attn.o_proj.weight.quant_map": "model-00002-of-00002.safetensors",
284
- "model.layers.15.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
285
- "model.layers.15.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
286
- "model.layers.15.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
287
- "model.layers.15.self_attn.q_proj.weight.absmax": "model-00002-of-00002.safetensors",
288
- "model.layers.15.self_attn.q_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
289
- "model.layers.15.self_attn.q_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
290
- "model.layers.15.self_attn.q_proj.weight.quant_map": "model-00002-of-00002.safetensors",
291
- "model.layers.15.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
292
- "model.layers.15.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
293
- "model.layers.15.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
294
- "model.layers.15.self_attn.v_proj.weight.absmax": "model-00002-of-00002.safetensors",
295
- "model.layers.15.self_attn.v_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
296
- "model.layers.15.self_attn.v_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
297
- "model.layers.15.self_attn.v_proj.weight.quant_map": "model-00002-of-00002.safetensors",
298
- "model.layers.15.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
299
  "model.layers.16.input_layernorm.weight": "model-00002-of-00002.safetensors",
300
  "model.layers.16.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
301
  "model.layers.16.mlp.down_proj.weight.absmax": "model-00002-of-00002.safetensors",
@@ -826,11 +826,31 @@
826
  "model.layers.4.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
827
  "model.layers.4.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
828
  "model.layers.4.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
 
 
 
 
 
829
  "model.layers.4.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
 
 
 
 
 
830
  "model.layers.4.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
831
  "model.layers.4.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
 
 
 
 
 
832
  "model.layers.4.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
833
  "model.layers.4.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
 
 
 
 
 
834
  "model.layers.5.input_layernorm.weight": "model-00001-of-00002.safetensors",
835
  "model.layers.5.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
836
  "model.layers.5.mlp.down_proj.weight.absmax": "model-00001-of-00002.safetensors",
@@ -880,8 +900,23 @@
880
  "model.layers.5.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
881
  "model.layers.6.input_layernorm.weight": "model-00001-of-00002.safetensors",
882
  "model.layers.6.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
 
 
 
 
 
883
  "model.layers.6.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
 
 
 
 
 
884
  "model.layers.6.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
 
 
 
 
 
885
  "model.layers.6.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
886
  "model.layers.6.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
887
  "model.layers.6.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
 
1
  {
2
  "metadata": {
3
+ "total_size": 8485578781
4
  },
5
  "weight_map": {
6
  "lm_head.weight": "model-00002-of-00002.safetensors",
 
155
  "model.layers.12.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
156
  "model.layers.12.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
157
  "model.layers.12.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
158
+ "model.layers.13.input_layernorm.weight": "model-00001-of-00002.safetensors",
159
+ "model.layers.13.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
160
+ "model.layers.13.mlp.down_proj.weight.absmax": "model-00001-of-00002.safetensors",
161
+ "model.layers.13.mlp.down_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
162
+ "model.layers.13.mlp.down_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
163
+ "model.layers.13.mlp.down_proj.weight.quant_map": "model-00001-of-00002.safetensors",
164
+ "model.layers.13.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
165
+ "model.layers.13.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
166
+ "model.layers.13.mlp.gate_proj.weight.absmax": "model-00001-of-00002.safetensors",
167
+ "model.layers.13.mlp.gate_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
168
+ "model.layers.13.mlp.gate_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
169
+ "model.layers.13.mlp.gate_proj.weight.quant_map": "model-00001-of-00002.safetensors",
170
+ "model.layers.13.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
171
+ "model.layers.13.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
172
+ "model.layers.13.mlp.up_proj.weight.absmax": "model-00001-of-00002.safetensors",
173
+ "model.layers.13.mlp.up_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
174
+ "model.layers.13.mlp.up_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
175
+ "model.layers.13.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
176
+ "model.layers.13.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
177
+ "model.layers.13.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
178
  "model.layers.13.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
179
  "model.layers.13.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
180
  "model.layers.13.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
 
182
  "model.layers.13.self_attn.k_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
183
  "model.layers.13.self_attn.k_proj.weight.quant_map": "model-00001-of-00002.safetensors",
184
  "model.layers.13.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
185
+ "model.layers.13.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
186
+ "model.layers.13.self_attn.o_proj.weight.absmax": "model-00001-of-00002.safetensors",
187
+ "model.layers.13.self_attn.o_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
188
+ "model.layers.13.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
189
+ "model.layers.13.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
190
+ "model.layers.13.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
191
  "model.layers.13.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
192
  "model.layers.13.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
193
  "model.layers.13.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
 
202
  "model.layers.13.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
203
  "model.layers.13.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
204
  "model.layers.13.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
205
+ "model.layers.14.input_layernorm.weight": "model-00001-of-00002.safetensors",
206
+ "model.layers.14.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
207
+ "model.layers.14.mlp.down_proj.weight.absmax": "model-00001-of-00002.safetensors",
208
+ "model.layers.14.mlp.down_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
209
+ "model.layers.14.mlp.down_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
210
+ "model.layers.14.mlp.down_proj.weight.quant_map": "model-00001-of-00002.safetensors",
211
+ "model.layers.14.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
212
+ "model.layers.14.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
213
+ "model.layers.14.mlp.gate_proj.weight.absmax": "model-00001-of-00002.safetensors",
214
+ "model.layers.14.mlp.gate_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
215
+ "model.layers.14.mlp.gate_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
216
+ "model.layers.14.mlp.gate_proj.weight.quant_map": "model-00001-of-00002.safetensors",
217
+ "model.layers.14.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
218
+ "model.layers.14.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
219
+ "model.layers.14.mlp.up_proj.weight.absmax": "model-00001-of-00002.safetensors",
220
+ "model.layers.14.mlp.up_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
221
+ "model.layers.14.mlp.up_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
222
+ "model.layers.14.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
223
+ "model.layers.14.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
224
+ "model.layers.14.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
225
+ "model.layers.14.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
226
+ "model.layers.14.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
227
+ "model.layers.14.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
228
+ "model.layers.14.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
229
+ "model.layers.14.self_attn.k_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
230
+ "model.layers.14.self_attn.k_proj.weight.quant_map": "model-00001-of-00002.safetensors",
231
+ "model.layers.14.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
232
+ "model.layers.14.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
233
+ "model.layers.14.self_attn.o_proj.weight.absmax": "model-00001-of-00002.safetensors",
234
+ "model.layers.14.self_attn.o_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
235
+ "model.layers.14.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
236
+ "model.layers.14.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
237
+ "model.layers.14.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
238
+ "model.layers.14.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
239
+ "model.layers.14.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
240
+ "model.layers.14.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
241
+ "model.layers.14.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
242
+ "model.layers.14.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
243
+ "model.layers.14.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
244
+ "model.layers.14.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
245
+ "model.layers.14.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
246
+ "model.layers.14.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
247
+ "model.layers.14.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
248
+ "model.layers.14.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
249
+ "model.layers.14.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
250
+ "model.layers.14.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
251
+ "model.layers.14.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
252
  "model.layers.15.input_layernorm.weight": "model-00002-of-00002.safetensors",
253
  "model.layers.15.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
254
  "model.layers.15.mlp.down_proj.weight.absmax": "model-00002-of-00002.safetensors",
 
256
  "model.layers.15.mlp.down_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
257
  "model.layers.15.mlp.down_proj.weight.quant_map": "model-00002-of-00002.safetensors",
258
  "model.layers.15.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
259
+ "model.layers.15.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
260
+ "model.layers.15.mlp.gate_proj.weight.absmax": "model-00001-of-00002.safetensors",
261
+ "model.layers.15.mlp.gate_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
262
+ "model.layers.15.mlp.gate_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
263
+ "model.layers.15.mlp.gate_proj.weight.quant_map": "model-00001-of-00002.safetensors",
264
+ "model.layers.15.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
265
+ "model.layers.15.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
266
+ "model.layers.15.mlp.up_proj.weight.absmax": "model-00001-of-00002.safetensors",
267
+ "model.layers.15.mlp.up_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
268
+ "model.layers.15.mlp.up_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
269
+ "model.layers.15.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
270
+ "model.layers.15.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
271
  "model.layers.15.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
272
+ "model.layers.15.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
273
+ "model.layers.15.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
274
+ "model.layers.15.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
275
+ "model.layers.15.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
276
+ "model.layers.15.self_attn.k_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
277
+ "model.layers.15.self_attn.k_proj.weight.quant_map": "model-00001-of-00002.safetensors",
278
+ "model.layers.15.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
279
+ "model.layers.15.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
280
+ "model.layers.15.self_attn.o_proj.weight.absmax": "model-00001-of-00002.safetensors",
281
+ "model.layers.15.self_attn.o_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
282
+ "model.layers.15.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
283
+ "model.layers.15.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
284
+ "model.layers.15.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
285
+ "model.layers.15.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
286
+ "model.layers.15.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
287
+ "model.layers.15.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
288
+ "model.layers.15.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
289
+ "model.layers.15.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
290
+ "model.layers.15.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
291
+ "model.layers.15.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
292
+ "model.layers.15.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
293
+ "model.layers.15.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
294
+ "model.layers.15.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
295
+ "model.layers.15.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
296
+ "model.layers.15.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
297
+ "model.layers.15.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
298
+ "model.layers.15.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
299
  "model.layers.16.input_layernorm.weight": "model-00002-of-00002.safetensors",
300
  "model.layers.16.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
301
  "model.layers.16.mlp.down_proj.weight.absmax": "model-00002-of-00002.safetensors",
 
826
  "model.layers.4.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
827
  "model.layers.4.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
828
  "model.layers.4.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
829
+ "model.layers.4.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
830
+ "model.layers.4.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
831
+ "model.layers.4.self_attn.k_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
832
+ "model.layers.4.self_attn.k_proj.weight.quant_map": "model-00001-of-00002.safetensors",
833
+ "model.layers.4.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
834
  "model.layers.4.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
835
+ "model.layers.4.self_attn.o_proj.weight.absmax": "model-00001-of-00002.safetensors",
836
+ "model.layers.4.self_attn.o_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
837
+ "model.layers.4.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
838
+ "model.layers.4.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
839
+ "model.layers.4.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
840
  "model.layers.4.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
841
  "model.layers.4.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
842
+ "model.layers.4.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
843
+ "model.layers.4.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
844
+ "model.layers.4.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
845
+ "model.layers.4.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
846
+ "model.layers.4.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
847
  "model.layers.4.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
848
  "model.layers.4.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
849
+ "model.layers.4.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
850
+ "model.layers.4.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
851
+ "model.layers.4.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
852
+ "model.layers.4.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
853
+ "model.layers.4.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
854
  "model.layers.5.input_layernorm.weight": "model-00001-of-00002.safetensors",
855
  "model.layers.5.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
856
  "model.layers.5.mlp.down_proj.weight.absmax": "model-00001-of-00002.safetensors",
 
900
  "model.layers.5.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
901
  "model.layers.6.input_layernorm.weight": "model-00001-of-00002.safetensors",
902
  "model.layers.6.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
903
+ "model.layers.6.mlp.down_proj.weight.absmax": "model-00001-of-00002.safetensors",
904
+ "model.layers.6.mlp.down_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
905
+ "model.layers.6.mlp.down_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
906
+ "model.layers.6.mlp.down_proj.weight.quant_map": "model-00001-of-00002.safetensors",
907
+ "model.layers.6.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
908
  "model.layers.6.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
909
+ "model.layers.6.mlp.gate_proj.weight.absmax": "model-00001-of-00002.safetensors",
910
+ "model.layers.6.mlp.gate_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
911
+ "model.layers.6.mlp.gate_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
912
+ "model.layers.6.mlp.gate_proj.weight.quant_map": "model-00001-of-00002.safetensors",
913
+ "model.layers.6.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
914
  "model.layers.6.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
915
+ "model.layers.6.mlp.up_proj.weight.absmax": "model-00001-of-00002.safetensors",
916
+ "model.layers.6.mlp.up_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
917
+ "model.layers.6.mlp.up_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
918
+ "model.layers.6.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
919
+ "model.layers.6.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
920
  "model.layers.6.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
921
  "model.layers.6.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
922
  "model.layers.6.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",