rojasdiego
commited on
Upload LlamaForCausalLM
Browse files
model-00003-of-00004.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7d49942a618082108e12263c1e8bc213e65b7114948b5e0ae3768f2fb9e47be6
|
3 |
+
size 4915916176
|
model-00004-of-00004.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:92ecfe1a2414458b4821ac8c13cf8cb70aed66b5eea8dc5ad9eeb4ff309d6d7b
|
3 |
+
size 1168138808
|
model.safetensors.index.json
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
{
|
2 |
"metadata": {
|
3 |
-
"total_size":
|
4 |
},
|
5 |
"weight_map": {
|
6 |
"lm_head.weight": "model-00004-of-00004.safetensors",
|
@@ -230,11 +230,11 @@
|
|
230 |
"model.layers.30.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
|
231 |
"model.layers.30.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
|
232 |
"model.layers.30.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
|
233 |
-
"model.layers.31.input_layernorm.weight": "model-
|
234 |
-
"model.layers.31.mlp.down_proj.weight": "model-
|
235 |
"model.layers.31.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
|
236 |
"model.layers.31.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
|
237 |
-
"model.layers.31.post_attention_layernorm.weight": "model-
|
238 |
"model.layers.31.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
|
239 |
"model.layers.31.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
|
240 |
"model.layers.31.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
|
@@ -293,6 +293,6 @@
|
|
293 |
"model.layers.9.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
|
294 |
"model.layers.9.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
|
295 |
"model.layers.9.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
|
296 |
-
"model.norm.weight": "model-
|
297 |
}
|
298 |
}
|
|
|
1 |
{
|
2 |
"metadata": {
|
3 |
+
"total_size": 16060522496
|
4 |
},
|
5 |
"weight_map": {
|
6 |
"lm_head.weight": "model-00004-of-00004.safetensors",
|
|
|
230 |
"model.layers.30.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
|
231 |
"model.layers.30.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
|
232 |
"model.layers.30.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
|
233 |
+
"model.layers.31.input_layernorm.weight": "model-00004-of-00004.safetensors",
|
234 |
+
"model.layers.31.mlp.down_proj.weight": "model-00004-of-00004.safetensors",
|
235 |
"model.layers.31.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
|
236 |
"model.layers.31.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
|
237 |
+
"model.layers.31.post_attention_layernorm.weight": "model-00004-of-00004.safetensors",
|
238 |
"model.layers.31.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
|
239 |
"model.layers.31.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
|
240 |
"model.layers.31.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
|
|
|
293 |
"model.layers.9.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
|
294 |
"model.layers.9.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
|
295 |
"model.layers.9.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
|
296 |
+
"model.norm.weight": "model-00004-of-00004.safetensors"
|
297 |
}
|
298 |
}
|