junlinw commited on
Commit
b7d02d7
·
verified ·
1 Parent(s): c60efec

Upload LlamaForSequenceClassification

Browse files
config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "_name_or_path": "llama3.2-3b-lr5e-5-ep1",
3
  "architectures": [
4
- "LlamaForCausalLM"
5
  ],
6
  "attention_bias": false,
7
  "attention_dropout": 0.0,
 
1
  {
2
  "_name_or_path": "llama3.2-3b-lr5e-5-ep1",
3
  "architectures": [
4
+ "LlamaForSequenceClassification"
5
  ],
6
  "attention_bias": false,
7
  "attention_dropout": 0.0,
model-00003-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dc242d4858a643958115d2bd99761466050c1a7560314ca74aae52c884c66d44
3
- size 2919452168
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6dcb2e60d62e55f1574300cc4045a2ef310ade75fc65d9dba4fb42f4cd5c34f7
3
+ size 2919476832
model.safetensors.index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "metadata": {
3
- "total_size": 12850999296
4
  },
5
  "weight_map": {
6
  "model.embed_tokens.weight": "model-00001-of-00003.safetensors",
@@ -256,6 +256,7 @@
256
  "model.layers.9.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
257
  "model.layers.9.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
258
  "model.layers.9.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
259
- "model.norm.weight": "model-00003-of-00003.safetensors"
 
260
  }
261
  }
 
1
  {
2
  "metadata": {
3
+ "total_size": 12851023872
4
  },
5
  "weight_map": {
6
  "model.embed_tokens.weight": "model-00001-of-00003.safetensors",
 
256
  "model.layers.9.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
257
  "model.layers.9.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
258
  "model.layers.9.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
259
+ "model.norm.weight": "model-00003-of-00003.safetensors",
260
+ "score.weight": "model-00003-of-00003.safetensors"
261
  }
262
  }