alexis779 commited on
Commit
5c7695e
1 Parent(s): 599d221

Upload folder using huggingface_hub

Browse files
adapter_config.json CHANGED
@@ -20,14 +20,14 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
23
  "up_proj",
24
  "lm_head",
25
- "down_proj",
26
  "gate_proj",
 
27
  "o_proj",
28
  "v_proj",
29
- "embed_tokensq_proj",
30
- "k_proj"
31
  ],
32
  "task_type": "CAUSAL_LM",
33
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "k_proj",
24
  "up_proj",
25
  "lm_head",
 
26
  "gate_proj",
27
+ "down_proj",
28
  "o_proj",
29
  "v_proj",
30
+ "embed_tokensq_proj"
 
31
  ],
32
  "task_type": "CAUSAL_LM",
33
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fefe9032b96144ce75ff74a166ef1a6b8aa0cdada498ed9d3600d394b9d8e435
3
- size 690277408
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5766f52b6a0a1c55c0ea594d6ea6804d4b15f58251f907d342b7a99abe30950f
3
+ size 345164832
config.json CHANGED
@@ -20,7 +20,7 @@
20
  "sliding_window": null,
21
  "tie_word_embeddings": false,
22
  "torch_dtype": "float16",
23
- "transformers_version": "4.41.2",
24
  "use_cache": true,
25
  "vocab_size": 32768
26
  }
 
20
  "sliding_window": null,
21
  "tie_word_embeddings": false,
22
  "torch_dtype": "float16",
23
+ "transformers_version": "4.42.3",
24
  "use_cache": true,
25
  "vocab_size": 32768
26
  }
generation_config.json CHANGED
@@ -2,5 +2,5 @@
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
- "transformers_version": "4.41.2"
6
  }
 
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
+ "transformers_version": "4.42.3"
6
  }
model-00001-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0529284280d83ed959ac60fae64dd7a0bb46c9633846d9d456b47bf91cf84735
3
  size 4949453696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bd783bd055a0caf769b031952a5a6e36feeacaaeb7333f8d1ad90886e71fa504
3
  size 4949453696
model-00002-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6ec40c8ff98710657fcc5833125e27f2643e36218eeaddd7753157d9f1565abb
3
  size 4999819232
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0ecc23c06ec4d3caea535607803ebfb211ece88eb70ceddaa95c7596e7611baf
3
  size 4999819232
model-00003-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5e63c76f9a80ddfc4f628d735d9a0b00280f27ca3cc053f107a5012e64b31f17
3
  size 4546807712
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:27a37423c497807c5c516cab8bc22baa1502ee90ae15115ab8407b35ea0be151
3
  size 4546807712
tokenizer_config.json CHANGED
@@ -6173,6 +6173,7 @@
6173
  }
6174
  },
6175
  "bos_token": "<s>",
 
6176
  "clean_up_tokenization_spaces": false,
6177
  "device_map": "auto",
6178
  "eos_token": "</s>",
 
6173
  }
6174
  },
6175
  "bos_token": "<s>",
6176
+ "chat_template": "{{bos_token}}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + '[/INST]' }}{% elif message['role'] == 'assistant' %}{{ ' ' + message['content'] + eos_token}}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}",
6177
  "clean_up_tokenization_spaces": false,
6178
  "device_map": "auto",
6179
  "eos_token": "</s>",