Gustavo de Rosa commited on
Commit
b15fb48
·
1 Parent(s): f957856

fix(root): Adds proposed changes by unsloth.ai.

Browse files
config.json CHANGED
@@ -8,7 +8,7 @@
8
  "auto_map": {},
9
  "bos_token_id": 100257,
10
  "embd_pdrop": 0.0,
11
- "eos_token_id": 100257,
12
  "hidden_act": "silu",
13
  "hidden_size": 5120,
14
  "initializer_range": 0.02,
@@ -19,7 +19,7 @@
19
  "num_hidden_layers": 40,
20
  "num_key_value_heads": 10,
21
  "original_max_position_embeddings": 16384,
22
- "pad_token_id": 100257,
23
  "resid_pdrop": 0.0,
24
  "rms_norm_eps": 1e-05,
25
  "rope_scaling": null,
 
8
  "auto_map": {},
9
  "bos_token_id": 100257,
10
  "embd_pdrop": 0.0,
11
+ "eos_token_id": 100265,
12
  "hidden_act": "silu",
13
  "hidden_size": 5120,
14
  "initializer_range": 0.02,
 
19
  "num_hidden_layers": 40,
20
  "num_key_value_heads": 10,
21
  "original_max_position_embeddings": 16384,
22
+ "pad_token_id": 100351,
23
  "resid_pdrop": 0.0,
24
  "rms_norm_eps": 1e-05,
25
  "rope_scaling": null,
generation_config.json CHANGED
@@ -1,9 +1,7 @@
1
  {
2
  "_from_model_config": true,
3
  "bos_token_id": 100257,
4
- "eos_token_id": [
5
- 100257,
6
- 100265
7
- ],
8
  "transformers_version": "4.47.0"
9
  }
 
1
  {
2
  "_from_model_config": true,
3
  "bos_token_id": 100257,
4
+ "eos_token_id": 100265,
5
+ "pad_token_id": 100351,
 
 
6
  "transformers_version": "4.47.0"
7
  }
special_tokens_map.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
  "bos_token": "<|endoftext|>",
3
- "eos_token": "<|endoftext|>",
4
- "pad_token": "<|endoftext|>"
5
  }
 
1
  {
2
  "bos_token": "<|endoftext|>",
3
+ "eos_token": "<|im_end|>",
4
+ "pad_token": "<|dummy_87|>"
5
  }
tokenizer_config.json CHANGED
@@ -771,10 +771,10 @@
771
  }
772
  },
773
  "bos_token": "<|endoftext|>",
774
- "chat_template": "{% for message in messages %}{% if (message['role'] == 'system') %}{{'<|im_start|>system<|im_sep|>' + message['content'] + '<|im_end|>'}}{% elif (message['role'] == 'user') %}{{'<|im_start|>user<|im_sep|>' + message['content'] + '<|im_end|><|im_start|>assistant<|im_sep|>'}}{% elif (message['role'] == 'assistant') %}{{message['content'] + '<|im_end|>'}}{% endif %}{% endfor %}",
775
  "clean_up_tokenization_spaces": false,
776
- "eos_token": "<|endoftext|>",
777
  "model_max_length": 16384,
778
- "pad_token": "<|endoftext|>",
779
  "tokenizer_class": "GPT2Tokenizer"
780
  }
 
771
  }
772
  },
773
  "bos_token": "<|endoftext|>",
774
+ "chat_template": "{% for message in messages %}{% if (message['role'] == 'system') %}{{'<|im_start|>system<|im_sep|>' + message['content'] + '<|im_end|>'}}{% elif (message['role'] == 'user') %}{{'<|im_start|>user<|im_sep|>' + message['content'] + '<|im_end|>'}}{% elif (message['role'] == 'assistant') %}{{'<|im_start|>assistant<|im_sep|>' + message['content'] + '<|im_end|>'}}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant<|im_sep|>' }}{% endif %}",
775
  "clean_up_tokenization_spaces": false,
776
+ "eos_token": "<|im_end|>",
777
  "model_max_length": 16384,
778
+ "pad_token": "<|dummy_87|>",
779
  "tokenizer_class": "GPT2Tokenizer"
780
  }