Text Generation
Transformers
Safetensors
llama
text-generation-inference
Inference Endpoints
Daniel Steinigen commited on
Commit
8e19e38
·
1 Parent(s): e2576d2

update model weights

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "/home/Shared/align/models/teuken-instruct/7B_EU24_4T_EU24/train_run_24EU_teuken_EU24_lr_1e-5_2024_09_18_10_29_51/opengptx_checkpoints/checkpoint-1521",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
@@ -31,7 +31,7 @@
31
  "tie_word_embeddings": true,
32
  "tokenizer_class": "SPTokenizer",
33
  "torch_dtype": "bfloat16",
34
- "transformers_version": "4.43.2",
35
  "use_cache": false,
36
  "vocab_size": 250680
37
  }
 
1
  {
2
+ "_name_or_path": "/raid/s3/opengptx/models/7B_EU24_4T_fw_iter_0238500_honey_2024_08_14_ckp-1350/checkpoint-1350",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
 
31
  "tie_word_embeddings": true,
32
  "tokenizer_class": "SPTokenizer",
33
  "torch_dtype": "bfloat16",
34
+ "transformers_version": "4.40.0",
35
  "use_cache": false,
36
  "vocab_size": 250680
37
  }
generation_config.json CHANGED
@@ -3,6 +3,6 @@
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
  "pad_token_id": 3,
6
- "transformers_version": "4.43.2",
7
  "use_cache": false
8
  }
 
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
  "pad_token_id": 3,
6
+ "transformers_version": "4.40.0",
7
  "use_cache": false
8
  }
gptx_tokenizer.py CHANGED
@@ -17,7 +17,7 @@ PAD_TOKEN = "<pad>"
17
  BOS_TOKEN = "<s>"
18
  EOS_TOKEN = "</s>"
19
  UNK_TOKEN = "<unk>"
20
- REPO_ID = "openGPT-X/Teuken-7B-instruct-v0.4"
21
 
22
  class HFGPTXTokenizer(PreTrainedTokenizer):
23
  """
@@ -451,7 +451,7 @@ class SPTokenizer(HFGPTXTokenizer):
451
  "SL": "Pogovor med človekom in pomočnikom z umetno inteligenco. Pomočnik človeku prijazno in vljudno odgovarja na njegova vprašanja.", # noqa
452
  "SV": "En chatt mellan en människa och en assistent med artificiell intelligens. Assistenten ger hjälpsamma och artiga svar på människans frågor.", # noqa
453
  }
454
- chat_template = "{%- for message in messages %}\n{%- if (message['role']|lower == 'user') != (loop.index0 % 2 == 0) %}\n{{- raise_exception('Roles must alternate User/Assistant/User/Assistant/...') }}\n{%- endif %}\n{%-if message['role']|lower == 'user' %}\n{{- message['role']|capitalize + ': ' + message['content'] + '\\n' }}\n{%- elif message['role']|lower == 'assistant' %}\n{{- message['role']|capitalize + ': ' + message['content'] + eos_token + '\\n' }}\n{%- else %}\n{{- raise_exception('Only user and assistant roles are supported!') }}\n {%- endif %}\n{%- endfor %}{%-if add_generation_prompt %}\n{{- 'Assistant: '}}\n{%- endif %}\n"
455
  self.chat_template = {
456
  lang: f"System: {sys_msg}" + "{{- '\\n'}}\n" + chat_template
457
  for lang, sys_msg in self.system_messages_by_lang.items()
 
17
  BOS_TOKEN = "<s>"
18
  EOS_TOKEN = "</s>"
19
  UNK_TOKEN = "<unk>"
20
+ REPO_ID = "openGPT-X/Teuken-7B-instruct-research-v0.4"
21
 
22
  class HFGPTXTokenizer(PreTrainedTokenizer):
23
  """
 
451
  "SL": "Pogovor med človekom in pomočnikom z umetno inteligenco. Pomočnik človeku prijazno in vljudno odgovarja na njegova vprašanja.", # noqa
452
  "SV": "En chatt mellan en människa och en assistent med artificiell intelligens. Assistenten ger hjälpsamma och artiga svar på människans frågor.", # noqa
453
  }
454
+ chat_template = "{%- for message in messages %}\n{%- if (message['role']|lower == 'user') != (loop.index0 % 2 == 0) %}\n{{- raise_exception('Roles must alternate User/Assistant/User/Assistant/...') }}\n{%- endif %}\n{%-if message['role']|lower == 'user' %}\n{{- message['role']|capitalize + ': ' + message['content'] + '\\n' }}\n{%- elif message['role']|lower == 'assistant' %}\n{{- message['role']|capitalize + ': ' + message['content'] + eos_token + '\\n' }}\n{%- else %}\n{{- raise_exception('Only user and assistant roles are supported!') }}\n {%- endif %}\n{%- endfor %}{%-if add_generation_prompt %}\n{{- 'Assistant:<s> '}}\n{%- endif %}\n"
455
  self.chat_template = {
456
  lang: f"System: {sys_msg}" + "{{- '\\n'}}\n" + chat_template
457
  for lang, sys_msg in self.system_messages_by_lang.items()
model-00001-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:04b97eb2a8835844994809bcafc969a556205ddfa9246888ac39d9716924467b
3
  size 4936228560
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b422ed0f2f3a97c4035860044f4de5b3b4b37786b9ba5835f2dd7ee354cea03
3
  size 4936228560
model-00002-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b2bad10b52f1b636827366f60cdea9e982d2c16cfd60bfe7081e49740bc47942
3
  size 4929565048
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:462d1cb02ef01bb190125fc8882bc5e86437968cdb640befb3695be197cbd102
3
  size 4929565048
model-00003-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:48176bb53fb336cb352019363fc6fb7d39ab40769f62d339d8e080e7dbc74933
3
  size 4929565072
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:63cc533509aee497a75b359f2034b4ffa5eb20c0f2afe38792c2362cd1711319
3
  size 4929565072
model-00004-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e9507cf5b19bb2b46f4aa38e5e17cc6b99007d3900a63e5ec1d2de7c6d55576f
3
  size 110125512
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:542c2c24d51ffa9849ed9415e0bd78168ca5536d26ba9c8fa6911d07f8154011
3
  size 110125512