Echo9Zulu commited on
Commit
9c87f52
·
verified ·
1 Parent(s): 20ffa20

Upload 15 files

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ openvino_model.xml filter=lfs diff=lfs merge=lfs -text
added_tokens.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "<|end_of_role|>": 49153,
3
+ "<|start_of_role|>": 49152,
4
+ "<|tool_call|>": 49154
5
+ }
config.json ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "ibm-granite/granite-3.1-3b-a800m-instruct",
3
+ "architectures": [
4
+ "GraniteMoeForCausalLM"
5
+ ],
6
+ "attention_bias": false,
7
+ "attention_dropout": 0.0,
8
+ "attention_multiplier": 0.015625,
9
+ "bos_token_id": 0,
10
+ "embedding_multiplier": 12.0,
11
+ "eos_token_id": 0,
12
+ "hidden_act": "silu",
13
+ "hidden_size": 1536,
14
+ "initializer_range": 0.02,
15
+ "intermediate_size": 512,
16
+ "logits_scaling": 6.0,
17
+ "max_position_embeddings": 131072,
18
+ "model_type": "granitemoe",
19
+ "num_attention_heads": 24,
20
+ "num_experts_per_tok": 8,
21
+ "num_hidden_layers": 32,
22
+ "num_key_value_heads": 8,
23
+ "num_local_experts": 40,
24
+ "output_router_logits": false,
25
+ "pad_token_id": 0,
26
+ "residual_multiplier": 0.22,
27
+ "rms_norm_eps": 1e-06,
28
+ "rope_scaling": null,
29
+ "rope_theta": 10000000.0,
30
+ "router_aux_loss_coef": 0.0,
31
+ "tie_word_embeddings": true,
32
+ "torch_dtype": "bfloat16",
33
+ "transformers_version": "4.46.3",
34
+ "use_cache": true,
35
+ "vocab_size": 49155
36
+ }
generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 0,
4
+ "eos_token_id": 0,
5
+ "pad_token_id": 0,
6
+ "transformers_version": "4.46.3"
7
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
openvino_config.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "compression": null,
3
+ "dtype": "int4",
4
+ "input_info": null,
5
+ "optimum_version": "1.24.0.dev0",
6
+ "quantization_config": {
7
+ "all_layers": null,
8
+ "backup_precision": null,
9
+ "bits": 4,
10
+ "dataset": "wikitext2",
11
+ "gptq": null,
12
+ "group_size": 128,
13
+ "ignored_scope": null,
14
+ "lora_correction": null,
15
+ "num_samples": null,
16
+ "processor": null,
17
+ "quant_method": "awq",
18
+ "ratio": 1.0,
19
+ "scale_estimation": null,
20
+ "sensitivity_metric": null,
21
+ "sym": false,
22
+ "tokenizer": null,
23
+ "trust_remote_code": false,
24
+ "weight_format": "int4"
25
+ },
26
+ "save_onnx_model": false,
27
+ "transformers_version": "4.46.3"
28
+ }
openvino_detokenizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:32be371868f1de11802bd5bb4498575f206d26717077766543848451d8e28815
3
+ size 493005
openvino_detokenizer.xml ADDED
@@ -0,0 +1,264 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ <?xml version="1.0"?>
2
+ <net name="detokenizer" version="11">
3
+ <layers>
4
+ <layer id="0" name="Parameter_657689" type="Parameter" version="opset1">
5
+ <data shape="?,?" element_type="i64" />
6
+ <output>
7
+ <port id="0" precision="I64" names="Parameter_657689">
8
+ <dim>-1</dim>
9
+ <dim>-1</dim>
10
+ </port>
11
+ </output>
12
+ </layer>
13
+ <layer id="1" name="Convert_657706" type="Convert" version="opset1">
14
+ <data destination_type="i32" />
15
+ <input>
16
+ <port id="0" precision="I64">
17
+ <dim>-1</dim>
18
+ <dim>-1</dim>
19
+ </port>
20
+ </input>
21
+ <output>
22
+ <port id="1" precision="I32">
23
+ <dim>-1</dim>
24
+ <dim>-1</dim>
25
+ </port>
26
+ </output>
27
+ </layer>
28
+ <layer id="2" name="Constant_657664" type="Const" version="opset1">
29
+ <data element_type="u8" shape="492952" offset="0" size="492952" />
30
+ <output>
31
+ <port id="0" precision="U8">
32
+ <dim>492952</dim>
33
+ </port>
34
+ </output>
35
+ </layer>
36
+ <layer id="3" name="StringTensorUnpack_657665" type="StringTensorUnpack" version="extension">
37
+ <data mode="begins_ends" />
38
+ <input>
39
+ <port id="0" precision="U8">
40
+ <dim>492952</dim>
41
+ </port>
42
+ </input>
43
+ <output>
44
+ <port id="1" precision="I32">
45
+ <dim>-1</dim>
46
+ </port>
47
+ <port id="2" precision="I32">
48
+ <dim>-1</dim>
49
+ </port>
50
+ <port id="3" precision="U8">
51
+ <dim>-1</dim>
52
+ </port>
53
+ </output>
54
+ </layer>
55
+ <layer id="4" name="VocabDecoder_657690" type="VocabDecoder" version="extension">
56
+ <data skip_tokens="0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 49152, 49153, 49154" />
57
+ <input>
58
+ <port id="0" precision="I32">
59
+ <dim>-1</dim>
60
+ <dim>-1</dim>
61
+ </port>
62
+ <port id="1" precision="I32">
63
+ <dim>-1</dim>
64
+ </port>
65
+ <port id="2" precision="I32">
66
+ <dim>-1</dim>
67
+ </port>
68
+ <port id="3" precision="U8">
69
+ <dim>-1</dim>
70
+ </port>
71
+ </input>
72
+ <output>
73
+ <port id="4" precision="I32">
74
+ <dim>-1</dim>
75
+ </port>
76
+ <port id="5" precision="I32">
77
+ <dim>-1</dim>
78
+ </port>
79
+ <port id="6" precision="I32">
80
+ <dim>-1</dim>
81
+ </port>
82
+ <port id="7" precision="I32">
83
+ <dim>-1</dim>
84
+ </port>
85
+ <port id="8" precision="U8">
86
+ <dim>-1</dim>
87
+ </port>
88
+ </output>
89
+ </layer>
90
+ <layer id="5" name="FuzeRagged_657691" type="FuzeRagged" version="extension">
91
+ <input>
92
+ <port id="0" precision="I32">
93
+ <dim>-1</dim>
94
+ </port>
95
+ <port id="1" precision="I32">
96
+ <dim>-1</dim>
97
+ </port>
98
+ <port id="2" precision="I32">
99
+ <dim>-1</dim>
100
+ </port>
101
+ <port id="3" precision="I32">
102
+ <dim>-1</dim>
103
+ </port>
104
+ </input>
105
+ <output>
106
+ <port id="4" precision="I32">
107
+ <dim>-1</dim>
108
+ </port>
109
+ <port id="5" precision="I32">
110
+ <dim>-1</dim>
111
+ </port>
112
+ </output>
113
+ </layer>
114
+ <layer id="6" name="UTF8Validate_657692" type="UTF8Validate" version="extension">
115
+ <data replace_mode="true" />
116
+ <input>
117
+ <port id="0" precision="I32">
118
+ <dim>-1</dim>
119
+ </port>
120
+ <port id="1" precision="I32">
121
+ <dim>-1</dim>
122
+ </port>
123
+ <port id="2" precision="U8">
124
+ <dim>-1</dim>
125
+ </port>
126
+ </input>
127
+ <output>
128
+ <port id="3" precision="I32">
129
+ <dim>-1</dim>
130
+ </port>
131
+ <port id="4" precision="I32">
132
+ <dim>-1</dim>
133
+ </port>
134
+ <port id="5" precision="U8">
135
+ <dim>-1</dim>
136
+ </port>
137
+ </output>
138
+ </layer>
139
+ <layer id="7" name="Constant_657694" type="Const" version="opset1">
140
+ <data element_type="u8" shape="51" offset="492952" size="51" />
141
+ <output>
142
+ <port id="0" precision="U8">
143
+ <dim>51</dim>
144
+ </port>
145
+ </output>
146
+ </layer>
147
+ <layer id="8" name="Constant_657696" type="Const" version="opset1">
148
+ <data element_type="u8" shape="2" offset="493003" size="2" />
149
+ <output>
150
+ <port id="0" precision="U8">
151
+ <dim>2</dim>
152
+ </port>
153
+ </output>
154
+ </layer>
155
+ <layer id="9" name="RegexNormalization_657697" type="RegexNormalization" version="extension">
156
+ <data global_replace="true" />
157
+ <input>
158
+ <port id="0" precision="I32">
159
+ <dim>-1</dim>
160
+ </port>
161
+ <port id="1" precision="I32">
162
+ <dim>-1</dim>
163
+ </port>
164
+ <port id="2" precision="U8">
165
+ <dim>-1</dim>
166
+ </port>
167
+ <port id="3" precision="U8">
168
+ <dim>51</dim>
169
+ </port>
170
+ <port id="4" precision="U8">
171
+ <dim>2</dim>
172
+ </port>
173
+ </input>
174
+ <output>
175
+ <port id="5" precision="I32">
176
+ <dim>-1</dim>
177
+ </port>
178
+ <port id="6" precision="I32">
179
+ <dim>-1</dim>
180
+ </port>
181
+ <port id="7" precision="U8">
182
+ <dim>-1</dim>
183
+ </port>
184
+ </output>
185
+ </layer>
186
+ <layer id="10" name="StringTensorPack_657698" type="StringTensorPack" version="extension">
187
+ <data mode="begins_ends" />
188
+ <input>
189
+ <port id="0" precision="I32">
190
+ <dim>-1</dim>
191
+ </port>
192
+ <port id="1" precision="I32">
193
+ <dim>-1</dim>
194
+ </port>
195
+ <port id="2" precision="U8">
196
+ <dim>-1</dim>
197
+ </port>
198
+ </input>
199
+ <output>
200
+ <port id="3" precision="STRING" names="string_output">
201
+ <dim>-1</dim>
202
+ </port>
203
+ </output>
204
+ </layer>
205
+ <layer id="11" name="Result_657699" type="Result" version="opset1">
206
+ <input>
207
+ <port id="0" precision="STRING">
208
+ <dim>-1</dim>
209
+ </port>
210
+ </input>
211
+ </layer>
212
+ </layers>
213
+ <edges>
214
+ <edge from-layer="0" from-port="0" to-layer="1" to-port="0" />
215
+ <edge from-layer="1" from-port="1" to-layer="4" to-port="0" />
216
+ <edge from-layer="2" from-port="0" to-layer="3" to-port="0" />
217
+ <edge from-layer="3" from-port="1" to-layer="4" to-port="1" />
218
+ <edge from-layer="3" from-port="2" to-layer="4" to-port="2" />
219
+ <edge from-layer="3" from-port="3" to-layer="4" to-port="3" />
220
+ <edge from-layer="4" from-port="4" to-layer="5" to-port="0" />
221
+ <edge from-layer="4" from-port="5" to-layer="5" to-port="1" />
222
+ <edge from-layer="4" from-port="6" to-layer="5" to-port="2" />
223
+ <edge from-layer="4" from-port="7" to-layer="5" to-port="3" />
224
+ <edge from-layer="4" from-port="8" to-layer="6" to-port="2" />
225
+ <edge from-layer="5" from-port="5" to-layer="6" to-port="1" />
226
+ <edge from-layer="5" from-port="4" to-layer="6" to-port="0" />
227
+ <edge from-layer="6" from-port="3" to-layer="9" to-port="0" />
228
+ <edge from-layer="6" from-port="4" to-layer="9" to-port="1" />
229
+ <edge from-layer="6" from-port="5" to-layer="9" to-port="2" />
230
+ <edge from-layer="7" from-port="0" to-layer="9" to-port="3" />
231
+ <edge from-layer="8" from-port="0" to-layer="9" to-port="4" />
232
+ <edge from-layer="9" from-port="5" to-layer="10" to-port="0" />
233
+ <edge from-layer="9" from-port="6" to-layer="10" to-port="1" />
234
+ <edge from-layer="9" from-port="7" to-layer="10" to-port="2" />
235
+ <edge from-layer="10" from-port="3" to-layer="11" to-port="0" />
236
+ </edges>
237
+ <rt_info>
238
+ <add_attention_mask value="True" />
239
+ <add_prefix_space />
240
+ <add_special_tokens value="True" />
241
+ <bos_token_id value="0" />
242
+ <chat_template value="{%- if messages[0]['role'] == 'system' %}&#10; {%- set system_message = messages[0]['content'] %}&#10; {%- set loop_messages = messages[1:] %}&#10;{%- else %}&#10; {%- set system_message = &quot;Knowledge Cutoff Date: April 2024.&#10;Today's Date: &quot; + strftime_now('%B %d, %Y') + &quot;.&#10;You are Granite, developed by IBM.&quot; %}&#10; {%- if tools and documents %}&#10; {%- set system_message = system_message + &quot; You are a helpful AI assistant with access to the following tools. When a tool is required to answer the user's query, respond with &lt;|tool_call|> followed by a JSON list of tools used. If a tool does not exist in the provided list of tools, notify the user that you do not have the ability to fulfill the request.&#10;&#10;Write the response to the user's input by strictly aligning with the facts in the provided documents. If the information needed to answer the question is not available in the documents, inform the user that the question cannot be answered based on the available data.&quot; %}&#10; {%- elif tools %}&#10; {%- set system_message = system_message + &quot; You are a helpful AI assistant with access to the following tools. When a tool is required to answer the user's query, respond with &lt;|tool_call|> followed by a JSON list of tools used. If a tool does not exist in the provided list of tools, notify the user that you do not have the ability to fulfill the request.&quot; %}&#10; {%- elif documents %}&#10; {%- set system_message = system_message + &quot; Write the response to the user's input by strictly aligning with the facts in the provided documents. If the information needed to answer the question is not available in the documents, inform the user that the question cannot be answered based on the available data.&quot; %}&#10; {%- else %}&#10; {%- set system_message = system_message + &quot; You are a helpful AI assistant.&quot; %} &#10; {%- endif %}&#10; {%- if 'citations' in controls and documents %}&#10; {%- set system_message = system_message + '&#10;&#10;In your response, use the symbols &lt;co> and &lt;/co> to indicate when a fact comes from a document in the search result, e.g &lt;co>0&lt;/co> for a fact from document 0. Afterwards, list all the citations with their corresponding documents in an ordered list.' %}&#10; {%- endif %}&#10; {%- if 'hallucinations' in controls and documents %}&#10; {%- set system_message = system_message + '&#10;&#10;Finally, after the response is written, include a numbered list of sentences from the response that are potentially hallucinated and not based in the documents.' %}&#10; {%- endif %}&#10; {%- set loop_messages = messages %}&#10;{%- endif %}&#10;{{- '&lt;|start_of_role|>system&lt;|end_of_role|>' + system_message + '&lt;|end_of_text|>&#10;' }}&#10;{%- if tools %}&#10; {{- '&lt;|start_of_role|>tools&lt;|end_of_role|>' }}&#10; {{- tools | tojson(indent=4) }}&#10; {{- '&lt;|end_of_text|>&#10;' }}&#10;{%- endif %}&#10;{%- if documents %}&#10; {{- '&lt;|start_of_role|>documents&lt;|end_of_role|>' }}&#10; {%- for document in documents %}&#10; {{- 'Document ' + loop.index0 | string + '&#10;' }}&#10; {{- document['text'] }}&#10; {%- if not loop.last %}&#10; {{- '&#10;&#10;'}}&#10; {%- endif%}&#10; {%- endfor %}&#10; {{- '&lt;|end_of_text|>&#10;' }}&#10;{%- endif %}&#10;{%- for message in loop_messages %}&#10; {{- '&lt;|start_of_role|>' + message['role'] + '&lt;|end_of_role|>' + message['content'] + '&lt;|end_of_text|>&#10;' }}&#10; {%- if loop.last and add_generation_prompt %}&#10; {{- '&lt;|start_of_role|>assistant' }}&#10; {%- if controls %}&#10; {{- ' ' + controls | tojson()}}&#10; {%- endif %}&#10; {{- '&lt;|end_of_role|>' }}&#10; {%- endif %}&#10;{%- endfor %}" />
243
+ <clean_up_tokenization_spaces />
244
+ <detokenizer_input_type value="i64" />
245
+ <eos_token_id value="0" />
246
+ <handle_special_tokens_with_re />
247
+ <number_of_inputs value="1" />
248
+ <openvino_tokenizers_version value="2024.6.0.0" />
249
+ <openvino_version value="2024.6.0" />
250
+ <original_tokenizer_class value="&lt;class 'transformers.models.gpt2.tokenization_gpt2_fast.GPT2TokenizerFast'>" />
251
+ <pad_token_id value="0" />
252
+ <sentencepiece_version value="0.2.0" />
253
+ <skip_special_tokens value="True" />
254
+ <streaming_detokenizer value="False" />
255
+ <tiktoken_version value="0.7.0" />
256
+ <tokenizer_output_type value="i64" />
257
+ <tokenizers_version value="0.20.3" />
258
+ <transformers_version value="4.46.3" />
259
+ <use_max_padding value="False" />
260
+ <use_sentencepiece_backend value="False" />
261
+ <utf8_replace_mode value="replace" />
262
+ <with_detokenizer value="True" />
263
+ </rt_info>
264
+ </net>
openvino_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d3c0d8bdf2f44447093f106e08e9cdf01b23572b162e2f1f2b91234be7df7eb
3
+ size 1862370129
openvino_model.xml ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:795b7da2efdf9f8229ea592f8719488f333a027a65f58cae517d3236ce420d28
3
+ size 23233881
openvino_tokenizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:98d3a1e8b53eb01ff74fb2dfe6f07a3db20b6f0517b983df8a23380769f79f4c
3
+ size 1180835
openvino_tokenizer.xml ADDED
@@ -0,0 +1,702 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ <?xml version="1.0"?>
2
+ <net name="tokenizer" version="11">
3
+ <layers>
4
+ <layer id="0" name="Parameter_657583" type="Parameter" version="opset1">
5
+ <data shape="?" element_type="string" />
6
+ <output>
7
+ <port id="0" precision="STRING" names="Parameter_657583">
8
+ <dim>-1</dim>
9
+ </port>
10
+ </output>
11
+ </layer>
12
+ <layer id="1" name="Constant_657589" type="Const" version="opset1">
13
+ <data element_type="i64" shape="" offset="0" size="8" />
14
+ <output>
15
+ <port id="0" precision="I64" />
16
+ </output>
17
+ </layer>
18
+ <layer id="2" name="StringTensorUnpack_657584" type="StringTensorUnpack" version="extension">
19
+ <data mode="begins_ends" />
20
+ <input>
21
+ <port id="0" precision="STRING">
22
+ <dim>-1</dim>
23
+ </port>
24
+ </input>
25
+ <output>
26
+ <port id="1" precision="I32">
27
+ <dim>-1</dim>
28
+ </port>
29
+ <port id="2" precision="I32">
30
+ <dim>-1</dim>
31
+ </port>
32
+ <port id="3" precision="U8">
33
+ <dim>-1</dim>
34
+ </port>
35
+ </output>
36
+ </layer>
37
+ <layer id="3" name="ShapeOf_657585" type="ShapeOf" version="opset3">
38
+ <data output_type="i64" />
39
+ <input>
40
+ <port id="0" precision="I32">
41
+ <dim>-1</dim>
42
+ </port>
43
+ </input>
44
+ <output>
45
+ <port id="1" precision="I64">
46
+ <dim>1</dim>
47
+ </port>
48
+ </output>
49
+ </layer>
50
+ <layer id="4" name="Constant_657586" type="Const" version="opset1">
51
+ <data element_type="i64" shape="" offset="0" size="8" />
52
+ <output>
53
+ <port id="0" precision="I64" />
54
+ </output>
55
+ </layer>
56
+ <layer id="5" name="Constant_657587" type="Const" version="opset1">
57
+ <data element_type="i64" shape="" offset="0" size="8" />
58
+ <output>
59
+ <port id="0" precision="I64" />
60
+ </output>
61
+ </layer>
62
+ <layer id="6" name="Gather_657588" type="Gather" version="opset8">
63
+ <data batch_dims="0" />
64
+ <input>
65
+ <port id="0" precision="I64">
66
+ <dim>1</dim>
67
+ </port>
68
+ <port id="1" precision="I64" />
69
+ <port id="2" precision="I64" />
70
+ </input>
71
+ <output>
72
+ <port id="3" precision="I64" />
73
+ </output>
74
+ </layer>
75
+ <layer id="7" name="Constant_657590" type="Const" version="opset1">
76
+ <data element_type="i64" shape="" offset="8" size="8" />
77
+ <output>
78
+ <port id="0" precision="I64" />
79
+ </output>
80
+ </layer>
81
+ <layer id="8" name="Range_657591" type="Range" version="opset4">
82
+ <data output_type="i32" />
83
+ <input>
84
+ <port id="0" precision="I64" />
85
+ <port id="1" precision="I64" />
86
+ <port id="2" precision="I64" />
87
+ </input>
88
+ <output>
89
+ <port id="3" precision="I32">
90
+ <dim>-1</dim>
91
+ </port>
92
+ </output>
93
+ </layer>
94
+ <layer id="9" name="Constant_657592" type="Const" version="opset1">
95
+ <data element_type="i64" shape="" offset="8" size="8" />
96
+ <output>
97
+ <port id="0" precision="I64" />
98
+ </output>
99
+ </layer>
100
+ <layer id="10" name="Constant_657593" type="Const" version="opset1">
101
+ <data element_type="i64" shape="" offset="8" size="8" />
102
+ <output>
103
+ <port id="0" precision="I64" />
104
+ </output>
105
+ </layer>
106
+ <layer id="11" name="Add_657594" type="Add" version="opset1">
107
+ <data auto_broadcast="numpy" />
108
+ <input>
109
+ <port id="0" precision="I64" />
110
+ <port id="1" precision="I64" />
111
+ </input>
112
+ <output>
113
+ <port id="2" precision="I64" />
114
+ </output>
115
+ </layer>
116
+ <layer id="12" name="Constant_657595" type="Const" version="opset1">
117
+ <data element_type="i64" shape="" offset="8" size="8" />
118
+ <output>
119
+ <port id="0" precision="I64" />
120
+ </output>
121
+ </layer>
122
+ <layer id="13" name="Range_657596" type="Range" version="opset4">
123
+ <data output_type="i32" />
124
+ <input>
125
+ <port id="0" precision="I64" />
126
+ <port id="1" precision="I64" />
127
+ <port id="2" precision="I64" />
128
+ </input>
129
+ <output>
130
+ <port id="3" precision="I32">
131
+ <dim>-1</dim>
132
+ </port>
133
+ </output>
134
+ </layer>
135
+ <layer id="14" name="Constant_657658" type="Const" version="opset1">
136
+ <data element_type="u8" shape="408" offset="16" size="408" />
137
+ <output>
138
+ <port id="0" precision="U8">
139
+ <dim>408</dim>
140
+ </port>
141
+ </output>
142
+ </layer>
143
+ <layer id="15" name="SpecialTokensSplit_657659" type="SpecialTokensSplit" version="extension">
144
+ <input>
145
+ <port id="0" precision="I32">
146
+ <dim>-1</dim>
147
+ </port>
148
+ <port id="1" precision="I32">
149
+ <dim>-1</dim>
150
+ </port>
151
+ <port id="2" precision="I32">
152
+ <dim>-1</dim>
153
+ </port>
154
+ <port id="3" precision="I32">
155
+ <dim>-1</dim>
156
+ </port>
157
+ <port id="4" precision="U8">
158
+ <dim>-1</dim>
159
+ </port>
160
+ <port id="5" precision="U8">
161
+ <dim>408</dim>
162
+ </port>
163
+ </input>
164
+ <output>
165
+ <port id="6" precision="I32">
166
+ <dim>-1</dim>
167
+ </port>
168
+ <port id="7" precision="I32">
169
+ <dim>-1</dim>
170
+ </port>
171
+ <port id="8" precision="I32">
172
+ <dim>-1</dim>
173
+ </port>
174
+ <port id="9" precision="I32">
175
+ <dim>-1</dim>
176
+ </port>
177
+ <port id="10" precision="U8">
178
+ <dim>-1</dim>
179
+ </port>
180
+ <port id="11" precision="BOOL">
181
+ <dim>-1</dim>
182
+ </port>
183
+ </output>
184
+ </layer>
185
+ <layer id="16" name="Constant_657661" type="Const" version="opset1">
186
+ <data element_type="u8" shape="64" offset="424" size="64" />
187
+ <output>
188
+ <port id="0" precision="U8">
189
+ <dim>64</dim>
190
+ </port>
191
+ </output>
192
+ </layer>
193
+ <layer id="17" name="RegexSplit_657662" type="RegexSplit" version="extension">
194
+ <data behaviour="isolate" invert="false" max_splits="-1" />
195
+ <input>
196
+ <port id="0" precision="I32">
197
+ <dim>-1</dim>
198
+ </port>
199
+ <port id="1" precision="I32">
200
+ <dim>-1</dim>
201
+ </port>
202
+ <port id="2" precision="I32">
203
+ <dim>-1</dim>
204
+ </port>
205
+ <port id="3" precision="I32">
206
+ <dim>-1</dim>
207
+ </port>
208
+ <port id="4" precision="U8">
209
+ <dim>-1</dim>
210
+ </port>
211
+ <port id="5" precision="BOOL">
212
+ <dim>-1</dim>
213
+ </port>
214
+ <port id="6" precision="U8">
215
+ <dim>64</dim>
216
+ </port>
217
+ </input>
218
+ <output>
219
+ <port id="7" precision="I32">
220
+ <dim>-1</dim>
221
+ </port>
222
+ <port id="8" precision="I32">
223
+ <dim>-1</dim>
224
+ </port>
225
+ <port id="9" precision="I32">
226
+ <dim>-1</dim>
227
+ </port>
228
+ <port id="10" precision="I32">
229
+ <dim>-1</dim>
230
+ </port>
231
+ <port id="11" precision="U8">
232
+ <dim>-1</dim>
233
+ </port>
234
+ <port id="12" precision="BOOL">
235
+ <dim>-1</dim>
236
+ </port>
237
+ </output>
238
+ </layer>
239
+ <layer id="18" name="Constant_657664" type="Const" version="opset1">
240
+ <data element_type="u8" shape="492952" offset="488" size="492952" />
241
+ <output>
242
+ <port id="0" precision="U8">
243
+ <dim>492952</dim>
244
+ </port>
245
+ </output>
246
+ </layer>
247
+ <layer id="19" name="StringTensorUnpack_657665" type="StringTensorUnpack" version="extension">
248
+ <data mode="begins_ends" />
249
+ <input>
250
+ <port id="0" precision="U8">
251
+ <dim>492952</dim>
252
+ </port>
253
+ </input>
254
+ <output>
255
+ <port id="1" precision="I32">
256
+ <dim>-1</dim>
257
+ </port>
258
+ <port id="2" precision="I32">
259
+ <dim>-1</dim>
260
+ </port>
261
+ <port id="3" precision="U8">
262
+ <dim>-1</dim>
263
+ </port>
264
+ </output>
265
+ </layer>
266
+ <layer id="20" name="Constant_657670" type="Const" version="opset1">
267
+ <data element_type="u8" shape="340606" offset="493440" size="340606" />
268
+ <output>
269
+ <port id="0" precision="U8">
270
+ <dim>340606</dim>
271
+ </port>
272
+ </output>
273
+ </layer>
274
+ <layer id="21" name="StringTensorUnpack_657671" type="StringTensorUnpack" version="extension">
275
+ <data mode="begins_ends" />
276
+ <input>
277
+ <port id="0" precision="U8">
278
+ <dim>340606</dim>
279
+ </port>
280
+ </input>
281
+ <output>
282
+ <port id="1" precision="I32">
283
+ <dim>-1</dim>
284
+ </port>
285
+ <port id="2" precision="I32">
286
+ <dim>-1</dim>
287
+ </port>
288
+ <port id="3" precision="U8">
289
+ <dim>-1</dim>
290
+ </port>
291
+ </output>
292
+ </layer>
293
+ <layer id="22" name="Constant_657673" type="Const" version="opset1">
294
+ <data element_type="u8" shape="346329" offset="834046" size="346329" />
295
+ <output>
296
+ <port id="0" precision="U8">
297
+ <dim>346329</dim>
298
+ </port>
299
+ </output>
300
+ </layer>
301
+ <layer id="23" name="StringTensorUnpack_657674" type="StringTensorUnpack" version="extension">
302
+ <data mode="begins_ends" />
303
+ <input>
304
+ <port id="0" precision="U8">
305
+ <dim>346329</dim>
306
+ </port>
307
+ </input>
308
+ <output>
309
+ <port id="1" precision="I32">
310
+ <dim>-1</dim>
311
+ </port>
312
+ <port id="2" precision="I32">
313
+ <dim>-1</dim>
314
+ </port>
315
+ <port id="3" precision="U8">
316
+ <dim>-1</dim>
317
+ </port>
318
+ </output>
319
+ </layer>
320
+ <layer id="24" name="Constant_657667" type="Const" version="opset1">
321
+ <data element_type="u8" shape="368" offset="1180375" size="368" />
322
+ <output>
323
+ <port id="0" precision="U8">
324
+ <dim>368</dim>
325
+ </port>
326
+ </output>
327
+ </layer>
328
+ <layer id="25" name="StringTensorUnpack_657668" type="StringTensorUnpack" version="extension">
329
+ <data mode="begins_ends" />
330
+ <input>
331
+ <port id="0" precision="U8">
332
+ <dim>368</dim>
333
+ </port>
334
+ </input>
335
+ <output>
336
+ <port id="1" precision="I32">
337
+ <dim>-1</dim>
338
+ </port>
339
+ <port id="2" precision="I32">
340
+ <dim>-1</dim>
341
+ </port>
342
+ <port id="3" precision="U8">
343
+ <dim>-1</dim>
344
+ </port>
345
+ </output>
346
+ </layer>
347
+ <layer id="26" name="Constant_657675" type="Const" version="opset1">
348
+ <data element_type="i32" shape="21" offset="1180743" size="84" />
349
+ <output>
350
+ <port id="0" precision="I32">
351
+ <dim>21</dim>
352
+ </port>
353
+ </output>
354
+ </layer>
355
+ <layer id="27" name="BPETokenizer_657676" type="BPETokenizer" version="extension">
356
+ <data unk_token="" fuse_unk="false" suffix_indicator="" end_suffix="" byte_fallback="false" cache_capacity="20000" />
357
+ <input>
358
+ <port id="0" precision="I32">
359
+ <dim>-1</dim>
360
+ </port>
361
+ <port id="1" precision="I32">
362
+ <dim>-1</dim>
363
+ </port>
364
+ <port id="2" precision="I32">
365
+ <dim>-1</dim>
366
+ </port>
367
+ <port id="3" precision="I32">
368
+ <dim>-1</dim>
369
+ </port>
370
+ <port id="4" precision="U8">
371
+ <dim>-1</dim>
372
+ </port>
373
+ <port id="5" precision="I32">
374
+ <dim>-1</dim>
375
+ </port>
376
+ <port id="6" precision="I32">
377
+ <dim>-1</dim>
378
+ </port>
379
+ <port id="7" precision="U8">
380
+ <dim>-1</dim>
381
+ </port>
382
+ <port id="8" precision="I32">
383
+ <dim>-1</dim>
384
+ </port>
385
+ <port id="9" precision="I32">
386
+ <dim>-1</dim>
387
+ </port>
388
+ <port id="10" precision="U8">
389
+ <dim>-1</dim>
390
+ </port>
391
+ <port id="11" precision="I32">
392
+ <dim>-1</dim>
393
+ </port>
394
+ <port id="12" precision="I32">
395
+ <dim>-1</dim>
396
+ </port>
397
+ <port id="13" precision="U8">
398
+ <dim>-1</dim>
399
+ </port>
400
+ <port id="14" precision="I32">
401
+ <dim>-1</dim>
402
+ </port>
403
+ <port id="15" precision="I32">
404
+ <dim>-1</dim>
405
+ </port>
406
+ <port id="16" precision="U8">
407
+ <dim>-1</dim>
408
+ </port>
409
+ <port id="17" precision="I32">
410
+ <dim>21</dim>
411
+ </port>
412
+ </input>
413
+ <output>
414
+ <port id="18" precision="I32">
415
+ <dim>-1</dim>
416
+ </port>
417
+ <port id="19" precision="I32">
418
+ <dim>-1</dim>
419
+ </port>
420
+ <port id="20" precision="I32">
421
+ <dim>-1</dim>
422
+ </port>
423
+ </output>
424
+ </layer>
425
+ <layer id="28" name="Subtract_657677" type="Subtract" version="opset1">
426
+ <data auto_broadcast="numpy" />
427
+ <input>
428
+ <port id="0" precision="I32">
429
+ <dim>-1</dim>
430
+ </port>
431
+ <port id="1" precision="I32">
432
+ <dim>-1</dim>
433
+ </port>
434
+ </input>
435
+ <output>
436
+ <port id="2" precision="I32">
437
+ <dim>-1</dim>
438
+ </port>
439
+ </output>
440
+ </layer>
441
+ <layer id="29" name="Constant_657678" type="Const" version="opset1">
442
+ <data element_type="i32" shape="" offset="1180827" size="4" />
443
+ <output>
444
+ <port id="0" precision="I32" />
445
+ </output>
446
+ </layer>
447
+ <layer id="30" name="Minimum_657679" type="Minimum" version="opset1">
448
+ <data auto_broadcast="numpy" />
449
+ <input>
450
+ <port id="0" precision="I32">
451
+ <dim>-1</dim>
452
+ </port>
453
+ <port id="1" precision="I32" />
454
+ </input>
455
+ <output>
456
+ <port id="2" precision="I32">
457
+ <dim>-1</dim>
458
+ </port>
459
+ </output>
460
+ </layer>
461
+ <layer id="31" name="Subtract_657680" type="Subtract" version="opset1">
462
+ <data auto_broadcast="numpy" />
463
+ <input>
464
+ <port id="0" precision="I32">
465
+ <dim>-1</dim>
466
+ </port>
467
+ <port id="1" precision="I32">
468
+ <dim>-1</dim>
469
+ </port>
470
+ </input>
471
+ <output>
472
+ <port id="2" precision="I32">
473
+ <dim>-1</dim>
474
+ </port>
475
+ </output>
476
+ </layer>
477
+ <layer id="32" name="Subtract_657681" type="Subtract" version="opset1">
478
+ <data auto_broadcast="numpy" />
479
+ <input>
480
+ <port id="0" precision="I32">
481
+ <dim>-1</dim>
482
+ </port>
483
+ <port id="1" precision="I32">
484
+ <dim>-1</dim>
485
+ </port>
486
+ </input>
487
+ <output>
488
+ <port id="2" precision="I32">
489
+ <dim>-1</dim>
490
+ </port>
491
+ </output>
492
+ </layer>
493
+ <layer id="33" name="Constant_657682" type="Const" version="opset1">
494
+ <data element_type="i32" shape="" offset="1180831" size="4" />
495
+ <output>
496
+ <port id="0" precision="I32" />
497
+ </output>
498
+ </layer>
499
+ <layer id="34" name="ReduceMax_657683" type="ReduceMax" version="opset1">
500
+ <data keep_dims="false" />
501
+ <input>
502
+ <port id="0" precision="I32">
503
+ <dim>-1</dim>
504
+ </port>
505
+ <port id="1" precision="I32" />
506
+ </input>
507
+ <output>
508
+ <port id="2" precision="I32" />
509
+ </output>
510
+ </layer>
511
+ <layer id="35" name="Constant_657684" type="Const" version="opset1">
512
+ <data element_type="i32" shape="" offset="1180831" size="4" />
513
+ <output>
514
+ <port id="0" precision="I32" />
515
+ </output>
516
+ </layer>
517
+ <layer id="36" name="RaggedToDense_657685" type="RaggedToDense" version="extension">
518
+ <data pad_right="false" />
519
+ <input>
520
+ <port id="0" precision="I32">
521
+ <dim>-1</dim>
522
+ </port>
523
+ <port id="1" precision="I32">
524
+ <dim>-1</dim>
525
+ </port>
526
+ <port id="2" precision="I32">
527
+ <dim>-1</dim>
528
+ </port>
529
+ <port id="3" precision="I32" />
530
+ <port id="4" precision="I32" />
531
+ </input>
532
+ <output>
533
+ <port id="5" precision="I32">
534
+ <dim>-1</dim>
535
+ <dim>-1</dim>
536
+ </port>
537
+ <port id="6" precision="BOOL">
538
+ <dim>-1</dim>
539
+ <dim>-1</dim>
540
+ </port>
541
+ </output>
542
+ </layer>
543
+ <layer id="37" name="Convert_657686" type="Convert" version="opset1">
544
+ <data destination_type="i32" />
545
+ <input>
546
+ <port id="0" precision="BOOL">
547
+ <dim>-1</dim>
548
+ <dim>-1</dim>
549
+ </port>
550
+ </input>
551
+ <output>
552
+ <port id="1" precision="I32">
553
+ <dim>-1</dim>
554
+ <dim>-1</dim>
555
+ </port>
556
+ </output>
557
+ </layer>
558
+ <layer id="38" name="Convert_657686" type="Convert" version="opset1">
559
+ <data destination_type="i64" />
560
+ <input>
561
+ <port id="0" precision="I32">
562
+ <dim>-1</dim>
563
+ <dim>-1</dim>
564
+ </port>
565
+ </input>
566
+ <output>
567
+ <port id="1" precision="I64" names="attention_mask">
568
+ <dim>-1</dim>
569
+ <dim>-1</dim>
570
+ </port>
571
+ </output>
572
+ </layer>
573
+ <layer id="40" name="RaggedToDense_657685.0" type="Convert" version="opset1">
574
+ <data destination_type="i64" />
575
+ <input>
576
+ <port id="0" precision="I32">
577
+ <dim>-1</dim>
578
+ <dim>-1</dim>
579
+ </port>
580
+ </input>
581
+ <output>
582
+ <port id="1" precision="I64" names="input_ids">
583
+ <dim>-1</dim>
584
+ <dim>-1</dim>
585
+ </port>
586
+ </output>
587
+ </layer>
588
+ <layer id="41" name="Result_657687" type="Result" version="opset1">
589
+ <input>
590
+ <port id="0" precision="I64">
591
+ <dim>-1</dim>
592
+ <dim>-1</dim>
593
+ </port>
594
+ </input>
595
+ </layer>
596
+ <layer id="39" name="Result_657688" type="Result" version="opset1">
597
+ <input>
598
+ <port id="0" precision="I64">
599
+ <dim>-1</dim>
600
+ <dim>-1</dim>
601
+ </port>
602
+ </input>
603
+ </layer>
604
+ </layers>
605
+ <edges>
606
+ <edge from-layer="0" from-port="0" to-layer="2" to-port="0" />
607
+ <edge from-layer="1" from-port="0" to-layer="8" to-port="0" />
608
+ <edge from-layer="2" from-port="1" to-layer="3" to-port="0" />
609
+ <edge from-layer="2" from-port="3" to-layer="15" to-port="4" />
610
+ <edge from-layer="2" from-port="2" to-layer="15" to-port="3" />
611
+ <edge from-layer="2" from-port="1" to-layer="15" to-port="2" />
612
+ <edge from-layer="3" from-port="1" to-layer="6" to-port="0" />
613
+ <edge from-layer="4" from-port="0" to-layer="6" to-port="1" />
614
+ <edge from-layer="5" from-port="0" to-layer="6" to-port="2" />
615
+ <edge from-layer="6" from-port="3" to-layer="11" to-port="0" />
616
+ <edge from-layer="6" from-port="3" to-layer="8" to-port="1" />
617
+ <edge from-layer="7" from-port="0" to-layer="8" to-port="2" />
618
+ <edge from-layer="8" from-port="3" to-layer="15" to-port="0" />
619
+ <edge from-layer="9" from-port="0" to-layer="13" to-port="0" />
620
+ <edge from-layer="10" from-port="0" to-layer="11" to-port="1" />
621
+ <edge from-layer="11" from-port="2" to-layer="13" to-port="1" />
622
+ <edge from-layer="12" from-port="0" to-layer="13" to-port="2" />
623
+ <edge from-layer="13" from-port="3" to-layer="15" to-port="1" />
624
+ <edge from-layer="14" from-port="0" to-layer="15" to-port="5" />
625
+ <edge from-layer="15" from-port="6" to-layer="17" to-port="0" />
626
+ <edge from-layer="15" from-port="7" to-layer="17" to-port="1" />
627
+ <edge from-layer="15" from-port="8" to-layer="17" to-port="2" />
628
+ <edge from-layer="15" from-port="9" to-layer="17" to-port="3" />
629
+ <edge from-layer="15" from-port="10" to-layer="17" to-port="4" />
630
+ <edge from-layer="15" from-port="11" to-layer="17" to-port="5" />
631
+ <edge from-layer="16" from-port="0" to-layer="17" to-port="6" />
632
+ <edge from-layer="17" from-port="11" to-layer="27" to-port="4" />
633
+ <edge from-layer="17" from-port="10" to-layer="27" to-port="3" />
634
+ <edge from-layer="17" from-port="9" to-layer="27" to-port="2" />
635
+ <edge from-layer="17" from-port="8" to-layer="27" to-port="1" />
636
+ <edge from-layer="17" from-port="7" to-layer="27" to-port="0" />
637
+ <edge from-layer="18" from-port="0" to-layer="19" to-port="0" />
638
+ <edge from-layer="19" from-port="1" to-layer="27" to-port="5" />
639
+ <edge from-layer="19" from-port="2" to-layer="27" to-port="6" />
640
+ <edge from-layer="19" from-port="3" to-layer="27" to-port="7" />
641
+ <edge from-layer="20" from-port="0" to-layer="21" to-port="0" />
642
+ <edge from-layer="21" from-port="1" to-layer="27" to-port="8" />
643
+ <edge from-layer="21" from-port="2" to-layer="27" to-port="9" />
644
+ <edge from-layer="21" from-port="3" to-layer="27" to-port="10" />
645
+ <edge from-layer="22" from-port="0" to-layer="23" to-port="0" />
646
+ <edge from-layer="23" from-port="1" to-layer="27" to-port="11" />
647
+ <edge from-layer="23" from-port="2" to-layer="27" to-port="12" />
648
+ <edge from-layer="23" from-port="3" to-layer="27" to-port="13" />
649
+ <edge from-layer="24" from-port="0" to-layer="25" to-port="0" />
650
+ <edge from-layer="25" from-port="1" to-layer="27" to-port="14" />
651
+ <edge from-layer="25" from-port="2" to-layer="27" to-port="15" />
652
+ <edge from-layer="25" from-port="3" to-layer="27" to-port="16" />
653
+ <edge from-layer="26" from-port="0" to-layer="27" to-port="17" />
654
+ <edge from-layer="27" from-port="20" to-layer="36" to-port="2" />
655
+ <edge from-layer="27" from-port="18" to-layer="28" to-port="1" />
656
+ <edge from-layer="27" from-port="19" to-layer="36" to-port="1" />
657
+ <edge from-layer="27" from-port="19" to-layer="32" to-port="0" />
658
+ <edge from-layer="27" from-port="19" to-layer="31" to-port="0" />
659
+ <edge from-layer="27" from-port="19" to-layer="28" to-port="0" />
660
+ <edge from-layer="28" from-port="2" to-layer="30" to-port="0" />
661
+ <edge from-layer="29" from-port="0" to-layer="30" to-port="1" />
662
+ <edge from-layer="30" from-port="2" to-layer="31" to-port="1" />
663
+ <edge from-layer="31" from-port="2" to-layer="32" to-port="1" />
664
+ <edge from-layer="31" from-port="2" to-layer="36" to-port="0" />
665
+ <edge from-layer="32" from-port="2" to-layer="34" to-port="0" />
666
+ <edge from-layer="33" from-port="0" to-layer="34" to-port="1" />
667
+ <edge from-layer="34" from-port="2" to-layer="36" to-port="3" />
668
+ <edge from-layer="35" from-port="0" to-layer="36" to-port="4" />
669
+ <edge from-layer="36" from-port="6" to-layer="37" to-port="0" />
670
+ <edge from-layer="36" from-port="5" to-layer="40" to-port="0" />
671
+ <edge from-layer="37" from-port="1" to-layer="38" to-port="0" />
672
+ <edge from-layer="38" from-port="1" to-layer="39" to-port="0" />
673
+ <edge from-layer="40" from-port="1" to-layer="41" to-port="0" />
674
+ </edges>
675
+ <rt_info>
676
+ <add_attention_mask value="True" />
677
+ <add_prefix_space />
678
+ <add_special_tokens value="True" />
679
+ <bos_token_id value="0" />
680
+ <chat_template value="{%- if messages[0]['role'] == 'system' %}&#10; {%- set system_message = messages[0]['content'] %}&#10; {%- set loop_messages = messages[1:] %}&#10;{%- else %}&#10; {%- set system_message = &quot;Knowledge Cutoff Date: April 2024.&#10;Today's Date: &quot; + strftime_now('%B %d, %Y') + &quot;.&#10;You are Granite, developed by IBM.&quot; %}&#10; {%- if tools and documents %}&#10; {%- set system_message = system_message + &quot; You are a helpful AI assistant with access to the following tools. When a tool is required to answer the user's query, respond with &lt;|tool_call|> followed by a JSON list of tools used. If a tool does not exist in the provided list of tools, notify the user that you do not have the ability to fulfill the request.&#10;&#10;Write the response to the user's input by strictly aligning with the facts in the provided documents. If the information needed to answer the question is not available in the documents, inform the user that the question cannot be answered based on the available data.&quot; %}&#10; {%- elif tools %}&#10; {%- set system_message = system_message + &quot; You are a helpful AI assistant with access to the following tools. When a tool is required to answer the user's query, respond with &lt;|tool_call|> followed by a JSON list of tools used. If a tool does not exist in the provided list of tools, notify the user that you do not have the ability to fulfill the request.&quot; %}&#10; {%- elif documents %}&#10; {%- set system_message = system_message + &quot; Write the response to the user's input by strictly aligning with the facts in the provided documents. If the information needed to answer the question is not available in the documents, inform the user that the question cannot be answered based on the available data.&quot; %}&#10; {%- else %}&#10; {%- set system_message = system_message + &quot; You are a helpful AI assistant.&quot; %} &#10; {%- endif %}&#10; {%- if 'citations' in controls and documents %}&#10; {%- set system_message = system_message + '&#10;&#10;In your response, use the symbols &lt;co> and &lt;/co> to indicate when a fact comes from a document in the search result, e.g &lt;co>0&lt;/co> for a fact from document 0. Afterwards, list all the citations with their corresponding documents in an ordered list.' %}&#10; {%- endif %}&#10; {%- if 'hallucinations' in controls and documents %}&#10; {%- set system_message = system_message + '&#10;&#10;Finally, after the response is written, include a numbered list of sentences from the response that are potentially hallucinated and not based in the documents.' %}&#10; {%- endif %}&#10; {%- set loop_messages = messages %}&#10;{%- endif %}&#10;{{- '&lt;|start_of_role|>system&lt;|end_of_role|>' + system_message + '&lt;|end_of_text|>&#10;' }}&#10;{%- if tools %}&#10; {{- '&lt;|start_of_role|>tools&lt;|end_of_role|>' }}&#10; {{- tools | tojson(indent=4) }}&#10; {{- '&lt;|end_of_text|>&#10;' }}&#10;{%- endif %}&#10;{%- if documents %}&#10; {{- '&lt;|start_of_role|>documents&lt;|end_of_role|>' }}&#10; {%- for document in documents %}&#10; {{- 'Document ' + loop.index0 | string + '&#10;' }}&#10; {{- document['text'] }}&#10; {%- if not loop.last %}&#10; {{- '&#10;&#10;'}}&#10; {%- endif%}&#10; {%- endfor %}&#10; {{- '&lt;|end_of_text|>&#10;' }}&#10;{%- endif %}&#10;{%- for message in loop_messages %}&#10; {{- '&lt;|start_of_role|>' + message['role'] + '&lt;|end_of_role|>' + message['content'] + '&lt;|end_of_text|>&#10;' }}&#10; {%- if loop.last and add_generation_prompt %}&#10; {{- '&lt;|start_of_role|>assistant' }}&#10; {%- if controls %}&#10; {{- ' ' + controls | tojson()}}&#10; {%- endif %}&#10; {{- '&lt;|end_of_role|>' }}&#10; {%- endif %}&#10;{%- endfor %}" />
681
+ <clean_up_tokenization_spaces />
682
+ <detokenizer_input_type value="i64" />
683
+ <eos_token_id value="0" />
684
+ <handle_special_tokens_with_re />
685
+ <number_of_inputs value="1" />
686
+ <openvino_tokenizers_version value="2024.6.0.0" />
687
+ <openvino_version value="2024.6.0" />
688
+ <original_tokenizer_class value="&lt;class 'transformers.models.gpt2.tokenization_gpt2_fast.GPT2TokenizerFast'>" />
689
+ <pad_token_id value="0" />
690
+ <sentencepiece_version value="0.2.0" />
691
+ <skip_special_tokens value="True" />
692
+ <streaming_detokenizer value="False" />
693
+ <tiktoken_version value="0.7.0" />
694
+ <tokenizer_output_type value="i64" />
695
+ <tokenizers_version value="0.20.3" />
696
+ <transformers_version value="4.46.3" />
697
+ <use_max_padding value="False" />
698
+ <use_sentencepiece_backend value="False" />
699
+ <utf8_replace_mode value="replace" />
700
+ <with_detokenizer value="True" />
701
+ </rt_info>
702
+ </net>
special_tokens_map.json ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|start_of_role|>",
4
+ "<|end_of_role|>",
5
+ "<|tool_call|>"
6
+ ],
7
+ "bos_token": {
8
+ "content": "<|end_of_text|>",
9
+ "lstrip": false,
10
+ "normalized": false,
11
+ "rstrip": false,
12
+ "single_word": false
13
+ },
14
+ "eos_token": {
15
+ "content": "<|end_of_text|>",
16
+ "lstrip": false,
17
+ "normalized": false,
18
+ "rstrip": false,
19
+ "single_word": false
20
+ },
21
+ "pad_token": {
22
+ "content": "<|end_of_text|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false
27
+ },
28
+ "unk_token": {
29
+ "content": "<|end_of_text|>",
30
+ "lstrip": false,
31
+ "normalized": false,
32
+ "rstrip": false,
33
+ "single_word": false
34
+ }
35
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,199 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "0": {
6
+ "content": "<|end_of_text|>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "1": {
14
+ "content": "<fim_prefix>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "2": {
22
+ "content": "<fim_middle>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "3": {
30
+ "content": "<fim_suffix>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "4": {
38
+ "content": "<fim_pad>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "5": {
46
+ "content": "<filename>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "6": {
54
+ "content": "<gh_stars>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "7": {
62
+ "content": "<issue_start>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "8": {
70
+ "content": "<issue_comment>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ },
77
+ "9": {
78
+ "content": "<issue_closed>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": true
84
+ },
85
+ "10": {
86
+ "content": "<jupyter_start>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": true
92
+ },
93
+ "11": {
94
+ "content": "<jupyter_text>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": true
100
+ },
101
+ "12": {
102
+ "content": "<jupyter_code>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": true
108
+ },
109
+ "13": {
110
+ "content": "<jupyter_output>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": true
116
+ },
117
+ "14": {
118
+ "content": "<empty_output>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": true
124
+ },
125
+ "15": {
126
+ "content": "<commit_before>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": true
132
+ },
133
+ "16": {
134
+ "content": "<commit_msg>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": true
140
+ },
141
+ "17": {
142
+ "content": "<commit_after>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": true
148
+ },
149
+ "18": {
150
+ "content": "<reponame>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": true
156
+ },
157
+ "49152": {
158
+ "content": "<|start_of_role|>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": true
164
+ },
165
+ "49153": {
166
+ "content": "<|end_of_role|>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": true
172
+ },
173
+ "49154": {
174
+ "content": "<|tool_call|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": true
180
+ }
181
+ },
182
+ "additional_special_tokens": [
183
+ "<|start_of_role|>",
184
+ "<|end_of_role|>",
185
+ "<|tool_call|>"
186
+ ],
187
+ "bos_token": "<|end_of_text|>",
188
+ "chat_template": "{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content'] %}\n {%- set loop_messages = messages[1:] %}\n{%- else %}\n {%- set system_message = \"Knowledge Cutoff Date: April 2024.\nToday's Date: \" + strftime_now('%B %d, %Y') + \".\nYou are Granite, developed by IBM.\" %}\n {%- if tools and documents %}\n {%- set system_message = system_message + \" You are a helpful AI assistant with access to the following tools. When a tool is required to answer the user's query, respond with <|tool_call|> followed by a JSON list of tools used. If a tool does not exist in the provided list of tools, notify the user that you do not have the ability to fulfill the request.\n\nWrite the response to the user's input by strictly aligning with the facts in the provided documents. If the information needed to answer the question is not available in the documents, inform the user that the question cannot be answered based on the available data.\" %}\n {%- elif tools %}\n {%- set system_message = system_message + \" You are a helpful AI assistant with access to the following tools. When a tool is required to answer the user's query, respond with <|tool_call|> followed by a JSON list of tools used. If a tool does not exist in the provided list of tools, notify the user that you do not have the ability to fulfill the request.\" %}\n {%- elif documents %}\n {%- set system_message = system_message + \" Write the response to the user's input by strictly aligning with the facts in the provided documents. If the information needed to answer the question is not available in the documents, inform the user that the question cannot be answered based on the available data.\" %}\n {%- else %}\n {%- set system_message = system_message + \" You are a helpful AI assistant.\" %} \n {%- endif %}\n {%- if 'citations' in controls and documents %}\n {%- set system_message = system_message + '\n\nIn your response, use the symbols <co> and </co> to indicate when a fact comes from a document in the search result, e.g <co>0</co> for a fact from document 0. Afterwards, list all the citations with their corresponding documents in an ordered list.' %}\n {%- endif %}\n {%- if 'hallucinations' in controls and documents %}\n {%- set system_message = system_message + '\n\nFinally, after the response is written, include a numbered list of sentences from the response that are potentially hallucinated and not based in the documents.' %}\n {%- endif %}\n {%- set loop_messages = messages %}\n{%- endif %}\n{{- '<|start_of_role|>system<|end_of_role|>' + system_message + '<|end_of_text|>\n' }}\n{%- if tools %}\n {{- '<|start_of_role|>tools<|end_of_role|>' }}\n {{- tools | tojson(indent=4) }}\n {{- '<|end_of_text|>\n' }}\n{%- endif %}\n{%- if documents %}\n {{- '<|start_of_role|>documents<|end_of_role|>' }}\n {%- for document in documents %}\n {{- 'Document ' + loop.index0 | string + '\n' }}\n {{- document['text'] }}\n {%- if not loop.last %}\n {{- '\n\n'}}\n {%- endif%}\n {%- endfor %}\n {{- '<|end_of_text|>\n' }}\n{%- endif %}\n{%- for message in loop_messages %}\n {{- '<|start_of_role|>' + message['role'] + '<|end_of_role|>' + message['content'] + '<|end_of_text|>\n' }}\n {%- if loop.last and add_generation_prompt %}\n {{- '<|start_of_role|>assistant' }}\n {%- if controls %}\n {{- ' ' + controls | tojson()}}\n {%- endif %}\n {{- '<|end_of_role|>' }}\n {%- endif %}\n{%- endfor %}",
189
+ "clean_up_tokenization_spaces": true,
190
+ "eos_token": "<|end_of_text|>",
191
+ "errors": "replace",
192
+ "extra_special_tokens": {},
193
+ "model_max_length": 9223372036854775807,
194
+ "pad_token": "<|end_of_text|>",
195
+ "padding_side": "left",
196
+ "tokenizer_class": "GPT2Tokenizer",
197
+ "unk_token": "<|end_of_text|>",
198
+ "vocab_size": 49152
199
+ }
vocab.json ADDED
The diff for this file is too large to render. See raw diff