|
{ |
|
"added_tokens_decoder": { |
|
"0": { |
|
"content": "<|endoftext|>", |
|
"lstrip": false, |
|
"normalized": false, |
|
"rstrip": false, |
|
"single_word": false, |
|
"special": true |
|
}, |
|
"4709": { |
|
"content": "user", |
|
"lstrip": false, |
|
"normalized": true, |
|
"rstrip": false, |
|
"single_word": false, |
|
"special": false |
|
}, |
|
"27673": { |
|
"content": "system", |
|
"lstrip": false, |
|
"normalized": true, |
|
"rstrip": false, |
|
"single_word": false, |
|
"special": false |
|
}, |
|
"81912": { |
|
"content": "<think>", |
|
"lstrip": false, |
|
"normalized": true, |
|
"rstrip": false, |
|
"single_word": false, |
|
"special": false |
|
}, |
|
"81913": { |
|
"content": "</think>", |
|
"lstrip": false, |
|
"normalized": true, |
|
"rstrip": false, |
|
"single_word": false, |
|
"special": false |
|
}, |
|
"81914": { |
|
"content": "<|im_end|>", |
|
"lstrip": false, |
|
"normalized": true, |
|
"rstrip": false, |
|
"single_word": false, |
|
"special": false |
|
}, |
|
"81915": { |
|
"content": "<|im_start|>", |
|
"lstrip": false, |
|
"normalized": true, |
|
"rstrip": false, |
|
"single_word": false, |
|
"special": false |
|
}, |
|
"81916": { |
|
"content": "assistant", |
|
"lstrip": false, |
|
"normalized": true, |
|
"rstrip": false, |
|
"single_word": false, |
|
"special": false |
|
}, |
|
"81917": { |
|
"content": "BiBo", |
|
"lstrip": false, |
|
"normalized": true, |
|
"rstrip": false, |
|
"single_word": false, |
|
"special": false |
|
}, |
|
"81918": { |
|
"content": "aloobun", |
|
"lstrip": false, |
|
"normalized": true, |
|
"rstrip": false, |
|
"single_word": false, |
|
"special": false |
|
}, |
|
"81919": { |
|
"content": "LowIQGenAI", |
|
"lstrip": false, |
|
"normalized": true, |
|
"rstrip": false, |
|
"single_word": false, |
|
"special": false |
|
} |
|
}, |
|
"clean_up_tokenization_spaces": false, |
|
"eos_token": "<|endoftext|>", |
|
"extra_special_tokens": {}, |
|
"model_max_length": 8192, |
|
"pad_token": "<|endoftext|>", |
|
"padding_side": "right", |
|
"tokenizer_class": "PreTrainedTokenizerFast" |
|
} |
|
|