{ "added_tokens_decoder": { "0": { "content": "[UNK]", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "1": { "content": "<|begin_of_text|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "2": { "content": "<|end_of_text|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "3": { "content": "[PAD]", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "65525": { "content": "<|text_start|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "65526": { "content": "<|thinking_start|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "65527": { "content": "<|thinking_end|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "65528": { "content": "<|strategy_start|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "65529": { "content": "<|strategy_end|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "65530": { "content": "<|data_start|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "65531": { "content": "<|text_end|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "65532": { "content": "<|data_end|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "65533": { "content": "<|subject|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "65534": { "content": "<|property|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "65535": { "content": "<|object|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true } }, "additional_special_tokens": [ "<|text_start|>", "<|thinking_start|>", "<|thinking_end|>", "<|strategy_start|>", "<|strategy_end|>", "<|data_start|>", "<|text_end|>", "<|data_end|>", "<|subject|>", "<|property|>", "<|object|>" ], "clean_up_tokenization_spaces": true, "model_max_length": 1000000000000000019884624838656, "tokenizer_class": "PreTrainedTokenizerFast" }