--- library_name: transformers tags: [] --- # llm-jp-13b-OpenWebMathInstruct_2_v1.1 開発停止 数学タスクにおける精度向上が見込めなかったため ## Overview This model is an instruction-tuned variant of [llm-jp/llm-jp-3-13b-instruct](https://huggingface.co./llm-jp/llm-jp-3-13b-instruct), further fine-tuned on a subset of [nvidia/OpenMathInstruct-2](https://huggingface.co./datasets/nvidia/OpenMathInstruct-2) with 256,000 samples. The fine-tuning process followed a parameter-efficient strategy, updating only selected layers while freezing most of the model parameters. ## Key Features - **Base Model**: [llm-jp/llm-jp-3-13b-instruct](https://huggingface.co./llm-jp/llm-jp-3-13b-instruct) - **Fine-Tuning Data**: 256,000 samples from [nvidia/OpenMathInstruct-2](https://huggingface.co./datasets/nvidia/OpenMathInstruct-2) - **Updated Parameters**: - All parameters were frozen except for: ```python for param in model.parameters(): param.requires_grad = False for param in model.lm_head.parameters(): param.requires_grad = True ``` - **Macro-o1 Tokens Added**: To align with [Marco-o1](https://arxiv.org/pdf/2411.14405v1), we introduced the following special tokens: - ``, `` - ``, `` - **Reasoning Model Integration**: Uses the implementation from [Hajime-Y/reasoning-model](https://github.com/Hajime-Y/reasoning-model) ## Usage Below is an example of using the model with Monte Carlo Tree Search (MCTS) for reasoning: ```python import sys import torch sys.path.append('./reasoning-model') from reasoning_model import ReasoningModelForCausalLM from tree_utils import print_tree_with_best_path from transformers import AutoTokenizer # tokenizerとmodelの準備 model_name = "doshisha-mil/llm-jp-13b-OpenMathInstruct-2-v1.1" tokenizer = AutoTokenizer.from_pretrained( model_name, trust_remote_code=True, ) # パディングトークンを明示的に設定 if tokenizer.pad_token is None: tokenizer.pad_token = tokenizer.eos_token # モデルのロード model = ReasoningModelForCausalLM.from_pretrained( model_name, torch_dtype="auto", device_map="auto" ) # 入力テキスト prompt = "Find the number of positive integers $x$ that satisfy $x^{-1}>x$." text = f"あなたは優秀で論理的なアシスタントです。まずはタグの中であなたの思考の過程を記載し、タグの中に最終的にユーザーに提供する出力を記載します。\n\n### 指示: {prompt}\n\n### 応答: \n" # Tokenize with explicit attention_mask model_inputs = tokenizer([text], return_tensors="pt", padding=True, truncation=True) model_inputs["attention_mask"] = (model_inputs["input_ids"] != tokenizer.pad_token_id).long() # デバイスをモデルのデバイスに統一 model_inputs = {key: val.to(model.device) for key, val in model_inputs.items()} # MCTSを用いて生成 final_tokens, final_node = model.generate( input_ids=model_inputs["input_ids"], attention_mask=model_inputs["attention_mask"], # 明示的に attention_mask を渡す iterations_per_step=3, max_iterations=30, mini_step_size=32, expand_threshold=0, step_separator_ids=None, ) # 結果をテキスト化 final_text = tokenizer.decode(final_tokens, skip_special_tokens=True) print("=== 最終生成テキスト ===") print(final_text) ``` ## Model Applications - Mathematical problem-solving with structured reasoning - Chain-of-Thought (CoT) enhanced reasoning - Integration with Monte Carlo Tree Search (MCTS) - Instruction-based question answering ## References - **Base Model**: [llm-jp/llm-jp-3-13b-instruct](https://huggingface.co./llm-jp/llm-jp-3-13b-instruct) - **Dataset**: [nvidia/OpenMathInstruct-2](https://huggingface.co./datasets/nvidia/OpenMathInstruct-2) - **Marco-o1 Paper**: [arXiv:2411.14405v1](https://arxiv.org/pdf/2411.14405v1) - **Reasoning Model Code**: [Hajime-Y/reasoning-model](https://github.com/Hajime-Y/reasoning-model) ## Citation If you use this model, please cite the original base model and relevant datasets. ```bibtex @article{llm-jp3-13b-instruct, title={LLM-JP 3-13B Instruct}, author={LLM-JP Team}, year={2024}, journal={Hugging Face Repository}, url={https://huggingface.co./llm-jp/llm-jp-3-13b-instruct} } @article{marco-o1, title={Marco-o1: Towards Open Reasoning Models for Open-Ended Solutions}, author={Yu Zhao, Huifeng Yin, Bo Zeng, Hao Wang, Tianqi Shi, Chenyang Lyu, Longyue Wang, Weihua Luo, Kaifu Zhang}, year={2024}, journal={arXiv}, eprint={2411.14405v1}, archivePrefix={arXiv}, primaryClass={cs.CL} } ``` ## License Refer to the base model's license at [llm-jp/llm-jp-3-13b-instruct](https://huggingface.co./llm-jp/llm-jp-3-13b-instruct) for details. --- This README provides clear documentation on how to use the model while crediting its sources. Let me know if you need modifications!