--- library_name: transformers tags: - math license: apache-2.0 datasets: - openai/gsm8k language: - en metrics: - accuracy base_model: - deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B pipeline_tag: text-generation --- # DeepMath-7B-L ## Model Overview DeepMath-7B-L are fine-tuned versions of [DeepSeek-R1-Distill-Qwen-1.5B](https://huggingface.co./deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B) on the [GSM8K dataset](https://huggingface.co./datasets/gsm8k). These models are designed for mathematical reasoning and problem-solving, excelling in arithmetic, algebra, and word problems. ## Model Details - **Base Model:** DeepSeek-R1-Distill-Qwen-1.5B - **Fine-Tuning Dataset:** GSM8K - **Parameters:** 1.5 Billion - **Task:** Mathematical Question Answering (Math QA) - **Repositories:** - [DeepMath-7B-L](https://huggingface.co./codewithdark/deepmath-7b-l) (LoRA adapter-enhanced model) - **Commit Messages:** - "Full merged model for math QA" - "Added LoRA adapters for math reasoning" ## Training Details - **Dataset:** GSM8K (Grade School Math 8K) - a high-quality dataset for mathematical reasoning - **Fine-Tuning Framework:** Hugging Face Transformers & PyTorch - **Optimization Techniques:** - AdamW Optimizer - Learning rate scheduling - Gradient accumulation - Mixed precision training (FP16) - **Training Steps:** Multiple epochs on a high-performance GPU cluster ## Capabilities & Performance DeepMath-7B-L excel in: - Solving word problems with step-by-step reasoning - Performing algebraic and arithmetic computations - Understanding complex problem structures - Generating structured solutions with explanations ### DeepMath-7B-L (LoRA Adapter-Enhanced Model) ```python from transformers import AutoModelForCausalLM, AutoTokenizer tokenizer = AutoTokenizer.from_pretrained("codewithdark/deepmath-7b-l") model = AutoModelForCausalLM.from_pretrained("codewithdark/deepmath-7b-l") input_text = "Solve: 2x + 3 = 7" inputs = tokenizer(input_text, return_tensors="pt") outputs = model.generate(**inputs, max_length=100) print(tokenizer.decode(outputs[0], skip_special_tokens=True)) ``` ## Limitations - May struggle with extremely complex mathematical proofs - Performance is limited to the scope of GSM8K-type problems - Potential biases in training data ## Future Work - Extending training to more diverse math datasets - Exploring larger models for improved accuracy - Fine-tuning on physics and higher-level mathematical reasoning datasets ## License This model is released under the Apache 2.0 License. ## Citation If you use these models, please cite: ``` @misc{DeepMath-7B-L, author = {Ahsan}, title = {DeepMath-7B-L: LoRA Adapter Enhanced Model for Math Reasoning}, year = {2025}, url = {https://huggingface.co./codewithdark/deepmath-7b-l} } ```