| | --- |
| | library_name: transformers |
| | tags: |
| | - math |
| | license: apache-2.0 |
| | datasets: |
| | - openai/gsm8k |
| | language: |
| | - en |
| | metrics: |
| | - accuracy |
| | base_model: |
| | - deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B |
| | pipeline_tag: text-generation |
| | --- |
| | |
| | # DeepMath-7B-L |
| |
|
| | ## Model Overview |
| | DeepMath-7B-L are fine-tuned versions of [DeepSeek-R1-Distill-Qwen-1.5B](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B) on the [GSM8K dataset](https://huggingface.co/datasets/gsm8k). These models are designed for mathematical reasoning and problem-solving, excelling in arithmetic, algebra, and word problems. |
| |
|
| | ## Model Details |
| | - **Base Model:** DeepSeek-R1-Distill-Qwen-1.5B |
| | - **Fine-Tuning Dataset:** GSM8K |
| | - **Parameters:** 1.5 Billion |
| | - **Task:** Mathematical Question Answering (Math QA) |
| | - **Repositories:** |
| | - [DeepMath-7B-L](https://huggingface.co/codewithdark/deepmath-7b-l) (LoRA adapter-enhanced model) |
| | - **Commit Messages:** |
| | - "Full merged model for math QA" |
| | - "Added LoRA adapters for math reasoning" |
| |
|
| | ## Training Details |
| | - **Dataset:** GSM8K (Grade School Math 8K) - a high-quality dataset for mathematical reasoning |
| | - **Fine-Tuning Framework:** Hugging Face Transformers & PyTorch |
| | - **Optimization Techniques:** |
| | - AdamW Optimizer |
| | - Learning rate scheduling |
| | - Gradient accumulation |
| | - Mixed precision training (FP16) |
| | - **Training Steps:** Multiple epochs on a high-performance GPU cluster |
| |
|
| | ## Capabilities & Performance |
| | DeepMath-7B-L excel in: |
| | - Solving word problems with step-by-step reasoning |
| | - Performing algebraic and arithmetic computations |
| | - Understanding complex problem structures |
| | - Generating structured solutions with explanations |
| |
|
| |
|
| |
|
| | ### DeepMath-7B-L (LoRA Adapter-Enhanced Model) |
| | ```python |
| | from transformers import AutoModelForCausalLM, AutoTokenizer |
| | |
| | tokenizer = AutoTokenizer.from_pretrained("codewithdark/deepmath-7b-l") |
| | model = AutoModelForCausalLM.from_pretrained("codewithdark/deepmath-7b-l") |
| | |
| | input_text = "Solve: 2x + 3 = 7" |
| | inputs = tokenizer(input_text, return_tensors="pt") |
| | outputs = model.generate(**inputs, max_length=100) |
| | print(tokenizer.decode(outputs[0], skip_special_tokens=True)) |
| | ``` |
| |
|
| | ## Limitations |
| | - May struggle with extremely complex mathematical proofs |
| | - Performance is limited to the scope of GSM8K-type problems |
| | - Potential biases in training data |
| |
|
| | ## Future Work |
| | - Extending training to more diverse math datasets |
| | - Exploring larger models for improved accuracy |
| | - Fine-tuning on physics and higher-level mathematical reasoning datasets |
| |
|
| | ## License |
| | This model is released under the Apache 2.0 License. |
| |
|
| | ## Citation |
| | If you use these models, please cite: |
| | ``` |
| | @misc{DeepMath-7B-L, |
| | author = {Ahsan}, |
| | title = {DeepMath-7B-L: LoRA Adapter Enhanced Model for Math Reasoning}, |
| | year = {2025}, |
| | url = {https://huggingface.co/codewithdark/deepmath-7b-l} |
| | } |
| | ``` |