DeepSeek Math Tutor

This model is a fine-tuned version of DeepSeek-R1-Distill-Llama-8B optimized for mathematics education. It's designed to provide step-by-step explanations for mathematical problems in a beginner-friendly way.

Model Details

  • Base Model: DeepSeek-R1-Distill-Llama-8B
  • Training Data: Math reasoning dataset with 7000 examples
  • Task: Mathematics education and problem-solving
  • Training Method: LoRA fine-tuning

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained("analist/deepseek-math-tutor-fine-tuned")
tokenizer = AutoTokenizer.from_pretrained("analist/deepseek-math-tutor-fine-tuned")

# Example prompt format
prompt = '''Below is an instruction that describes a task, paired with an input that provides further context.
Write a response that appropriately completes the request.

### Instruction:
You are a maths expert with advanced knowledge in pedagogy, arithmetics, geometry, analysis, calculus.
Please answer the following questions.

### Question:
{your_math_question}

### Response:
'''

# Generate response
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=1200)
response = tokenizer.decode(outputs[0])

Training Details

  • Training framework: Unsloth
  • Optimizer: AdamW 8-bit
  • Learning rate: 2e-4
  • Batch size: 2 (per device)
  • Gradient accumulation steps: 4
  • Training steps: 60
  • Warmup steps: 5

Limitations

This model is specifically tuned for mathematics education and may not perform as well on other tasks. It's designed to provide explanations suitable for beginners learning mathematics.

Downloads last month
7
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.