Merge
Added chat template and eos_token to original model
This is a merge of pre-trained language models created using mergekit.
Merge Details
Qwen2.5 Instruct tied to Reasoning LORA.
Merge Method
This model was merged using the Passthrough merge method using unsloth/Qwen2.5-3B-Instruct + bunnycore/Qwen-2.5-3b-R1-lora_model-v.1 as a base.
Models Merged
The following models were included in the merge:
Configuration
The following YAML configuration was used to produce this model:
base_model: unsloth/Qwen2.5-3B-Instruct+bunnycore/Qwen-2.5-3b-R1-lora_model-v.1
dtype: bfloat16
merge_method: passthrough
models:
- model: unsloth/Qwen2.5-3B-Instruct+bunnycore/Qwen-2.5-3b-R1-lora_model-v.1
tokenizer_source: unsloth/Qwen2.5-3B
Open LLM Leaderboard Evaluation Results
Detailed results can be found here
Metric | Value |
---|---|
Avg. | 24.67 |
IFEval (0-Shot) | 42.14 |
BBH (3-Shot) | 27.20 |
MATH Lvl 5 (4-Shot) | 26.74 |
GPQA (0-shot) | 7.94 |
MuSR (0-shot) | 12.73 |
MMLU-PRO (5-shot) | 31.26 |
- Downloads last month
- 5
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support
Model tree for dbaeka/Qwen2.5-R1-3B-Instruct
Evaluation results
- strict accuracy on IFEval (0-Shot)Open LLM Leaderboard42.140
- normalized accuracy on BBH (3-Shot)Open LLM Leaderboard27.200
- exact match on MATH Lvl 5 (4-Shot)Open LLM Leaderboard26.740
- acc_norm on GPQA (0-shot)Open LLM Leaderboard7.940
- acc_norm on MuSR (0-shot)Open LLM Leaderboard12.730
- accuracy on MMLU-PRO (5-shot)test set Open LLM Leaderboard31.260