Qwen2 fine-tune

MaziyarPanahi/calme-2.3-qwen2-7b

This is a fine-tuned version of the Qwen/Qwen2-7B model. It aims to improve the base model across all benchmarks.

⚑ Quantized GGUF

All GGUF models are available here: MaziyarPanahi/calme-2.3-qwen2-7b-GGUF

πŸ† Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 22.74
IFEval (0-Shot) 38.25
BBH (3-Shot) 30.96
MATH Lvl 5 (4-Shot) 18.66
GPQA (0-shot) 6.26
MuSR (0-shot) 13.31
MMLU-PRO (5-shot) 29.01

Prompt Template

This model uses ChatML prompt template:

<|im_start|>system
{System}
<|im_end|>
<|im_start|>user
{User}
<|im_end|>
<|im_start|>assistant
{Assistant}

How to use


# Use a pipeline as a high-level helper

from transformers import pipeline

messages = [
    {"role": "user", "content": "Who are you?"},
]
pipe = pipeline("text-generation", model="MaziyarPanahi/calme-2.3-qwen2-7b")
pipe(messages)


# Load model directly

from transformers import AutoTokenizer, AutoModelForCausalLM

tokenizer = AutoTokenizer.from_pretrained("MaziyarPanahi/calme-2.3-qwen2-7b")
model = AutoModelForCausalLM.from_pretrained("MaziyarPanahi/calme-2.3-qwen2-7b")

Downloads last month
47,670
Safetensors
Model size
7.62B params
Tensor type
BF16
Β·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model authors have turned it off explicitly.

Model tree for MaziyarPanahi/calme-2.3-qwen2-7b

Base model

Qwen/Qwen2-7B
Finetuned
(58)
this model
Quantizations
1 model

Datasets used to train MaziyarPanahi/calme-2.3-qwen2-7b

Collections including MaziyarPanahi/calme-2.3-qwen2-7b

Evaluation results