EXL2 Quantizations of Qwen2.5-3B-Instruct

Using exllamav2 release 0.2.6 for quantization.

Original model: https://huggingface.co/Qwen/Qwen2.5-3B-Instruct

Bits 8.0, lm_head 8.0

Downloads last month
29
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.

Model tree for Zenabius/Qwen2.5-3B-Instruct-exl2

Base model

Qwen/Qwen2.5-3B
Quantized
(106)
this model