pcuenq's picture
pcuenq HF staff
Create README.md (#1)
c88d73e verified
metadata
license: mit
library_name: transformers
pipeline_tag: text-generation
tags:
  - conversational
  - mlx
base_model: deepseek-ai/DeepSeek-R1-Distill-Qwen-32B

DeepSeek-R1-Distill-Qwen-32B-Q2-6

This model was converted to MLX from deepseek-ai/DeepSeek-R1-Distill-Qwen-32B, using mixed 2/6 bit quantization. This scheme preserves quality much more than a standard 2-bit quantization.

Use with mlx

pip install mlx-lm
python -m mlx_lm.chat --model pcuenq/DeepSeek-R1-Distill-Qwen-32B-Q2-6 --max-tokens 10000 --temp 0.6 --top-p 0.7