Description

This repo contains fp8 model files for aya-expanse-32b.

Quantization parameter

  • activation_scheme : dynamic
  • quant_method : fp8
Downloads last month
16
Safetensors
Model size
32.3B params
Tensor type
FP16
·
F8_E4M3
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.

Model tree for minyichen/aya-expanse-32b-Dynamic-fp8

Quantized
(20)
this model