This is a FP16.GGUF model, converted to GGUF using llama.cpp/convert-hf-to-gguf.py

Downloads last month
10
GGUF
Model size
7.25B params
Architecture
llama
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.