ggml-mllm / config.json
matthoffner's picture
Update config.json
539065a verified
raw
history blame
430 Bytes
{
"host": "0.0.0.0",
"port": 8080,
"models": [
{
"model": "llava-v1.6-mistral-7b.Q3_K_XS.gguf",
"model_alias": "llava-1.6",
"chat_format": "llava-1-5",
"clip_model_path": "mmproj-model-f16.gguf",
"n_gpu_layers": -1,
"offload_kqv": false, // Assuming default value, adjust as needed
"n_threads": 12,
"n_batch": 1,
"n_ctx": 2048,
"logits_all": true
}
]
}