sksameermujahid's picture
Upload config.json
8c3a8de verified
raw
history blame contribute delete
212 Bytes
{
"architectures": [
"LlamaForCausalLM"
],
"model_type": "llama",
"torch_dtype": "float16",
"transformers_version": "4.49.0",
"use_cache": true,
"vocab_size": 32000
}