---
license: apache-2.0
---
## Impish_LLAMA_3B is available at the following quantizations:
- Original: [FP16](https://huggingface.co/SicariusSicariiStuff/Impish_LLAMA_3B)
- GGUF: [Static Quants](https://huggingface.co/SicariusSicariiStuff/Impish_LLAMA_3B_GGUF) | [iMatrix_GGUF](https://huggingface.co/SicariusSicariiStuff/Impish_LLAMA_3B_iMatrix)
- EXL2: [3.0 bpw](3333333) | [4.0 bpw](44444) | [5.0 bpw](5555555) | [6.0 bpw](6666666) | [7.0 bpw](7777777) | [8.0 bpw](888888)
- Specialized: [FP8](https://huggingface.co/SicariusSicariiStuff/Impish_LLAMA_3B_FP8)
- Mobile (ARM): [Q4_0_X_X](https://huggingface.co/SicariusSicariiStuff/Impish_LLAMA_3B_ARM)
# Model instruction template: Llama-3-Instruct
```
<|begin_of_text|><|start_header_id|>system<|end_header_id|>
{system_prompt}<|eot_id|><|start_header_id|>user<|end_header_id|>
{input}<|eot_id|><|start_header_id|>assistant<|end_header_id|>
{output}<|eot_id|>
```
**Recommended generation Presets:**
Midnight Enigma
max_new_tokens: 512
temperature: 0.98
top_p: 0.37
top_k: 100
typical_p: 1
min_p: 0
repetition_penalty: 1.18
do_sample: True
min_p
max_new_tokens: 512
temperature: 1
top_p: 1
top_k: 0
typical_p: 1
min_p: 0.05
repetition_penalty: 1
do_sample: True
Divine Intellect
max_new_tokens: 512
temperature: 1.31
top_p: 0.14
top_k: 49
typical_p: 1
min_p: 0
repetition_penalty: 1.17
do_sample: True
simple-1
max_new_tokens: 512
temperature: 0.7
top_p: 0.9
top_k: 20
typical_p: 1
min_p: 0
repetition_penalty: 1.15
do_sample: True
# Model Details