GGUF quantization for the InternLM-20b-chat model.
Requires a recent version of llama.cpp: https://github.com/ggerganov/llama.cpp/pull/4283
Link to the original model: https://huggingface.co/internlm/internlm-chat-20b
Github repository: https://github.com/InternLM/InternLM
Prompt format:
<|User|>: Your input.<eoh>
<|Bot|>:
License:
Model weights are fully open for academic research and also allow free commercial usage. To apply for a commercial license, please fill in the application form (English)/申请表(ä¸æ–‡ï¼‰. For other questions or collaborations, please contact [email protected].
- Downloads last month
- 54
Inference Providers
NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API:
The model has no library tag.