Spaces:
Build error
Build error
## Working models | |
As of **2023-07-19**, the following GPTQ models on HuggingFace all appear to be working: | |
- iambestfeed/open_llama_3b_4bit_128g | |
- Neko-Institute-of-Science/LLaMA-7B-4bit-128g | |
- Neko-Institute-of-Science/LLaMA-13B-4bit-128g | |
- Neko-Institute-of-Science/LLaMA-30B-4bit-32g | |
- Neko-Institute-of-Science/LLaMA-30B-4bit-128g | |
- Neko-Institute-of-Science/LLaMA-65B-4bit-32g | |
- Neko-Institute-of-Science/LLaMA-65B-4bit-128g | |
- Panchovix/LLaMA-2-70B-GPTQ-transformers4.32.0.dev0 | |
- reeducator/bluemoonrp-13b | |
- reeducator/bluemoonrp-30b | |
- TehVenom/Metharme-13b-4bit-GPTQ | |
- TheBloke/airoboros-13B-GPTQ | |
- TheBloke/gpt4-x-vicuna-13B-GPTQ | |
- TheBloke/GPT4All-13B-snoozy-GPTQ | |
- TheBloke/guanaco-33B-GPTQ | |
- TheBloke/guanaco-65B-GPTQ | |
- TheBloke/h2ogpt-oasst1-512-30B-GPTQ | |
- TheBloke/koala-13B-GPTQ-4bit-128g | |
- TheBloke/Llama-2-13B-chat-GPTQ (128g) | |
- TheBloke/Llama-2-13B-GPTQ (32g, 64g, 128g) | |
- TheBloke/Llama-2-70B-GPTQ (32g, 128g) | |
- TheBloke/Manticore-13B-GPTQ | |
- TheBloke/medalpaca-13B-GPTQ-4bit | |
- TheBloke/medalpaca-13B-GPTQ-4bit (compat version) | |
- TheBloke/Nous-Hermes-13B-GPTQ | |
- TheBloke/robin-65B-v2-GPTQ | |
- TheBloke/tulu-7B-GPTQ | |
- TheBloke/Tulu-13B-SuperHOT-8K-GPTQ | |
- TheBloke/tulu-30B-GPTQ | |
- TheBloke/vicuna-13B-1.1-GPTQ-4bit-128g | |
- TheBloke/VicUnlocked-30B-LoRA-GPTQ | |
- TheBloke/wizard-mega-13B-GPTQ | |
- TheBloke/Wizard-Vicuna-7B-Uncensored-GPTQ | |
- TheBloke/Wizard-Vicuna-13B-Uncensored-GPTQ | |
- TheBloke/WizardLM-7B-uncensored-GPTQ | |
- TheBloke/WizardLM-30B-Uncensored-GPTQ | |
- TheBloke/WizardLM-33B-V1.0-Uncensored-SuperHOT-8K-GPTQ | |
- tmpupload/superhot-30b-8k-no-rlhf-test-128g-GPTQ | |
- Yhyu13/chimera-inst-chat-13b-gptq-4bit | |
- Yhyu13/oasst-rlhf-2-llama-30b-7k-steps-gptq-4bit | |
## Non-working models | |
None as of **2023-07-19**. |