-
-
-
-
-
-
Inference Providers
Active filters:
gptq
TheBloke/saiga_mistral_7b-GPTQ
Text Generation
•
Updated
•
386
•
8
TheBloke/deepseek-llm-67b-chat-GPTQ
Text Generation
•
Updated
•
74
•
7
TheBloke/deepseek-llm-7B-chat-GPTQ
Text Generation
•
Updated
•
577
•
1
Pi3141/alpaca-7b-native-enhanced-GPTQ
Text Generation
•
Updated
•
2
TheBloke/dolphin-2.5-mixtral-8x7b-GPTQ
Text Generation
•
Updated
•
161
•
110
TheBloke/GEITje-7B-chat-GPTQ
Text Generation
•
Updated
•
36
•
4
astronomer/Llama-3-8B-Instruct-GPTQ-4-Bit
Text Generation
•
Updated
•
8.86k
•
25
neuralmagic/Mistral-7B-Instruct-v0.3-GPTQ-4bit
Text Generation
•
Updated
•
1.1k
•
18
allganize/Llama-3-Alpha-Ko-8B-Instruct-marlin
Text Generation
•
Updated
•
19
•
5
Qwen/Qwen2-7B-Instruct-GPTQ-Int4
Text Generation
•
Updated
•
1.97k
•
24
neuralmagic/Meta-Llama-3-70B-Instruct-quantized.w8a16
Text Generation
•
Updated
•
493
•
4
AI-MO/NuminaMath-7B-TIR-GPTQ
Text Generation
•
Updated
•
301
•
7
pentagoniac/SEMIKONG-8b-GPTQ
Text Generation
•
Updated
•
859
•
26
ModelCloud/Meta-Llama-3.1-8B-Instruct-gptq-4bit
Text Generation
•
Updated
•
1.38k
•
4
shuyuej/Mistral-Nemo-Instruct-2407-GPTQ
Text Generation
•
Updated
•
1.33k
•
5
neuralmagic/Meta-Llama-3.1-8B-Instruct-quantized.w4a16
Text Generation
•
Updated
•
488k
•
24
neuralmagic/Meta-Llama-3.1-70B-Instruct-quantized.w4a16
Text Generation
•
Updated
•
21.6k
•
30
Qwen/Qwen2-VL-7B-Instruct-GPTQ-Int4
Image-Text-to-Text
•
Updated
•
133k
•
32
Qwen/Qwen2-VL-7B-Instruct-GPTQ-Int8
Image-Text-to-Text
•
Updated
•
6.5k
•
26
Qwen/Qwen2-VL-2B-Instruct-GPTQ-Int8
Image-Text-to-Text
•
Updated
•
4.05k
•
13
Qwen/Qwen2-VL-72B-Instruct-GPTQ-Int4
Image-Text-to-Text
•
Updated
•
173k
•
23
Qwen/Qwen2-VL-72B-Instruct-GPTQ-Int8
Image-Text-to-Text
•
Updated
•
2.05k
•
10
Qwen/Qwen2.5-0.5B-Instruct-GPTQ-Int8
Text Generation
•
Updated
•
37.3k
•
8
Qwen/Qwen2.5-7B-Instruct-GPTQ-Int4
Text Generation
•
Updated
•
11.9k
•
14
Qwen/Qwen2.5-7B-Instruct-GPTQ-Int8
Text Generation
•
Updated
•
11.6k
•
12
Qwen/Qwen2.5-14B-Instruct-GPTQ-Int4
Text Generation
•
Updated
•
12.9k
•
14
Qwen/Qwen2.5-32B-Instruct-GPTQ-Int4
Text Generation
•
Updated
•
70.9k
•
24
Qwen/Qwen2.5-72B-Instruct-GPTQ-Int4
Text Generation
•
Updated
•
24.3k
•
32
Qwen/Qwen2.5-72B-Instruct-GPTQ-Int8
Text Generation
•
Updated
•
4.83k
•
18
Qwen/Qwen2.5-Coder-7B-Instruct-GPTQ-Int4
Text Generation
•
Updated
•
8.78k
•
4