-
-
-
-
-
-
Inference Providers
Active filters:
gptq
pentagoniac/SEMIKONG-8b-GPTQ
Text Generation
•
Updated
•
768
•
27
ChenMnZ/Llama-2-7b-EfficientQAT-w2g64-GPTQ
Text Generation
•
Updated
•
80
•
1
shuyuej/Mixtral-8x22B-Instruct-v0.1-GPTQ
Text Generation
•
Updated
•
8
•
1
ModelCloud/Meta-Llama-3.1-8B-Instruct-gptq-4bit
Text Generation
•
Updated
•
1.45k
•
4
shuyuej/Mistral-Nemo-Instruct-2407-GPTQ
Text Generation
•
Updated
•
12k
•
5
hugging-quants/Meta-Llama-3.1-70B-Instruct-GPTQ-INT4
Text Generation
•
Updated
•
5.35k
•
23
hugging-quants/Meta-Llama-3.1-8B-Instruct-GPTQ-INT4
Text Generation
•
Updated
•
12.5k
•
22
neuralmagic/Meta-Llama-3.1-8B-Instruct-quantized.w4a16
Text Generation
•
Updated
•
133k
•
24
IntelLabs/sqft-qa-sparsepeft-mistral-7b-v0.3-50-gptq-gsm8k-heu
Text Generation
•
Updated
•
316
•
2
IntelLabs/sqft-qa-sparsepeft-mistral-7b-v0.3-50-gptq-math-heu
Text Generation
•
Updated
•
174
•
3
IntelLabs/sqft-qa-sparsepeft-phi-3-mini-4k-50-gptq-math-heu
Text Generation
•
Updated
•
175
•
2
IntelLabs/sqft-qa-sparsepeft-phi-3-mini-4k-50-gptq-cs-heu
Text Generation
•
Updated
•
448
•
2
Qwen/Qwen2-VL-7B-Instruct-GPTQ-Int4
Image-Text-to-Text
•
Updated
•
54.4k
•
35
Qwen/Qwen2-VL-7B-Instruct-GPTQ-Int8
Image-Text-to-Text
•
Updated
•
3.82k
•
29
Qwen/Qwen2-VL-2B-Instruct-GPTQ-Int4
Image-Text-to-Text
•
Updated
•
6.63k
•
24
Qwen/Qwen2-VL-2B-Instruct-GPTQ-Int8
Image-Text-to-Text
•
Updated
•
1.5k
•
14
alexwww94/glm-4v-9b-gptq-4bit
Updated
•
245
•
7
Qwen/Qwen2-VL-72B-Instruct-GPTQ-Int4
Image-Text-to-Text
•
Updated
•
78.8k
•
27
Qwen/Qwen2-VL-72B-Instruct-GPTQ-Int8
Image-Text-to-Text
•
Updated
•
1.88k
•
11
Qwen/Qwen2.5-0.5B-Instruct-GPTQ-Int4
Text Generation
•
Updated
•
2.52k
•
6
Qwen/Qwen2.5-0.5B-Instruct-GPTQ-Int8
Text Generation
•
Updated
•
2.36k
•
9
Qwen/Qwen2.5-1.5B-Instruct-GPTQ-Int4
Text Generation
•
Updated
•
2.31k
•
1
Qwen/Qwen2.5-1.5B-Instruct-GPTQ-Int8
Text Generation
•
Updated
•
737
•
3
Qwen/Qwen2.5-3B-Instruct-GPTQ-Int4
Text Generation
•
Updated
•
49.4k
•
2
Qwen/Qwen2.5-3B-Instruct-GPTQ-Int8
Text Generation
•
Updated
•
1.09k
•
3
Qwen/Qwen2.5-7B-Instruct-GPTQ-Int4
Text Generation
•
Updated
•
50.1k
•
17
Qwen/Qwen2.5-7B-Instruct-GPTQ-Int8
Text Generation
•
Updated
•
29.4k
•
14
Qwen/Qwen2.5-14B-Instruct-GPTQ-Int4
Text Generation
•
Updated
•
69.3k
•
16
Qwen/Qwen2.5-14B-Instruct-GPTQ-Int8
Text Generation
•
Updated
•
21.9k
•
17
Qwen/Qwen2.5-32B-Instruct-GPTQ-Int8
Text Generation
•
Updated
•
97.2k
•
10