-
-
-
-
-
-
Inference Providers
Active filters:
4-bit
Intel/GLM-4.7-int4-mixed-AutoRound
Text Generation
•
2B
•
Updated
•
89
•
14
tencent/HY-MT1.5-1.8B-GPTQ-Int4
Translation
•
2B
•
Updated
•
291
•
9
QuantTrio/MiniMax-M2.1-AWQ
Text Generation
•
229B
•
Updated
•
1.17k
•
5
tencent/HY-MT1.5-7B-GPTQ-Int4
Translation
•
8B
•
Updated
•
226
•
5
Text Generation
•
358B
•
Updated
•
10.6k
•
12
QuantTrio/GLM-4.7-GPTQ-Int4-Int8Mix
Text Generation
•
390B
•
Updated
•
126
•
4
mlx-community/MiniMax-M2.1-4bit
Text Generation
•
229B
•
Updated
•
866
•
4
Disty0/Qwen-Image-Edit-2511-SDNQ-uint4-svd-r32
Image-to-Image
•
Updated
•
281
•
6
TheBloke/CapybaraHermes-2.5-Mistral-7B-GPTQ
7B
•
Updated
•
192
•
61
unsloth/Phi-3-mini-4k-instruct-bnb-4bit
Text Generation
•
4B
•
Updated
•
35k
•
42
lmstudio-community/Qwen2.5-Coder-7B-Instruct-MLX-4bit
Text Generation
•
1B
•
Updated
•
1.94k
•
3
ICEPVP8977/Uncensored_Qwen2.5_Coder_7B_4_bit_quantized_Seaftensors
8B
•
Updated
•
54
•
3
lmstudio-community/Devstral-Small-2507-MLX-4bit
Text Generation
•
24B
•
Updated
•
26.4k
•
5
Intel/Qwen3-Next-80B-A3B-Instruct-int4-mixed-AutoRound
Text Generation
•
Updated
•
18.9k
•
23
nota-ai/Qwen3-30B-A3B-NotaMoEQuant-Int4
Text Generation
•
0.6B
•
Updated
•
136
•
4
nota-ai/GLM-4.5-Air-NotaMoeQuant-Int4
Text Generation
•
1B
•
Updated
•
63
•
2
nightmedia/Qwen3-4B-Agent-F32-dwq4-mlx
Text Generation
•
0.8B
•
Updated
•
212
•
2
Text-to-Speech
•
0.5B
•
Updated
•
34
•
2
TheBloke/WizardLM-33B-V1-0-Uncensored-SuperHOT-8K-GPTQ
Text Generation
•
33B
•
Updated
•
39
•
93
MaziyarPanahi/TheTop-5x7B-Instruct-S5-v0.1-GGUF
Text Generation
•
7B
•
Updated
•
32
•
3
MaziyarPanahi/gemma-7b-GGUF
Text Generation
•
9B
•
Updated
•
1.35k
•
12
CohereLabs/c4ai-command-r-v01-4bit
Text Generation
•
35B
•
Updated
•
33
•
176
Qwen/Qwen1.5-MoE-A2.7B-Chat-GPTQ-Int4
Text Generation
•
14B
•
Updated
•
1.19k
•
49
SweatyCrayfish/llama-3-8b-quantized
Text Generation
•
8B
•
Updated
•
32
•
•
12
solidrust/Llama-3-8B-Lexi-Uncensored-AWQ
Text Generation
•
8B
•
Updated
•
99.5k
•
4
MaziyarPanahi/Mistral-7B-Instruct-v0.3-GGUF
Text Generation
•
7B
•
Updated
•
169k
•
129
Intel/Qwen2-0.5B-Instuct-int4-inc
Text Generation
•
0.6B
•
Updated
•
4
•
1
Intel/Qwen2-1.5B-Instuct-int4-inc
Text Generation
•
2B
•
Updated
•
4
•
3
MaziyarPanahi/Mistral-Nemo-Instruct-2407-GGUF
Text Generation
•
12B
•
Updated
•
165k
•
50
hugging-quants/Meta-Llama-3.1-8B-Instruct-AWQ-INT4
Text Generation
•
8B
•
Updated
•
177k
•
82