-
-
-
-
-
-
Inference Providers
Active filters:
4-bit
0xSero/GLM-4.7-REAP-50-W4A16
Text Generation
•
2B
•
Updated
•
1.84k
•
50
0xSero/MiniMax-M2.1-REAP-50-W4A16-REPAIR-IN-PROGRESS
Text Generation
•
17B
•
Updated
•
1.96k
•
27
unsloth/gemma-3-12b-it-bnb-4bit
Any-to-Any
•
13B
•
Updated
•
6.92k
•
23
mlx-community/GLM-4.7-REAP-50-mxfp4
Text Generation
•
185B
•
Updated
•
1.47k
•
22
mlx-community/IQuest-Coder-V1-40B-Loop-Instruct-4bit
Text Generation
•
40B
•
Updated
•
1.38k
•
10
Text Generation
•
358B
•
Updated
•
20.1k
•
18
Intel/GLM-4.7-int4-mixed-AutoRound
Text Generation
•
2B
•
Updated
•
193
•
24
LiquidAI/LFM2.5-1.2B-Instruct-MLX-4bit
Text Generation
•
0.2B
•
Updated
•
220
•
5
LiquidAI/LFM2.5-1.2B-JP-MLX-4bit
Text Generation
•
0.2B
•
Updated
•
104
•
4
Disty0/LTX-2-SDNQ-4bit-dynamic
Updated
•
92
•
4
Disty0/Z-Image-Turbo-SDNQ-uint4-svd-r32
Text-to-Image
•
Updated
•
56.7k
•
52
mlx-community/Youtu-LLM-2B-mlx-4bit
Text Generation
•
0.3B
•
Updated
•
111
•
3
mlx-community/Falcon-H1R-7B-4bit
Text Generation
•
1B
•
Updated
•
190
•
3
MaziyarPanahi/gemma-7b-GGUF
Text Generation
•
9B
•
Updated
•
1.33k
•
15
MaziyarPanahi/Mistral-7B-Instruct-v0.3-GGUF
Text Generation
•
7B
•
Updated
•
137k
•
131
hugging-quants/Meta-Llama-3.1-8B-Instruct-AWQ-INT4
Text Generation
•
8B
•
Updated
•
143k
•
84
MaziyarPanahi/gemma-3-4b-it-GGUF
Text Generation
•
4B
•
Updated
•
166k
•
16
gaunernst/gemma-3-4b-it-int4-awq
Image-Text-to-Text
•
Updated
•
40.3k
•
5
unsloth/gemma-3-12b-it-qat-bnb-4bit
Image-Text-to-Text
•
13B
•
Updated
•
773
•
4
unsloth/Qwen3-14B-unsloth-bnb-4bit
Text Generation
•
15B
•
Updated
•
69k
•
14
unsloth/Qwen3-8B-unsloth-bnb-4bit
8B
•
Updated
•
169k
•
14
unsloth/Qwen3-VL-8B-Instruct-unsloth-bnb-4bit
Image-Text-to-Text
•
9B
•
Updated
•
52.3k
•
15
MaziyarPanahi/Nemotron-Orchestrator-8B-GGUF
Text Generation
•
8B
•
Updated
•
58.3k
•
4
Text-to-Image
•
Updated
•
4
mbakgun/Qwen2.5-Coder-14B-n8n-Workflow-Generator
Text Generation
•
15B
•
Updated
•
941
•
4
QuantTrio/MiniMax-M2.1-AWQ
Text Generation
•
229B
•
Updated
•
6.45k
•
8
tencent/HY-MT1.5-1.8B-GPTQ-Int4
Translation
•
2B
•
Updated
•
820
•
11
tencent/HY-MT1.5-7B-GPTQ-Int4
Translation
•
8B
•
Updated
•
583
•
7
mlx-community/Youtu-LLM-2B-4bit
Text Generation
•
0.3B
•
Updated
•
200
•
3
mlx-community/IQuest-Coder-V1-40B-Instruct-4bit
Text Generation
•
40B
•
Updated
•
925
•
2