-
-
-
-
-
-
Inference Providers
Active filters: vLLM
QuantTrio/Qwen3.5-122B-A10B-AWQ
Image-Text-to-Text
• 125B • Updated
• 16.2k
• 8
QuantTrio/Qwen3.5-35B-A3B-AWQ
Image-Text-to-Text
• 36B • Updated
• 14.8k
• 5
QuantTrio/Qwen3.5-397B-A17B-AWQ
Image-Text-to-Text
• Updated
• 3.11k
• 4
QuantTrio/Qwen3.5-27B-AWQ
Image-Text-to-Text
• 28B • Updated
• 15.2k
• 4
QuantTrio/MiniMax-M2.5-AWQ
Text Generation
• 229B • Updated
• 43.9k
• 10
Text Generation
• 586B • Updated
• 50
• 2
QuantTrio/Qwen3-30B-A3B-Thinking-2507-AWQ
Text Generation
• 31B • Updated
• 5.04k
• 4
JunHowie/Qwen3-4B-Instruct-2507-GPTQ-Int4
Text Generation
• 4B • Updated
• 21.9k
• 2
QuantTrio/Qwen3-VL-30B-A3B-Thinking-AWQ
Text Generation
• 31B • Updated
• 3.47k
• 12
QuantTrio/Qwen3-VL-32B-Thinking-AWQ
Image-Text-to-Text
• 33B • Updated
• 1.41k
• 7
QuantTrio/GLM-4.7-Flash-AWQ
Text Generation
• 31B • Updated
• 120k
• 7
QuantTrio/Qwen3-Coder-Next-E336
Text Generation
• 53B • Updated
• 107
• 1
QuantTrio/Qwen3-Coder-Next-E400
Text Generation
• 63B • Updated
• 1.22k
• 2
model-scope/glm-4-9b-chat-GPTQ-Int4
Text Generation
• 9B • Updated
• 79
• 6
model-scope/glm-4-9b-chat-GPTQ-Int8
Text Generation
• 9B • Updated
• 6
• 2
tclf90/qwen2.5-72b-instruct-gptq-int4
Text Generation
• 73B • Updated
• 45
• 2
tclf90/qwen2.5-72b-instruct-gptq-int3
Text Generation
• 69B • Updated
• 73
prithivMLmods/Nu2-Lupi-Qwen-14B
Text Generation
• 15B • Updated
• 2
• 2
mradermacher/Nu2-Lupi-Qwen-14B-GGUF
15B • Updated
• 230
• 1
mradermacher/Nu2-Lupi-Qwen-14B-i1-GGUF
15B • Updated
• 215
• 1
JunHowie/Qwen3-0.6B-GPTQ-Int4
Text Generation
• 0.6B • Updated
• 219
• 1
JunHowie/Qwen3-0.6B-GPTQ-Int8
Text Generation
• 0.6B • Updated
• 35
JunHowie/Qwen3-1.7B-GPTQ-Int4
Text Generation
• 2B • Updated
• 600
• 1
JunHowie/Qwen3-1.7B-GPTQ-Int8
Text Generation
• 2B • Updated
• 5
JunHowie/Qwen3-32B-GPTQ-Int4
Text Generation
• 33B • Updated
• 11.7k
• 4
JunHowie/Qwen3-32B-GPTQ-Int8
Text Generation
• 33B • Updated
• 1.7k
• 4
JunHowie/Qwen3-30B-A3B-GPTQ-Int4
Text Generation
• 5B • Updated
• 8
• 1
JunHowie/Qwen3-14B-GPTQ-Int8
Text Generation
• 15B • Updated
• 417
• 1
JunHowie/Qwen3-14B-GPTQ-Int4
Text Generation
• 15B • Updated
• 1.7k
• 4
JunHowie/Qwen3-8B-GPTQ-Int8
Text Generation
• 8B • Updated
• 106