-
-
-
-
-
-
Inference Providers
Active filters: quantllm
codewithdark/Llama-3.2-3B-4bit
3B • Updated
• 8
codewithdark/Llama-3.2-3B-GGUF-4bit
3B • Updated
• 5
codewithdark/Llama-3.2-3B-4bit-mlx
Text Generation
• 3B • Updated
• 65
QuantLLM/Llama-3.2-3B-4bit-mlx
Text Generation
• 3B • Updated
• 15
QuantLLM/Llama-3.2-3B-2bit-mlx
Text Generation
• 3B • Updated
• 17
QuantLLM/Llama-3.2-3B-8bit-mlx
Text Generation
• 3B • Updated
• 25
QuantLLM/Llama-3.2-3B-5bit-mlx
Text Generation
• 3B • Updated
• 14
QuantLLM/Llama-3.2-3B-5bit-gguf
3B • Updated
• 12
QuantLLM/Llama-3.2-3B-2bit-gguf
3B • Updated
• 3
QuantLLM/functiongemma-270m-it-8bit-gguf
0.3B • Updated
• 3
• 1
QuantLLM/functiongemma-270m-it-4bit-gguf
0.3B • Updated
• 4
QuantLLM/functiongemma-270m-it-4bit-mlx
Text Generation
• 0.3B • Updated
• 10