-
-
-
-
-
-
Inference status
Active filters:
8-bit
MaziyarPanahi/Meta-Llama-3-8B-Instruct-GGUF
Text Generation
•
Updated
•
1.94M
•
83
mlx-community/phi-4-8bit
Text Generation
•
Updated
•
32.6k
•
10
MaziyarPanahi/Mistral-7B-Instruct-v0.3-GGUF
Text Generation
•
Updated
•
1.92M
•
74
MaziyarPanahi/Llama-3.3-70B-Instruct-GGUF
Text Generation
•
Updated
•
1.38M
•
10
huihui-ai/Llama-3.3-70B-Instruct-abliterated-finetuned-GPTQ-Int8
Text Generation
•
Updated
•
235k
•
5
MaziyarPanahi/Qwen2.5-Coder-0.5B-QwQ-draft-GGUF
Text Generation
•
Updated
•
390
•
3
nejumi/phi-4-GPTQ-Int8-calib-ja-1k
Updated
•
69
•
2
lumolabs-ai/Lumo-8B-Instruct
Updated
•
498
•
2
CyberNative/CyberBase-13b
Text Generation
•
Updated
•
229
•
26
MaziyarPanahi/Mixtral-8x22B-v0.1-GGUF
Text Generation
•
Updated
•
1.11M
•
74
MaziyarPanahi/WizardLM-2-7B-GGUF
Text Generation
•
Updated
•
1.86M
•
75
MaziyarPanahi/WizardLM-2-8x22B-GGUF
Text Generation
•
Updated
•
34.8k
•
126
Zoyd/mlabonne_NeuralDaredevil-8B-abliterated-8_0bpw_exl2
Text Generation
•
Updated
•
27
•
2
MaziyarPanahi/Qwen2-7B-Instruct-GGUF
Text Generation
•
Updated
•
1.84M
•
11
kim512/Llama-3-70b-Arimas-story-RP-V1.6-8.0bpw-h8-exl2
Text Generation
•
Updated
•
25
•
1
MaziyarPanahi/firefunction-v2-GGUF
Text Generation
•
Updated
•
1.83M
•
16
neuralmagic/Meta-Llama-3-70B-Instruct-quantized.w8a16
Text Generation
•
Updated
•
303
•
4
MaziyarPanahi/Mistral-Nemo-Instruct-2407-GGUF
Text Generation
•
Updated
•
1.89M
•
40
meta-llama/Llama-Guard-3-8B-INT8
Text Generation
•
Updated
•
1.96k
•
32
MaziyarPanahi/Meta-Llama-3.1-8B-Instruct-GGUF
Text Generation
•
Updated
•
1.83M
•
15
neuralmagic/Meta-Llama-3.1-70B-Instruct-quantized.w8a8
Text Generation
•
Updated
•
7.61k
•
19
LoneStriker/Hermes-3-Llama-3.1-8B-8.0bpw-h8-exl2
Updated
•
9
•
1
Qwen/Qwen2-VL-7B-Instruct-GPTQ-Int8
Image-Text-to-Text
•
Updated
•
6.25k
•
21
Qwen/Qwen2-VL-2B-Instruct-GPTQ-Int8
Image-Text-to-Text
•
Updated
•
3.2k
•
13
HF1BitLLM/Llama3-8B-1.58-100B-tokens
Text Generation
•
Updated
•
1.84k
•
167
MaziyarPanahi/solar-pro-preview-instruct-GGUF
Text Generation
•
Updated
•
1.83M
•
24
Qwen/Qwen2-VL-72B-Instruct-GPTQ-Int8
Image-Text-to-Text
•
Updated
•
5.3k
•
8
Qwen/Qwen2.5-7B-Instruct-GPTQ-Int8
Text Generation
•
Updated
•
23.3k
•
12
Qwen/Qwen2.5-14B-Instruct-GPTQ-Int8
Text Generation
•
Updated
•
25.6k
•
13
Qwen/Qwen2.5-72B-Instruct-GPTQ-Int8
Text Generation
•
Updated
•
5.18k
•
16