EXL3 quantization of Qwen3-14B, 6 bits per weight.

HumanEval (argmax)

Model Q4 Q6 Q8 FP16
Qwen3-14B-exl3-4bpw 88.4 89.0 89.0 89.0
Qwen3-14B-exl3-6bpw 89.6 88.4 89.6 89.0
Qwen3-8B-exl3-4bpw 86.0 85.4 86.0 87.2
Qwen3-8B-exl3-6bpw 84.8 86.0 87.2 87.2
Qwen3-8B-exl3-8bpw-h8 86.0 87.2 86.6 86.6
Qwen3-30B-A3B-exl3-2.25bpw 88.4
Qwen3-30B-A3B-exl3-3bpw 89.6
Qwen3-30B-A3B-exl3-4bpw 92.1
Qwen3-32B-exl3-4bpw 91.5
Downloads last month
70
Safetensors
Model size
6.03B params
Tensor type
FP16
·
I16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for isogen/Qwen3-14B-exl3-6bpw

Finetuned
Qwen/Qwen3-14B
Quantized
(83)
this model