8-bit Quantization of the Qwen3 30B A3B Model

Quantized using GPTQModel

quantiziation config:

quant_config = QuantizeConfig(
    bits=4,
    group_size=32,
    sym=True,
    desc_act=False,
    true_sequential=True,
    pack_dtype=torch.int32,
    damp_percent=0.1       
)
Downloads last month
50
Safetensors
Model size
5.44B params
Tensor type
I32
·
F16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for btbtyler09/Qwen3-30B-A3B-gptq-4bit

Finetuned
Qwen/Qwen3-30B-A3B
Quantized
(97)
this model