GreenBitAI/Qwen-3-30B-A3B-layer-mix-bpw-4.0-mlx

This quantized low-bit model GreenBitAI/Qwen-3-30B-A3B-layer-mix-bpw-4.0-mlx was converted to MLX format from GreenBitAI/Qwen-3-30B-A3B-layer-mix-bpw-4.0 using gbx-lm version 0.4.0. Refer to the original model card for more details on the model.

Use with mlx

pip install gbx-lm
from gbx_lm import load, generate

model, tokenizer = load("GreenBitAI/Qwen-3-30B-A3B-layer-mix-bpw-4.0-mlx")

prompt = "hello"

if tokenizer.chat_template is not None:
    messages = [{"role": "user", "content": prompt}]
    prompt = tokenizer.apply_chat_template(
        messages, add_generation_prompt=True
    )

response = generate(model, tokenizer, prompt=prompt, verbose=True)
Downloads last month
20
Safetensors
Model size
4.83B params
Tensor type
BF16
U32
I16
Inference Providers NEW
This model isn't deployed by any Inference Provider. 馃檵 Ask for provider support

Model tree for GreenBitAI/Qwen-3-30B-A3B-layer-mix-bpw-4.0-mlx

Finetuned
(1)
this model