exllamav3 quantizations of ArliAI's QwQ-32B-ArliAI-RpR-v4.

2.25bpw h6 (10.201 GiB)
3.00bpw h6 (12.925 GiB)
4.00bpw h6 (16.558 GiB)
6.00bpw h6 (23.824 GiB)
8.00bpw h8 (31.271 GiB)

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for MikeRoz/QwQ-32B-ArliAI-RpR-v4-exl3

Base model

Qwen/Qwen2.5-32B
Finetuned
Qwen/QwQ-32B
Quantized
(18)
this model