exllamav2 quantizations of ArliAI's QwQ-32B-ArliAI-RpR-v4.

2.25bpw h6 (10.213 GiB)
3.00bpw h6 (12.938 GiB)
4.00bpw h6 (16.571 GiB)
6.00bpw h6 (23.837 GiB)
8.00bpw h8 (31.254 GiB)
measurement.json

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for MikeRoz/QwQ-32B-ArliAI-RpR-v4-exl2

Base model

Qwen/Qwen2.5-32B
Finetuned
Qwen/QwQ-32B
Quantized
(20)
this model