Using turboderp-org's ExLlamaV3 v0.0.1 for quantization.
2.00 bits per weight
2.25 bits per weight
2.50 bits per weight
3.00 bits per weight
3.50 bits per weight
4.00 bits per weight
5.00 bits per weight
6.00 bits per weight
8.00 bits per weight
8.00 bits per weight / H8
ELYZA-Thinking-1.0-Qwen-32B-exl3
- Model creator: elyza
- Original model: ELYZA-Thinking-1.0-Qwen-32B
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support
Model tree for RioShiina/ELYZA-Thinking-1.0-Qwen-32B-exl3
Base model
Qwen/Qwen2.5-32B
Finetuned
Qwen/Qwen2.5-32B-Instruct
Finetuned
elyza/ELYZA-Thinking-1.0-Qwen-32B