This is basically a test to see if the conversion and inference in llama.cpp works fine It seems to work though i wont add more quant sizes for now

Since this is merely a quantization of the original model the license of the original model still applies!

Downloads last month
92
GGUF
Model size
30.5B params
Architecture
qwen3moe
Hardware compatibility
Log In to view the estimation

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for QuantStack/InternVL3_5-30B-A3B-gguf

Collection including QuantStack/InternVL3_5-30B-A3B-gguf