Qwen3-4B-ShiningValiant3-GGUF

Shining Valiant 3 is a specialized model in science, AI design, and general reasoning, built on Qwen 3. It has been fine-tuned using our latest high-difficulty science reasoning dataset, generated with Deepseek R1 0528. Designed as an "AI to build AI," Shining Valiant 3 excels at supporting advanced AI development, innovation, and problem-solving. With enhanced general and creative reasoning capabilities, it offers strong performance in both technical tasks and everyday conversations. Despite its power, its compact size makes it ideal for local desktop and mobile use, while also enabling ultra-fast inference on servers.

Model Files

File Name Size Precision
Qwen3-4B-ShiningValiant3.BF16.gguf 8.05 GB BF16
Qwen3-4B-ShiningValiant3.F16.gguf 8.05 GB F16
Qwen3-4B-ShiningValiant3.F32.gguf 16.1 GB F32
Qwen3-4B-ShiningValiant3.Q2_K.gguf 1.67 GB Q2_K
Qwen3-4B-ShiningValiant3.Q3_K_L.gguf 2.24 GB Q3_K_L
Qwen3-4B-ShiningValiant3.Q3_K_M.gguf 2.08 GB Q3_K_M
Qwen3-4B-ShiningValiant3.Q3_K_S.gguf 1.89 GB Q3_K_S
Qwen3-4B-ShiningValiant3.Q4_K_M.gguf 2.5 GB Q4_K_M
Qwen3-4B-ShiningValiant3.Q4_K_S.gguf 2.38 GB Q4_K_S
Qwen3-4B-ShiningValiant3.Q5_K_M.gguf 2.89 GB Q5_K_M
Qwen3-4B-ShiningValiant3.Q5_K_S.gguf 2.82 GB Q5_K_S
Qwen3-4B-ShiningValiant3.Q6_K.gguf 3.31 GB Q6_K
Qwen3-4B-ShiningValiant3.Q8_0.gguf 4.28 GB Q8_0

Quants Usage

(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants)

Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

image.png

Downloads last month
175
GGUF
Model size
4.02B params
Architecture
qwen3
Hardware compatibility
Log In to view the estimation

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

16-bit

32-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for prithivMLmods/Qwen3-4B-ShiningValiant3-GGUF

Base model

Qwen/Qwen3-4B-Base
Finetuned
Qwen/Qwen3-4B
Quantized
(1)
this model