Cortex Dual ~ DiMind
Collection
(direct, reactive, retrieval-based responses), (reasoning, planning, deeper analysis)
•
4 items
•
Updated
•
1
Magpie-Qwen-CortexDual-0.6B is a specialized, general-purpose model designed for math, code, and structured reasoning. Built with CortexDual thinking mode, it dynamically adapts to the complexity of a problem, automatically shifting into a stepwise reasoning mode for intricate logic or math tasks. This 0.6B parameter model leverages 80% of the Magpie Pro 330k dataset and a modular blend of datasets for general-purpose proficiency and domain versatility.
File Name | Size | Source |
---|---|---|
Magpie-Qwen-0.6B.BF16.gguf | 1.2 GB | xet |
Magpie-Qwen-0.6B.F16.gguf | 1.2 GB | xet |
Magpie-Qwen-0.6B.F32.gguf | 2.39 GB | xet |
Magpie-Qwen-0.6B.Q4_K_M.gguf | 397 MB | xet |
Magpie-Qwen-0.6B.Q5_K_M.gguf | 444 MB | xet |
Magpie-Qwen-0.6B.Q8_0.gguf | 639 MB | xet |
.gitattributes | 1.97 kB | - |
README.md | 723 Bytes | - |
config.json | 31 Bytes | - |
(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants)
Link | Type | Size/GB | Notes |
---|---|---|---|
GGUF | Q2_K | 0.4 | |
GGUF | Q3_K_S | 0.5 | |
GGUF | Q3_K_M | 0.5 | lower quality |
GGUF | Q3_K_L | 0.5 | |
GGUF | IQ4_XS | 0.6 | |
GGUF | Q4_K_S | 0.6 | fast, recommended |
GGUF | Q4_K_M | 0.6 | fast, recommended |
GGUF | Q5_K_S | 0.6 | |
GGUF | Q5_K_M | 0.7 | |
GGUF | Q6_K | 0.7 | very good quality |
GGUF | Q8_0 | 0.9 | fast, best quality |
GGUF | f16 | 1.6 | 16 bpw, overkill |
Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):
4-bit
5-bit
8-bit
16-bit
32-bit
Base model
Qwen/Qwen3-0.6B-Base