--- language: - en - fr - de - es - pt - it - ja - ko - ru - zh - ar - fa - id - ms - ne - pl - ro - sr - sv - tr - uk - vi - hi - bn tags: - Magistral-Small-2506 - bfloat16 - thinking - reasoning - all use cases - creative use cases - creative - creative writing - fiction writing - plot generation - sub-plot generation - fiction writing - story generation - scene continue - storytelling - fiction story - science fiction - romance - all genres - story - writing - vivid prose - vivid writing - fiction - roleplaying - bfloat16 - swearing - rp - NEO Imatrix - GGUFs - Maxed Output Tensor license: apache-2.0 base_model: - mistralai/Mistral-Small-3.1-24B-Instruct-2503 --- (quants uploading, examples to be added, model card updates to follow.)

Magistral-Small-2506-Reasoning-24B-NEO-MAX-Imatrix-GGUF

NEO Imatrix quants, with MAX Output tensor (bf16 // full precision) to improve reasoning / output generation of Mistral's new reasoning model "Magistral-Small-2506": https://huggingface.co/mistralai/Magistral-Small-2506/ About these GGUFS: - Quanted using NEO Imatrix Dataset - The Output Tensor is set at BF16 / 16 bit full precision. - Correct Jijna template which includes "System Prompt" embedded for reasoning. - 32K / 32,768 context max (default/set at org repo) - Suggest min context of 4k-8k for reasoning/output.- An additional repo of GGUFs set at 128k / 131,072 context will follow, as per Mistrals notes that the model was trained at 128k max context. Please see notes at: https://huggingface.co/mistralai/Magistral-Small-2506/ For temp, topk, top p and other suggested parameter settings. Special thanks to "MLX-Community" for correct config/tokenizer files. https://huggingface.co/mlx-community/Magistral-Small-2506-bf16 ---