Text Generation
GGUF
Magistral-Small-2506
bfloat16
thinking
reasoning
all use cases
creative use cases
creative
creative writing
fiction writing
plot generation
sub-plot generation
story generation
scene continue
storytelling
fiction story
science fiction
romance
all genres
story
writing
vivid prose
vivid writing
fiction
roleplaying
swearing
rp
NEO Imatrix
GGUFs
Maxed Output Tensor
imatrix
conversational
Update README.md
Browse files
README.md
CHANGED
@@ -70,6 +70,14 @@ of Mistral's new reasoning model "Magistral-Small-2506":
|
|
70 |
|
71 |
https://huggingface.co/mistralai/Magistral-Small-2506/
|
72 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
73 |
Special thanks to "MLX-Community" for correct config/tokenizer files.
|
74 |
|
75 |
https://huggingface.co/mlx-community/Magistral-Small-2506-bf16
|
|
|
70 |
|
71 |
https://huggingface.co/mistralai/Magistral-Small-2506/
|
72 |
|
73 |
+
These GGUFS are:
|
74 |
+
- Quanted using NEO Imatrix Dataset
|
75 |
+
- The Output Tensor is set at BF16 / 16 bit full precision.
|
76 |
+
- Correct Jijna template which includes "System Prompt" embedded for reasoning.
|
77 |
+
- 32K / 32,768 context max
|
78 |
+
|
79 |
+
An additional repo of GGUFs set at 128k / 131,072 context will follow.
|
80 |
+
|
81 |
Special thanks to "MLX-Community" for correct config/tokenizer files.
|
82 |
|
83 |
https://huggingface.co/mlx-community/Magistral-Small-2506-bf16
|