Triangle104 commited on
Commit
c700546
·
verified ·
1 Parent(s): 32af944

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -2
README.md CHANGED
@@ -39,9 +39,15 @@ This model was converted to GGUF format from [`unsloth/Mistral-Small-3.2-24B-Ins
39
  Refer to the [original model card](https://huggingface.co/unsloth/Mistral-Small-3.2-24B-Instruct-2506) for more details on the model.
40
 
41
  ---
42
- Building upon Mistral Small 3.1 (2503), with added reasoning capabilities, undergoing SFT from Magistral Medium traces and RL on top, it's a small, efficient reasoning model with 24B parameters.
43
 
44
- Magistral Small can be deployed locally, fitting within a single RTX 4090 or a 32GB RAM MacBook once quantized.
 
 
 
 
 
 
45
 
46
  ---
47
  ## Use with llama.cpp
 
39
  Refer to the [original model card](https://huggingface.co/unsloth/Mistral-Small-3.2-24B-Instruct-2506) for more details on the model.
40
 
41
  ---
42
+ Mistral-Small-3.2-24B-Instruct-2506 is a minor update of Mistral-Small-3.1-24B-Instruct-2503.
43
 
44
+ Small-3.2 improves in the following categories:
45
+
46
+ - Instruction following: Small-3.2 is better at following precise instructions
47
+ - Repetition errors: Small-3.2 produces less infinite generations or repetitive answers
48
+ - Function calling: Small-3.2's function calling template is more robust (see here and examples)
49
+
50
+ In all other categories Small-3.2 should match or slightly improve compared to Mistral-Small-3.1-24B-Instruct-2503.
51
 
52
  ---
53
  ## Use with llama.cpp