patrickvonplaten b-a-s-e-d commited on
Commit
9151ac4
·
verified ·
1 Parent(s): a708d64

Some minor readme fixes (#1)

Browse files

- Some minor readme fixes (5ae3eec6fee052d4155f2969952bc22cdb293ddc)


Co-authored-by: based <[email protected]>

Files changed (1) hide show
  1. README.md +9 -9
README.md CHANGED
@@ -37,22 +37,22 @@ pipeline_tag: image-text-to-text
37
 
38
  # Mistral-Small-3.2-24B-Instruct-2506
39
 
40
- Mistral-Small-3.2-24B-Instruct-2506 is a minor update of [Mistral-Small-3.1-24B-Instruct-2503](https://huggingface.co/mistralai/Mistral-Small-3.1-24B-Base-2503).
41
 
42
  Small-3.2 improves in the following categories:
43
  - **Instruction following**: Small-3.2 is better at following precise instructions
44
  - **Repetition errors**: Small-3.2 produces less infinite generations or repetitive answers
45
  - **Function calling**: Small-3.2's function calling template is more robust (see [here](https://github.com/mistralai/mistral-common/blob/535b4d0a0fc94674ea17db6cf8dc2079b81cbcfa/src/mistral_common/tokens/tokenizers/instruct.py#L778) and [examples](#function-calling))
46
 
47
- In all other categories Small-3.2 should match or slightly improve compared to [Mistral-Small-3.1-24B-Instruct-2503](https://huggingface.co/mistralai/Mistral-Small-3.1-24B-Base-2503).
48
 
49
  ## Key Features
50
- - same as [Mistral-Small-3.1-24B-Instruct-2503](https://huggingface.co/mistralai/Mistral-Small-3.1-24B-Base-2503#key-features)
51
 
52
  ## Benchmark Results
53
 
54
- We compare Mistral-Small-3.2-24B to [Mistral-Small-3.1-24B-Instruct-2503](https://huggingface.co/mistralai/Mistral-Small-3.1-24B-Base-2503).
55
- For more comparison against other models of similar size, please check [Mistral-Small-3.1's Benchmarks'](https://huggingface.co/mistralai/Mistral-Small-3.1-24B-Base-2503#benchmark-results)
56
 
57
  ### Text
58
 
@@ -65,7 +65,7 @@ For more comparison against other models of similar size, please check [Mistral-
65
 
66
  #### Infinite Generations
67
 
68
- Small 3.2 reduces infitine generations by 2x on challenging, long and repetitive prompts.
69
 
70
  | Model | Infinite Generations (Internal; Lower is better) |
71
  |-------|-------|
@@ -95,7 +95,7 @@ The model can be used with the following frameworks;
95
 
96
  **Note 1**: We recommend using a relatively low temperature, such as `temperature=0.15`.
97
 
98
- **Note 2**: Make sure to add a system prompt to the model to best tailer it for your needs. If you want to use the model as a general assistant, we recommend to use the one provided in the [SYSTEM_PROMPT.txt](https://huggingface.co/mistralai/Mistral-Small-3.2-24B-Instruct-2506/blob/main/SYSTEM_PROMPT.txt) file.
99
 
100
  ### vLLM (recommended)
101
 
@@ -120,7 +120,7 @@ You can also make use of a ready-to-go [docker image](https://github.com/vllm-pr
120
 
121
  #### Serve
122
 
123
- We recommand that you use Mistral-Small-3.2-24B-Instruct-2506 in a server/client setting.
124
 
125
  1. Spin up a server:
126
 
@@ -136,7 +136,7 @@ vllm serve mistralai/Mistral-Small-3.2-24B-Instruct-2506 --tokenizer_mode mistra
136
 
137
  #### Vision reasoning
138
 
139
- Take leverage of the vision capabilities of Mistral-Small-3.2-24B-Instruct-2506 to take the best choice given a scenario, go catch them all !
140
 
141
  <details>
142
  <summary>Python snippet</summary>
 
37
 
38
  # Mistral-Small-3.2-24B-Instruct-2506
39
 
40
+ Mistral-Small-3.2-24B-Instruct-2506 is a minor update of [Mistral-Small-3.1-24B-Instruct-2503](https://huggingface.co/mistralai/Mistral-Small-3.1-24B-Instruct-2503).
41
 
42
  Small-3.2 improves in the following categories:
43
  - **Instruction following**: Small-3.2 is better at following precise instructions
44
  - **Repetition errors**: Small-3.2 produces less infinite generations or repetitive answers
45
  - **Function calling**: Small-3.2's function calling template is more robust (see [here](https://github.com/mistralai/mistral-common/blob/535b4d0a0fc94674ea17db6cf8dc2079b81cbcfa/src/mistral_common/tokens/tokenizers/instruct.py#L778) and [examples](#function-calling))
46
 
47
+ In all other categories Small-3.2 should match or slightly improve compared to [Mistral-Small-3.1-24B-Instruct-2503](https://huggingface.co/mistralai/Mistral-Small-3.1-24B-Instruct-2503).
48
 
49
  ## Key Features
50
+ - same as [Mistral-Small-3.1-24B-Instruct-2503](https://huggingface.co/mistralai/Mistral-Small-3.1-24B-Instruct-2503#key-features)
51
 
52
  ## Benchmark Results
53
 
54
+ We compare Mistral-Small-3.2-24B to [Mistral-Small-3.1-24B-Instruct-2503](https://huggingface.co/mistralai/Mistral-Small-3.1-24B-Instruct-2503).
55
+ For more comparison against other models of similar size, please check [Mistral-Small-3.1's Benchmarks'](https://huggingface.co/mistralai/Mistral-Small-3.1-24B-Instruct-2503#benchmark-results)
56
 
57
  ### Text
58
 
 
65
 
66
  #### Infinite Generations
67
 
68
+ Small 3.2 reduces infinite generations by 2x on challenging, long and repetitive prompts.
69
 
70
  | Model | Infinite Generations (Internal; Lower is better) |
71
  |-------|-------|
 
95
 
96
  **Note 1**: We recommend using a relatively low temperature, such as `temperature=0.15`.
97
 
98
+ **Note 2**: Make sure to add a system prompt to the model to best tailor it to your needs. If you want to use the model as a general assistant, we recommend to use the one provided in the [SYSTEM_PROMPT.txt](https://huggingface.co/mistralai/Mistral-Small-3.2-24B-Instruct-2506/blob/main/SYSTEM_PROMPT.txt) file.
99
 
100
  ### vLLM (recommended)
101
 
 
120
 
121
  #### Serve
122
 
123
+ We recommend that you use Mistral-Small-3.2-24B-Instruct-2506 in a server/client setting.
124
 
125
  1. Spin up a server:
126
 
 
136
 
137
  #### Vision reasoning
138
 
139
+ Leverage the vision capabilities of Mistral-Small-3.2-24B-Instruct-2506 to make the best choice given a scenario, go catch them all !
140
 
141
  <details>
142
  <summary>Python snippet</summary>