Triangle104 commited on
Commit
46e1de2
·
verified ·
1 Parent(s): fe671c5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +53 -0
README.md CHANGED
@@ -40,6 +40,59 @@ extra_gated_description: If you want to learn more about how we process your per
40
  This model was converted to GGUF format from [`mistralai/Mistral-Small-3.1-24B-Instruct-2503`](https://huggingface.co/mistralai/Mistral-Small-3.1-24B-Instruct-2503) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
41
  Refer to the [original model card](https://huggingface.co/mistralai/Mistral-Small-3.1-24B-Instruct-2503) for more details on the model.
42
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
43
  ## Use with llama.cpp
44
  Install llama.cpp through brew (works on Mac and Linux)
45
 
 
40
  This model was converted to GGUF format from [`mistralai/Mistral-Small-3.1-24B-Instruct-2503`](https://huggingface.co/mistralai/Mistral-Small-3.1-24B-Instruct-2503) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
41
  Refer to the [original model card](https://huggingface.co/mistralai/Mistral-Small-3.1-24B-Instruct-2503) for more details on the model.
42
 
43
+ ---
44
+ Building upon Mistral Small 3 (2501), Mistral Small 3.1 (2503) adds state-of-the-art vision understanding and enhances long context capabilities up to 128k tokens without compromising text performance.
45
+ With 24 billion parameters, this model achieves top-tier capabilities in both text and vision tasks.
46
+ This model is an instruction-finetuned version of: Mistral-Small-3.1-24B-Base-2503.
47
+
48
+
49
+ Mistral Small 3.1 can be deployed locally and is exceptionally
50
+ "knowledge-dense," fitting within a single RTX 4090 or a 32GB RAM
51
+ MacBook once quantized.
52
+
53
+
54
+ It is ideal for:
55
+
56
+ -Fast-response conversational agents.
57
+
58
+ -Low-latency function calling.
59
+
60
+ -Subject matter experts via fine-tuning.
61
+
62
+ -Local inference for hobbyists and organizations handling sensitive data.
63
+
64
+ -Programming and math reasoning.
65
+
66
+ -Long document understanding.
67
+
68
+ -Visual understanding.
69
+
70
+
71
+ For enterprises requiring specialized capabilities (increased
72
+ context, specific modalities, domain-specific knowledge, etc.), we will
73
+ release commercial models beyond what Mistral AI contributes to the
74
+ community.
75
+
76
+ Key Features
77
+ -
78
+
79
+ -Vision: Vision capabilities enable the model to analyze images and provide insights based on visual content in addition to text.
80
+
81
+ -Multilingual: Supports dozens of languages,including English, French, German, Greek, Hindi, Indonesian, Italian, Japanese, Korean, Malay, Nepali, Polish, Portuguese, Romanian, Russian, Serbian, Spanish, Swedish, Turkish, Ukrainian, Vietnamese, Arabic, Bengali, Chinese, Farsi.
82
+
83
+ -Agent-Centric: Offers best-in-class agentic capabilities with native function calling and JSON outputting.
84
+
85
+ -Advanced Reasoning: State-of-the-art conversational and reasoning capabilities.
86
+
87
+ -Apache 2.0 License: Open license allowing usage and modification for both commercial and non-commercial purposes.
88
+
89
+ -Context Window: A 128k context window.
90
+
91
+ -System Prompt: Maintains strong adherence and support for system prompts.
92
+
93
+ -Tokenizer: Utilizes a Tekken tokenizer with a 131k vocabulary size.
94
+
95
+ ---
96
  ## Use with llama.cpp
97
  Install llama.cpp through brew (works on Mac and Linux)
98