Triangle104 commited on
Commit
40c33d6
·
verified ·
1 Parent(s): 9df6bfb

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +52 -0
README.md CHANGED
@@ -40,6 +40,58 @@ extra_gated_description: If you want to learn more about how we process your per
40
  This model was converted to GGUF format from [`mistralai/Mistral-Small-3.1-24B-Instruct-2503`](https://huggingface.co/mistralai/Mistral-Small-3.1-24B-Instruct-2503) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
41
  Refer to the [original model card](https://huggingface.co/mistralai/Mistral-Small-3.1-24B-Instruct-2503) for more details on the model.
42
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
43
  ## Use with llama.cpp
44
  Install llama.cpp through brew (works on Mac and Linux)
45
 
 
40
  This model was converted to GGUF format from [`mistralai/Mistral-Small-3.1-24B-Instruct-2503`](https://huggingface.co/mistralai/Mistral-Small-3.1-24B-Instruct-2503) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
41
  Refer to the [original model card](https://huggingface.co/mistralai/Mistral-Small-3.1-24B-Instruct-2503) for more details on the model.
42
 
43
+ ---
44
+ Building upon Mistral Small 3 (2501), Mistral Small 3.1 (2503) adds state-of-the-art vision understanding and enhances long context capabilities up to 128k tokens without compromising text performance.
45
+ With 24 billion parameters, this model achieves top-tier capabilities in both text and vision tasks.
46
+ This model is an instruction-finetuned version of: Mistral-Small-3.1-24B-Base-2503.
47
+
48
+
49
+ Mistral Small 3.1 can be deployed locally and is exceptionally
50
+ "knowledge-dense," fitting within a single RTX 4090 or a 32GB RAM
51
+ MacBook once quantized.
52
+
53
+
54
+ It is ideal for:
55
+
56
+ -Fast-response conversational agents.
57
+
58
+ -Low-latency function calling.
59
+
60
+ -Subject matter experts via fine-tuning.
61
+
62
+ -Local inference for hobbyists and organizations handling sensitive data.
63
+
64
+ -Programming and math reasoning.
65
+
66
+ -Long document understanding.
67
+
68
+ -Visual understanding.
69
+
70
+ For enterprises requiring specialized capabilities (increased
71
+ context, specific modalities, domain-specific knowledge, etc.), we will
72
+ release commercial models beyond what Mistral AI contributes to the
73
+ community.
74
+
75
+ Key Features
76
+ -
77
+
78
+ -Vision: Vision capabilities enable the model to analyze images and provide insights based on visual content in addition to text.
79
+
80
+ -Multilingual: Supports dozens of languages,including English, French, German, Greek, Hindi, Indonesian, Italian, Japanese, Korean, Malay, Nepali, Polish, Portuguese, Romanian, Russian, Serbian, Spanish, Swedish, Turkish, Ukrainian, Vietnamese, Arabic, Bengali, Chinese, Farsi.
81
+
82
+ -Agent-Centric: Offers best-in-class agentic capabilities with native function calling and JSON outputting.
83
+
84
+ -Advanced Reasoning: State-of-the-art conversational and reasoning capabilities.
85
+
86
+ -Apache 2.0 License: Open license allowing usage and modification for both commercial and non-commercial purposes.
87
+
88
+ -Context Window: A 128k context window.
89
+
90
+ -System Prompt: Maintains strong adherence and support for system prompts.
91
+
92
+ -Tokenizer: Utilizes a Tekken tokenizer with a 131k vocabulary size.
93
+
94
+ ---
95
  ## Use with llama.cpp
96
  Install llama.cpp through brew (works on Mac and Linux)
97