Create README.md
Browse filesUpdated the README.md file
README.md
ADDED
@@ -0,0 +1,77 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
library_name: onnxruntime_genai
|
3 |
+
license: apache-2.0
|
4 |
+
language:
|
5 |
+
- en
|
6 |
+
- bn
|
7 |
+
- hi
|
8 |
+
- kn
|
9 |
+
- gu
|
10 |
+
- mr
|
11 |
+
- ml
|
12 |
+
- or
|
13 |
+
- pa
|
14 |
+
- ta
|
15 |
+
- te
|
16 |
+
tags:
|
17 |
+
- misteral3
|
18 |
+
- indic
|
19 |
+
- onnx
|
20 |
+
- onnxruntime-genai
|
21 |
+
- sarvam
|
22 |
+
- text-generation-inference
|
23 |
+
- cuda
|
24 |
+
base_model_relation: quantized
|
25 |
+
base_model:
|
26 |
+
- sarvamai/sarvam-m
|
27 |
+
---
|
28 |
+
|
29 |
+
# Sarvam-M
|
30 |
+
<p align="center">
|
31 |
+
<a href="https://dashboard.sarvam.ai/playground"
|
32 |
+
target="_blank" rel="noopener noreferrer">
|
33 |
+
<img
|
34 |
+
src="https://img.shields.io/badge/🚀 Chat on Sarvam Playground-1488CC?style=for-the-badge&logo=rocket"
|
35 |
+
alt="Chat on Sarvam Playground"
|
36 |
+
/>
|
37 |
+
</a>
|
38 |
+
</p>
|
39 |
+
|
40 |
+
|
41 |
+
# Model Information
|
42 |
+
|
43 |
+
`sarvam-m` is a multilingual, hybrid-reasoning, text-only language model built on Mistral-Small. This post-trained version delivers exceptional improvements over the base model:
|
44 |
+
|
45 |
+
- +20% average improvement on Indian language benchmarks
|
46 |
+
- +21.6% enhancement on math benchmarks
|
47 |
+
- +17.6% boost on programming benchmarks
|
48 |
+
|
49 |
+
Performance gains are even more impressive at the intersection of Indian languages and mathematics, with an outstanding +86% improvement in romanized Indian language GSM-8K benchmarks.
|
50 |
+
|
51 |
+
Learn more about sarvam-m in our detailed [blog post](https://www.sarvam.ai/blogs/sarvam-m).
|
52 |
+
|
53 |
+
# Key Features
|
54 |
+
|
55 |
+
- **Hybrid Thinking Mode**: A single versatile model supporting both "think" and "non-think" modes. Use the think mode for complex logical reasoning, mathematical problems, and coding tasks, or switch to non-think mode for efficient, general-purpose conversation.
|
56 |
+
|
57 |
+
- **Advanced Indic Skills**: Specifically post-trained on Indian languages alongside English, embodying a character that authentically reflects and emphasizes Indian cultural values.
|
58 |
+
|
59 |
+
- **Superior Reasoning Capabilities**: Outperforms most similarly-sized models on coding and math benchmarks, demonstrating exceptional reasoning abilities.
|
60 |
+
|
61 |
+
- **Seamless Chatting Experience**: Full support for both Indic scripts and romanized versions of Indian languages, providing a smooth and accessible multilingual conversation experience.
|
62 |
+
|
63 |
+
# Convertion
|
64 |
+
|
65 |
+
The original model is converted to `onnx` using [OnnxRuntime-GenAI](https://github.com/microsoft/onnxruntime-genai) develop by Microsoft .
|
66 |
+
|
67 |
+
# Quickstart
|
68 |
+
|
69 |
+
The following code snippet demonstrates how to use `sarvam-m` using Onnx.
|
70 |
+
|
71 |
+
```python
|
72 |
+
|
73 |
+
```
|
74 |
+
|
75 |
+
> [!NOTE]
|
76 |
+
> For thinking mode, we recommend `temperature=0.5`; for no-think mode, `temperature=0.2`.
|
77 |
+
|