Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,114 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
tags:
|
3 |
+
- unsloth
|
4 |
+
license: mit
|
5 |
+
library_name: transformers
|
6 |
+
base_model:
|
7 |
+
- tngtech/DeepSeek-TNG-R1T2-Chimera
|
8 |
+
pipeline_tag: text-generation
|
9 |
+
---
|
10 |
+
<div>
|
11 |
+
<p style="margin-top: 0;margin-bottom: 0;">
|
12 |
+
<em><a href="https://docs.unsloth.ai/basics/unsloth-dynamic-v2.0-gguf">Unsloth Dynamic 2.0</a> achieves superior accuracy & outperforms other leading quants.</em>
|
13 |
+
</p>
|
14 |
+
<div style="display: flex; gap: 5px; align-items: center; ">
|
15 |
+
<a href="https://github.com/unslothai/unsloth/">
|
16 |
+
<img src="https://github.com/unslothai/unsloth/raw/main/images/unsloth%20new%20logo.png" width="133">
|
17 |
+
</a>
|
18 |
+
<a href="https://discord.gg/unsloth">
|
19 |
+
<img src="https://github.com/unslothai/unsloth/raw/main/images/Discord%20button.png" width="173">
|
20 |
+
</a>
|
21 |
+
<a href="https://docs.unsloth.ai/">
|
22 |
+
<img src="https://raw.githubusercontent.com/unslothai/unsloth/refs/heads/main/images/documentation%20green%20button.png" width="143">
|
23 |
+
</a>
|
24 |
+
</div>
|
25 |
+
</div>
|
26 |
+
|
27 |
+
# DeepSeek-TNG-R1T2-Chimera
|
28 |
+
|
29 |
+
<div align="center">
|
30 |
+
<img src="https://354918363417-runtime-assets.s3.eu-central-1.amazonaws.com/company_logo_light.svg"
|
31 |
+
alt="TNG Logo"
|
32 |
+
width="400"
|
33 |
+
style="display: inline-block; vertical-align: middle;"/>
|
34 |
+
</div>
|
35 |
+
<br>
|
36 |
+
<div align="center">
|
37 |
+
<a href="https://huggingface.co/tngtech/DeepSeek-TNG-R1T2-Chimera/blob/main/LICENSE.DeepSeek" style="margin: 2px;">
|
38 |
+
<img alt="License" src="https://img.shields.io/badge/License-MIT-f5de53?&color=f5de53" style="display: inline-block; vertical-align: middle;"/>
|
39 |
+
</a>
|
40 |
+
</div>
|
41 |
+
<br>
|
42 |
+
<div align="center">
|
43 |
+
<img alt="Intelligence Score" src="intelligence_score_vs_output_tokens.png" style="display: inline-block; vertical-align: middle;" width="750"/>
|
44 |
+
</div>
|
45 |
+
|
46 |
+
**Assembly of Experts Chimera model constructed with the DeepSeek [R1-0528](https://huggingface.co/deepseek-ai/DeepSeek-R1-0528), [R1](https://huggingface.co/deepseek-ai/DeepSeek-R1) and [V3-0324](https://huggingface.co/deepseek-ai/DeepSeek-V3-0324) parent models**
|
47 |
+
|
48 |
+
We present our new **DeepSeek-TNG R1T2 Chimera** 671B model, the first successor to our original [*DeepSeek R1T Chimera*](https://huggingface.co/tngtech/DeepSeek-R1T-Chimera) that was released on April 26th. Unlike the original Chimera, which was based on the *two parent models* V3-0324 and R1, the new Chimera is a **Tri-Mind** *with three parents*, namely additionally R1-0528. It is constructed using the Assembly of Experts-method with relatively fine-granular direct brain edits. This more refined assembly allowed, among other improvements, the fixing of the <think> token consistency issue, which was a weakness of R1T and is now solved for R1T2.
|
49 |
+
|
50 |
+
**Sweet spot**
|
51 |
+
|
52 |
+
R1T2 operates at a new sweet spot in intelligence vs. output token length. It appears to be...
|
53 |
+
|
54 |
+
- about **20% faster than** the regular **R1**, and more than **twice as fast as R1-0528**
|
55 |
+
- significantly **more intelligent than** the regular **R1** in benchmarks such as **GPQA** and **AIME-24**
|
56 |
+
- much **more intelligent** and also **think-token consistent** compared to the first **R1T Chimera** 0426
|
57 |
+
- and generally well-behaved and a **nice persona** to talk to, even without any system prompt.
|
58 |
+
|
59 |
+
**Recommendations for your model decision**
|
60 |
+
|
61 |
+
*R1T2* compared...
|
62 |
+
- *vs R1:* We hope that R1T2 is a very desirable, almost universal **better and drop-in replacement for R1**
|
63 |
+
- *vs R1-0528:* R1T2 is a much **cheaper alternative to full R1-0528**, if the fullest 0528-level intelligence is not required
|
64 |
+
- *vs R1T:* R1T2 is usually **recommended over R1T**, unless the specific personality of R1T was optimal, the think-token issue not important, or R1T's higher speed crucial
|
65 |
+
- *vs V3-0324:* V3 is so much faster that if you can live with the **lower intelligence, take V3**, however, if you **need reasoning, R1T2** is the go-to model
|
66 |
+
|
67 |
+
**Limitations**
|
68 |
+
|
69 |
+
- **R1-0528** is thinking much longer, but also is achieving **better hard benchmark results** than R1T2
|
70 |
+
- As measured by SpeechMap.ai (courtesy of xlr8harder), **R1T2** is significantly **more reserved** than R1T, but not as much as R1-0528
|
71 |
+
- Due to the influence of its R1 parent, which does not support function calling, **R1T2 is not yet recommended for function-calling** intensive applications at this stage (this may be fixed at a later stage)
|
72 |
+
- When switching from R1T to R1T2 development, we changed from AIME24 and MT-Bench to AIME24, AIME25 and GPQA-Diamond for the intelligence score. With the new benchmark set, there is a larger score difference between R1 and the original R1T Chimera than published earlier.
|
73 |
+
|
74 |
+
**Technological background**
|
75 |
+
|
76 |
+
For details on the AoE construction process, you can read our [Paper on arXiV](https://arxiv.org/abs/2506.14794).
|
77 |
+
|
78 |
+
|
79 |
+
## Model Details
|
80 |
+
|
81 |
+
- **Architecture**: DeepSeek-MoE transformer-based language model
|
82 |
+
- **Combination Method**: Assembly of Experts from the three DeepSeek parent models R1-0528, R1 and V3-0324
|
83 |
+
- **Release Date**: 2025-07-02
|
84 |
+
- **Design Team**: Robert Dahlke, Henrik Klagges, Benjamin Merkel, Fabian Klemm and David Reiss, Munich, Germany
|
85 |
+
- **Extra Thanks**: Big thanks to DeepSeek for their great models and open-source generosity, and to the other researchers that have published on model merging methodologies.
|
86 |
+
|
87 |
+
|
88 |
+
## Use, Out-of-scope Use, Other Limitations, Risks, Recommendations et al.
|
89 |
+
Regarding the R1T/R1T2-Chimeras, we ask you to follow the careful guidelines that Microsoft has created for their "MAI-DS-R1" DeepSeek-based model.
|
90 |
+
These professional guidelines are available [here on Hugging Face](https://huggingface.co/microsoft/MAI-DS-R1).
|
91 |
+
|
92 |
+
## EU AI Act
|
93 |
+
|
94 |
+
Due to the strict new guidelines of the EU AI Act that take effect on August 2nd 2025, we recommend that each R1T/R1T2 user in the EU either familiarizes themselves with these requirements and assess their compliance, or ceases using the model in the EU after August 1st, 2025.
|
95 |
+
|
96 |
+
## Contact, especially for your user feedback
|
97 |
+
|
98 |
+
Please give us your feedback, especially if you find deficiencies in the model:
|
99 |
+
- Email: [email protected]
|
100 |
+
- X.com: @tngtech
|
101 |
+
|
102 |
+
## Citation
|
103 |
+
|
104 |
+
```
|
105 |
+
@misc{tng_technology_consulting_gmbh_2025_07_0x,
|
106 |
+
author = { TNG Technology Consulting GmbH },
|
107 |
+
title = { DeepSeek-TNG-R1T2-Chimera },
|
108 |
+
year = 2025,
|
109 |
+
month = { July },
|
110 |
+
url = { https://huggingface.co/tngtech/DeepSeek-TNG-R1T2-Chimera },
|
111 |
+
doi = { xxx },
|
112 |
+
publisher = { Hugging Face }
|
113 |
+
}
|
114 |
+
```
|