Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,143 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: mit
|
3 |
+
language:
|
4 |
+
- en
|
5 |
+
- zh
|
6 |
+
metrics:
|
7 |
+
- accuracy
|
8 |
+
base_model:
|
9 |
+
- Qwen/Qwen3-32B
|
10 |
+
pipeline_tag: text-generation
|
11 |
+
library_name: transformers
|
12 |
+
tags:
|
13 |
+
- blockchain
|
14 |
+
- conversational
|
15 |
+
- web3
|
16 |
+
- qwen3
|
17 |
+
|
18 |
+
eval_results:
|
19 |
+
- task: domain-specific evaluation
|
20 |
+
dataset: DMindAI/DMind_Benchmark
|
21 |
+
metric: normalized web3 score
|
22 |
+
score: 77.44
|
23 |
+
model: DMind-1
|
24 |
+
model_rank: 1 / 24
|
25 |
+
---
|
26 |
+
|
27 |
+
|
28 |
+
|
29 |
+
|
30 |
+
<div align="center">
|
31 |
+
<img src="figures/dmind-ai-logo.png" width="60%" alt="DMind-1" />
|
32 |
+
</div>
|
33 |
+
|
34 |
+
<hr>
|
35 |
+
|
36 |
+
<div align="center">
|
37 |
+
<a href="https://img.shields.io/badge/DMind-Homepage-blue?logo=data:image/svg+xml;base64,)">
|
38 |
+
<img alt="DMind Website" src="https://img.shields.io/badge/DMind-Homepage-blue?logo=data:image/svg+xml;base64,)"/>
|
39 |
+
</a>
|
40 |
+
<a href="https://huggingface.co/datasets/DMindAI/DMind_Benchmark">
|
41 |
+
<img alt="Hugging Face" src="https://img.shields.io/badge/HuggingFace-DMind_Benchmark-yellow?logo=huggingface"/>
|
42 |
+
</a>
|
43 |
+
<a href="https://x.com/dmindai">
|
44 |
+
<img alt="X" src="https://img.shields.io/badge/X-@dmindai-1DA1F2?logo=x"/>
|
45 |
+
</a>
|
46 |
+
<a href="https://www.apache.org/licenses/LICENSE-2.0">
|
47 |
+
<img alt="Apache 2.0" src="https://img.shields.io/badge/Apache%202.0-blue.svg"/>
|
48 |
+
</a>
|
49 |
+
<a href="https://opensource.org/licenses/MIT">
|
50 |
+
<img alt="Code License: MIT" src="https://img.shields.io/badge/Code%20License-MIT-yellow.svg"/>
|
51 |
+
</a>
|
52 |
+
<a href="MODEL-LICENSE">
|
53 |
+
<img alt="Model License: Model Agreement" src="https://img.shields.io/badge/Model%20License-Model%20Agreement-yellow.svg"/>
|
54 |
+
</a>
|
55 |
+
</div>
|
56 |
+
|
57 |
+
## Table of Contents
|
58 |
+
- [Introduction](#introduction)
|
59 |
+
- [1. Model Overview](#1-model-overview)
|
60 |
+
- [2. Evaluation Results](#2-evaluation-results)
|
61 |
+
- [3. Use Cases](#3-use-cases)
|
62 |
+
- [4. Quickstart](#4-quickstart)
|
63 |
+
- [4.1 Model Downloads](#41-model-downloads)
|
64 |
+
- [4.2 OpenRouter API](#42-openrouter-api)
|
65 |
+
- [Contact](#contact)
|
66 |
+
|
67 |
+
## Introduction
|
68 |
+
The rapid growth of Web3 technologies—blockchain, DeFi, and smart contracts—demands specialized AI large language models (LLMs) with precise domain alignment and advanced reasoning capabilities. However, General-purpose LLMs often lack the domain-specific accuracy, nuanced reasoning, and instruction-following aligned with expert expectations.
|
69 |
+
|
70 |
+
To address these limitations, we introduce **DMind-1**, a domain-specialized LLM fine-tuned for the Web3 ecosystem via supervised instruction tuning and reinforcement learning from human feedback (RLHF). Built on a powerful base model, DMind-1 achieves strong improvements in task accuracy, content safety, and expert-aligned interaction, significantly surpassing general-purpose models. DMind-1 represents a robust foundation for intelligent agents in the Web3 ecosystem.
|
71 |
+
|
72 |
+
To support real-time and resource-constrained applications, we further release **DMind-1-mini**, a compact variant distilled from both DMind-1 and a generalist LLM using a multi-level distillation framework. It retains key domain reasoning abilities while operating with significantly lower computational overhead.
|
73 |
+
|
74 |
+
## 1. Model Overview
|
75 |
+
|
76 |
+
### DMind-1
|
77 |
+
DMind-1 is a specialized Web3 expert model built on the Qwen3-32B base. Leveraging a state-of-the-art transformer architecture, it integrates deep domain knowledge through a novel two-stage fine-tuning pipeline, establishing its distinctive strengths in Web3-specific applications.
|
78 |
+
|
79 |
+
**Key Points:**
|
80 |
+
- **Comprehensive Domain Expertise Data**: In the first stage, DMind-1 underwent Supervised Fine-Tuning (SFT) on 13,276 expert-curated knowledge items distilled from 32.7GB of Web3 documentation, covering 8 key subdomains including DeFi, tokenomics, governance, and smart contracts. These data points were extracted and structured by a team of domain experts to ensure both depth and accuracy. To enable efficient and scalable training, we employed Low-Rank Adaptation (LoRA) during the SFT stage, allowing DMind-1 to internalize specialized Web3 knowledge while preserving the general-language capabilities of its base model.
|
81 |
+
|
82 |
+
|
83 |
+
- **Reinforcement Learning from Human Feedback (RLHF)**
|
84 |
+
To further align the model with expert expectations in realistic interaction scenarios and accuracy, we implemented an RLHF phase composed of:
|
85 |
+
- **Reward Model Training**: We trained a domain-specific reward model using preference-ranked outputs collected from human experts across diverse Web3-specific question-answer and interaction scenarios. This model learned to assess which responses best reflect factual accuracy and expert-level reasoning in the Web3 domain.
|
86 |
+
- **Policy Optimization with PPO**: Building on the SFT model, we fine-tuned Qwen3-32B using Proximal Policy Optimization (PPO), guided by the trained reward model. The policy network was optimized based on feedback from simulated Web3 dialogue environments, while LoRA ensured resource-efficient parameter updates and significantly reduced compute and memory requirements. This dual-stage approach enabled efficient fine-tuning of a larger model on Web3-specific tasks while achieving high alignment with human intent.
|
87 |
+
|
88 |
+
|
89 |
+
- **Domain-Aligned Reasoning and Interaction**:
|
90 |
+
DMind-1 exhibits advanced web3-aligned reasoning and interactive capabilities in the following fields:
|
91 |
+
- **Natural Dialogue Fluency**: Coherent, context-aware conversations on complex Web3 topics, with strong multi-turn consistency.
|
92 |
+
|
93 |
+
- **Complex Instruction Following**: Reliable execution of multi-step instructions and conditional logic, supporting agent-driven workflows.
|
94 |
+
|
95 |
+
- **Safe and Compliant Content Generation**: Outputs are aligned with domain-specific safety, ethics, and regulatory standards.
|
96 |
+
|
97 |
+
### DMind-1-mini
|
98 |
+
|
99 |
+
To address scenarios requiring lower latency and faster inference, we also introduce **DMind-1-mini**, a lightweight distilled version of DMind-1 based on Qwen3-14B.
|
100 |
+
DMind-1-mini is trained using knowledge distillation and our custom **DeepResearch** framework, drawing from two teacher models:
|
101 |
+
- **DMind-1** (Qwen3-32B): Our in-house Web3 expert model.
|
102 |
+
- **GPT-o3 + DeepResearch**: A general-purpose SOTA LLM with broad capabilities.
|
103 |
+
|
104 |
+
The **Distillation pipeline** combines:
|
105 |
+
|
106 |
+
- **Web3-specific data distillation**, using filtered instruction-following and QA samples generated by the teachers
|
107 |
+
|
108 |
+
- **Soft-label supervision**, encouraging the student to match the teachers’ output distributions and uncertainty
|
109 |
+
|
110 |
+
- **Intermediate representation alignment**, transferring structural knowledge from selected internal layers
|
111 |
+
|
112 |
+
|
113 |
+
This multi-level distillation strategy allows DMind-1-mini to maintain high Web3 task performance with significantly reduced computational overhead and latency, making it suitable for real-time applications such as instant Q&A and on-chain analytics, and lightweight agent deployment.
|
114 |
+
|
115 |
+
|
116 |
+
## 2. Evaluation Results
|
117 |
+
|
118 |
+

|
119 |
+
|
120 |
+
We evaluate DMind-1 using the **DMind Benchmark**, a domain-specific evaluation suite tailored to assess large language models in the Web3 context. The benchmark spans 1,917 expert-reviewed questions across nine critical categories—including Blockchain Fundamentals, Infrastructure, Smart Contracts, DeFi, DAO, NFT, Token Economics, Meme, and Security. It combines multiple-choice and subjective open-ended tasks, simulating real-world challenges and requiring deep contextual understanding, which provides a comprehensive assessment of both factual knowledge and advanced reasoning.
|
121 |
+
|
122 |
+
Under this rigorous evaluation, DMind-1 ranked 1st among 24 leading models, outperforming both proprietary (e.g., Grok-3) and open-source (e.g., DeepSeek-R1) LLMs. Notably, our distilled variant DMind-1-mini also performed strongly, ranking 2nd overall. This demonstrates the effectiveness of our compact distillation pipeline.
|
123 |
+
|
124 |
+
|
125 |
+
## 3. Use Cases
|
126 |
+
- **Expert-Level Question & Answering**: Provides accurate, context-aware answers on blockchain, DeFi, smart contracts, and related Web3 topics
|
127 |
+
- **Compliance-Aware Support**: Assists in drafting or reviewing content within regulatory and legal contexts
|
128 |
+
- **Content Generation in Domain**: Produces Web3-specific blog posts, documentation, and tutorials tailored to developers and users
|
129 |
+
- **DeFi Strategy Suggestions**: Generates insights and recommendations for yield farming, liquidity provision, and portfolio strategies based on user-provided data
|
130 |
+
- **Risk Management**: Suggests strategies aligned with user risk profiles for more informed decision-making in volatile markets
|
131 |
+
|
132 |
+
## 4. Quickstart
|
133 |
+
|
134 |
+
### 4.1 Model Downloads
|
135 |
+
|
136 |
+
| **Model** | **#Total Params** | **#Activated Params** | **Context Length** | **Download** |
|
137 |
+
|:--------------:|:----------------:|:--------------------:|:-----------------:|:----------------------------------------------------------------------------:|
|
138 |
+
| DMind-1 | | | | [Hugging Face Link](https://huggingface.co/dmind-ai/dmind-1-base) |
|
139 |
+
| DMind-1-mini | | | | [Hugging Face Link](https://huggingface.co/dmind-ai/dmind-1) |
|
140 |
+
|
141 |
+
### 4.2 OpenRouter API
|
142 |
+
## Contact
|
143 |
+
For questions or support, please contact [email protected]
|