SynLogic-Mix-3-32B / README.md
Junteng's picture
Upload README.md with huggingface_hub
ee8bf5d verified
metadata
license: mit
language:
  - en
tags:
  - LLM
library_name: transformers
base_model:
  - Qwen/Qwen2.5-32B
datasets:
  - MiniMaxAI/SynLogic

SynLogic Zero-Mix-3: Large-Scale Multi-Domain Reasoning Model

Model Overview

Zero-Mix-3 is an advanced multi-domain reasoning model trained using Zero-RL (reinforcement learning from scratch) on a diverse mixture of logical reasoning, mathematical, and coding data. Built on Qwen2.5-32B-Base, this model demonstrates the power of combining diverse verifiable reasoning tasks in a unified training framework.

Key Features

  • Multi-Domain Training: Jointly trained on logical reasoning (SynLogic), mathematics, and coding tasks
  • Zero-RL Training: Trained from base model without instruction tuning, using pure reinforcement learning
  • Diverse Data Mixture: 35k mathematical samples + 9k coding samples + 17k SynLogic samples
  • Enhanced Generalization: Superior cross-domain transfer compared to single-domain training

Performance Highlights

Model BBEH KOR-Bench LiveCodeBench AIME 2024 GPQA Diamond
DeepSeek-R1-Distill-Qwen-32B 19.2 66.6 57.2 72.6 63.1
DeepSeek-R1-Zero-Qwen-32B - - 40.2 47.0 55.0
Zero-Mix-2 (Math+Coding) 18.5 58.6 39.5 34.5 55.2
Zero-Mix-3 (SynLogic+Math+Coding) 28.6 65.0 40.7 35.8 57.5

Key Achievements:

  • Matches or Surpasses DeepSeek-R1-Distill-Qwen-32B on KOR-Bench and BBEH (+9.4 points)
  • Outperforms DeepSeek-R1-Zero-Qwen-32B on LiveCodeBench and GPQA-Diamond (+2.5 points)

Training Details

  • Base Model: Qwen2.5-32B-Base
  • Training Algorithm: GRPO (Group Relative Policy Optimization)
  • Training Data:
    • 35k mathematical reasoning samples
    • 9k coding problem samples
    • 17k SynLogic logical reasoning samples

Ablation Insights

Comparison with Zero-Mix-2 (Math+Coding only) demonstrates that adding SynLogic logical reasoning data:

  • +10.1 points on logical reasoning (BBEH)
  • +6.4 points on logical reasoning (KOR-Bench)
  • +2.3 points on out-of-domain reasoning (GPQA-Diamond)
  • +1.2 points on coding (LiveCodeBench)

Citation

@misc{liu2025synlogic,
      title={SynLogic: Synthesizing Verifiable Reasoning Data at Scale for Learning Logical Reasoning and Beyond}, 
      author={Junteng Liu and Yuanxiang Fan and Zhuo Jiang and Han Ding and Yongyi Hu and Chi Zhang and Yiqi Shi and Shitong Weng and Aili Chen and Shiqi Chen and Yunan Huang and Mozhi Zhang and Pengyu Zhao and Junjie Yan and Junxian He},
      year={2025},
      eprint={2505.19641},
      archivePrefix={arXiv},
      primaryClass={cs.AI},
      url={https://arxiv.org/abs/2505.19641}, 
}