SynLogic
Collection
Synthesizing Verifiable Reasoning Data at Scale for Learning Logical Reasoning and Beyond
•
5 items
•
Updated
•
7
Zero-Mix-3 is an advanced multi-domain reasoning model trained using Zero-RL (reinforcement learning from scratch) on a diverse mixture of logical reasoning, mathematical, and coding data. Built on Qwen2.5-32B-Base, this model demonstrates the power of combining diverse verifiable reasoning tasks in a unified training framework.
Model | BBEH | KOR-Bench | LiveCodeBench | AIME 2024 | GPQA Diamond |
---|---|---|---|---|---|
DeepSeek-R1-Distill-Qwen-32B | 19.2 | 66.6 | 57.2 | 72.6 | 63.1 |
DeepSeek-R1-Zero-Qwen-32B | - | - | 40.2 | 47.0 | 55.0 |
Zero-Mix-2 (Math+Coding) | 18.5 | 58.6 | 39.5 | 34.5 | 55.2 |
Zero-Mix-3 (SynLogic+Math+Coding) | 28.6 | 65.0 | 40.7 | 35.8 | 57.5 |
Key Achievements:
Comparison with Zero-Mix-2 (Math+Coding only) demonstrates that adding SynLogic logical reasoning data:
@misc{liu2025synlogic,
title={SynLogic: Synthesizing Verifiable Reasoning Data at Scale for Learning Logical Reasoning and Beyond},
author={Junteng Liu and Yuanxiang Fan and Zhuo Jiang and Han Ding and Yongyi Hu and Chi Zhang and Yiqi Shi and Shitong Weng and Aili Chen and Shiqi Chen and Yunan Huang and Mozhi Zhang and Pengyu Zhao and Junjie Yan and Junxian He},
year={2025},
eprint={2505.19641},
archivePrefix={arXiv},
primaryClass={cs.AI},
url={https://arxiv.org/abs/2505.19641},
}