Abstract
The $\alpha$-Flow framework improves few-step generative modeling by unifying and disentangling conflicting objectives, leading to better convergence and state-of-the-art performance on ImageNet-1K.
MeanFlow has recently emerged as a powerful framework for few-step generative modeling trained from scratch, but its success is not yet fully understood. In this work, we show that the MeanFlow objective naturally decomposes into two parts: trajectory flow matching and trajectory consistency. Through gradient analysis, we find that these terms are strongly negatively correlated, causing optimization conflict and slow convergence. Motivated by these insights, we introduce alpha-Flow, a broad family of objectives that unifies trajectory flow matching, Shortcut Model, and MeanFlow under one formulation. By adopting a curriculum strategy that smoothly anneals from trajectory flow matching to MeanFlow, alpha-Flow disentangles the conflicting objectives, and achieves better convergence. When trained from scratch on class-conditional ImageNet-1K 256x256 with vanilla DiT backbones, alpha-Flow consistently outperforms MeanFlow across scales and settings. Our largest alpha-Flow-XL/2+ model achieves new state-of-the-art results using vanilla DiT backbones, with FID scores of 2.58 (1-NFE) and 2.15 (2-NFE).
Community
AlphaFlow introduces a unified framework that generalizes MeanFlow and other few-step generative models, resolving optimization conflicts to achieve faster convergence and state-of-the-art image generation quality on ImageNet.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Transition Models: Rethinking the Generative Learning Objective (2025)
- CMT: Mid-Training for Efficient Learning of Consistency, Mean Flow, and Flow Map Models (2025)
- Advancing End-to-End Pixel Space Generative Modeling via Self-supervised Pre-training (2025)
- Shortcutting Pre-trained Flow Matching Diffusion Models is Almost Free Lunch (2025)
- Score Distillation of Flow Matching Models (2025)
- LayerSync: Self-aligning Intermediate Layers (2025)
- Align Your Tangent: Training Better Consistency Models via Manifold-Aligned Tangents (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper