Leanabell-Prover: Posttraining Scaling in Formal Reasoning
Abstract
Recent advances in automated theorem proving (ATP) through LLMs have highlighted the potential of formal reasoning with Lean 4 codes. However, ATP has not yet be revolutionized by the recent posttraining scaling as demonstrated by Open AI O1/O3 and Deepseek R1. In this work, we investigate the entire posttraining of ATP, aiming to align it with breakthroughs in reasoning models in natural languages.To begin, we continual train current ATP models with a hybrid dataset, which consists of numerous statement-proof pairs, and additional data aimed at incorporating cognitive behaviors that emulate human reasoning and hypothesis refinement. Next, we explore reinforcement learning with the use of outcome reward returned by Lean 4 compiler. Through our designed continual training and reinforcement learning processes, we have successfully improved existing formal provers, including both DeepSeek-Prover-v1.5 and Goedel-Prover, achieving state-of-the-art performance in the field of whole-proof generation. For example, we achieve a 59.8% pass rate (pass@32) on MiniF2F. This is an on-going project and we will progressively update our findings, release our data and training details.
Community
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Theorem Prover as a Judge for Synthetic Data Generation (2025)
- Local Look-Ahead Guidance via Verifier-in-the-Loop for Automated Theorem Proving (2025)
- FANS -- Formal Answer Selection for Natural Language Math Reasoning Using Lean4 (2025)
- LeanProgress: Guiding Search for Neural Theorem Proving via Proof Progress Prediction (2025)
- Goedel-Prover: A Frontier Model for Open-Source Automated Theorem Proving (2025)
- OpenCodeReasoning: Advancing Data Distillation for Competitive Coding (2025)
- Reinforcement Learning for Reasoning in Small LLMs: What Works and What Doesn't (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper