ABEL: Sample Efficient Online Reinforcement Learning for Neural Theorem Proving
Abstract
We propose a scalable and efficient reinforcement learning framework as a strong baseline for theorem proving with limited data. This baseline reaches performances comparable to the current state-of-the-art in theorem proving, while only training on a few hundred examples. This a first step toward an efficient and easily reproducible combination of autoformalization, synthetic data generation and reinforcement learning, which could unlock significant advancements in neural theorem proving.
Cite
Text
Gloeckle et al. "ABEL: Sample Efficient Online Reinforcement Learning for Neural Theorem Proving." NeurIPS 2024 Workshops: MATH-AI, 2024.Markdown
[Gloeckle et al. "ABEL: Sample Efficient Online Reinforcement Learning for Neural Theorem Proving." NeurIPS 2024 Workshops: MATH-AI, 2024.](https://mlanthology.org/neuripsw/2024/gloeckle2024neuripsw-abel/)BibTeX
@inproceedings{gloeckle2024neuripsw-abel,
title = {{ABEL: Sample Efficient Online Reinforcement Learning for Neural Theorem Proving}},
author = {Gloeckle, Fabian and Limperg, Jannis and Synnaeve, Gabriel and Hayat, Amaury},
booktitle = {NeurIPS 2024 Workshops: MATH-AI},
year = {2024},
url = {https://mlanthology.org/neuripsw/2024/gloeckle2024neuripsw-abel/}
}