Magnushammer: A Transformer-Based Approach to Premise Selection
Abstract
We present Magnushammer: a novel approach to premise selection -- a crucial task in automated theorem proving. Traditionally, symbolic methods that rely on domain knowledge and engineering effort are applied to this task. In contrast, this work demonstrates that contrastive training with the transformer architecture can achieve higher-quality retrieval of relevant premises, without the domain knowledge or feature engineering overhead. Magnushammer outperforms the most advanced and widely used automation tool in interactive theorem proving: Sledgehammer. On the PISA and miniF2F benchmarks Magnushammer achieves $59.5\%$ (against $38.3\%$) and $34.0\%$ (against $20.9\%$) success rates, respectively. By combining Magnushammer with a language-model-based theorem prover, we further improve the state-of-the-art proof success rate from $57.0\%$ to $71.0\%$ on the PISA benchmark. Moreover, we develop and open source a novel, large dataset for premise selection.
Cite
Text
Mikuła et al. "Magnushammer: A Transformer-Based Approach to Premise Selection." NeurIPS 2023 Workshops: MATH-AI, 2023.Markdown
[Mikuła et al. "Magnushammer: A Transformer-Based Approach to Premise Selection." NeurIPS 2023 Workshops: MATH-AI, 2023.](https://mlanthology.org/neuripsw/2023/mikua2023neuripsw-magnushammer/)BibTeX
@inproceedings{mikua2023neuripsw-magnushammer,
title = {{Magnushammer: A Transformer-Based Approach to Premise Selection}},
author = {Mikuła, Maciej and Antoniak, Szymon and Tworkowski, Szymon and Piotrowski, Bartosz and Jiang, Albert and Zhou, Jin Peng and Szegedy, Christian and Kuciński, Łukasz and Miłoś, Piotr and Wu, Yuhuai},
booktitle = {NeurIPS 2023 Workshops: MATH-AI},
year = {2023},
url = {https://mlanthology.org/neuripsw/2023/mikua2023neuripsw-magnushammer/}
}