FormalAlign: Automated Alignment Evaluation for Autoformalization

Abstract

Autoformalization aims to convert informal mathematical proofs into machine-verifiable formats, bridging the gap between natural and formal languages. However, ensuring semantic alignment between the informal and formalized statements remains challenging. Existing approaches heavily rely on manual verification, hindering scalability. To address this, we introduce FormalAlign, a framework for automatically evaluating the alignment between natural and formal languages in autoformalization. FormalAlign trains on both the autoformalization sequence generation task and the representational alignment between input and output, employing a dual loss that combines a pair of mutually enhancing autoformalization and alignment tasks. Evaluated across four benchmarks augmented by our proposed misalignment strategies, FormalAlign demonstrates superior performance. In our experiments, FormalAlign outperforms GPT-4, achieving an Alignment-Selection Score 11.58\% higher on \forml-Basic (99.21\% vs. 88.91\%) and 3.19\% higher on MiniF2F-Valid (66.39\% vs. 64.34\%). This effective alignment evaluation significantly reduces the need for manual verification.

Cite

Text

Lu et al. "FormalAlign: Automated Alignment Evaluation for Autoformalization." International Conference on Learning Representations, 2025.

Markdown

[Lu et al. "FormalAlign: Automated Alignment Evaluation for Autoformalization." International Conference on Learning Representations, 2025.](https://mlanthology.org/iclr/2025/lu2025iclr-formalalign/)

BibTeX

@inproceedings{lu2025iclr-formalalign,
  title     = {{FormalAlign: Automated Alignment Evaluation for Autoformalization}},
  author    = {Lu, Jianqiao and Wan, Yingjia and Huang, Yinya and Xiong, Jing and Liu, Zhengying and Guo, Zhijiang},
  booktitle = {International Conference on Learning Representations},
  year      = {2025},
  url       = {https://mlanthology.org/iclr/2025/lu2025iclr-formalalign/}
}