Program Synthesis via Test-Time Transduction

Abstract

We introduce transductive program synthesis, a new formulation of the program synthesis task that explicitly leverages test inputs during synthesis. While prior approaches to program synthesis--whether based on natural language descriptions or input-output examples--typically aim to generalize from training examples, they often struggle with robustness, especially in real-world settings where training examples are limited and test inputs involve various edge cases. To address this, we propose a novel framework that improves robustness by treating synthesis as an active learning over a finite hypothesis class defined by programs' outputs. We use an LLM to predict outputs for selected test inputs and eliminate inconsistent hypotheses, where the inputs are chosen via a greedy maximin algorithm to minimize the number of LLM queries required. We evaluate our approach on four benchmarks: Playgol, MBPP+, 1D-ARC, and programmatic world modeling on MiniGrid. We demonstrate that our method significantly improves program synthesis in both accuracy and efficiency. We release our code at https://github.com/klee972/SYNTRA.

Cite

Text

Lee et al. "Program Synthesis via Test-Time Transduction." Advances in Neural Information Processing Systems, 2025.

Markdown

[Lee et al. "Program Synthesis via Test-Time Transduction." Advances in Neural Information Processing Systems, 2025.](https://mlanthology.org/neurips/2025/lee2025neurips-program/)

BibTeX

@inproceedings{lee2025neurips-program,
  title     = {{Program Synthesis via Test-Time Transduction}},
  author    = {Lee, Kang-il and Koo, Jahyun and Yoon, Seunghyun and Kim, Minbeom and Koh, Hyukhun and Lee, Dongryeol and Jung, Kyomin},
  booktitle = {Advances in Neural Information Processing Systems},
  year      = {2025},
  url       = {https://mlanthology.org/neurips/2025/lee2025neurips-program/}
}