LLM Program Optimization via Retrieval Augmented Search

Abstract

With the advent of large language models (LLMs), there has been a great deal of interest in applying them to solve difficult programming tasks. Recent work has demonstrated their potential at program optimization, a key challenge in programming languages research. We propose a blackbox adaptation method called Retrieval Augmented Search (RAS) that performs beam search over candidate optimizations; at each step, it retrieves in-context examples from a given training dataset of slow-fast program pairs to guide the LLM. Critically, we find that performing contextual retrieval based on an LLM-generated natural language description significantly outperforms retrieval based on the source code. In addition, we propose a method called AEGIS for improving interpretability by decomposing training examples into "atomic edits" that are significantly more incremental in nature. We show that RAS performs 1.8 $\times$ better than prior state-of-the-art blackbox adaptation strategies, and that AEGIS performs 1.37 $\times$ better while performing significantly smaller edits.

Cite

Text

Anupam et al. "LLM Program Optimization via Retrieval Augmented Search." ICLR 2025 Workshops: DL4C, 2025.

Markdown

[Anupam et al. "LLM Program Optimization via Retrieval Augmented Search." ICLR 2025 Workshops: DL4C, 2025.](https://mlanthology.org/iclrw/2025/anupam2025iclrw-llm/)

BibTeX

@inproceedings{anupam2025iclrw-llm,
  title     = {{LLM Program Optimization via Retrieval Augmented Search}},
  author    = {Anupam, Sagnik and Shypula, Alexander and Bastani, Osbert},
  booktitle = {ICLR 2025 Workshops: DL4C},
  year      = {2025},
  url       = {https://mlanthology.org/iclrw/2025/anupam2025iclrw-llm/}
}