CAPO: Cost-Aware Prompt Optimization
Abstract
Large language models (LLMs) have revolutionized natural language processing by solving a wide range of tasks simply guided by a prompt. Yet their performance is highly sensitive to prompt formulation. While automatic prompt optimization addresses this challenge by finding optimal prompts, current methods require a substantial number of LLM calls and input tokens, making prompt optimization expensive. We introduce CAPO (Cost-Aware Prompt Optimization), an algorithm that enhances prompt optimization efficiency by integrating AutoML techniques. CAPO is an evolutionary approach with LLMs as operators, incorporating racing to save evaluations and multi-objective optimization to balance performance with prompt length. It jointly optimizes instructions and few-shot examples while leveraging task descriptions for improved robustness. Our extensive experiments across diverse datasets and LLMs demonstrate that CAPO outperforms state-of-the-art discrete prompt optimization methods in 11/15 cases with improvements up to 21% in accuracy. Our algorithm achieves better performances already with smaller budgets, saves evaluations through racing, and decreases average prompt length via a length penalty, making it both cost-efficient and cost-aware. Even without few-shot examples, CAPO outperforms its competitors and generally remains robust to initial prompts. CAPO represents an important step toward making prompt optimization more powerful and accessible by improving cost-efficiency.
Cite
Text
Zehle et al. "CAPO: Cost-Aware Prompt Optimization." Proceedings of the Fourth International Conference on Automated Machine Learning, 2025. doi:10.48550/arXiv.2504.16005Markdown
[Zehle et al. "CAPO: Cost-Aware Prompt Optimization." Proceedings of the Fourth International Conference on Automated Machine Learning, 2025.](https://mlanthology.org/automl/2025/zehle2025automl-capo/) doi:10.48550/arXiv.2504.16005BibTeX
@inproceedings{zehle2025automl-capo,
title = {{CAPO: Cost-Aware Prompt Optimization}},
author = {Zehle, Tom and Schlager, Moritz and Heiß, Timo and Feurer, Matthias},
booktitle = {Proceedings of the Fourth International Conference on Automated Machine Learning},
year = {2025},
pages = {18/1-45},
doi = {10.48550/arXiv.2504.16005},
volume = {293},
url = {https://mlanthology.org/automl/2025/zehle2025automl-capo/}
}