metaTextGrad: Learning to Learn with Language Models as Optimizers
Abstract
Large language models (LLMs) are increasingly used in learning algorithms, evaluations, and optimization tasks. Recent studies have shown that incorporating self-criticism into LLMs can significantly enhance model performance, with frameworks such as TextGrad illustrating this approach by iteratively refining model outputs through prompting. However, these frameworks often require extensive hand-crafting and are sensitive to instruction wording. To mitigate these challenges, we propose metaTextGrad, a meta-learning approach for LLM-based optimizers, focusing on learning loss functions and templates for inference-time optimization. Our method significantly improves performance across multiple benchmarks, achieving 5-27% gains on question-answering tasks. These results demonstrate the potential of meta-learning to enhance LLM-based systems, reducing manual tuning and improving generalizability.
Cite
Text
Xu et al. "metaTextGrad: Learning to Learn with Language Models as Optimizers." NeurIPS 2024 Workshops: AFM, 2024.Markdown
[Xu et al. "metaTextGrad: Learning to Learn with Language Models as Optimizers." NeurIPS 2024 Workshops: AFM, 2024.](https://mlanthology.org/neuripsw/2024/xu2024neuripsw-metatextgrad/)BibTeX
@inproceedings{xu2024neuripsw-metatextgrad,
title = {{metaTextGrad: Learning to Learn with Language Models as Optimizers}},
author = {Xu, Guowei and Yuksekgonul, Mert and Guestrin, Carlos and Zou, James},
booktitle = {NeurIPS 2024 Workshops: AFM},
year = {2024},
url = {https://mlanthology.org/neuripsw/2024/xu2024neuripsw-metatextgrad/}
}