metaTextGrad: Learning to Learn with Language Models as Optimizers

Abstract

Large language models (LLMs) are increasingly used in learning algorithms, evaluations, and optimization tasks. Recent studies have shown that incorporating self-criticism into LLMs can significantly enhance model performance, with frameworks such as TextGrad illustrating this approach by iteratively refining model outputs through prompting. However, these frameworks often require extensive hand-crafting and are sensitive to instruction wording. To mitigate these challenges, we propose metaTextGrad, a meta-learning approach for LLM-based optimizers, focusing on learning loss functions and templates for inference-time optimization. Our method significantly improves performance across multiple benchmarks, achieving 5-27% gains on question-answering tasks. These results demonstrate the potential of meta-learning to enhance LLM-based systems, reducing manual tuning and improving generalizability.

Cite

Text

Xu et al. "metaTextGrad: Learning to Learn with Language Models as Optimizers." NeurIPS 2024 Workshops: AFM, 2024.

Markdown

[Xu et al. "metaTextGrad: Learning to Learn with Language Models as Optimizers." NeurIPS 2024 Workshops: AFM, 2024.](https://mlanthology.org/neuripsw/2024/xu2024neuripsw-metatextgrad/)

BibTeX

@inproceedings{xu2024neuripsw-metatextgrad,
  title     = {{metaTextGrad: Learning to Learn with Language Models as Optimizers}},
  author    = {Xu, Guowei and Yuksekgonul, Mert and Guestrin, Carlos and Zou, James},
  booktitle = {NeurIPS 2024 Workshops: AFM},
  year      = {2024},
  url       = {https://mlanthology.org/neuripsw/2024/xu2024neuripsw-metatextgrad/}
}