InstructZero: Efficient Instruction Optimization for Black-Box Large Language Models

Abstract

Large language models (LLMs) are instruction followers but the performance varies under different instructions. It is challenging to create the best instruction, especially for black-box LLMs on which backpropagation is forbidden. Instead of directly optimizing the discrete instruction, we optimize a low-dimensional soft prompt applied to an open-source LLM to generate the instruction for the black-box LLM. In each optimization step of the proposed method InstructZero, a soft prompt is converted into an instruction by the open-source LLM, which is then submitted to the black-box LLM for zero-shot evaluation, whose result is sent to Bayesian optimization to produce new soft prompts improving the zero-shot performance. We evaluate InstructZero on different combinations of open-source LLMs and APIs including Vicuna and ChatGPT. InstructZero outperforms SOTA auto-instruction methods across a variety of downstream tasks.

Cite

Text

Chen et al. "InstructZero: Efficient Instruction Optimization for Black-Box Large Language Models." International Conference on Machine Learning, 2024.

Markdown

[Chen et al. "InstructZero: Efficient Instruction Optimization for Black-Box Large Language Models." International Conference on Machine Learning, 2024.](https://mlanthology.org/icml/2024/chen2024icml-instructzero/)

BibTeX

@inproceedings{chen2024icml-instructzero,
  title     = {{InstructZero: Efficient Instruction Optimization for Black-Box Large Language Models}},
  author    = {Chen, Lichang and Chen, Jiuhai and Goldstein, Tom and Huang, Heng and Zhou, Tianyi},
  booktitle = {International Conference on Machine Learning},
  year      = {2024},
  pages     = {6503-6518},
  volume    = {235},
  url       = {https://mlanthology.org/icml/2024/chen2024icml-instructzero/}
}