Self-Alignment with Instruction Backtranslation

Abstract

We present a scalable method to build a high quality instruction following language model by automatically labelling human-written text with corresponding instructions. Our approach, named instruction backtranslation, starts with a language model finetuned on a small amount of seed data, and a given web corpus. The seed model is used to construct training examples by generating instruction prompts for web documents (self-augmentation), and then selecting high quality examples from among these candidates (self-curation). This data is then used to finetune a stronger model. Finetuning LLaMa on two iterations of our approach yields a model that outperforms all other LLaMa-based models on the Alpaca leaderboard not relying on distillation data, demonstrating highly effective self-alignment.

Cite

Text

Li et al. "Self-Alignment with Instruction Backtranslation." International Conference on Learning Representations, 2024.

Markdown

[Li et al. "Self-Alignment with Instruction Backtranslation." International Conference on Learning Representations, 2024.](https://mlanthology.org/iclr/2024/li2024iclr-selfalignment/)

BibTeX

@inproceedings{li2024iclr-selfalignment,
  title     = {{Self-Alignment with Instruction Backtranslation}},
  author    = {Li, Xian and Yu, Ping and Zhou, Chunting and Schick, Timo and Levy, Omer and Zettlemoyer, Luke and Weston, Jason E and Lewis, Mike},
  booktitle = {International Conference on Learning Representations},
  year      = {2024},
  url       = {https://mlanthology.org/iclr/2024/li2024iclr-selfalignment/}
}