SELF-DISCOVER: Large Language Models Self-Compose Reasoning Structures
Abstract
We introduce SELF-DISCOVER, a general framework for LLMs to self-discover the task-intrinsic reasoning structures to tackle complex reasoning problems that are challenging for typical prompting methods. Core to the framework is a self-discovery process where LLMs select multiple atomic reasoning modules such as critical thinking and step-by-step thinking, and compose them into an explicit reasoning structure for LLMs to follow during decoding. SELF-DISCOVER substantially improves GPT-4 and PaLM 2’s performance on challenging reasoning benchmarks such as BigBench-Hard, grounded agent reasoning, and MATH, by as much as 32% compared to Chain of Thought (CoT). Furthermore, SELF-DISCOVER outperforms inference-intensive methods such as CoT-Self-Consistency by more than 20%, while requiring 10-40x fewer inference compute. Finally, we show that the self-discovered reasoning structures are universally applicable across model families: from PaLM 2-L to GPT-4, and from GPT-4 to Llama2, and share commonalities with human reasoning patterns.
Cite
Text
Zhou et al. "SELF-DISCOVER: Large Language Models Self-Compose Reasoning Structures." Neural Information Processing Systems, 2024. doi:10.52202/079017-4004Markdown
[Zhou et al. "SELF-DISCOVER: Large Language Models Self-Compose Reasoning Structures." Neural Information Processing Systems, 2024.](https://mlanthology.org/neurips/2024/zhou2024neurips-selfdiscover/) doi:10.52202/079017-4004BibTeX
@inproceedings{zhou2024neurips-selfdiscover,
title = {{SELF-DISCOVER: Large Language Models Self-Compose Reasoning Structures}},
author = {Zhou, Pei and Pujara, Jay and Ren, Xiang and Chen, Xinyun and Cheng, Heng-Tze and Le, Quoc V. and Chi, Ed H. and Zhou, Denny and Mishra, Swaroop and Zheng, Huaixiu Steven},
booktitle = {Neural Information Processing Systems},
year = {2024},
doi = {10.52202/079017-4004},
url = {https://mlanthology.org/neurips/2024/zhou2024neurips-selfdiscover/}
}