Semantic Self-Consistency: Enhancing Language Model Reasoning via Semantic Weighting
Abstract
While large language models (LLMs) have rapidly improved their performance on a broad number of tasks, they still often fall short on reasoning tasks. As LLMs become more integrated in diverse real-world tasks, advancing their reasoning capabilities is crucial to their effectiveness in nuanced, complex problems. \citet{wang2023selfconsistency}’s \textit{self-consistency} framework reveals that sampling multiple rationales before taking a majority vote reliably improves model performance across various closed-answer reasoning tasks. Standard methods based on this framework aggregate the final decisions of these rationales but fail to utilize the semantic information detailed in the step-by-step reasoning paths. Our work introduces \textit{semantic self-consistency}, enhancing this approach by incorporating and analyzing both the reasoning paths of these rationales in addition to their final decisions before taking a majority vote. These methods not only improve the reliability of reasoning paths but also cause more robust performance on complex reasoning tasks.
Cite
Text
Knappe et al. "Semantic Self-Consistency: Enhancing Language Model Reasoning via Semantic Weighting." NeurIPS 2024 Workshops: MATH-AI, 2024.Markdown
[Knappe et al. "Semantic Self-Consistency: Enhancing Language Model Reasoning via Semantic Weighting." NeurIPS 2024 Workshops: MATH-AI, 2024.](https://mlanthology.org/neuripsw/2024/knappe2024neuripsw-semantic/)BibTeX
@inproceedings{knappe2024neuripsw-semantic,
title = {{Semantic Self-Consistency: Enhancing Language Model Reasoning via Semantic Weighting}},
author = {Knappe, Tim and Li, Ryan Luo and Chauhan, Ayush and Chhua, Kaylee and Zhu, Kevin and O'Brien, Sean},
booktitle = {NeurIPS 2024 Workshops: MATH-AI},
year = {2024},
url = {https://mlanthology.org/neuripsw/2024/knappe2024neuripsw-semantic/}
}