Enhancing LLM’s Cognition via Structurization
Abstract
When reading long-form text, human cognition is complex and structurized. While large language models (LLMs) process input contexts through a causal and sequential perspective, this approach can potentially limit their ability to handle intricate and complex inputs effectively. To enhance LLM’s cognition capability, this paper presents a novel concept of context structurization. Specifically, we transform the plain, unordered contextual sentences into well-ordered and hierarchically structurized elements. By doing so, LLMs can better grasp intricate and extended contexts through precise attention and information-seeking along the organized structures. Extensive evaluations are conducted across various model architectures and sizes (including a series of auto-regressive LLMs as well as BERT-like masking models) on a diverse set of NLP tasks (e.g., context-based question-answering, exhaustive hallucination evaluation, and passage-level dense retrieval). Empirical results show consistent and significant performance gains afforded by a single-round structurization. In particular, we boost the open-sourced LLaMA2-70B model to achieve comparable performance against GPT-3.5-Turbo as the halluci- nation evaluator. Besides, we show the feasibility of distilling advanced LLMs’ language processing abilities to a smaller yet effective StruXGPT-7B to execute structurization, addressing the practicality of our approach. Code is available at https://github.com/alibaba/struxgpt.
Cite
Text
Liu et al. "Enhancing LLM’s Cognition via Structurization." Neural Information Processing Systems, 2024. doi:10.52202/079017-4249Markdown
[Liu et al. "Enhancing LLM’s Cognition via Structurization." Neural Information Processing Systems, 2024.](https://mlanthology.org/neurips/2024/liu2024neurips-enhancing/) doi:10.52202/079017-4249BibTeX
@inproceedings{liu2024neurips-enhancing,
title = {{Enhancing LLM’s Cognition via Structurization}},
author = {Liu, Kai and Fu, Zhihang and Chen, Chao and Zhang, Wei and Jiang, Rongxin and Zhou, Fan and Chen, Yaowu and Wu, Yue and Ye, Jieping},
booktitle = {Neural Information Processing Systems},
year = {2024},
doi = {10.52202/079017-4249},
url = {https://mlanthology.org/neurips/2024/liu2024neurips-enhancing/}
}