MemReasoner: A Memory-Augmented LLM Architecture for Multi-Hop Reasoning

Abstract

Recent benchmarks suggest that there remains significant room to improve large language models’ ability to robustly reason across facts distributed in extremely long documents. In this work, we propose MemReasoner, a new memory-augmented LLM architecture that is trained to perform temporal reasoning, along with multiple computational steps, over the context stored in the memory. Experiments show that MemReasoner trained on the core reasoning facts generalizes better, when compared to off-the-shelf large language models and existing recurrent models, on a test distribution where the required facts are scattered across long natural text up to 128k tokens. Further, MemReasoner demonstrates robust reasoning performance relative to the baselines, when the answer distribution in test samples differs from that in the training set.

Cite

Text

Ko et al. "MemReasoner: A Memory-Augmented LLM Architecture for Multi-Hop Reasoning." NeurIPS 2024 Workshops: Sys2-Reasoning, 2024.

Markdown

[Ko et al. "MemReasoner: A Memory-Augmented LLM Architecture for Multi-Hop Reasoning." NeurIPS 2024 Workshops: Sys2-Reasoning, 2024.](https://mlanthology.org/neuripsw/2024/ko2024neuripsw-memreasoner/)

BibTeX

@inproceedings{ko2024neuripsw-memreasoner,
  title     = {{MemReasoner: A Memory-Augmented LLM Architecture for Multi-Hop Reasoning}},
  author    = {Ko, Ching-Yun and Dai, Sihui and Das, Payel and Kollias, Georgios and Chaudhury, Subhajit and Lozano, Aurelie},
  booktitle = {NeurIPS 2024 Workshops: Sys2-Reasoning},
  year      = {2024},
  url       = {https://mlanthology.org/neuripsw/2024/ko2024neuripsw-memreasoner/}
}