CAMELoT: Towards Large Language Models with Training-Free Consolidated Associative Memory
Abstract
Large Language Models (LLMs) struggle with long input sequences due to high memory and runtime costs. Memory-augmented models offer a promising solution to this problem, but existing methods have limited memory capacity and require costly re-training to integrate with the LLM. In this work, we introduce CAMELoT, a **C**onsolidated **A**ssociative **M**emory **E**nhanced **Lo**ng **T**ransformer, which has an associative memory (AM) module integrated with any pre-trained attention-based LLM. The AM module in CAMELoT consolidates token representations into a non-parametric distribution model, balancing novelty and recency, therefore giving the LLM the capability to process the long input sequences without any re-training. By retrieving information from AM, CAMELoT achieves a significant perplexity reduction in long-context modeling benchmarks, e.g.,~29.7\% on Arxiv, even with a tiny context window of 128 tokens.
Cite
Text
He et al. "CAMELoT: Towards Large Language Models with Training-Free Consolidated Associative Memory." ICML 2024 Workshops: LCFM, 2024.Markdown
[He et al. "CAMELoT: Towards Large Language Models with Training-Free Consolidated Associative Memory." ICML 2024 Workshops: LCFM, 2024.](https://mlanthology.org/icmlw/2024/he2024icmlw-camelot/)BibTeX
@inproceedings{he2024icmlw-camelot,
title = {{CAMELoT: Towards Large Language Models with Training-Free Consolidated Associative Memory}},
author = {He, Zexue and Karlinsky, Leonid and Kim, Donghyun and McAuley, Julian and Krotov, Dmitry and Feris, Rogerio},
booktitle = {ICML 2024 Workshops: LCFM},
year = {2024},
url = {https://mlanthology.org/icmlw/2024/he2024icmlw-camelot/}
}