Titans: Learning to Memorize at Test Time
Abstract
Over more than a decade there has been an extensive research effort on how to effectively utilize recurrent models and attention. While recurrent models aim to compress the data into a fixed-size memory (called hidden state), attention allows attending to the entire context window, capturing the direct dependencies of all tokens. This more accurate modeling of dependencies, however, comes with a quadratic cost, limiting the model to a fixed-length context. We present a neural long-term memory module that learns to memorize historical context and helps attention to attend to the current context while utilizing long-past information. We show that this neural memory has the advantage of fast parallelizable training. From a memory perspective, we argue that attention due to its limited context but accurate dependency modeling performs as a short-term memory, while neural memory due to its ability to memorize the data, acts as a long-term, more persistent, memory. Based on these two modules, we introduce a new family of architectures, called Titans, and present three variants to address how one can effectively incorporate memory into this architecture. Our experimental results on language modeling, common-sense reasoning, and time series tasks show that Titans are effective compared to baselines, while they can effectively scale to larger context window in needle-in-haystack tasks.
Cite
Text
Behrouz et al. "Titans: Learning to Memorize at Test Time." Advances in Neural Information Processing Systems, 2025.Markdown
[Behrouz et al. "Titans: Learning to Memorize at Test Time." Advances in Neural Information Processing Systems, 2025.](https://mlanthology.org/neurips/2025/behrouz2025neurips-titans/)BibTeX
@inproceedings{behrouz2025neurips-titans,
title = {{Titans: Learning to Memorize at Test Time}},
author = {Behrouz, Ali and Zhong, Peilin and Mirrokni, Vahab},
booktitle = {Advances in Neural Information Processing Systems},
year = {2025},
url = {https://mlanthology.org/neurips/2025/behrouz2025neurips-titans/}
}