Efficient Length-Generalizable Attention via Causal Retrieval for Long-Context Language Modeling

Abstract

Despite the success of Transformers, handling longer contexts remains challenging due to the limited length generalization and quadratic complexity of self-attention, which often requires post-training with a larger attention window, significantly increasing computational and memory costs. In this paper, we propose a novel attention mechanism based on dynamic context, Grouped Cross Attention (GCA), which can generalize to 1000 $\times$ the pre-training context length while maintaining the ability to access distant information with a constant attention window size. For a given input sequence, we split it into chunks and use each chunk to retrieve top-$k$ relevant past chunks for subsequent text generation. Specifically, unlike most previous works that use an off-the-shelf retriever, our key innovation allows the retriever to learn how to retrieve past chunks that better minimize the auto-regressive loss of subsequent tokens in an end-to-end manner, which adapts better to causal language models. Such a mechanism accommodates retrieved chunks with a fixed-size attention window to achieve long-range information access, significantly reducing computational and memory costs during training and inference. Experiments show that GCA-based models achieve near-perfect accuracy in passkey retrieval for 16M context lengths, which is $1000 \times$ the training length.

Cite

Text

Hu et al. "Efficient Length-Generalizable Attention via Causal Retrieval for Long-Context Language Modeling." Proceedings of the 42nd International Conference on Machine Learning, 2025.

Markdown

[Hu et al. "Efficient Length-Generalizable Attention via Causal Retrieval for Long-Context Language Modeling." Proceedings of the 42nd International Conference on Machine Learning, 2025.](https://mlanthology.org/icml/2025/hu2025icml-efficient/)

BibTeX

@inproceedings{hu2025icml-efficient,
  title     = {{Efficient Length-Generalizable Attention via Causal Retrieval for Long-Context Language Modeling}},
  author    = {Hu, Xiang and Teng, Zhihao and Zhao, Jun and Wu, Wei and Tu, Kewei},
  booktitle = {Proceedings of the 42nd International Conference on Machine Learning},
  year      = {2025},
  pages     = {24727-24743},
  volume    = {267},
  url       = {https://mlanthology.org/icml/2025/hu2025icml-efficient/}
}