Quantifying Memory Utilization with Effective State-Size

Abstract

As the space of causal sequence modeling architectures continues to grow, the need to develop a general framework for their analysis becomes increasingly important. With this aim, we draw insights from classical signal processing and control theory, to develop a quantitative measure of memory utilization: the internal mechanisms through which a model stores past information to produce future outputs. This metric, which we call effective state-size (ESS), is tailored to the fundamental class of systems with input-invariant and input-varying linear operators, encompassing a variety of computational units such as variants of attention, convolutions, and recurrences. Unlike prior work on memory utilization, which either relies on raw operator visualizations (e.g. attention maps), or simply the total memory capacity (i.e. cache size) of a model, our metrics provide highly interpretable and actionable measurements. In particular, we show how ESS can be leveraged to improve initialization strategies, inform novel regularizers and advance the performance-efficiency frontier through model distillation. Furthermore, we demonstrate that the effect of context delimiters (such as end-of-speech tokens) on ESS highlights cross-architectural differences in how large language models utilize their available memory to recall information. Overall, we find that ESS provides valuable insights into the dynamics that dictate memory utilization, enabling the design of more efficient and effective sequence models.

Cite

Text

Parnichkun et al. "Quantifying Memory Utilization with Effective State-Size." Proceedings of the 42nd International Conference on Machine Learning, 2025.

Markdown

[Parnichkun et al. "Quantifying Memory Utilization with Effective State-Size." Proceedings of the 42nd International Conference on Machine Learning, 2025.](https://mlanthology.org/icml/2025/parnichkun2025icml-quantifying/)

BibTeX

@inproceedings{parnichkun2025icml-quantifying,
  title     = {{Quantifying Memory Utilization with Effective State-Size}},
  author    = {Parnichkun, Rom and Tumma, Neehal and Thomas, Armin W and Moro, Alessandro and An, Qi and Suzuki, Taiji and Yamashita, Atsushi and Poli, Michael and Massaroli, Stefano},
  booktitle = {Proceedings of the 42nd International Conference on Machine Learning},
  year      = {2025},
  pages     = {48276-48334},
  volume    = {267},
  url       = {https://mlanthology.org/icml/2025/parnichkun2025icml-quantifying/}
}