Logarithmic Width Suffices for Robust Memorization

Abstract

The memorization capacity of neural networks with a given architecture has been thoroughly studied in many works. Specifically, it is well-known that memorizing $N$ samples can be done using a network of constant width, independent of $N$. However, the required constructions are often quite delicate. In this paper, we consider the natural question of how well feedforward ReLU neural networks can memorize \emph{robustly}, namely while being able to withstand adversarial perturbations of a given radius. We establish both upper and lower bounds on the possible radius for general $l_p$ norms, implying (among other things) that width \emph{logarithmic} in the number of input samples is necessary and sufficient to achieve robust memorization (with robustness radius independent of $N$).

Cite

Text

Egosi et al. "Logarithmic Width Suffices for Robust Memorization." Proceedings of Thirty Eighth Conference on Learning Theory, 2025.

Markdown

[Egosi et al. "Logarithmic Width Suffices for Robust Memorization." Proceedings of Thirty Eighth Conference on Learning Theory, 2025.](https://mlanthology.org/colt/2025/egosi2025colt-logarithmic/)

BibTeX

@inproceedings{egosi2025colt-logarithmic,
  title     = {{Logarithmic Width Suffices for Robust Memorization}},
  author    = {Egosi, Amitsour and Yehudai, Gilad and Shamir, Ohad},
  booktitle = {Proceedings of Thirty Eighth Conference on Learning Theory},
  year      = {2025},
  pages     = {1638-1690},
  volume    = {291},
  url       = {https://mlanthology.org/colt/2025/egosi2025colt-logarithmic/}
}