Logarithmic Width Suffices for Robust Memorization
Abstract
The memorization capacity of neural networks with a given architecture has been thoroughly studied in many works. Specifically, it is well-known that memorizing $N$ samples can be done using a network of constant width, independent of $N$. However, the required constructions are often quite delicate. In this paper, we consider the natural question of how well feedforward ReLU neural networks can memorize \emph{robustly}, namely while being able to withstand adversarial perturbations of a given radius. We establish both upper and lower bounds on the possible radius for general $l_p$ norms, implying (among other things) that width \emph{logarithmic} in the number of input samples is necessary and sufficient to achieve robust memorization (with robustness radius independent of $N$).
Cite
Text
Egosi et al. "Logarithmic Width Suffices for Robust Memorization." Proceedings of Thirty Eighth Conference on Learning Theory, 2025.Markdown
[Egosi et al. "Logarithmic Width Suffices for Robust Memorization." Proceedings of Thirty Eighth Conference on Learning Theory, 2025.](https://mlanthology.org/colt/2025/egosi2025colt-logarithmic/)BibTeX
@inproceedings{egosi2025colt-logarithmic,
title = {{Logarithmic Width Suffices for Robust Memorization}},
author = {Egosi, Amitsour and Yehudai, Gilad and Shamir, Ohad},
booktitle = {Proceedings of Thirty Eighth Conference on Learning Theory},
year = {2025},
pages = {1638-1690},
volume = {291},
url = {https://mlanthology.org/colt/2025/egosi2025colt-logarithmic/}
}