Finding Nearly Everything Within Random Binary Networks
Abstract
A recent work by Ramanujan et al. (2020) provides significant empirical evidence that sufficiently overparameterized, random neural networks contain untrained subnetworks that achieve state-of-the-art accuracy on several predictive tasks. A follow-up line of theoretical work provides justification of these findings by proving that slightly overparameterized neural networks, with commonly used continuous-valued random initializations can indeed be pruned to approximate any target network. In this work, we show that the amplitude of those random weights does not even matter. We prove that any target network of width $d$ and depth $l$ can be approximated up to arbitrary accuracy $\varepsilon$ by simply pruning a random network of binary $\{\pm1\}$ weights that is wider and deeper than the target network only by a polylogarithmic factor of $d, l$ and $\varepsilon$.
Cite
Text
Sreenivasan et al. "Finding Nearly Everything Within Random Binary Networks." Artificial Intelligence and Statistics, 2022.Markdown
[Sreenivasan et al. "Finding Nearly Everything Within Random Binary Networks." Artificial Intelligence and Statistics, 2022.](https://mlanthology.org/aistats/2022/sreenivasan2022aistats-finding/)BibTeX
@inproceedings{sreenivasan2022aistats-finding,
title = {{Finding Nearly Everything Within Random Binary Networks}},
author = {Sreenivasan, Kartik and Rajput, Shashank and Sohn, Jy-Yong and Papailiopoulos, Dimitris},
booktitle = {Artificial Intelligence and Statistics},
year = {2022},
pages = {3531-3541},
volume = {151},
url = {https://mlanthology.org/aistats/2022/sreenivasan2022aistats-finding/}
}