Strong Lottery Ticket Hypothesis with $\epsilon$–perturbation
Abstract
The strong Lottery Ticket Hypothesis (LTH) claims that there exists a subnetwork in a sufficiently large, randomly initialized neural network that approximates some target neural networks without the need of training. This work extends the theoretical guarantee of the strong LTH literature to a scenario more similar to the original LTH, by generalizing the weight change achieved in the pre-training step to some perturbation around the initialization. In particular, we focus on the following open questions: By allowing an $\varepsilon$-scale perturbation on the random initial weights, can we reduce the over-parameterization requirement for the candidate network in the strong LTH? Furthermore, does the weight change by SGD coincide with a good set of such perturbation?
Cite
Text
Liao et al. "Strong Lottery Ticket Hypothesis with $\epsilon$–perturbation." NeurIPS 2022 Workshops: OPT, 2022.Markdown
[Liao et al. "Strong Lottery Ticket Hypothesis with $\epsilon$–perturbation." NeurIPS 2022 Workshops: OPT, 2022.](https://mlanthology.org/neuripsw/2022/liao2022neuripsw-strong/)BibTeX
@inproceedings{liao2022neuripsw-strong,
title = {{Strong Lottery Ticket Hypothesis with $\epsilon$–perturbation}},
author = {Liao, Fangshuo and Xiong, Zheyang and Kyrillidis, Anastasios},
booktitle = {NeurIPS 2022 Workshops: OPT},
year = {2022},
url = {https://mlanthology.org/neuripsw/2022/liao2022neuripsw-strong/}
}