Strong Lottery Ticket Hypothesis with $\epsilon$–perturbation

Abstract

The strong Lottery Ticket Hypothesis (LTH) claims that there exists a subnetwork in a sufficiently large, randomly initialized neural network that approximates some target neural networks without the need of training. This work extends the theoretical guarantee of the strong LTH literature to a scenario more similar to the original LTH, by generalizing the weight change achieved in the pre-training step to some perturbation around the initialization. In particular, we focus on the following open questions: By allowing an $\varepsilon$-scale perturbation on the random initial weights, can we reduce the over-parameterization requirement for the candidate network in the strong LTH? Furthermore, does the weight change by SGD coincide with a good set of such perturbation?

Cite

Text

Liao et al. "Strong Lottery Ticket Hypothesis with $\epsilon$–perturbation." NeurIPS 2022 Workshops: OPT, 2022.

Markdown

[Liao et al. "Strong Lottery Ticket Hypothesis with $\epsilon$–perturbation." NeurIPS 2022 Workshops: OPT, 2022.](https://mlanthology.org/neuripsw/2022/liao2022neuripsw-strong/)

BibTeX

@inproceedings{liao2022neuripsw-strong,
  title     = {{Strong Lottery Ticket Hypothesis with $\epsilon$–perturbation}},
  author    = {Liao, Fangshuo and Xiong, Zheyang and Kyrillidis, Anastasios},
  booktitle = {NeurIPS 2022 Workshops: OPT},
  year      = {2022},
  url       = {https://mlanthology.org/neuripsw/2022/liao2022neuripsw-strong/}
}