Does Preprocessing Help Training Over-Parameterized Neural Networks?

Abstract

Deep neural networks have achieved impressive performance in many areas. Designing a fast and provable method for training neural networks is a fundamental question in machine learning. The classical training method requires paying $\Omega(mnd)$ cost for both forward computation and backward computation, where $m$ is the width of the neural network, and we are given $n$ training points in $d$-dimensional space. In this paper, we propose two novel preprocessing ideas to bypass this $\Omega(mnd)$ barrier:* First, by preprocessing the initial weights of the neural networks, we can train the neural network in $\widetilde{O}(m^{1-\Theta(1/d)} n d)$ cost per iteration.* Second, by preprocessing the input data points, we can train neural network in $\widetilde{O} (m^{4/5} nd )$ cost per iteration.From the technical perspective, our result is a sophisticated combination of tools in different fields, greedy-type convergence analysis in optimization, sparsity observation in practical work, high-dimensional geometric search in data structure, concentration and anti-concentration in probability. Our results also provide theoretical insights for a large number of previously established fast training methods.In addition, our classical algorithm can be generalized to the Quantum computation model. Interestingly, we can get a similar sublinear cost per iteration but avoid preprocessing initial weights or input data points.

Cite

Text

Song et al. "Does Preprocessing Help Training Over-Parameterized Neural Networks?." Neural Information Processing Systems, 2021.

Markdown

[Song et al. "Does Preprocessing Help Training Over-Parameterized Neural Networks?." Neural Information Processing Systems, 2021.](https://mlanthology.org/neurips/2021/song2021neurips-preprocessing/)

BibTeX

@inproceedings{song2021neurips-preprocessing,
  title     = {{Does Preprocessing Help Training Over-Parameterized Neural Networks?}},
  author    = {Song, Zhao and Yang, Shuo and Zhang, Ruizhe},
  booktitle = {Neural Information Processing Systems},
  year      = {2021},
  url       = {https://mlanthology.org/neurips/2021/song2021neurips-preprocessing/}
}