Training Data Size Induced Double Descent for Denoising Feedforward Neural Networks and the Role of Training Noise
Abstract
When training an unregularized denoising feedforward neural network, we show that the generalization error versus the number of training data points is a double descent curve. We formalize the question of how many training data points should be used by looking at the generalization error for denoising noisy test data. Prior work on computing the generalization error focuses on adding noise to target outputs. However, adding noise to the input is more in line with current pre-training practices. In the linear (in the inputs) regime, we provide an asymptotically exact formula for the generalization error for rank 1 data and an approximation for the generalization error for rank $r$ data. From this, we derive a formula for the amount of noise that needs to be added to the training data to minimize the denoising error. This results in the emergence of a shrinkage phenomenon for improving the performance of denoising DNNs by making the training SNR smaller than the test SNR. Further, we see that the amount of shrinkage (ratio of the train to test SNR) also follows a double descent curve.
Cite
Text
Sonthalia and Nadakuditi. "Training Data Size Induced Double Descent for Denoising Feedforward Neural Networks and the Role of Training Noise." Transactions on Machine Learning Research, 2023.Markdown
[Sonthalia and Nadakuditi. "Training Data Size Induced Double Descent for Denoising Feedforward Neural Networks and the Role of Training Noise." Transactions on Machine Learning Research, 2023.](https://mlanthology.org/tmlr/2023/sonthalia2023tmlr-training/)BibTeX
@article{sonthalia2023tmlr-training,
title = {{Training Data Size Induced Double Descent for Denoising Feedforward Neural Networks and the Role of Training Noise}},
author = {Sonthalia, Rishi and Nadakuditi, Raj Rao},
journal = {Transactions on Machine Learning Research},
year = {2023},
url = {https://mlanthology.org/tmlr/2023/sonthalia2023tmlr-training/}
}