Stable Gradient Descent
Abstract
The goal of many machine learning tasks is to learn a model that has small population risk. While mini-batch stochastic gradient descent (SGD) and variants are popular approaches for achieving this goal, it is hard to prescribe a clear stopping criterion and to establish high probability convergence bounds to the population risk. In this paper, we introduce Stable Gradient Descent which validates stochastic gradient computations by splitting data into training and validation sets and reuses samples using a differential private mechanism. StGD comes with a natural upper bound on the number of iterations and has high-probability convergence to the population risk. Experimental results illustrate that StGD is empirically competitive and often better than SGD and GD.
Cite
Text
Zhou et al. "Stable Gradient Descent." Conference on Uncertainty in Artificial Intelligence, 2018.Markdown
[Zhou et al. "Stable Gradient Descent." Conference on Uncertainty in Artificial Intelligence, 2018.](https://mlanthology.org/uai/2018/zhou2018uai-stable/)BibTeX
@inproceedings{zhou2018uai-stable,
title = {{Stable Gradient Descent}},
author = {Zhou, Yingxue and Chen, Sheng and Banerjee, Arindam},
booktitle = {Conference on Uncertainty in Artificial Intelligence},
year = {2018},
pages = {766-775},
url = {https://mlanthology.org/uai/2018/zhou2018uai-stable/}
}