Stochastic Convex Optimization
Abstract
For supervised classification problems, it is well known that learnability is equivalent to uniform convergence of the empirical risks and thus to learnability by empirical minimization. Inspired by recent regret bounds for online convex optimization, we study stochastic convex optimization, and uncover a surprisingly different situation in the more general setting: although the stochastic convex optimization problem is learnable (e.g. using online-to-batch conversions), no uniform convergence holds in the general case, and empirical minimization might fail. Rather then being a difference between online methods and a global minimization approach, we show that the key ingredient is strong convexity and regularization. Our results demonstrate that the celebrated theorem of Alon et al on the equivalence of learnability and uniform convergence does not extend to Vapnik's General Setting of Learning, that in the General Setting considering only empirical minimization is not enough, and that despite Vanpnik's result on the equivalence of strict consistency and uniform convergence, uniform convergence is only a sufficient, but not necessary, condition for meaningful non-trivial learnability.
Cite
Text
Shalev-Shwartz et al. "Stochastic Convex Optimization." Annual Conference on Computational Learning Theory, 2009.Markdown
[Shalev-Shwartz et al. "Stochastic Convex Optimization." Annual Conference on Computational Learning Theory, 2009.](https://mlanthology.org/colt/2009/shalevshwartz2009colt-stochastic/)BibTeX
@inproceedings{shalevshwartz2009colt-stochastic,
title = {{Stochastic Convex Optimization}},
author = {Shalev-Shwartz, Shai and Shamir, Ohad and Srebro, Nathan and Sridharan, Karthik},
booktitle = {Annual Conference on Computational Learning Theory},
year = {2009},
url = {https://mlanthology.org/colt/2009/shalevshwartz2009colt-stochastic/}
}