Thinking Inside the Ball: Near-Optimal Minimization of the Maximal Loss
Abstract
We characterize the complexity of minimizing $\max_{i\in[N]} f_i(x)$ for convex, Lipschitz functions $f_1,\ldots, f_N$. For non-smooth functions, existing methods require $O(N\epsilon^{-2})$ queries to a first-order oracle to compute an $\epsilon$-suboptimal point and $\widetilde{O}(N\epsilon^{-1})$ queries if the $f_i$ are $O(1/\epsilon)$-smooth. We develop methods with improved complexity bounds of $\widetilde{O}(N\epsilon^{-2/3} + \epsilon^{-8/3})$ in the non-smooth case and $\widetilde{O}(N\epsilon^{-2/3} + \sqrt{N}\epsilon^{-1})$ in the $O(1/\epsilon)$-smooth case. Our methods consist of a recently proposed ball optimization oracle acceleration algorithm (which we refine) and a careful implementation of said oracle for the softmax function. We also prove an oracle complexity lower bound scaling as $\Omega(N\epsilon^{-2/3})$, showing that our dependence on $N$ is optimal up to polylogarithmic factors.
Cite
Text
Carmon et al. "Thinking Inside the Ball: Near-Optimal Minimization of the Maximal Loss." Conference on Learning Theory, 2021.Markdown
[Carmon et al. "Thinking Inside the Ball: Near-Optimal Minimization of the Maximal Loss." Conference on Learning Theory, 2021.](https://mlanthology.org/colt/2021/carmon2021colt-thinking/)BibTeX
@inproceedings{carmon2021colt-thinking,
title = {{Thinking Inside the Ball: Near-Optimal Minimization of the Maximal Loss}},
author = {Carmon, Yair and Jambulapati, Arun and Jin, Yujia and Sidford, Aaron},
booktitle = {Conference on Learning Theory},
year = {2021},
pages = {866-882},
volume = {134},
url = {https://mlanthology.org/colt/2021/carmon2021colt-thinking/}
}