Convergence of $\text{log}(1/\epsilon)$ for Gradient-Based Algorithms in Zero-Sum Games Without the Condition Number: A Smoothed Analysis
Abstract
Gradient-based algorithms have shown great promise in solving large (two-player) zero-sum games. However, their success has been mostly confined to the low-precision regime since the number of iterations grows polynomially in $1/\epsilon$, where $\epsilon > 0$ is the duality gap. While it has been well-documented that linear convergence---an iteration complexity scaling as $\text{log}(1/\epsilon)$---can be attained even with gradient-based algorithms, that comes at the cost of introducing a dependency on certain condition number-like quantities which can be exponentially large in the description of the game. To address this shortcoming, we examine the iteration complexity of several gradient-based algorithms in the celebrated framework of smoothed analysis, and we show that they have polynomial smoothed complexity, in that their number of iterations grows as a polynomial in the dimensions of the game, $\text{log}(1/\epsilon)$, and $1/\sigma$, where $\sigma$ measures the magnitude of the smoothing perturbation. Our result applies to optimistic gradient and extra-gradient descent/ascent, as well as a certain iterative variant of Nesterov's smoothing technique. From a technical standpoint, the proof proceeds by characterizing and performing a smoothed analysis of a certain error bound, the key ingredient driving linear convergence in zero-sum games. En route, our characterization also makes a natural connection between the convergence rate of such algorithms and perturbation-stability properties of the equilibrium, which is of interest beyond the model of smoothed complexity.
Cite
Text
Anagnostides and Sandholm. "Convergence of $\text{log}(1/\epsilon)$ for Gradient-Based Algorithms in Zero-Sum Games Without the Condition Number: A Smoothed Analysis." Neural Information Processing Systems, 2024. doi:10.52202/079017-3998Markdown
[Anagnostides and Sandholm. "Convergence of $\text{log}(1/\epsilon)$ for Gradient-Based Algorithms in Zero-Sum Games Without the Condition Number: A Smoothed Analysis." Neural Information Processing Systems, 2024.](https://mlanthology.org/neurips/2024/anagnostides2024neurips-convergence/) doi:10.52202/079017-3998BibTeX
@inproceedings{anagnostides2024neurips-convergence,
title = {{Convergence of $\text{log}(1/\epsilon)$ for Gradient-Based Algorithms in Zero-Sum Games Without the Condition Number: A Smoothed Analysis}},
author = {Anagnostides, Ioannis and Sandholm, Tuomas},
booktitle = {Neural Information Processing Systems},
year = {2024},
doi = {10.52202/079017-3998},
url = {https://mlanthology.org/neurips/2024/anagnostides2024neurips-convergence/}
}