Distributed Online and Bandit Convex Optimization
Abstract
We study the problems of distributed online and bandit convex optimization against an adaptive adversary. Our goal is to minimize the average regret on M machines working in parallel over T rounds that can communicate R times intermittently. Assuming the underlying cost functions are convex, our results show collaboration is not beneficial if the machines have access to the first-order gradient information at the queried points. We show that in this setting, simple non-collaborative algorithms are min-max optimal, as opposed to the case for stochastic functions, where each machine samples the cost functions from a fixed distribution. Next, we consider the more challenging setting of federated optimization with bandit (zeroth-order) feedback, where the machines can only access values of the cost functions at the queried points. The key finding here is to identify the high-dimensional regime where collaboration is beneficial and may even lead to a linear speedup in the number of machines. Our results are the first attempts towards bridging the gap between distributed online optimization against stochastic and adaptive adversaries.
Cite
Text
Patel et al. "Distributed Online and Bandit Convex Optimization." NeurIPS 2022 Workshops: OPT, 2022.Markdown
[Patel et al. "Distributed Online and Bandit Convex Optimization." NeurIPS 2022 Workshops: OPT, 2022.](https://mlanthology.org/neuripsw/2022/patel2022neuripsw-distributed/)BibTeX
@inproceedings{patel2022neuripsw-distributed,
title = {{Distributed Online and Bandit Convex Optimization}},
author = {Patel, Kumar Kshitij and Saha, Aadirupa and Wang, Lingxiao and Srebro, Nathan},
booktitle = {NeurIPS 2022 Workshops: OPT},
year = {2022},
url = {https://mlanthology.org/neuripsw/2022/patel2022neuripsw-distributed/}
}