Federated Minimax Optimization with Client Heterogeneity

Abstract

Minimax optimization has seen a surge in interest with the advent of modern applications such as GANs, and it is inherently more challenging than simple minimization. The difficulty is exacerbated by the training data residing at multiple edge devices or \textit{clients}, especially when these clients can have heterogeneous datasets and heterogeneous local computation capabilities. We propose a general federated minimax optimization framework that subsumes such settings and several existing methods like Local SGDA. We show that naive aggregation of model updates made by clients running unequal number of local steps can result in optimizing a mismatched objective function -- a phenomenon previously observed in standard federated minimization. To fix this problem, we propose normalizing the client updates by the number of local steps. We analyze the convergence of the proposed algorithm for classes of nonconvex-concave and nonconvex-nonconcave functions and characterize the impact of heterogeneous client data, partial client participation, and heterogeneous local computations. For all the function classes considered, we significantly improve the existing computation and communication complexity results. Experimental results support our theoretical claims.

Cite

Text

Sharma et al. "Federated Minimax Optimization with Client Heterogeneity." Transactions on Machine Learning Research, 2023.

Markdown

[Sharma et al. "Federated Minimax Optimization with Client Heterogeneity." Transactions on Machine Learning Research, 2023.](https://mlanthology.org/tmlr/2023/sharma2023tmlr-federated/)

BibTeX

@article{sharma2023tmlr-federated,
  title     = {{Federated Minimax Optimization with Client Heterogeneity}},
  author    = {Sharma, Pranay and Panda, Rohan and Joshi, Gauri},
  journal   = {Transactions on Machine Learning Research},
  year      = {2023},
  url       = {https://mlanthology.org/tmlr/2023/sharma2023tmlr-federated/}
}