Efficient Adaptive Federated Optimization
Abstract
Adaptive optimization plays a pivotal role in federated learning, where simultaneous server- and client-side adaptivity have been shown to be essential for maximizing its performance. However, the scalability of jointly adaptive systems are often constrained by limited resources in communication and memory. In this paper, we introduce a class of efficient adaptive algorithms, named FedAda$^2$, designed specifically for large-scale, cross-device federated environments. FedAda$^2$ optimizes communication efficiency by avoiding the transfer of preconditioners between the server and clients, while simultaneously utilizing memory-efficient adaptive optimizers on the client-side to reduce extra on-device memory cost. Theoretically, we demonstrate that FedAda$^2$ achieves the same convergence rates for general, non-convex objectives as its more resource-intensive counterparts that naively integrate joint adaptivity. Empirically, we showcase the benefits of joint adaptivity and the effectiveness of FedAda$^2$ on several image datasets.
Cite
Text
Lee et al. "Efficient Adaptive Federated Optimization." ICML 2024 Workshops: WANT, 2024.Markdown
[Lee et al. "Efficient Adaptive Federated Optimization." ICML 2024 Workshops: WANT, 2024.](https://mlanthology.org/icmlw/2024/lee2024icmlw-efficient/)BibTeX
@inproceedings{lee2024icmlw-efficient,
title = {{Efficient Adaptive Federated Optimization}},
author = {Lee, Su Hyeong and Sharma, Sidharth and Zaheer, Manzil and Li, Tian},
booktitle = {ICML 2024 Workshops: WANT},
year = {2024},
url = {https://mlanthology.org/icmlw/2024/lee2024icmlw-efficient/}
}