Decentralized Learning with Random Walks and Communication-Efficient Adaptive Optimization

Abstract

We tackle the problem of federated learning (FL) in a peer-to-peer fashion without a central server. While prior work mainly considered gossip-style protocols for learning, our solution is based on random walks. This allows to communicate only to a single peer at a time, thereby reducing the total communication and enabling asynchronous execution. To improve convergence and reduce the need for extensive tuning, we consider an adaptive optimization method -- Adam. Two extensions reduce its communication costs: state compression and multiple local updates on each client. We theoretically analyse the convergence behaviour of the proposed algorithm and its modifications in the non-convex setting. We show that our method can achieve performance comparable to centralized FL without communication overhead. Empirical results are reported on a variety of tasks (vision, text), neural network architectures and large-scale federations (up to $\sim342$k clients).

Cite

Text

Triastcyn et al. "Decentralized Learning with Random Walks and Communication-Efficient Adaptive Optimization." NeurIPS 2022 Workshops: Federated_Learning, 2022.

Markdown

[Triastcyn et al. "Decentralized Learning with Random Walks and Communication-Efficient Adaptive Optimization." NeurIPS 2022 Workshops: Federated_Learning, 2022.](https://mlanthology.org/neuripsw/2022/triastcyn2022neuripsw-decentralized/)

BibTeX

@inproceedings{triastcyn2022neuripsw-decentralized,
  title     = {{Decentralized Learning with Random Walks and Communication-Efficient Adaptive Optimization}},
  author    = {Triastcyn, Aleksei and Reisser, Matthias and Louizos, Christos},
  booktitle = {NeurIPS 2022 Workshops: Federated_Learning},
  year      = {2022},
  url       = {https://mlanthology.org/neuripsw/2022/triastcyn2022neuripsw-decentralized/}
}