Distributed Deep Learning in Open Collaborations

Abstract

Modern deep learning applications require increasingly more compute to train state-of-the-art models. To address this demand, large corporations and institutions use dedicated High-Performance Computing clusters, whose construction and maintenance are both environmentally costly and well beyond the budget of most organizations. As a result, some research directions become the exclusive domain of a few large industrial and even fewer academic actors. To alleviate this disparity, smaller groups may pool their computational resources and run collaborative experiments that benefit all participants. This paradigm, known as grid- or volunteer computing, has seen successful applications in numerous scientific areas. However, using this approach for machine learning is difficult due to high latency, asymmetric bandwidth, and several challenges unique to volunteer computing. In this work, we carefully analyze these constraints and propose a novel algorithmic framework designed specifically for collaborative training. We demonstrate the effectiveness of our approach for SwAV and ALBERT pretraining in realistic conditions and achieve performance comparable to traditional setups at a fraction of the cost. Finally, we provide a detailed report of successful collaborative language model pretraining with nearly 50 participants.

Cite

Text

Diskin et al. "Distributed Deep Learning in Open Collaborations." Neural Information Processing Systems, 2021.

Markdown

[Diskin et al. "Distributed Deep Learning in Open Collaborations." Neural Information Processing Systems, 2021.](https://mlanthology.org/neurips/2021/diskin2021neurips-distributed/)

BibTeX

@inproceedings{diskin2021neurips-distributed,
  title     = {{Distributed Deep Learning in Open Collaborations}},
  author    = {Diskin, Michael and Bukhtiyarov, Alexey and Ryabinin, Max and Saulnier, Lucile and Lhoest, Quentin and Sinitsin, Anton and Popov, Dmitry and Pyrkin, Dmitry V. and Kashirin, Maxim and Borzunov, Alexander and del Moral, Albert Villanova and Mazur, Denis and Kobelev, Ilia and Jernite, Yacine and Wolf, Thomas and Pekhimenko, Gennady},
  booktitle = {Neural Information Processing Systems},
  year      = {2021},
  url       = {https://mlanthology.org/neurips/2021/diskin2021neurips-distributed/}
}