CocktailSGD: Fine-Tuning Foundation Models over 500Mbps Networks
Abstract
Distributed training of foundation models, especially large language models (LLMs), is communication-intensive and so has heavily relied on centralized data centers with fast interconnects. Can we train on slow networks and unlock the potential of decentralized infrastructure for foundation models? In this paper, we propose CocktailSGD, a novel communication-efficient training framework that combines three distinct compression techniques – random sparsification, top-K sparsification, and quantization – to achieve much greater compression than each individual technique alone. We justify the benefit of such a hybrid approach through a theoretical analysis of convergence. Empirically, we show that CocktailSGD achieves up to 117$\times$ compression in fine-tuning LLMs up to 20 billion parameters without hurting convergence. On a 500Mbps network, CocktailSGD only incurs $\sim$1.2$\times$ slowdown compared with data center networks.
Cite
Text
Wang et al. "CocktailSGD: Fine-Tuning Foundation Models over 500Mbps Networks." International Conference on Machine Learning, 2023.Markdown
[Wang et al. "CocktailSGD: Fine-Tuning Foundation Models over 500Mbps Networks." International Conference on Machine Learning, 2023.](https://mlanthology.org/icml/2023/wang2023icml-cocktailsgd/)BibTeX
@inproceedings{wang2023icml-cocktailsgd,
title = {{CocktailSGD: Fine-Tuning Foundation Models over 500Mbps Networks}},
author = {Wang, Jue and Lu, Yucheng and Yuan, Binhang and Chen, Beidi and Liang, Percy and De Sa, Christopher and Re, Christopher and Zhang, Ce},
booktitle = {International Conference on Machine Learning},
year = {2023},
pages = {36058-36076},
volume = {202},
url = {https://mlanthology.org/icml/2023/wang2023icml-cocktailsgd/}
}