GLaM: Efficient Scaling of Language Models with Mixture-of-Experts

Abstract

Scaling language models with more data, compute and parameters has driven significant progress in natural language processing. For example, thanks to scaling, GPT-3 was able to achieve strong results on in-context learning tasks. However, training these large dense models requires significant amounts of computing resources. In this paper, we propose and develop a family of language models named \glam (\textbf{G}eneralist \textbf{La}nguage \textbf{M}odel), which uses a sparsely activated mixture-of-experts architecture to scale the model capacity while also incurring substantially less training cost compared to dense variants. The largest \glam has 1.2 trillion parameters, which is approximately 7x larger than GPT-3. It consumes only 1/3 of the energy used to train GPT-3 and requires half of the computation flops for inference, while still achieving better overall fewshot performance across 29 NLP tasks.

Cite

Text

Du et al. "GLaM: Efficient Scaling of Language Models with Mixture-of-Experts." International Conference on Machine Learning, 2022.

Markdown

[Du et al. "GLaM: Efficient Scaling of Language Models with Mixture-of-Experts." International Conference on Machine Learning, 2022.](https://mlanthology.org/icml/2022/du2022icml-glam/)

BibTeX

@inproceedings{du2022icml-glam,
  title     = {{GLaM: Efficient Scaling of Language Models with Mixture-of-Experts}},
  author    = {Du, Nan and Huang, Yanping and Dai, Andrew M and Tong, Simon and Lepikhin, Dmitry and Xu, Yuanzhong and Krikun, Maxim and Zhou, Yanqi and Yu, Adams Wei and Firat, Orhan and Zoph, Barret and Fedus, Liam and Bosma, Maarten P and Zhou, Zongwei and Wang, Tao and Wang, Emma and Webster, Kellie and Pellat, Marie and Robinson, Kevin and Meier-Hellstern, Kathleen and Duke, Toju and Dixon, Lucas and Zhang, Kun and Le, Quoc and Wu, Yonghui and Chen, Zhifeng and Cui, Claire},
  booktitle = {International Conference on Machine Learning},
  year      = {2022},
  pages     = {5547-5569},
  volume    = {162},
  url       = {https://mlanthology.org/icml/2022/du2022icml-glam/}
}