Global Convergence in Training Large-Scale Transformers

Abstract

Despite the widespread success of Transformers across various domains, their optimization guarantees in large-scale model settings are not well-understood. This paper rigorously analyzes the convergence properties of gradient flow in training Transformers with weight decay regularization. First, we construct the mean-field limit of large-scale Transformers, showing that as the model width and depth go to infinity, gradient flow converges to the Wasserstein gradient flow, which is represented by a partial differential equation. Then, we demonstrate that the gradient flow reaches a global minimum consistent with the PDE solution when the weight decay regularization parameter is sufficiently small. Our analysis is based on a series of novel mean-field techniques that adapt to Transformers. Compared with existing tools for deep networks (Lu et al., 2020) that demand homogeneity and global Lipschitz smoothness, we utilize a refined analysis assuming only $\textit{partial homogeneity}$ and $\textit{local Lipschitz smoothness}$. These new techniques may be of independent interest.

Cite

Text

Gao et al. "Global Convergence in Training Large-Scale Transformers." Neural Information Processing Systems, 2024. doi:10.52202/079017-0921

Markdown

[Gao et al. "Global Convergence in Training Large-Scale Transformers." Neural Information Processing Systems, 2024.](https://mlanthology.org/neurips/2024/gao2024neurips-global/) doi:10.52202/079017-0921

BibTeX

@inproceedings{gao2024neurips-global,
  title     = {{Global Convergence in Training Large-Scale Transformers}},
  author    = {Gao, Cheng and Cao, Yuan and Li, Zihao and He, Yihan and Wang, Mengdi and Liu, Han and Klusowski, Jason M. and Fan, Jianqing},
  booktitle = {Neural Information Processing Systems},
  year      = {2024},
  doi       = {10.52202/079017-0921},
  url       = {https://mlanthology.org/neurips/2024/gao2024neurips-global/}
}