LocMoE: A Low-Overhead MoE for Large Language Model Training

Cite

Text

Li et al. "LocMoE: A Low-Overhead MoE for Large Language Model Training." International Joint Conference on Artificial Intelligence, 2024.

Markdown

[Li et al. "LocMoE: A Low-Overhead MoE for Large Language Model Training." International Joint Conference on Artificial Intelligence, 2024.](https://mlanthology.org/ijcai/2024/li2024ijcai-locmoe/)

BibTeX

@inproceedings{li2024ijcai-locmoe,
  title     = {{LocMoE: A Low-Overhead MoE for Large Language Model Training}},
  author    = {Li, Jing and Sun, Zhijie and He, Xuan and Zeng, Li and Lin, Yi and Li, Entong and Zheng, Binfan and Zhao, Rongqian and Chen, Xin},
  booktitle = {International Joint Conference on Artificial Intelligence},
  year      = {2024},
  pages     = {6377-6387},
  url       = {https://mlanthology.org/ijcai/2024/li2024ijcai-locmoe/}
}