LocMoE: A Low-Overhead MoE for Large Language Model Training
Cite
Text
Li et al. "LocMoE: A Low-Overhead MoE for Large Language Model Training." International Joint Conference on Artificial Intelligence, 2024.Markdown
[Li et al. "LocMoE: A Low-Overhead MoE for Large Language Model Training." International Joint Conference on Artificial Intelligence, 2024.](https://mlanthology.org/ijcai/2024/li2024ijcai-locmoe/)BibTeX
@inproceedings{li2024ijcai-locmoe,
title = {{LocMoE: A Low-Overhead MoE for Large Language Model Training}},
author = {Li, Jing and Sun, Zhijie and He, Xuan and Zeng, Li and Lin, Yi and Li, Entong and Zheng, Binfan and Zhao, Rongqian and Chen, Xin},
booktitle = {International Joint Conference on Artificial Intelligence},
year = {2024},
pages = {6377-6387},
url = {https://mlanthology.org/ijcai/2024/li2024ijcai-locmoe/}
}