Routing Experts: Learning to Route Dynamic Experts in Existing Multi-Modal Large Language Models

Abstract

Recently, mixture of experts (MoE) has become a popular paradigm for achieving the trade-off between modal capacity and efficiency of multimodal large language models (MLLMs). Different from previous efforts, we are dedicated to exploring the dynamic experts in existing MLLMs and showing that a standard MLLM can also be a mixture of experts. However, achieving this target is still notoriously challenging. The well-trained MLLMs are more accustomed to the fixed pathway and a drastic change in its inference manner also greatly impedes its performance. To address these issues, we propose a novel dynamic expert routing method for existing MLLMs, termed Routing Experts (RoE), which can achieve example-dependent optimal path routing without obvious structure tweaks. Meanwhile, a new structure sparsity regularization is also introduced to force the well-trained MLLMs to learn more short-cut pathways. In addition, we also address the alignment of the training and inference of MLLMs in terms of network routing. To validate RoE, we apply it to a set of existing MLLMs, including LLaVA-1.5, LLaVA-HR and VILA, and conduct extensive experiments on a bunch of VL benchmarks. The experiment results not only show the effectiveness of our RoE in improving MLLMs' efficiency, but also yield obvious advantages over MoE-LLaVA in both performance and speed, e.g., an average performance gain of 3.3% on 5 benchmarks while being 1.61 times faster. Our code is anonymously released at https://github.com/DoubtedSteam/RoE

Cite

Text

Wu et al. "Routing Experts: Learning to Route Dynamic Experts in Existing Multi-Modal Large Language Models." International Conference on Learning Representations, 2025.

Markdown

[Wu et al. "Routing Experts: Learning to Route Dynamic Experts in Existing Multi-Modal Large Language Models." International Conference on Learning Representations, 2025.](https://mlanthology.org/iclr/2025/wu2025iclr-routing/)

BibTeX

@inproceedings{wu2025iclr-routing,
  title     = {{Routing Experts: Learning to Route Dynamic Experts in Existing Multi-Modal Large Language Models}},
  author    = {Wu, Qiong and Ke, Zhaoxi and Zhou, Yiyi and Sun, Xiaoshuai and Ji, Rongrong},
  booktitle = {International Conference on Learning Representations},
  year      = {2025},
  url       = {https://mlanthology.org/iclr/2025/wu2025iclr-routing/}
}