MoLE: Decoding by Mixture of Layer Experts Alleviates Hallucination in Large Vision-Language Models
Abstract
Recent advancements in Large Vision-Language Models (LVLMs) highlight their ability to integrate and process multi-modal information. However, hallucinations—where generated content is inconsistent with input vision and instructions—remain a challenge. In this paper, we analyze LVLMs' layer-wise decoding and identify that hallucinations can arise during the reasoning and factual information injection process. Additionally, as the number of generated tokens increases, the forgetting of the original prompt may also lead to hallucinations.To address this, we propose a training-free decoding method called Mixture of Layer Experts (MoLE). MoLE leverages a heuristic gating mechanism to dynamically select multiple layers of LVLMs as expert layers: the Final Expert, the Second Opinion expert, and the Prompt Retention Expert. By the cooperation of each expert, MoLE enhances the robustness and faithfulness of the generation process. Our extensive experiments demonstrate that MoLE significantly reduces hallucinations, outperforming the current state-of-the-art decoding techniques across three mainstream LVLMs and two established hallucination benchmarks. Moreover, our method reveals the potential of LVLMs to independently produce more reliable and accurate outputs.
Cite
Text
Liang et al. "MoLE: Decoding by Mixture of Layer Experts Alleviates Hallucination in Large Vision-Language Models." AAAI Conference on Artificial Intelligence, 2025. doi:10.1609/AAAI.V39I18.34056Markdown
[Liang et al. "MoLE: Decoding by Mixture of Layer Experts Alleviates Hallucination in Large Vision-Language Models." AAAI Conference on Artificial Intelligence, 2025.](https://mlanthology.org/aaai/2025/liang2025aaai-mole/) doi:10.1609/AAAI.V39I18.34056BibTeX
@inproceedings{liang2025aaai-mole,
title = {{MoLE: Decoding by Mixture of Layer Experts Alleviates Hallucination in Large Vision-Language Models}},
author = {Liang, Tian and Du, Yuetian and Huang, Jing and Kong, Ming and Chen, Luyuan and Li, Yadong and Chen, Siye and Zhu, Qiang},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2025},
pages = {18684-18692},
doi = {10.1609/AAAI.V39I18.34056},
url = {https://mlanthology.org/aaai/2025/liang2025aaai-mole/}
}