Detecting and Mitigating Hallucination in Large Vision Language Models via Fine-Grained AI Feedback

Abstract

The rapidly developing Large Vision Language Models (LVLMs) still face the hallucination phenomena where the generated responses do not align with the given contexts, significantly restricting the usages of LVLMs. Most previous work detects and mitigates hallucination at the coarse-grained level or requires expensive annotation (e.g., labeling by human experts or proprietary models). To address these issues, we propose detecting and mitigating hallucinations in LVLMs via fine-grained AI feedback. The basic idea is that we generate a small-size sentence-level hallucination annotation dataset by proprietary models, whereby we train a detection model which can perform sentence-level hallucination detection. Then, we propose a detect-then-rewrite pipeline to automatically construct preference dataset for hallucination mitigation training. Furthermore, we propose differentiating the severity of hallucinations, and introducing a Hallucination Severity-Aware Direct Preference Optimization (HSA-DPO) which prioritizes the mitigation of critical hallucination in LVLMs by incorporating the severity of hallucinations into preference learning. Extensive experiments on hallucination detection and mitigation benchmarks demonstrate that our method sets a new state-of-the-art in hallucination detection on MHaluBench, surpassing GPT-4V and Gemini, and reduces the hallucination rate by 36.1% on AMBER and 76.3% on Object HalBench compared to the base model.

Cite

Text

Xiao et al. "Detecting and Mitigating Hallucination in Large Vision Language Models via Fine-Grained AI Feedback." AAAI Conference on Artificial Intelligence, 2025. doi:10.1609/AAAI.V39I24.34744

Markdown

[Xiao et al. "Detecting and Mitigating Hallucination in Large Vision Language Models via Fine-Grained AI Feedback." AAAI Conference on Artificial Intelligence, 2025.](https://mlanthology.org/aaai/2025/xiao2025aaai-detecting/) doi:10.1609/AAAI.V39I24.34744

BibTeX

@inproceedings{xiao2025aaai-detecting,
  title     = {{Detecting and Mitigating Hallucination in Large Vision Language Models via Fine-Grained AI Feedback}},
  author    = {Xiao, Wenyi and Huang, Ziwei and Gan, Leilei and He, Wanggui and Li, Haoyuan and Yu, Zhelun and Shu, Fangxun and Jiang, Hao and Zhu, Linchao},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2025},
  pages     = {25543-25551},
  doi       = {10.1609/AAAI.V39I24.34744},
  url       = {https://mlanthology.org/aaai/2025/xiao2025aaai-detecting/}
}