Union Is Strength! Unite the Power of LLMs and MLLMs for Chart Question Answering
Abstract
Chart Question Answering (CQA) requires models to perform chart perception and reasoning. Recent studies driven by Large Language Models (LLMs) have dominated CQA. These include employing more cognitively capable LLMs for indirectly reasoning over transformed charts, i.e., tables, and directly perceiving charts utilizing Multimodal Large Language Models (MLLMs) with a wider perceptual range. Yet, they often encounter bottlenecks due to the limitation of the receptive field of LLMs and the fragility of the complex reasoning of some MLLMs. To unite the strengths of LLMs and MLLMs to complement each other's limitations, we propose Synergy, a framework that unites the power of both models for CQA. Synergy first unites the chart with a table as the augmented perceptual signal. Next, it unites LLMs and MLLMs, scheduling the former to decompose a question into subquestions and the latter to answer these by perceiving the chart. Lastly, it operates LLMs to summarize the subquestion-answer pairs to refine the final answer. Extensive experimental results on popular CharQA and PlotQA benchmarks reveal that, with the power of union, Synergy outperforms strong competitors and achieves superior boosts over naive MLLMs by uniting them with a smaller LLM.
Cite
Text
Liu et al. "Union Is Strength! Unite the Power of LLMs and MLLMs for Chart Question Answering." AAAI Conference on Artificial Intelligence, 2025. doi:10.1609/AAAI.V39I5.32584Markdown
[Liu et al. "Union Is Strength! Unite the Power of LLMs and MLLMs for Chart Question Answering." AAAI Conference on Artificial Intelligence, 2025.](https://mlanthology.org/aaai/2025/liu2025aaai-union/) doi:10.1609/AAAI.V39I5.32584BibTeX
@inproceedings{liu2025aaai-union,
title = {{Union Is Strength! Unite the Power of LLMs and MLLMs for Chart Question Answering}},
author = {Liu, Jiapeng and Li, Liang and Rao, Shihao and Gao, Xiyan and Guan, Weixin and Li, Bing and Ma, Can},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2025},
pages = {5487-5495},
doi = {10.1609/AAAI.V39I5.32584},
url = {https://mlanthology.org/aaai/2025/liu2025aaai-union/}
}