kgMBQA: Quality Knowledge Graph-Driven Multimodal Blind Image Assessment
Abstract
Blind image assessment aims to simulate human prediction of image quality distortion levels and provide quality scores. However, existing unimodal quality indicators have limited representational ability when facing complex contents and distortion types, and the predicted scores also fail to provide explanatory reasons, which further affects the credibility of their prediction results. To address these challenges, we propose a multimodal quality indicator with explanatory text descriptions, called kgMBQA. Specifically, we construct an image quality knowledge graph and conduct in-depth mining to generate explanatory texts. The text modality is further aligned and fused with the image modality, thereby improving the model performance while also outputting its corresponding quality explanatory description. The experimental results demonstrate that our kgMBQA achieves the best performance compared to recent representative methods on the KonIQ-10k, LIVE Challenge, BIQ2021, TID2013, and AIGC-3K datasets.
Cite
Text
Xie et al. "kgMBQA: Quality Knowledge Graph-Driven Multimodal Blind Image Assessment." International Joint Conference on Artificial Intelligence, 2025. doi:10.24963/IJCAI.2025/473Markdown
[Xie et al. "kgMBQA: Quality Knowledge Graph-Driven Multimodal Blind Image Assessment." International Joint Conference on Artificial Intelligence, 2025.](https://mlanthology.org/ijcai/2025/xie2025ijcai-kgmbqa/) doi:10.24963/IJCAI.2025/473BibTeX
@inproceedings{xie2025ijcai-kgmbqa,
title = {{kgMBQA: Quality Knowledge Graph-Driven Multimodal Blind Image Assessment}},
author = {Xie, Wuyuan and Bian, Tingcheng and Wang, Miaohui},
booktitle = {International Joint Conference on Artificial Intelligence},
year = {2025},
pages = {4245-4253},
doi = {10.24963/IJCAI.2025/473},
url = {https://mlanthology.org/ijcai/2025/xie2025ijcai-kgmbqa/}
}