II-Bench: An Image Implication Understanding Benchmark for Multimodal Large Language Models

Abstract

The rapid advancements in the development of multimodal large language models (MLLMs) have consistently led to new breakthroughs on various benchmarks. In response, numerous challenging and comprehensive benchmarks have been proposed to more accurately assess the capabilities of MLLMs. However, there is a dearth of exploration of the higher-order perceptual capabilities of MLLMs. To fill this gap, we propose the Image Implication understanding Benchmark, II-Bench, which aims to evaluate the model's higher-order perception of images. Through extensive experiments on II-Bench across multiple MLLMs, we have made significant findings. Initially, a substantial gap is observed between the performance of MLLMs and humans on II-Bench. The pinnacle accuracy of MLLMs attains 74.8%, whereas human accuracy averages 90%, peaking at an impressive 98%. Subsequently, MLLMs perform worse on abstract and complex images, suggesting limitations in their ability to understand high-level semantics and capture image details. Finally, it is observed that most models exhibit enhanced accuracy when image sentiment polarity hints are incorporated into the prompts. This observation underscores a notable deficiency in their inherent understanding of image sentiment. We believe that II-Bench will inspire the community to develop the next generation of MLLMs, advancing the journey towards expert artificial general intelligence (AGI). II-Bench is publicly available at https://huggingface.co/datasets/m-a-p/II-Bench.

Cite

Text

Liu et al. "II-Bench: An Image Implication Understanding Benchmark for Multimodal Large Language Models." Neural Information Processing Systems, 2024. doi:10.52202/079017-1474

Markdown

[Liu et al. "II-Bench: An Image Implication Understanding Benchmark for Multimodal Large Language Models." Neural Information Processing Systems, 2024.](https://mlanthology.org/neurips/2024/liu2024neurips-iibench/) doi:10.52202/079017-1474

BibTeX

@inproceedings{liu2024neurips-iibench,
  title     = {{II-Bench: An Image Implication Understanding Benchmark for Multimodal Large Language Models}},
  author    = {Liu, Ziqiang and Fang, Feiteng and Feng, Xi and Du, Xinrun and Zhang, Chenhao and Wang, Zekun and Bai, Yuelin and Zhao, Qixuan and Fan, Liyang and Gan, Chengguang and Lin, Hongquan and Li, Jiaming and Ni, Yuansheng and Wu, Haihong and Narsupalli, Yaswanth and Zheng, Zhigang and Li, Chengming and Hu, Xiping and Xu, Ruifeng and Chen, Xiaojun and Yang, Min and Liu, Jiaheng and Liu, Ruibo and Huang, Wenhao and Zhang, Ge and Ni, Shiwen},
  booktitle = {Neural Information Processing Systems},
  year      = {2024},
  doi       = {10.52202/079017-1474},
  url       = {https://mlanthology.org/neurips/2024/liu2024neurips-iibench/}
}