SCVBench: A Benchmark with Multi-Turn Dialogues for Story-Centric Video Understanding

Abstract

Video understanding seeks to enable machines to interpret visual content across three levels: action, event, and story. Existing models are limited in their ability to perform high-level long-term story understanding, due to (1) the oversimplified treatment of temporal information and (2) the training bias introduced by action/event-centric datasets. To address this, we introduce SCVBench, a novel benchmark for story-centric video understanding. SCVBench evaluates LVLMs through an event ordering task decomposed into sub-questions leading to a final question, quantitatively measuring historical dialogue exploration. We collected 1,253 final questions and 6,027 sub-question pairs from 925 videos, constructing continuous multi-turn dialogues. Experimental results show that while closed-source GPT-4o outperforms other models, most open-source LVLMs struggle with story-centric video understanding. Additionally, our StoryCoT model significantly surpasses open-source LVLMs on SCVBench. SCVBench aims to advance research by comprehensively analyzing LVLMs' temporal reasoning and comprehension capabilities. Code can be accessed at https://github.com/yuanrr/SCVBench.

Cite

Text

You et al. "SCVBench: A Benchmark with Multi-Turn Dialogues for Story-Centric Video Understanding." International Joint Conference on Artificial Intelligence, 2025. doi:10.24963/IJCAI.2025/255

Markdown

[You et al. "SCVBench: A Benchmark with Multi-Turn Dialogues for Story-Centric Video Understanding." International Joint Conference on Artificial Intelligence, 2025.](https://mlanthology.org/ijcai/2025/you2025ijcai-scvbench/) doi:10.24963/IJCAI.2025/255

BibTeX

@inproceedings{you2025ijcai-scvbench,
  title     = {{SCVBench: A Benchmark with Multi-Turn Dialogues for Story-Centric Video Understanding}},
  author    = {You, Sisi and Yuan, Bowen and Bao, Bing-Kun},
  booktitle = {International Joint Conference on Artificial Intelligence},
  year      = {2025},
  pages     = {2287-2295},
  doi       = {10.24963/IJCAI.2025/255},
  url       = {https://mlanthology.org/ijcai/2025/you2025ijcai-scvbench/}
}