RoadSocial: A Diverse VideoQA Dataset and Benchmark for Road Event Understanding from Social Video Narratives

Abstract

We introduce RoadSocial, a large-scale, diverse VideoQA dataset tailored for generic road event understanding from social media narratives. Unlike existing datasets limited by regional bias, viewpoint bias and expert-driven annotations, RoadSocial captures the global complexity of road events with varied geographies, camera viewpoints (CCTV, handheld, drones) and rich social discourse. Our scalable semi-automatic annotation framework leverages Text LLMs and Video LLMs to generate comprehensive question-answer pairs across 12 challenging QA tasks, pushing the boundaries of road event understanding. RoadSocial is derived from social media videos spanning 14M frames and 414K social comments, resulting in a dataset with 13.2K videos, 674 tags and 260K high-quality QA pairs. We evaluate 18 Video LLMs (open-source and proprietary, driving-specific and general-purpose) on our road event understanding benchmark. We also demonstrate RoadSocial's utility in improving road event understanding capabilities of general-purpose Video LLMs.

Cite

Text

Parikh et al. "RoadSocial: A Diverse VideoQA Dataset and Benchmark for Road Event Understanding from Social Video Narratives." Conference on Computer Vision and Pattern Recognition, 2025. doi:10.1109/CVPR52734.2025.01770

Markdown

[Parikh et al. "RoadSocial: A Diverse VideoQA Dataset and Benchmark for Road Event Understanding from Social Video Narratives." Conference on Computer Vision and Pattern Recognition, 2025.](https://mlanthology.org/cvpr/2025/parikh2025cvpr-roadsocial/) doi:10.1109/CVPR52734.2025.01770

BibTeX

@inproceedings{parikh2025cvpr-roadsocial,
  title     = {{RoadSocial: A Diverse VideoQA Dataset and Benchmark for Road Event Understanding from Social Video Narratives}},
  author    = {Parikh, Chirag and Rawat, Deepti and Rakshitha, R. T. and Ghosh, Tathagata and Sarvadevabhatla, Ravi Kiran},
  booktitle = {Conference on Computer Vision and Pattern Recognition},
  year      = {2025},
  pages     = {19002-19011},
  doi       = {10.1109/CVPR52734.2025.01770},
  url       = {https://mlanthology.org/cvpr/2025/parikh2025cvpr-roadsocial/}
}