Beyond Average: Individualized Visual Scanpath Prediction

Abstract

Understanding how attention varies across individuals has significant scientific and societal impacts. However existing visual scanpath models treat attention uniformly neglecting individual differences. To bridge this gap this paper focuses on individualized scanpath prediction (ISP) a new attention modeling task that aims to accurately predict how different individuals shift their attention in diverse visual tasks. It proposes an ISP method featuring three novel technical components: (1) an observer encoder to characterize and integrate an observer's unique attention traits (2) an observer-centric feature integration approach that holistically combines visual features task guidance and observer-specific characteristics and (3) an adaptive fixation prioritization mechanism that refines scanpath predictions by dynamically prioritizing semantic feature maps based on individual observers' attention traits. These novel components allow scanpath models to effectively address the attention variations across different observers. Our method is generally applicable to different datasets model architectures and visual tasks offering a comprehensive tool for transforming general scanpath models into individualized ones. Comprehensive evaluations using value-based and ranking-based metrics verify the method's effectiveness and generalizability.

Cite

Text

Chen et al. "Beyond Average: Individualized Visual Scanpath Prediction." Conference on Computer Vision and Pattern Recognition, 2024. doi:10.1109/CVPR52733.2024.02402

Markdown

[Chen et al. "Beyond Average: Individualized Visual Scanpath Prediction." Conference on Computer Vision and Pattern Recognition, 2024.](https://mlanthology.org/cvpr/2024/chen2024cvpr-beyond/) doi:10.1109/CVPR52733.2024.02402

BibTeX

@inproceedings{chen2024cvpr-beyond,
  title     = {{Beyond Average: Individualized Visual Scanpath Prediction}},
  author    = {Chen, Xianyu and Jiang, Ming and Zhao, Qi},
  booktitle = {Conference on Computer Vision and Pattern Recognition},
  year      = {2024},
  pages     = {25420-25431},
  doi       = {10.1109/CVPR52733.2024.02402},
  url       = {https://mlanthology.org/cvpr/2024/chen2024cvpr-beyond/}
}