Probabilistic Learning of Task-Specific Visual Attention

Abstract

Despite a considerable amount of previous work on bottom-up saliency modeling for predicting human fixations over static and dynamic stimuli, few studies have thus far attempted to model top-down and task-driven influences of visual attention. Here, taking advantage of the sequential nature of real-world tasks, we propose a unified Bayesian approach for modeling task-driven visual attention. Several sources of information, including global context of a scene, previous attended locations, and previous motor actions, are integrated over time to predict the next attended location. Recording eye movements while subjects engage in 5 contemporary 2D and 3D video games, as modest counterparts of everyday tasks, we show that our approach is able to predict human attention and gaze better than the state-of-the-art, with a large margin (about 15% increase in prediction accuracy). The advantage of our approach is that it is automatic and applicable to arbitrary visual tasks.

Cite

Text

Borji et al. "Probabilistic Learning of Task-Specific Visual Attention." IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2012. doi:10.1109/CVPR.2012.6247710

Markdown

[Borji et al. "Probabilistic Learning of Task-Specific Visual Attention." IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2012.](https://mlanthology.org/cvpr/2012/borji2012cvpr-probabilistic/) doi:10.1109/CVPR.2012.6247710

BibTeX

@inproceedings{borji2012cvpr-probabilistic,
  title     = {{Probabilistic Learning of Task-Specific Visual Attention}},
  author    = {Borji, Ali and Sihite, Dicky N. and Itti, Laurent},
  booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition},
  year      = {2012},
  pages     = {470-477},
  doi       = {10.1109/CVPR.2012.6247710},
  url       = {https://mlanthology.org/cvpr/2012/borji2012cvpr-probabilistic/}
}