Gaze Target Detection Based on Head-Local-Global Coordination

Abstract

This paper introduces a novel approach to gaze target detection leveraging a head-local-global coordination framework. Unlike traditional methods that rely heavily on estimating gaze direction and identifying salient objects in global view images, our method incorporates a FOV-based local view to more accurately predict gaze targets. We also propose a unique global-local position and representation consistency mechanism to integrate the features from head view, local view, and global view, significantly improving prediction accuracy. Through extensive experiments, our approach demonstrates state-of-the-art performance on multiple significant gaze target detection benchmarks, showcasing its scalability and the effectiveness of the local view and view-coordination mechanisms. The method’s scalability is further evidenced by enhancing the performance of existing gaze target detection methods within our proposed head-local-global coordination framework.

Cite

Text

Yang and Lu. "Gaze Target Detection Based on Head-Local-Global Coordination." Proceedings of the European Conference on Computer Vision (ECCV), 2024. doi:10.1007/978-3-031-73383-3_18

Markdown

[Yang and Lu. "Gaze Target Detection Based on Head-Local-Global Coordination." Proceedings of the European Conference on Computer Vision (ECCV), 2024.](https://mlanthology.org/eccv/2024/yang2024eccv-gaze/) doi:10.1007/978-3-031-73383-3_18

BibTeX

@inproceedings{yang2024eccv-gaze,
  title     = {{Gaze Target Detection Based on Head-Local-Global Coordination}},
  author    = {Yang, Yaokun and Lu, Feng},
  booktitle = {Proceedings of the European Conference on Computer Vision (ECCV)},
  year      = {2024},
  doi       = {10.1007/978-3-031-73383-3_18},
  url       = {https://mlanthology.org/eccv/2024/yang2024eccv-gaze/}
}