Unsupervised Learning of a Scene-Specific Coarse Gaze Estimator

Abstract

We present a method to estimate the coarse gaze directions of people from surveillance data. Unlike previous work we aim to do this without recourse to a large hand-labelled corpus of training data. In contrast we propose a method for learning a classifier without any hand labelled data using only the output from an automatic tracking system. A Conditional Random Field is used to model the interactions between the head motion, walking direction, and appearance to recover the gaze directions and simultaneously train randomised decision tree classifiers. Experiments demonstrate performance exceeding that of conventionally trained classifiers on two large surveillance datasets.

Cite

Text

Benfold and Reid. "Unsupervised Learning of a Scene-Specific Coarse Gaze Estimator." IEEE/CVF International Conference on Computer Vision, 2011. doi:10.1109/ICCV.2011.6126516

Markdown

[Benfold and Reid. "Unsupervised Learning of a Scene-Specific Coarse Gaze Estimator." IEEE/CVF International Conference on Computer Vision, 2011.](https://mlanthology.org/iccv/2011/benfold2011iccv-unsupervised/) doi:10.1109/ICCV.2011.6126516

BibTeX

@inproceedings{benfold2011iccv-unsupervised,
  title     = {{Unsupervised Learning of a Scene-Specific Coarse Gaze Estimator}},
  author    = {Benfold, Ben and Reid, Ian D.},
  booktitle = {IEEE/CVF International Conference on Computer Vision},
  year      = {2011},
  pages     = {2344-2351},
  doi       = {10.1109/ICCV.2011.6126516},
  url       = {https://mlanthology.org/iccv/2011/benfold2011iccv-unsupervised/}
}