Unsupervised Learning of a Scene-Specific Coarse Gaze Estimator
Abstract
We present a method to estimate the coarse gaze directions of people from surveillance data. Unlike previous work we aim to do this without recourse to a large hand-labelled corpus of training data. In contrast we propose a method for learning a classifier without any hand labelled data using only the output from an automatic tracking system. A Conditional Random Field is used to model the interactions between the head motion, walking direction, and appearance to recover the gaze directions and simultaneously train randomised decision tree classifiers. Experiments demonstrate performance exceeding that of conventionally trained classifiers on two large surveillance datasets.
Cite
Text
Benfold and Reid. "Unsupervised Learning of a Scene-Specific Coarse Gaze Estimator." IEEE/CVF International Conference on Computer Vision, 2011. doi:10.1109/ICCV.2011.6126516Markdown
[Benfold and Reid. "Unsupervised Learning of a Scene-Specific Coarse Gaze Estimator." IEEE/CVF International Conference on Computer Vision, 2011.](https://mlanthology.org/iccv/2011/benfold2011iccv-unsupervised/) doi:10.1109/ICCV.2011.6126516BibTeX
@inproceedings{benfold2011iccv-unsupervised,
title = {{Unsupervised Learning of a Scene-Specific Coarse Gaze Estimator}},
author = {Benfold, Ben and Reid, Ian D.},
booktitle = {IEEE/CVF International Conference on Computer Vision},
year = {2011},
pages = {2344-2351},
doi = {10.1109/ICCV.2011.6126516},
url = {https://mlanthology.org/iccv/2011/benfold2011iccv-unsupervised/}
}