Fixation Prediction in Videos Using Unsupervised Hierarchical Features
Abstract
This paper presents a framework for saliency estimation and fixation prediction in videos. The proposed framework is based on a hierarchical feature representation obtained by stacking convolutional layers of independent subspace analysis (ISA) filters. The feature learning is thus unsupervised and independent of the task. To compute the saliency, we then employ a multiresolution saliency architecture that exploits both local and global saliency. That is, for a given image, an image pyramid is initially built. After that, for each resolution, both local and global saliency measures are computed to obtain a saliency map. The integration of saliency maps over the image pyramid provides the final video saliency. We first show that combining local and global saliency improves the results. We then compare the proposed model with several video saliency models and demonstrate that the proposed framework is capable of predicting video saliency effectively, outperforming all the other models.
Cite
Text
Wang et al. "Fixation Prediction in Videos Using Unsupervised Hierarchical Features." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2017. doi:10.1109/CVPRW.2017.276Markdown
[Wang et al. "Fixation Prediction in Videos Using Unsupervised Hierarchical Features." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2017.](https://mlanthology.org/cvprw/2017/wang2017cvprw-fixation/) doi:10.1109/CVPRW.2017.276BibTeX
@inproceedings{wang2017cvprw-fixation,
title = {{Fixation Prediction in Videos Using Unsupervised Hierarchical Features}},
author = {Wang, Julius and Tavakoli, Hamed R. and Laaksonen, Jorma},
booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops},
year = {2017},
pages = {2225-2232},
doi = {10.1109/CVPRW.2017.276},
url = {https://mlanthology.org/cvprw/2017/wang2017cvprw-fixation/}
}