Fusing Visual Saliency for Material Recognition

Abstract

Material recognition is researched in both computer vision and vision science fields. In this paper, we investigated how humans observe material images and found the eye fixation information improves the performance of material image classification models. We first collected eye-tracking data from human observers and used it to fine-tune a generative adversarial network for saliency prediction (SalGAN). We then fused the predicted saliency map with material images and fed them to CNN models for material classification. The experiment results show that the classification accuracy is improved than those using original images. This indicates that human's visual cues could benefit computational models as priors.

Cite

Text

Qi et al. "Fusing Visual Saliency for Material Recognition." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2018. doi:10.1109/CVPRW.2018.00263

Markdown

[Qi et al. "Fusing Visual Saliency for Material Recognition." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2018.](https://mlanthology.org/cvprw/2018/qi2018cvprw-fusing/) doi:10.1109/CVPRW.2018.00263

BibTeX

@inproceedings{qi2018cvprw-fusing,
  title     = {{Fusing Visual Saliency for Material Recognition}},
  author    = {Qi, Lin and Xu, Ying and Shang, Xiaowei and Dong, Junyu},
  booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops},
  year      = {2018},
  pages     = {1965-1968},
  doi       = {10.1109/CVPRW.2018.00263},
  url       = {https://mlanthology.org/cvprw/2018/qi2018cvprw-fusing/}
}