Non-Local Deep Features for Salient Object Detection

Abstract

Saliency detection aims to highlight the most relevant objects in an image. Methods using conventional models struggle whenever salient objects are pictured on top of a cluttered background while deep neural nets suffer from excess complexity and slow evaluation speeds. In this paper, we propose a simplified convolutional neural network which combines local and global information through a multi-resolution 4x5 grid structure. Instead of enforcing spacial coherence with a CRF or superpixels as is usually the case, we implemented a loss function inspired by the Mumford-Shah functional which penalizes errors on the boundary. We trained our model on the MSRA-B dataset, and tested it on six different saliency benchmark datasets. Results show that our method is on par with the state-of-the-art while reducing computation time by a factor of 18 to 100 times, enabling near real-time, high performance saliency detection.

Cite

Text

Luo et al. "Non-Local Deep Features for Salient Object Detection." Conference on Computer Vision and Pattern Recognition, 2017. doi:10.1109/CVPR.2017.698

Markdown

[Luo et al. "Non-Local Deep Features for Salient Object Detection." Conference on Computer Vision and Pattern Recognition, 2017.](https://mlanthology.org/cvpr/2017/luo2017cvpr-nonlocal/) doi:10.1109/CVPR.2017.698

BibTeX

@inproceedings{luo2017cvpr-nonlocal,
  title     = {{Non-Local Deep Features for Salient Object Detection}},
  author    = {Luo, Zhiming and Mishra, Akshaya and Achkar, Andrew and Eichel, Justin and Li, Shaozi and Jodoin, Pierre-Marc},
  booktitle = {Conference on Computer Vision and Pattern Recognition},
  year      = {2017},
  doi       = {10.1109/CVPR.2017.698},
  url       = {https://mlanthology.org/cvpr/2017/luo2017cvpr-nonlocal/}
}