MVSalNet:Multi-View Augmentation for RGB-D Salient Object Detection
Abstract
RGB-D salient object detection (SOD) enjoys significant advantages in understanding 3D geometry of the scene. However, the geometry information conveyed by depth maps are mostly under-explored in existing RGB-D SOD methods. In this paper, we propose a new framework to address this issue. We augment the input image with multiple different views rendered using the depth maps, and cast the conventional single-view RGB-D SOD into a multi-view setting. Since different views captures complementary context of the 3D scene, the accuracy can be significantly improved through multi-view aggregation. We further design a multi-view saliency detection network (MVSalNet), which firstly performs saliency prediction for each view separately and incorporates multi-view outputs through a fusion model to produce final saliency prediction. A dynamic filtering module is also designed to facilitate more effective and flexible feature extraction. Extensive experiments on 6 widely used datasets demonstrate that our approach compares favorably against state-of-the-art approaches.
Cite
Text
Zhou et al. "MVSalNet:Multi-View Augmentation for RGB-D Salient Object Detection." Proceedings of the European Conference on Computer Vision (ECCV), 2022. doi:10.1007/978-3-031-19818-2_16Markdown
[Zhou et al. "MVSalNet:Multi-View Augmentation for RGB-D Salient Object Detection." Proceedings of the European Conference on Computer Vision (ECCV), 2022.](https://mlanthology.org/eccv/2022/zhou2022eccv-mvsalnet/) doi:10.1007/978-3-031-19818-2_16BibTeX
@inproceedings{zhou2022eccv-mvsalnet,
title = {{MVSalNet:Multi-View Augmentation for RGB-D Salient Object Detection}},
author = {Zhou, Jiayuan and Wang, Lijun and Lu, Huchuan and Huang, Kaining and Shi, Xinchu and Liu, Bocong},
booktitle = {Proceedings of the European Conference on Computer Vision (ECCV)},
year = {2022},
doi = {10.1007/978-3-031-19818-2_16},
url = {https://mlanthology.org/eccv/2022/zhou2022eccv-mvsalnet/}
}