A Multi-Modal Graphical Model for Scene Analysis
Abstract
In this paper, we introduce a multi-modal graphical model to address the problems of semantic segmentation using 2D-3D data exhibiting extensive many-to-one correspondences. Existing methods often impose a hard correspondence between the 2D and 3D data, where the 2D and 3D corresponding regions are forced to receive identical labels. This results in performance degradation due to misalignments, 3D-2D projection errors and occlusions. We address this issue by defining a graph over the entire set of data that models soft correspondences between the two modalities. This graph encourages each region in a modality to leverage the information from its corresponding regions in the other modality to better estimate its class label. We evaluate our method on a publicly available dataset and beat the state-of-the-art. Additionally, to demonstrate the ability of our model to support multiple correspondences for objects in 3D and 2D domains, we introduce a new multi-modal dataset, which is composed of panoramic images and LIDAR data, and features a rich set of many-to-one correspondences.
Cite
Text
Namin et al. "A Multi-Modal Graphical Model for Scene Analysis." IEEE/CVF Winter Conference on Applications of Computer Vision, 2015. doi:10.1109/WACV.2015.139Markdown
[Namin et al. "A Multi-Modal Graphical Model for Scene Analysis." IEEE/CVF Winter Conference on Applications of Computer Vision, 2015.](https://mlanthology.org/wacv/2015/namin2015wacv-multi/) doi:10.1109/WACV.2015.139BibTeX
@inproceedings{namin2015wacv-multi,
title = {{A Multi-Modal Graphical Model for Scene Analysis}},
author = {Namin, Sarah Taghavi and Najafi, Mohammad and Salzmann, Mathieu and Petersson, Lars},
booktitle = {IEEE/CVF Winter Conference on Applications of Computer Vision},
year = {2015},
pages = {1006-1013},
doi = {10.1109/WACV.2015.139},
url = {https://mlanthology.org/wacv/2015/namin2015wacv-multi/}
}