Multiple View Semantic Segmentation for Street View Images

Abstract

We propose a simple but powerful multi-view semantic segmentation framework for images captured by a camera mounted on a car driving along streets. In our approach, a pair-wise Markov Random Field (MRF) is laid out across multiple views. Both 2D and 3D features are extracted at a super-pixel level to train classifiers for the unary data terms of MRF. For smoothness terms, our approach makes use of color differences in the same image to identify accurate segmentation boundaries, and dense pixel-to-pixel correspondences to enforce consistency across different views. To speed up training and to improve the recognition quality, our approach adaptively selects the most similar training data for each scene from the label pool. Furthermore, we also propose a powerful approach within the same framework to enable large-scale labeling in both the 3D space and 2D images. We demonstrate our approach on more than 10,000 images from Google Maps Street View.

Cite

Text

Xiao and Quan. "Multiple View Semantic Segmentation for Street View Images." IEEE/CVF International Conference on Computer Vision, 2009. doi:10.1109/ICCV.2009.5459249

Markdown

[Xiao and Quan. "Multiple View Semantic Segmentation for Street View Images." IEEE/CVF International Conference on Computer Vision, 2009.](https://mlanthology.org/iccv/2009/xiao2009iccv-multiple/) doi:10.1109/ICCV.2009.5459249

BibTeX

@inproceedings{xiao2009iccv-multiple,
  title     = {{Multiple View Semantic Segmentation for Street View Images}},
  author    = {Xiao, Jianxiong and Quan, Long},
  booktitle = {IEEE/CVF International Conference on Computer Vision},
  year      = {2009},
  pages     = {686-693},
  doi       = {10.1109/ICCV.2009.5459249},
  url       = {https://mlanthology.org/iccv/2009/xiao2009iccv-multiple/}
}