Manhattan-World Urban Reconstruction from Point Clouds

Abstract

Manhattan-world urban scenes are common in the real world. We propose a fully automatic approach for reconstructing such scenes from 3D point samples. Our key idea is to represent the geometry of the buildings in the scene using a set of well-aligned boxes. We first extract plane hypothesis from the points followed by an iterative refinement step. Then, candidate boxes are obtained by partitioning the space of the point cloud into a non-uniform grid. After that, we choose an optimal subset of the candidate boxes to approximate the geometry of the buildings. The contribution of our work is that we transform scene reconstruction into a labeling problem that is solved based on a novel Markov Random Field formulation. Unlike previous methods designed for particular types of input point clouds, our method can obtain faithful reconstructions from a variety of data sources. Experiments demonstrate that our method is superior to state-of-the-art methods.

Cite

Text

Li et al. "Manhattan-World Urban Reconstruction from Point Clouds." European Conference on Computer Vision, 2016. doi:10.1007/978-3-319-46493-0_4

Markdown

[Li et al. "Manhattan-World Urban Reconstruction from Point Clouds." European Conference on Computer Vision, 2016.](https://mlanthology.org/eccv/2016/li2016eccv-manhattan/) doi:10.1007/978-3-319-46493-0_4

BibTeX

@inproceedings{li2016eccv-manhattan,
  title     = {{Manhattan-World Urban Reconstruction from Point Clouds}},
  author    = {Li, Minglei and Wonka, Peter and Nan, Liangliang},
  booktitle = {European Conference on Computer Vision},
  year      = {2016},
  pages     = {54-69},
  doi       = {10.1007/978-3-319-46493-0_4},
  url       = {https://mlanthology.org/eccv/2016/li2016eccv-manhattan/}
}