Manhattan-World Urban Reconstruction from Point Clouds
Abstract
Manhattan-world urban scenes are common in the real world. We propose a fully automatic approach for reconstructing such scenes from 3D point samples. Our key idea is to represent the geometry of the buildings in the scene using a set of well-aligned boxes. We first extract plane hypothesis from the points followed by an iterative refinement step. Then, candidate boxes are obtained by partitioning the space of the point cloud into a non-uniform grid. After that, we choose an optimal subset of the candidate boxes to approximate the geometry of the buildings. The contribution of our work is that we transform scene reconstruction into a labeling problem that is solved based on a novel Markov Random Field formulation. Unlike previous methods designed for particular types of input point clouds, our method can obtain faithful reconstructions from a variety of data sources. Experiments demonstrate that our method is superior to state-of-the-art methods.
Cite
Text
Li et al. "Manhattan-World Urban Reconstruction from Point Clouds." European Conference on Computer Vision, 2016. doi:10.1007/978-3-319-46493-0_4Markdown
[Li et al. "Manhattan-World Urban Reconstruction from Point Clouds." European Conference on Computer Vision, 2016.](https://mlanthology.org/eccv/2016/li2016eccv-manhattan/) doi:10.1007/978-3-319-46493-0_4BibTeX
@inproceedings{li2016eccv-manhattan,
title = {{Manhattan-World Urban Reconstruction from Point Clouds}},
author = {Li, Minglei and Wonka, Peter and Nan, Liangliang},
booktitle = {European Conference on Computer Vision},
year = {2016},
pages = {54-69},
doi = {10.1007/978-3-319-46493-0_4},
url = {https://mlanthology.org/eccv/2016/li2016eccv-manhattan/}
}