See and Think: Disentangling Semantic Scene Completion

Abstract

Semantic scene completion predicts volumetric occupancy and object category of a 3D scene, which helps intelligent agents to understand and interact with the surroundings. In this work, we propose a disentangled framework, sequentially carrying out 2D semantic segmentation, 2D-3D reprojection and 3D semantic scene completion. This three-stage framework has three advantages: (1) explicit semantic segmentation significantly boosts performance; (2) flexible fusion ways of sensor data bring good extensibility; (3) progress in any subtask will promote the holistic performance. Experimental results show that regardless of inputing a single depth or RGB-D, our framework can generate high-quality semantic scene completion, and outperforms state-of-the-art approaches on both synthetic and real datasets.

Cite

Text

Liu et al. "See and Think: Disentangling Semantic Scene Completion." Neural Information Processing Systems, 2018.

Markdown

[Liu et al. "See and Think: Disentangling Semantic Scene Completion." Neural Information Processing Systems, 2018.](https://mlanthology.org/neurips/2018/liu2018neurips-see/)

BibTeX

@inproceedings{liu2018neurips-see,
  title     = {{See and Think: Disentangling Semantic Scene Completion}},
  author    = {Liu, Shice and Hu, Yu and Zeng, Yiming and Tang, Qiankun and Jin, Beibei and Han, Yinhe and Li, Xiaowei},
  booktitle = {Neural Information Processing Systems},
  year      = {2018},
  pages     = {263-274},
  url       = {https://mlanthology.org/neurips/2018/liu2018neurips-see/}
}