LGT-Net: Indoor Panoramic Room Layout Estimation with Geometry-Aware Transformer Network

Abstract

3D room layout estimation by a single panorama using deep neural networks has made great progress. However, previous approaches can not obtain efficient geometry awareness of room layout with the only latitude of boundaries or horizon-depth. We present that using horizon-depth along with room height can obtain omnidirectional-geometry awareness of room layout in both horizontal and vertical directions. In addition, we propose a planar-geometry aware loss function with normals and gradients of normals to supervise the planeness of walls and turning of corners. We propose an efficient network, LGT-Net, for room layout estimation, which contains a novel Transformer architecture called SWG-Transformer to model geometry relations. SWG-Transformer consists of (Shifted) Window Blocks and Global Blocks to combine the local and global geometry relations. Moreover, we design a novel relative position embedding of Transformer to enhance the spatial identification ability for the panorama. Experiments show that the proposed LGT-Net achieves better performance than current state-of-the-arts (SOTA) on benchmark datasets.

Cite

Text

Jiang et al. "LGT-Net: Indoor Panoramic Room Layout Estimation with Geometry-Aware Transformer Network." Conference on Computer Vision and Pattern Recognition, 2022. doi:10.1109/CVPR52688.2022.00170

Markdown

[Jiang et al. "LGT-Net: Indoor Panoramic Room Layout Estimation with Geometry-Aware Transformer Network." Conference on Computer Vision and Pattern Recognition, 2022.](https://mlanthology.org/cvpr/2022/jiang2022cvpr-lgtnet/) doi:10.1109/CVPR52688.2022.00170

BibTeX

@inproceedings{jiang2022cvpr-lgtnet,
  title     = {{LGT-Net: Indoor Panoramic Room Layout Estimation with Geometry-Aware Transformer Network}},
  author    = {Jiang, Zhigang and Xiang, Zhongzheng and Xu, Jinhua and Zhao, Ming},
  booktitle = {Conference on Computer Vision and Pattern Recognition},
  year      = {2022},
  pages     = {1654-1663},
  doi       = {10.1109/CVPR52688.2022.00170},
  url       = {https://mlanthology.org/cvpr/2022/jiang2022cvpr-lgtnet/}
}