LGT-Net: Indoor Panoramic Room Layout Estimation with Geometry-Aware Transformer Network
Abstract
3D room layout estimation by a single panorama using deep neural networks has made great progress. However, previous approaches can not obtain efficient geometry awareness of room layout with the only latitude of boundaries or horizon-depth. We present that using horizon-depth along with room height can obtain omnidirectional-geometry awareness of room layout in both horizontal and vertical directions. In addition, we propose a planar-geometry aware loss function with normals and gradients of normals to supervise the planeness of walls and turning of corners. We propose an efficient network, LGT-Net, for room layout estimation, which contains a novel Transformer architecture called SWG-Transformer to model geometry relations. SWG-Transformer consists of (Shifted) Window Blocks and Global Blocks to combine the local and global geometry relations. Moreover, we design a novel relative position embedding of Transformer to enhance the spatial identification ability for the panorama. Experiments show that the proposed LGT-Net achieves better performance than current state-of-the-arts (SOTA) on benchmark datasets.
Cite
Text
Jiang et al. "LGT-Net: Indoor Panoramic Room Layout Estimation with Geometry-Aware Transformer Network." Conference on Computer Vision and Pattern Recognition, 2022. doi:10.1109/CVPR52688.2022.00170Markdown
[Jiang et al. "LGT-Net: Indoor Panoramic Room Layout Estimation with Geometry-Aware Transformer Network." Conference on Computer Vision and Pattern Recognition, 2022.](https://mlanthology.org/cvpr/2022/jiang2022cvpr-lgtnet/) doi:10.1109/CVPR52688.2022.00170BibTeX
@inproceedings{jiang2022cvpr-lgtnet,
title = {{LGT-Net: Indoor Panoramic Room Layout Estimation with Geometry-Aware Transformer Network}},
author = {Jiang, Zhigang and Xiang, Zhongzheng and Xu, Jinhua and Zhao, Ming},
booktitle = {Conference on Computer Vision and Pattern Recognition},
year = {2022},
pages = {1654-1663},
doi = {10.1109/CVPR52688.2022.00170},
url = {https://mlanthology.org/cvpr/2022/jiang2022cvpr-lgtnet/}
}