Self-Training Room Layout via Geometry-Aware Ray-Casting

Abstract

In this paper, we introduce a novel geometry-aware self-training framework for room layout estimation models on unseen scenes with unlabeled data. Our approach utilizes a ray-casting formulation to aggregate multiple estimates from different viewing positions, enabling the computation of reliable pseudo-labels for self-training. In particular, our ray-casting approach enforces multi-view consistency along all ray directions and prioritizes spatial proximity to the camera view for geometry reasoning. As a result, our geometry-aware pseudo-labels effectively handle complex room geometries and occluded walls without relying on assumptions such as Manhattan World or planar room walls. Evaluation on publicly available datasets, including synthetic and real-world scenarios, demonstrates significant improvements in current state-of-the-art layout models without using any human annotation.

Cite

Text

Solarte et al. "Self-Training Room Layout via Geometry-Aware Ray-Casting." Proceedings of the European Conference on Computer Vision (ECCV), 2024. doi:10.1007/978-3-031-72995-9_15

Markdown

[Solarte et al. "Self-Training Room Layout via Geometry-Aware Ray-Casting." Proceedings of the European Conference on Computer Vision (ECCV), 2024.](https://mlanthology.org/eccv/2024/solarte2024eccv-selftraining/) doi:10.1007/978-3-031-72995-9_15

BibTeX

@inproceedings{solarte2024eccv-selftraining,
  title     = {{Self-Training Room Layout via Geometry-Aware Ray-Casting}},
  author    = {Solarte, Bolivar and Wu, Chin-Hsuan and Jhang, Jin-Cheng and Lee, Jonathan and Tsai, Yi-Hsuan and Sun, Min},
  booktitle = {Proceedings of the European Conference on Computer Vision (ECCV)},
  year      = {2024},
  doi       = {10.1007/978-3-031-72995-9_15},
  url       = {https://mlanthology.org/eccv/2024/solarte2024eccv-selftraining/}
}