Self-Training Room Layout via Geometry-Aware Ray-Casting
Abstract
In this paper, we introduce a novel geometry-aware self-training framework for room layout estimation models on unseen scenes with unlabeled data. Our approach utilizes a ray-casting formulation to aggregate multiple estimates from different viewing positions, enabling the computation of reliable pseudo-labels for self-training. In particular, our ray-casting approach enforces multi-view consistency along all ray directions and prioritizes spatial proximity to the camera view for geometry reasoning. As a result, our geometry-aware pseudo-labels effectively handle complex room geometries and occluded walls without relying on assumptions such as Manhattan World or planar room walls. Evaluation on publicly available datasets, including synthetic and real-world scenarios, demonstrates significant improvements in current state-of-the-art layout models without using any human annotation.
Cite
Text
Solarte et al. "Self-Training Room Layout via Geometry-Aware Ray-Casting." Proceedings of the European Conference on Computer Vision (ECCV), 2024. doi:10.1007/978-3-031-72995-9_15Markdown
[Solarte et al. "Self-Training Room Layout via Geometry-Aware Ray-Casting." Proceedings of the European Conference on Computer Vision (ECCV), 2024.](https://mlanthology.org/eccv/2024/solarte2024eccv-selftraining/) doi:10.1007/978-3-031-72995-9_15BibTeX
@inproceedings{solarte2024eccv-selftraining,
title = {{Self-Training Room Layout via Geometry-Aware Ray-Casting}},
author = {Solarte, Bolivar and Wu, Chin-Hsuan and Jhang, Jin-Cheng and Lee, Jonathan and Tsai, Yi-Hsuan and Sun, Min},
booktitle = {Proceedings of the European Conference on Computer Vision (ECCV)},
year = {2024},
doi = {10.1007/978-3-031-72995-9_15},
url = {https://mlanthology.org/eccv/2024/solarte2024eccv-selftraining/}
}