ObjectGS: Object-Aware Scene Reconstruction and Scene Understanding via Gaussian Splatting

Abstract

3D Gaussian Splatting is renowned for its high-fidelity reconstructions and real-time novel view synthesis, yet its lack of semantic understanding limits object-level perception. In this work, we propose ObjectGS, an object-aware framework that unifies 3D scene reconstruction with semantic understanding. Instead of treating the scene as a unified whole, ObjectGS models individual objects as local anchors that generate neural Gaussians and share object IDs, enabling precise object-level reconstruction. During training, we dynamically grow or prune these anchors and optimize their features, while a one-hot ID encoding with a classification loss enforces clear semantic constraints. We show through extensive experiments that ObjectGS not only outperforms state-of-the-art methods on open-vocabulary and panoptic segmentation tasks, but also integrates seamlessly with applications like mesh extraction and scene editing.

Cite

Text

Zhu et al. "ObjectGS: Object-Aware Scene Reconstruction and Scene Understanding via Gaussian Splatting." International Conference on Computer Vision, 2025.

Markdown

[Zhu et al. "ObjectGS: Object-Aware Scene Reconstruction and Scene Understanding via Gaussian Splatting." International Conference on Computer Vision, 2025.](https://mlanthology.org/iccv/2025/zhu2025iccv-objectgs/)

BibTeX

@inproceedings{zhu2025iccv-objectgs,
  title     = {{ObjectGS: Object-Aware Scene Reconstruction and Scene Understanding via Gaussian Splatting}},
  author    = {Zhu, Ruijie and Yu, Mulin and Xu, Linning and Jiang, Lihan and Li, Yixuan and Zhang, Tianzhu and Pang, Jiangmiao and Dai, Bo},
  booktitle = {International Conference on Computer Vision},
  year      = {2025},
  pages     = {8350-8360},
  url       = {https://mlanthology.org/iccv/2025/zhu2025iccv-objectgs/}
}