MuGS: Multi-Baseline Generalizable Gaussian Splatting Reconstruction

Abstract

We present Multi-Baseline Gaussian Splatting (MuGS), a generalized feed-forward approach for novel view synthesis that effectively handles diverse baseline settings, including sparse input views with both small and large baselines. Specifically, we integrate features from Multi-View Stereo (MVS) and Monocular Depth Estimation (MDE) to enhance feature representations for generalizable reconstruction. Next, We propose a projection-and-sampling mechanism for deep depth fusion, which constructs a fine probability volume to guide the regression of the feature map. Furthermore, We introduce a reference-view loss to improve geometry and optimization efficiency. We leverage 3D Gaussian representations to accelerate training and inference time while enhancing rendering quality. MuGS achieves state-of-the-art performance across multiple baseline settings and diverse scenarios ranging from simple objects (DTU) to complex indoor and outdoor scenes (RealEstate10K). We also demonstrate promising zero-shot performance on the LLFF and Mip-NeRF 360 datasets. Code is available at https://github.com/EuclidLou/MuGS.

Cite

Text

Lou et al. "MuGS: Multi-Baseline Generalizable Gaussian Splatting Reconstruction." International Conference on Computer Vision, 2025.

Markdown

[Lou et al. "MuGS: Multi-Baseline Generalizable Gaussian Splatting Reconstruction." International Conference on Computer Vision, 2025.](https://mlanthology.org/iccv/2025/lou2025iccv-mugs/)

BibTeX

@inproceedings{lou2025iccv-mugs,
  title     = {{MuGS: Multi-Baseline Generalizable Gaussian Splatting Reconstruction}},
  author    = {Lou, Yaopeng and Shen, Liao and Liu, Tianqi and Li, Jiaqi and Huang, Zihao and Sun, Huiqiang and Cao, Zhiguo},
  booktitle = {International Conference on Computer Vision},
  year      = {2025},
  pages     = {25583-25593},
  url       = {https://mlanthology.org/iccv/2025/lou2025iccv-mugs/}
}