Novel View Synthesis from a Few Glimpses via Test-Time Natural Video Completion
Abstract
Given just a few glimpses of a scene, can you imagine the movie playing out as the camera glides through it? That’s the lens we take on sparse-input novel view synthesis, not only as filling spatial gaps between widely spaced views, but also as completing a natural video unfolding through space. We recast the task as test-time natural video completion, using powerful priors from pretrained video diffusion models to hallucinate plausible in-between views. Our zero-shot, generation-guided framework produces pseudo views at novel camera poses, modulated by an uncertainty-aware mechanism for spatial coherence. These synthesized frames densify supervision for 3D Gaussian Splatting (3D-GS) for scene reconstruction, especially in under-observed regions. An iterative feedback loop lets 3D geometry and 2D view synthesis inform each other, improving both the scene reconstruction and the generated views. The result is coherent, high-fidelity renderings from sparse inputs without any scene-specific training or fine-tuning. On LLFF, DTU, DL3DV, and MipNeRF-360, our method significantly outperforms strong 3D-GS baselines under extreme sparsity. Our project page is at https://decayale.github.io/project/SV2CGS.
Cite
Text
Xu et al. "Novel View Synthesis from a Few Glimpses via Test-Time Natural Video Completion." Advances in Neural Information Processing Systems, 2025.Markdown
[Xu et al. "Novel View Synthesis from a Few Glimpses via Test-Time Natural Video Completion." Advances in Neural Information Processing Systems, 2025.](https://mlanthology.org/neurips/2025/xu2025neurips-novel/)BibTeX
@inproceedings{xu2025neurips-novel,
title = {{Novel View Synthesis from a Few Glimpses via Test-Time Natural Video Completion}},
author = {Xu, Yan and Wang, Yixing and Yu, Stella X.},
booktitle = {Advances in Neural Information Processing Systems},
year = {2025},
url = {https://mlanthology.org/neurips/2025/xu2025neurips-novel/}
}