HyperNVD: Accelerating Neural Video Decomposition via Hypernetworks

Abstract

Decomposing a video into a layer-based representation is crucial for easy video editing for the creative industries, as it enables independent editing of specific layers. Existing video-layer decomposition models rely on implicit neural representations (INRs) trained independently for each video, making the process time-consuming when applied to new videos. Noticing this limitation, we propose a meta-learning strategy to learn a generic video decomposition model to speed up the training on new videos. Our model is based on a hypernetwork architecture which, given a video-encoder embedding, generates the parameters for a compact INR-based neural video decomposition model. Our strategy mitigates the problem of single-video overfitting and, importantly, shortens the convergence of video decomposition on new, unseen videos. Our code is available at: https://hypernvd.github.io/

Cite

Text

Pilligua et al. "HyperNVD: Accelerating Neural Video Decomposition via Hypernetworks." Conference on Computer Vision and Pattern Recognition, 2025. doi:10.1109/CVPR52734.2025.02135

Markdown

[Pilligua et al. "HyperNVD: Accelerating Neural Video Decomposition via Hypernetworks." Conference on Computer Vision and Pattern Recognition, 2025.](https://mlanthology.org/cvpr/2025/pilligua2025cvpr-hypernvd/) doi:10.1109/CVPR52734.2025.02135

BibTeX

@inproceedings{pilligua2025cvpr-hypernvd,
  title     = {{HyperNVD: Accelerating Neural Video Decomposition via Hypernetworks}},
  author    = {Pilligua, Maria and Xue, Danna and Vazquez-Corral, Javier},
  booktitle = {Conference on Computer Vision and Pattern Recognition},
  year      = {2025},
  pages     = {22933-22942},
  doi       = {10.1109/CVPR52734.2025.02135},
  url       = {https://mlanthology.org/cvpr/2025/pilligua2025cvpr-hypernvd/}
}