Joint Motion Detection in Neural Videos Training

Abstract

Neural radiance fields (NeRF) can produce photo realistic free-viewpoint images. Recently, incremental neural video training approaches took a step towards interactive streaming via a frame-by-frame approach naturally free of lag. Motion detection in neural videos via a frame-byframe approach can provide valuable cues to enable temporally stable neural videos suitable for interactive streaming. In addition, motion cues can be used to guide the ray sampling phase to model dynamic regions more efficiently. Hence, motion detection can be a key component in telepresence/social networking and immersive cloud gaming applications. In this paper, we propose a novel approach that computes static/dynamic separation masks with high accuracy and spatial coherency across different views together with NeRF optimization process. This is enabled by using explicit deformation network instead of implicit motions/structure layers (novel network architecture) as well as novel specifically designed training schedule. To the best of our knowledge, this is the first work that enables motion estimation via a frame-by-frame approach in a neural video training. The proposed work is desirable as it does not require buffer chunks of frames available before processing and hence is suitable for interactive streaming scenarios. Experimental results shows the effectiveness of the proposed motion detection approach in neural videos.

Cite

Text

Pourian and Supikov. "Joint Motion Detection in Neural Videos Training." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2024. doi:10.1109/CVPRW63382.2024.00578

Markdown

[Pourian and Supikov. "Joint Motion Detection in Neural Videos Training." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2024.](https://mlanthology.org/cvprw/2024/pourian2024cvprw-joint/) doi:10.1109/CVPRW63382.2024.00578

BibTeX

@inproceedings{pourian2024cvprw-joint,
  title     = {{Joint Motion Detection in Neural Videos Training}},
  author    = {Pourian, Niloufar and Supikov, Alexey},
  booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops},
  year      = {2024},
  pages     = {5693-5700},
  doi       = {10.1109/CVPRW63382.2024.00578},
  url       = {https://mlanthology.org/cvprw/2024/pourian2024cvprw-joint/}
}