Self-Supervised Multi-Frame Monocular Scene Flow

Abstract

Estimating 3D scene flow from a sequence of monocular images has been gaining increased attention due to the simple, economical capture setup. Owing to the severe ill-posedness of the problem, the accuracy of current methods has been limited, especially that of efficient, real-time approaches. In this paper, we introduce a multi-frame monocular scene flow network based on self-supervised learning, improving the accuracy over previous networks while retaining real-time efficiency. Based on an advanced two-frame baseline with a split-decoder design, we propose (i) a multi-frame model using a triple frame input and convolutional LSTM connections, (ii) an occlusion-aware census loss for better accuracy, and (iii) a gradient detaching strategy to improve training stability. On the KITTI dataset, we observe state-of-the-art accuracy among monocular scene flow methods based on self-supervised learning.

Cite

Text

Hur and Roth. "Self-Supervised Multi-Frame Monocular Scene Flow." Conference on Computer Vision and Pattern Recognition, 2021. doi:10.1109/CVPR46437.2021.00271

Markdown

[Hur and Roth. "Self-Supervised Multi-Frame Monocular Scene Flow." Conference on Computer Vision and Pattern Recognition, 2021.](https://mlanthology.org/cvpr/2021/hur2021cvpr-selfsupervised/) doi:10.1109/CVPR46437.2021.00271

BibTeX

@inproceedings{hur2021cvpr-selfsupervised,
  title     = {{Self-Supervised Multi-Frame Monocular Scene Flow}},
  author    = {Hur, Junhwa and Roth, Stefan},
  booktitle = {Conference on Computer Vision and Pattern Recognition},
  year      = {2021},
  pages     = {2684-2694},
  doi       = {10.1109/CVPR46437.2021.00271},
  url       = {https://mlanthology.org/cvpr/2021/hur2021cvpr-selfsupervised/}
}