Imposing Consistency for Optical Flow Estimation
Abstract
Imposing consistency through proxy tasks has been shown to enhance data-driven learning and enable self-supervision in various tasks. This paper introduces novel and effective consistency strategies for optical flow estimation, a problem where labels from real-world data are very challenging to derive. More specifically, we propose occlusion consistency and zero forcing in the forms of self-supervised learning and transformation consistency in the form of semi-supervised learning. We apply these consistency techniques in a way that the network model learns to describe pixel-level motions better while requiring no additional annotations. We demonstrate that our consistency strategies applied to a strong baseline network model using the original datasets and labels provide further improvements, attaining the state-of-the-art results on the KITTI-2015 scene flow benchmark in the non-stereo category. Our method achieves the best foreground accuracy (4.33% in Fl-all) over both the stereo and non-stereo categories, even though using only monocular image inputs.
Cite
Text
Jeong et al. "Imposing Consistency for Optical Flow Estimation." Conference on Computer Vision and Pattern Recognition, 2022. doi:10.1109/CVPR52688.2022.00318Markdown
[Jeong et al. "Imposing Consistency for Optical Flow Estimation." Conference on Computer Vision and Pattern Recognition, 2022.](https://mlanthology.org/cvpr/2022/jeong2022cvpr-imposing/) doi:10.1109/CVPR52688.2022.00318BibTeX
@inproceedings{jeong2022cvpr-imposing,
title = {{Imposing Consistency for Optical Flow Estimation}},
author = {Jeong, Jisoo and Lin, Jamie Menjay and Porikli, Fatih and Kwak, Nojun},
booktitle = {Conference on Computer Vision and Pattern Recognition},
year = {2022},
pages = {3181-3191},
doi = {10.1109/CVPR52688.2022.00318},
url = {https://mlanthology.org/cvpr/2022/jeong2022cvpr-imposing/}
}