VORNet: Spatio-Temporally Consistent Video Inpainting for Object Removal
Abstract
Video object removal is a challenging task in video processing that often requires massive human efforts. Given the mask of the foreground object in each frame, the goal is to complete (inpaint) the object region and generate a video without the target object. While recently deep learning based methods have achieved great success on the image inpainting task, they often lead to inconsistent results between frames when applied to videos. In this work, we propose a novel learning-based Video Object Removal Network (VORNet) to solve the video object removal task in a spatio-temporally consistent manner, by combining the optical flow warping and image-based inpainting model. Experiments are done on our Synthesized Video Object Removal (SVOR) dataset based on the YouTube-VOS video segmentation dataset, and both the objective and subjective evaluation demonstrate that our VORNet generates more spatially and temporally consistent videos compared with existing methods.
Cite
Text
Chang et al. "VORNet: Spatio-Temporally Consistent Video Inpainting for Object Removal." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2019. doi:10.1109/CVPRW.2019.00229Markdown
[Chang et al. "VORNet: Spatio-Temporally Consistent Video Inpainting for Object Removal." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2019.](https://mlanthology.org/cvprw/2019/chang2019cvprw-vornet/) doi:10.1109/CVPRW.2019.00229BibTeX
@inproceedings{chang2019cvprw-vornet,
title = {{VORNet: Spatio-Temporally Consistent Video Inpainting for Object Removal}},
author = {Chang, Ya-Liang and Liu, Zhe Yu and Hsu, Winston H.},
booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops},
year = {2019},
pages = {1785-1794},
doi = {10.1109/CVPRW.2019.00229},
url = {https://mlanthology.org/cvprw/2019/chang2019cvprw-vornet/}
}