Video-Rate Video Inpainting
Abstract
This paper considers the problem of video inpainting, i.e., to remove specified objects from an input video. Many methods have been developed for the problem so far, in which there is a trade-off between image quality and computational time. There was no method that can generate high-quality images in video rate. The key to video inpainting is how to establish correspondences from scene regions occluded in a frame to those observed in other frames. To break the trade-off, we propose to use CNNs as a solution to this key problem. We extend existing CNNs for the standard task of optical flow estimation to be able to estimate the flow of occluded background regions. The extension includes augmentation of their architecture and changes of their training method. We experimentally show that this approach works well despite its simplicity, and that a simple video inpainting method integrating this flow estimator runs in video rate (e.g., 32fps for 832 × 448 pixel videos on a standard PC with a GPU) while achieving image quality close to the state-of-the-art.
Cite
Text
Murase et al. "Video-Rate Video Inpainting." IEEE/CVF Winter Conference on Applications of Computer Vision, 2019. doi:10.1109/WACV.2019.00170Markdown
[Murase et al. "Video-Rate Video Inpainting." IEEE/CVF Winter Conference on Applications of Computer Vision, 2019.](https://mlanthology.org/wacv/2019/murase2019wacv-video/) doi:10.1109/WACV.2019.00170BibTeX
@inproceedings{murase2019wacv-video,
title = {{Video-Rate Video Inpainting}},
author = {Murase, Rito and Zhang, Yan and Okatani, Takayuki},
booktitle = {IEEE/CVF Winter Conference on Applications of Computer Vision},
year = {2019},
pages = {1553-1561},
doi = {10.1109/WACV.2019.00170},
url = {https://mlanthology.org/wacv/2019/murase2019wacv-video/}
}