Temporal Adaptive Alignment Network for Deep Video Inpainting
Abstract
Video inpainting aims to synthesize visually pleasant and temporally consistent content in missing regions of video. Due to a variety of motions across different frames, it is highly challenging to utilize effective temporal information to recover videos. Existing deep learning based methods usually estimate optical flow to align frames and thereby exploit useful information between frames. However, these methods tend to generate artifacts once the estimated optical flow is inaccurate. To alleviate above problem, we propose a novel end-to-end Temporal Adaptive Alignment Network(TAAN) for video inpainting. The TAAN aligns reference frames with target frame via implicit motion estimation at a feature level and then reconstruct target frame by taking the aggregated aligned reference frame features as input. In the proposed network, a Temporal Adaptive Alignment (TAA) module based on deformable convolutions is designed to perform temporal alignment in a local, dense and adaptive manner. Both quantitative and qualitative evaluation results show that our method significantly outperforms existing deep learning based methods.
Cite
Text
Liu et al. "Temporal Adaptive Alignment Network for Deep Video Inpainting." International Joint Conference on Artificial Intelligence, 2020. doi:10.24963/IJCAI.2020/129Markdown
[Liu et al. "Temporal Adaptive Alignment Network for Deep Video Inpainting." International Joint Conference on Artificial Intelligence, 2020.](https://mlanthology.org/ijcai/2020/liu2020ijcai-temporal/) doi:10.24963/IJCAI.2020/129BibTeX
@inproceedings{liu2020ijcai-temporal,
title = {{Temporal Adaptive Alignment Network for Deep Video Inpainting}},
author = {Liu, Ruixin and Weng, Zhenyu and Zhu, Yuesheng and Li, Bairong},
booktitle = {International Joint Conference on Artificial Intelligence},
year = {2020},
pages = {927-933},
doi = {10.24963/IJCAI.2020/129},
url = {https://mlanthology.org/ijcai/2020/liu2020ijcai-temporal/}
}