Parallel Multi-Resolution Fusion Network for Image Inpainting

Abstract

Conventional deep image inpainting methods are based on auto-encoder architecture, in which the spatial details of images will be lost in the down-sampling process, leading to the degradation of generated results. Also, the structure information in deep layers and texture information in shallow layers of the auto-encoder architecture can not be well integrated. Differing from the conventional image inpainting architecture, we design a parallel multi-resolution inpainting network with multi-resolution partial convolution, in which low-resolution branches focus on the global structure while high-resolution branches focus on the local texture details. All these high- and low-resolution streams are in parallel and fused repeatedly with multi-resolution masked representation fusion so that the reconstructed images are semantically robust and textually plausible. Experimental results show that our method can effectively fuse structure and texture information, producing more realistic results than state-of-the-art methods.

Cite

Text

Wang et al. "Parallel Multi-Resolution Fusion Network for Image Inpainting." International Conference on Computer Vision, 2021. doi:10.1109/ICCV48922.2021.01429

Markdown

[Wang et al. "Parallel Multi-Resolution Fusion Network for Image Inpainting." International Conference on Computer Vision, 2021.](https://mlanthology.org/iccv/2021/wang2021iccv-parallel/) doi:10.1109/ICCV48922.2021.01429

BibTeX

@inproceedings{wang2021iccv-parallel,
  title     = {{Parallel Multi-Resolution Fusion Network for Image Inpainting}},
  author    = {Wang, Wentao and Zhang, Jianfu and Niu, Li and Ling, Haoyu and Yang, Xue and Zhang, Liqing},
  booktitle = {International Conference on Computer Vision},
  year      = {2021},
  pages     = {14559-14568},
  doi       = {10.1109/ICCV48922.2021.01429},
  url       = {https://mlanthology.org/iccv/2021/wang2021iccv-parallel/}
}