Learning Attention from Attention: Efficient Self-Refinement Transformer for Face Super-Resolution

Abstract

Recently, Transformer-based architecture has been introduced into face super-resolution task due to its advantage in capturing long-range dependencies. However, these approaches tend to integrate global information in a large searching region, which neglect to focus on the most relevant information and induce blurry effect by the irrelevant textures. Some improved methods simply constrain self-attention in a local window to suppress the useless information. But it also limits the capability of recovering high-frequency details when flat areas dominate the local searching window. To improve the above issues, we propose a novel self-refinement mechanism which could adaptively achieve texture-aware reconstruction in a coarse-to-fine procedure. Generally, the primary self-attention is first conducted to reconstruct the coarse-grained textures and detect the fine-grained regions required further compensation. Then, region selection attention is performed to refine the textures on these key regions. Since self-attention considers the channel information on tokens equally, we employ a dual-branch feature integration module to privilege the important channels in feature extraction. Furthermore, we design the wavelet fusion module which integrate shallow-layer structure and deep-layer detailed feature to recover realistic face images in frequency domain. Extensive experiments demonstrate the effectiveness on a variety of datasets.

Cite

Text

Li et al. "Learning Attention from Attention: Efficient Self-Refinement Transformer for Face Super-Resolution." International Joint Conference on Artificial Intelligence, 2023. doi:10.24963/IJCAI.2023/115

Markdown

[Li et al. "Learning Attention from Attention: Efficient Self-Refinement Transformer for Face Super-Resolution." International Joint Conference on Artificial Intelligence, 2023.](https://mlanthology.org/ijcai/2023/li2023ijcai-learning/) doi:10.24963/IJCAI.2023/115

BibTeX

@inproceedings{li2023ijcai-learning,
  title     = {{Learning Attention from Attention: Efficient Self-Refinement Transformer for Face Super-Resolution}},
  author    = {Li, Guanxin and Shi, Jingang and Zong, Yuan and Wang, Fei and Wang, Tian and Gong, Yihong},
  booktitle = {International Joint Conference on Artificial Intelligence},
  year      = {2023},
  pages     = {1035-1043},
  doi       = {10.24963/IJCAI.2023/115},
  url       = {https://mlanthology.org/ijcai/2023/li2023ijcai-learning/}
}