Transformer-Based Selective Super-Resolution for Efficient Image Refinement
Abstract
Conventional super-resolution methods suffer from two drawbacks: substantial computational cost in upscaling an entire large image, and the introduction of extraneous or potentially detrimental information for downstream computer vision tasks during the refinement of the background. To solve these issues, we propose a novel transformer-based algorithm, Selective Super-Resolution (SSR), which partitions images into non-overlapping tiles, selects tiles of interest at various scales with a pyramid architecture, and exclusively reconstructs these selected tiles with deep features. Experimental results on three datasets demonstrate the efficiency and robust performance of our approach for super-resolution. Compared to the state-of-the-art methods, the FID score is reduced from 26.78 to 10.41 with 40% reduction in computation cost for the BDD100K dataset.
Cite
Text
Zhang et al. "Transformer-Based Selective Super-Resolution for Efficient Image Refinement." AAAI Conference on Artificial Intelligence, 2024. doi:10.1609/AAAI.V38I7.28560Markdown
[Zhang et al. "Transformer-Based Selective Super-Resolution for Efficient Image Refinement." AAAI Conference on Artificial Intelligence, 2024.](https://mlanthology.org/aaai/2024/zhang2024aaai-transformer/) doi:10.1609/AAAI.V38I7.28560BibTeX
@inproceedings{zhang2024aaai-transformer,
title = {{Transformer-Based Selective Super-Resolution for Efficient Image Refinement}},
author = {Zhang, Tianyi and Kasichainula, Kishore and Zhuo, Yaoxin and Li, Baoxin and Seo, Jae-Sun and Cao, Yu},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2024},
pages = {7305-7313},
doi = {10.1609/AAAI.V38I7.28560},
url = {https://mlanthology.org/aaai/2024/zhang2024aaai-transformer/}
}