Improving Dynamic HDR Imaging with Fusion Transformer

Abstract

Reconstructing a High Dynamic Range (HDR) image from several Low Dynamic Range (LDR) images with different exposures is a challenging task, especially in the presence of camera and object motion. Though existing models using convolutional neural networks (CNNs) have made great progress, challenges still exist, e.g., ghosting artifacts. Transformers, originating from the field of natural language processing, have shown success in computer vision tasks, due to their ability to address a large receptive field even within a single layer. In this paper, we propose a transformer model for HDR imaging. Our pipeline includes three steps: alignment, fusion, and reconstruction. The key component is the HDR transformer module. Through experiments and ablation studies, we demonstrate that our model outperforms the state-of-the-art by large margins on several popular public datasets.

Cite

Text

Chen et al. "Improving Dynamic HDR Imaging with Fusion Transformer." AAAI Conference on Artificial Intelligence, 2023. doi:10.1609/AAAI.V37I1.25107

Markdown

[Chen et al. "Improving Dynamic HDR Imaging with Fusion Transformer." AAAI Conference on Artificial Intelligence, 2023.](https://mlanthology.org/aaai/2023/chen2023aaai-improving-a/) doi:10.1609/AAAI.V37I1.25107

BibTeX

@inproceedings{chen2023aaai-improving-a,
  title     = {{Improving Dynamic HDR Imaging with Fusion Transformer}},
  author    = {Chen, Rufeng and Zheng, Bolun and Zhang, Hua and Chen, Quan and Yan, Chenggang and Slabaugh, Gregory G. and Yuan, Shanxin},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2023},
  pages     = {340-349},
  doi       = {10.1609/AAAI.V37I1.25107},
  url       = {https://mlanthology.org/aaai/2023/chen2023aaai-improving-a/}
}