Diffusion-Based Image Translation with Label Guidance for Domain Adaptive Semantic Segmentation

Abstract

Translating images from a source domain to a target domain for learning target models is one of the most common strategies in domain adaptive semantic segmentation (DASS). However, existing methods still struggle to preserve semantically-consistent local details between the original and translated images. In this work, we present an innovative approach that addresses this challenge by using source-domain labels as explicit guidance during image translation. Concretely, we formulate cross-domain image translation as a denoising diffusion process and utilize a novel Semantic Gradient Guidance (SGG) method to constrain the translation process, conditioning it on the pixel-wise source labels. Additionally, a Progressive Translation Learning (PTL) strategy is devised to enable the SGG method to work reliably across domains with large gaps. Extensive experiments demonstrate the superiority of our approach over state-of-the-art methods.

Cite

Text

Peng et al. "Diffusion-Based Image Translation with Label Guidance for Domain Adaptive Semantic Segmentation." International Conference on Computer Vision, 2023. doi:10.1109/ICCV51070.2023.00081

Markdown

[Peng et al. "Diffusion-Based Image Translation with Label Guidance for Domain Adaptive Semantic Segmentation." International Conference on Computer Vision, 2023.](https://mlanthology.org/iccv/2023/peng2023iccv-diffusionbased/) doi:10.1109/ICCV51070.2023.00081

BibTeX

@inproceedings{peng2023iccv-diffusionbased,
  title     = {{Diffusion-Based Image Translation with Label Guidance for Domain Adaptive Semantic Segmentation}},
  author    = {Peng, Duo and Hu, Ping and Ke, Qiuhong and Liu, Jun},
  booktitle = {International Conference on Computer Vision},
  year      = {2023},
  pages     = {808-820},
  doi       = {10.1109/ICCV51070.2023.00081},
  url       = {https://mlanthology.org/iccv/2023/peng2023iccv-diffusionbased/}
}