Decoder Denoising Pretraining for Semantic Segmentation

Abstract

Semantic segmentation labels are expensive and time consuming to acquire. Hence, pretraining is commonly used to improve the label-efficiency of segmentation models. Typically, the encoder of a segmentation model is pretrained as a classifier and the decoder is randomly initialized. Here, we argue that random initialization of the decoder can be suboptimal, especially when few labeled examples are available. We propose a decoder pretraining approach based on denoising, which can be combined with supervised pretraining of the encoder. We find that decoder denoising pretraining on the ImageNet dataset strongly outperforms encoder-only supervised pretraining. Despite its simplicity, decoder denoising pretraining achieves state-of-the-art results on label-efficient semantic segmentation and offers considerable gains on the Cityscapes, Pascal Context, and ADE20K datasets.

Cite

Text

Brempong et al. "Decoder Denoising Pretraining for Semantic Segmentation." Transactions on Machine Learning Research, 2022.

Markdown

[Brempong et al. "Decoder Denoising Pretraining for Semantic Segmentation." Transactions on Machine Learning Research, 2022.](https://mlanthology.org/tmlr/2022/brempong2022tmlr-decoder/)

BibTeX

@article{brempong2022tmlr-decoder,
  title     = {{Decoder Denoising Pretraining for Semantic Segmentation}},
  author    = {Brempong, Emmanuel Asiedu and Kornblith, Simon and Chen, Ting and Parmar, Niki and Minderer, Matthias and Norouzi, Mohammad},
  journal   = {Transactions on Machine Learning Research},
  year      = {2022},
  url       = {https://mlanthology.org/tmlr/2022/brempong2022tmlr-decoder/}
}