GMMSeg: Gaussian Mixture Based Generative Semantic Segmentation Models
Abstract
Prevalent semantic segmentation solutions are, in essence, a dense discriminative classifier of p(class|pixel feature). Though straightforward, this de facto paradigm neglects the underlying data distribution p(pixel feature|class), and struggles to identify out-of-distribution data. Going beyond this, we propose GMMSeg, a new family of segmentation models that rely on a dense generative classifier for the joint distribution p(pixel feature,class). For each class, GMMSeg builds Gaussian Mixture Models (GMMs) via Expectation-Maximization (EM), so as to capture class-conditional densities. Meanwhile, the deep dense representation is end-to-end trained in a discriminative manner, i.e., maximizing p(class|pixel feature). This endows GMMSeg with the strengths of both generative and discriminative models. With a variety of segmentation architectures and backbones, GMMSeg outperforms the discriminative counterparts on three closed-set datasets. More impressively, without any modification, GMMSeg even performs well on open-world datasets. We believe this work brings fundamental insights into the related fields.
Cite
Text
Liang et al. "GMMSeg: Gaussian Mixture Based Generative Semantic Segmentation Models." Neural Information Processing Systems, 2022.Markdown
[Liang et al. "GMMSeg: Gaussian Mixture Based Generative Semantic Segmentation Models." Neural Information Processing Systems, 2022.](https://mlanthology.org/neurips/2022/liang2022neurips-gmmseg/)BibTeX
@inproceedings{liang2022neurips-gmmseg,
title = {{GMMSeg: Gaussian Mixture Based Generative Semantic Segmentation Models}},
author = {Liang, Chen and Wang, Wenguan and Miao, Jiaxu and Yang, Yi},
booktitle = {Neural Information Processing Systems},
year = {2022},
url = {https://mlanthology.org/neurips/2022/liang2022neurips-gmmseg/}
}