Image Segmentation by Figure-Ground Composition into Maximal Cliques

Abstract

We propose a mid-level statistical model for image segmentation that composes multiple figure-ground hypotheses (FG) obtained by applying constraints at different locations and scales, into larger interpretations (tilings) of the entire image. Inference is cast as optimization over sets of maximal cliques sampled from a graph connecting all non-overlapping figure-ground segment hypotheses. Potential functions over cliques combine unary, Gestalt-based figure qualities, and pairwise compatibilities among spatially neighboring segments, constrained by T-junctions and the boundary interface statistics of real scenes. Learning the model parameters is based on maximum likelihood, alternating between sampling image tilings and optimizing their potential function parameters. State of the art results are reported on the Berkeley and Stanford segmentation datasets, as well as VOC2009, where a 28% improvement was achieved.

Cite

Text

Ion et al. "Image Segmentation by Figure-Ground Composition into Maximal Cliques." IEEE/CVF International Conference on Computer Vision, 2011. doi:10.1109/ICCV.2011.6126486

Markdown

[Ion et al. "Image Segmentation by Figure-Ground Composition into Maximal Cliques." IEEE/CVF International Conference on Computer Vision, 2011.](https://mlanthology.org/iccv/2011/ion2011iccv-image/) doi:10.1109/ICCV.2011.6126486

BibTeX

@inproceedings{ion2011iccv-image,
  title     = {{Image Segmentation by Figure-Ground Composition into Maximal Cliques}},
  author    = {Ion, Adrian and Carreira, João and Sminchisescu, Cristian},
  booktitle = {IEEE/CVF International Conference on Computer Vision},
  year      = {2011},
  pages     = {2110-2117},
  doi       = {10.1109/ICCV.2011.6126486},
  url       = {https://mlanthology.org/iccv/2011/ion2011iccv-image/}
}