SCNet: Learning Semantic Correspondence

Abstract

This paper addresses the problem of establishing semantic correspondences between images depicting different instances of the same object or scene category. Previous approaches focus on either combining a spatial regularizer with hand-crafted features, or learning a correspondence model for appearance only. We propose instead a convolutional neural network architecture, called SCNet, for learning a geometrically plausible model for semantic correspondence. SCNet uses region proposals as matching primitives, and explicitly incorporates geometric consistency in its loss function. It is trained on image pairs obtained from the PASCAL VOC 2007 keypoint dataset, and a comparative evaluation on several standard benchmarks demonstrates that the proposed approach substantially outperforms both recent deep learning architectures and previous methods based on hand-crafted features.

Cite

Text

Han et al. "SCNet: Learning Semantic Correspondence." International Conference on Computer Vision, 2017. doi:10.1109/ICCV.2017.203

Markdown

[Han et al. "SCNet: Learning Semantic Correspondence." International Conference on Computer Vision, 2017.](https://mlanthology.org/iccv/2017/han2017iccv-scnet/) doi:10.1109/ICCV.2017.203

BibTeX

@inproceedings{han2017iccv-scnet,
  title     = {{SCNet: Learning Semantic Correspondence}},
  author    = {Han, Kai and Rezende, Rafael S. and Ham, Bumsub and Wong, Kwan-Yee K. and Cho, Minsu and Schmid, Cordelia and Ponce, Jean},
  booktitle = {International Conference on Computer Vision},
  year      = {2017},
  doi       = {10.1109/ICCV.2017.203},
  url       = {https://mlanthology.org/iccv/2017/han2017iccv-scnet/}
}