Unsupervised Learning of Discriminative Attributes and Visual Representations

Abstract

Attributes offer useful mid-level features to interpret visual data. While most attribute learning methods are supervised by costly human-generated labels, we introduce a simple yet powerful unsupervised approach to learn and predict visual attributes directly from data. Given a large unlabeled image collection as input, we train deep Convolutional Neural Networks (CNNs) to output a set of discriminative, binary attributes often with semantic meanings. Specifically, we first train a CNN coupled with unsupervised discriminative clustering, and then use the cluster membership as a soft supervision to discover shared attributes from the clusters while maximizing their separability. The learned attributes are shown to be capable of encoding rich imagery properties from both natural images and contour patches. The visual representations learned in this way are also transferrable to other tasks such as object detection. We show other convincing results on the related tasks of image retrieval and classification, and contour detection.

Cite

Text

Huang et al. "Unsupervised Learning of Discriminative Attributes and Visual Representations." Conference on Computer Vision and Pattern Recognition, 2016. doi:10.1109/CVPR.2016.559

Markdown

[Huang et al. "Unsupervised Learning of Discriminative Attributes and Visual Representations." Conference on Computer Vision and Pattern Recognition, 2016.](https://mlanthology.org/cvpr/2016/huang2016cvpr-unsupervised/) doi:10.1109/CVPR.2016.559

BibTeX

@inproceedings{huang2016cvpr-unsupervised,
  title     = {{Unsupervised Learning of Discriminative Attributes and Visual Representations}},
  author    = {Huang, Chen and Loy, Chen Change and Tang, Xiaoou},
  booktitle = {Conference on Computer Vision and Pattern Recognition},
  year      = {2016},
  doi       = {10.1109/CVPR.2016.559},
  url       = {https://mlanthology.org/cvpr/2016/huang2016cvpr-unsupervised/}
}