Non-Local Contrastive Objectives

Abstract

Pseudo-likelihood and contrastive divergence are two well-known examples of contrastive methods. These algorithms trade off the probability of the correct label with the probabilities of other �nearby� instantiations. In this paper we explore more general types of contrastive objectives, which trade off the probability of the correct label against an arbitrary set of other instantiations. We prove that a large class of contrastive objectives are consistent with maximum likelihood, even for finite amounts of data. This result generalizes asymptotic consistency for pseudo-likelihood. The proof gives significant insight into contrastive objectives, suggesting that they enforce (soft) probability-ratio constraints between pairs of instantiations. Based on this insight, we propose Contrastive Constraint Generation (CCG),an iterative constraint-generation style algorithm that allows us to learn a log-linear model using only MAP inference. We evaluate CCG on a scene classification task, showing that it significantly outperforms pseudo-likelihood, contrastive divergence, and a well-known margin-based method.

Cite

Text

Vickrey et al. "Non-Local Contrastive Objectives." International Conference on Machine Learning, 2010.

Markdown

[Vickrey et al. "Non-Local Contrastive Objectives." International Conference on Machine Learning, 2010.](https://mlanthology.org/icml/2010/vickrey2010icml-non/)

BibTeX

@inproceedings{vickrey2010icml-non,
  title     = {{Non-Local Contrastive Objectives}},
  author    = {Vickrey, David and Lin, Cliff Chiung-Yu and Koller, Daphne},
  booktitle = {International Conference on Machine Learning},
  year      = {2010},
  pages     = {1103-1110},
  url       = {https://mlanthology.org/icml/2010/vickrey2010icml-non/}
}