Non-Local Contrastive Objectives
Abstract
Pseudo-likelihood and contrastive divergence are two well-known examples of contrastive methods. These algorithms trade off the probability of the correct label with the probabilities of other �nearby� instantiations. In this paper we explore more general types of contrastive objectives, which trade off the probability of the correct label against an arbitrary set of other instantiations. We prove that a large class of contrastive objectives are consistent with maximum likelihood, even for finite amounts of data. This result generalizes asymptotic consistency for pseudo-likelihood. The proof gives significant insight into contrastive objectives, suggesting that they enforce (soft) probability-ratio constraints between pairs of instantiations. Based on this insight, we propose Contrastive Constraint Generation (CCG),an iterative constraint-generation style algorithm that allows us to learn a log-linear model using only MAP inference. We evaluate CCG on a scene classification task, showing that it significantly outperforms pseudo-likelihood, contrastive divergence, and a well-known margin-based method.
Cite
Text
Vickrey et al. "Non-Local Contrastive Objectives." International Conference on Machine Learning, 2010.Markdown
[Vickrey et al. "Non-Local Contrastive Objectives." International Conference on Machine Learning, 2010.](https://mlanthology.org/icml/2010/vickrey2010icml-non/)BibTeX
@inproceedings{vickrey2010icml-non,
title = {{Non-Local Contrastive Objectives}},
author = {Vickrey, David and Lin, Cliff Chiung-Yu and Koller, Daphne},
booktitle = {International Conference on Machine Learning},
year = {2010},
pages = {1103-1110},
url = {https://mlanthology.org/icml/2010/vickrey2010icml-non/}
}