A Unifying Mutual Information View of Metric Learning: Cross-Entropy vs. Pairwise Losses
Abstract
Recently, substantial research efforts in Deep Metric Learning (DML) focused on designing complex pairwise-distance losses, which require convoluted schemes to ease optimization, such as sample mining or pair weighting. The standard cross-entropy loss for classification has been largely overlooked in DML. On the surface, the cross-entropy may seem unrelated and irrelevant to metric learning as it does not explicitly involve pairwise distances. However, we provide a theoretical analysis that links the cross-entropy to several well-known and recent pairwise losses. Our connections are drawn from two different perspectives: one based on an explicit optimization insight; the other on discriminative and generative views of the mutual information between the labels and the learned features. First, we explicitly demonstrate that the cross-entropy is an upper bound on a new pairwise loss, which has a structure similar to various pairwise losses: it minimizes intra-class distances while maximizing inter-class distances. As a result, minimizing the cross-entropy can be seen as an approximate bound-optimization (or Majorize-Minimize) algorithm for minimizing this pairwise loss. Second, we show that, more generally,minimizing the cross-entropy is actually equivalent to maximizing the mutual information, to which we connect several well-known pairwise losses. Furthermore, we show that various standard pairwise losses can be explicitly related to one another via bound relationships. Our findings indicate that the cross-entropy represents a proxy for maximizing the mutual information – as pairwise losses do – without the need for convoluted sample-mining heuristics. Our experiments†over four standard DML benchmarks strongly support our findings. We obtain state-of-the-art results, outperforming recent and complex DML methods.
Cite
Text
Boudiaf et al. "A Unifying Mutual Information View of Metric Learning: Cross-Entropy vs. Pairwise Losses." Proceedings of the European Conference on Computer Vision (ECCV), 2020. doi:10.1007/978-3-030-58539-6_33Markdown
[Boudiaf et al. "A Unifying Mutual Information View of Metric Learning: Cross-Entropy vs. Pairwise Losses." Proceedings of the European Conference on Computer Vision (ECCV), 2020.](https://mlanthology.org/eccv/2020/boudiaf2020eccv-unifying/) doi:10.1007/978-3-030-58539-6_33BibTeX
@inproceedings{boudiaf2020eccv-unifying,
title = {{A Unifying Mutual Information View of Metric Learning: Cross-Entropy vs. Pairwise Losses}},
author = {Boudiaf, Malik and Rony, Jérôme and Ziko, Imtiaz Masud and Granger, Eric and Pedersoli, Marco and Piantanida, Pablo and Ayed, Ismail Ben},
booktitle = {Proceedings of the European Conference on Computer Vision (ECCV)},
year = {2020},
doi = {10.1007/978-3-030-58539-6_33},
url = {https://mlanthology.org/eccv/2020/boudiaf2020eccv-unifying/}
}