Bayesian Triplet Loss: Uncertainty Quantification in Image Retrieval
Abstract
Uncertainty quantification in image retrieval is crucial for downstream decisions, yet it remains a challenging and largely unexplored problem. Current methods for estimating uncertainties are poorly calibrated, computationally expensive, or based on heuristics. We present a new method that views image embeddings as stochastic features rather than deterministic features. Our two main contributions are (1) a likelihood that matches the triplet constraint and that evaluates the probability of an anchor being closer to a positive than a negative; and (2) a prior over the feature space that justifies the conventional l2 normalization. To ensure computational efficiency, we derive a variational approximation of the posterior, called the Bayesian triplet loss, that produces state-of-the-art uncertainty estimates and matches the predictive performance of current state-of-the-art methods.
Cite
Text
Warburg et al. "Bayesian Triplet Loss: Uncertainty Quantification in Image Retrieval." International Conference on Computer Vision, 2021. doi:10.1109/ICCV48922.2021.01194Markdown
[Warburg et al. "Bayesian Triplet Loss: Uncertainty Quantification in Image Retrieval." International Conference on Computer Vision, 2021.](https://mlanthology.org/iccv/2021/warburg2021iccv-bayesian/) doi:10.1109/ICCV48922.2021.01194BibTeX
@inproceedings{warburg2021iccv-bayesian,
title = {{Bayesian Triplet Loss: Uncertainty Quantification in Image Retrieval}},
author = {Warburg, Frederik and Jørgensen, Martin and Civera, Javier and Hauberg, Søren},
booktitle = {International Conference on Computer Vision},
year = {2021},
pages = {12158-12168},
doi = {10.1109/ICCV48922.2021.01194},
url = {https://mlanthology.org/iccv/2021/warburg2021iccv-bayesian/}
}