Distance Metric Learning Revisited
Abstract
The success of many machine learning algorithms (e.g. the nearest neighborhood classification and k-means clustering) depends on the representation of the data as elements in a metric space. Learning an appropriate distance metric from data is usually superior to the default Euclidean distance. In this paper, we revisit the original model proposed by Xing et al. [25] and propose a general formulation of learning a Mahalanobis distance from data. We prove that this novel formulation is equivalent to a convex optimization problem over the spectrahedron. Then, a gradient-based optimization algorithm is proposed to obtain the optimal solution which only needs the computation of the largest eigenvalue of a matrix per iteration. Finally, experiments on various UCI datasets and a benchmark face verification dataset called Labeled Faces in the Wild (LFW) demonstrate that the proposed method compares competitively to those state-of-the-art methods.
Cite
Text
Cao et al. "Distance Metric Learning Revisited." European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, 2012. doi:10.1007/978-3-642-33460-3_24Markdown
[Cao et al. "Distance Metric Learning Revisited." European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, 2012.](https://mlanthology.org/ecmlpkdd/2012/cao2012ecmlpkdd-distance/) doi:10.1007/978-3-642-33460-3_24BibTeX
@inproceedings{cao2012ecmlpkdd-distance,
title = {{Distance Metric Learning Revisited}},
author = {Cao, Qiong and Ying, Yiming and Li, Peng},
booktitle = {European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases},
year = {2012},
pages = {283-298},
doi = {10.1007/978-3-642-33460-3_24},
url = {https://mlanthology.org/ecmlpkdd/2012/cao2012ecmlpkdd-distance/}
}