Distance Metric Learning Revisited

Abstract

The success of many machine learning algorithms (e.g. the nearest neighborhood classification and k-means clustering) depends on the representation of the data as elements in a metric space. Learning an appropriate distance metric from data is usually superior to the default Euclidean distance. In this paper, we revisit the original model proposed by Xing et al. [25] and propose a general formulation of learning a Mahalanobis distance from data. We prove that this novel formulation is equivalent to a convex optimization problem over the spectrahedron. Then, a gradient-based optimization algorithm is proposed to obtain the optimal solution which only needs the computation of the largest eigenvalue of a matrix per iteration. Finally, experiments on various UCI datasets and a benchmark face verification dataset called Labeled Faces in the Wild (LFW) demonstrate that the proposed method compares competitively to those state-of-the-art methods.

Cite

Text

Cao et al. "Distance Metric Learning Revisited." European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, 2012. doi:10.1007/978-3-642-33460-3_24

Markdown

[Cao et al. "Distance Metric Learning Revisited." European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, 2012.](https://mlanthology.org/ecmlpkdd/2012/cao2012ecmlpkdd-distance/) doi:10.1007/978-3-642-33460-3_24

BibTeX

@inproceedings{cao2012ecmlpkdd-distance,
  title     = {{Distance Metric Learning Revisited}},
  author    = {Cao, Qiong and Ying, Yiming and Li, Peng},
  booktitle = {European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases},
  year      = {2012},
  pages     = {283-298},
  doi       = {10.1007/978-3-642-33460-3_24},
  url       = {https://mlanthology.org/ecmlpkdd/2012/cao2012ecmlpkdd-distance/}
}