Parametric Local Multimodal Hashing for Cross-View Similarity Search

Abstract

Recent years have witnessed the growing popularity of hashing for efficient large-scale similarity search. It has been shown that the hashing quality could be boosted by hash function learning(HFL). In this paper, we study HFL in the context of multimodal data for cross-view similarity search. We present a novel multimodal HFL method, called Parametric Local Multimodal Hashing(PLMH), which learns a set of hash functions to locally adapt to the data structure of each modality. To balance locality and computational efficiency, the hashing projection matrix of each instance is parameterized, with guaranteed approximation errorbound, as a linear combination of basis hashing projections of a small set of anchor points. A local optimal conjugate gradient algorithm is designed to learn the hash functions for each bit, and the overall hash codes are learned in a sequential manner to progressively minimize the bias. Experimental evaluations on cross-media retrieval tasks demonstrate that PLMH performs competitively against the state-of-the-art methods.

Cite

Text

Zhai et al. "Parametric Local Multimodal Hashing for Cross-View Similarity Search." International Joint Conference on Artificial Intelligence, 2013.

Markdown

[Zhai et al. "Parametric Local Multimodal Hashing for Cross-View Similarity Search." International Joint Conference on Artificial Intelligence, 2013.](https://mlanthology.org/ijcai/2013/zhai2013ijcai-parametric/)

BibTeX

@inproceedings{zhai2013ijcai-parametric,
  title     = {{Parametric Local Multimodal Hashing for Cross-View Similarity Search}},
  author    = {Zhai, Deming and Chang, Hong and Zhen, Yi and Liu, Xianming and Chen, Xilin and Gao, Wen},
  booktitle = {International Joint Conference on Artificial Intelligence},
  year      = {2013},
  pages     = {2754-2760},
  url       = {https://mlanthology.org/ijcai/2013/zhai2013ijcai-parametric/}
}