Query-Aware Quantization for Maximum Inner Product Search

Abstract

Maximum Inner Product Search (MIPS) plays an essential role in many applications ranging from information retrieval, recommender systems to natural language processing. However, exhaustive MIPS is often expensive and impractical when there are a large number of candidate items. The state-of-the-art quantization method of approximated MIPS is product quantization with a score-aware loss, developed by assuming that queries are uniformly distributed in the unit sphere. However, in real-world datasets, the above assumption about queries does not necessarily hold. To this end, we propose a quantization method based on the distribution of queries combined with sampled softmax. Further, we introduce a general framework encompassing the proposed method and multiple quantization methods, and we develop an effective optimization for the proposed general framework. The proposed method is evaluated on three real-world datasets. The experimental results show that it outperforms the state-of-the-art baselines.

Cite

Text

Zhang et al. "Query-Aware Quantization for Maximum Inner Product Search." AAAI Conference on Artificial Intelligence, 2023. doi:10.1609/AAAI.V37I4.25613

Markdown

[Zhang et al. "Query-Aware Quantization for Maximum Inner Product Search." AAAI Conference on Artificial Intelligence, 2023.](https://mlanthology.org/aaai/2023/zhang2023aaai-query/) doi:10.1609/AAAI.V37I4.25613

BibTeX

@inproceedings{zhang2023aaai-query,
  title     = {{Query-Aware Quantization for Maximum Inner Product Search}},
  author    = {Zhang, Jin and Lian, Defu and Zhang, Haodi and Wang, Baoyun and Chen, Enhong},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2023},
  pages     = {4875-4883},
  doi       = {10.1609/AAAI.V37I4.25613},
  url       = {https://mlanthology.org/aaai/2023/zhang2023aaai-query/}
}