Sampling for Approximate Maximum Search in Factorized Tensor

Abstract

Factorization models have been extensively used for recovering the missing entries of a matrix or tensor. However, directly computing all of the entries using the learned factorization models is prohibitive when the size of the matrix/tensor is large. On the other hand, in many applications, such as collaborative filtering, we are only interested in a few entries that are the largest among them. In this work, we propose a sampling-based approach for finding the top entries of a tensor which is decomposed by the CANDECOMP/PARAFAC model. We develop an algorithm to sample the entries with probabilities proportional to their values. We further extend it to make the sampling proportional to the $k$-th power of the values, amplifying the focus on the top ones. We provide theoretical analysis of the sampling algorithm and evaluate its performance on several real-world data sets. Experimental results indicate that the proposed approach is orders of magnitude faster than exhaustive computing. When applied to the special case of searching in a matrix, it also requires fewer samples than the other state-of-the-art method.

Cite

Text

Lu et al. "Sampling for Approximate Maximum Search in Factorized Tensor." International Joint Conference on Artificial Intelligence, 2017. doi:10.24963/IJCAI.2017/334

Markdown

[Lu et al. "Sampling for Approximate Maximum Search in Factorized Tensor." International Joint Conference on Artificial Intelligence, 2017.](https://mlanthology.org/ijcai/2017/lu2017ijcai-sampling/) doi:10.24963/IJCAI.2017/334

BibTeX

@inproceedings{lu2017ijcai-sampling,
  title     = {{Sampling for Approximate Maximum Search in Factorized Tensor}},
  author    = {Lu, Zhi and Hu, Yang and Zeng, Bing},
  booktitle = {International Joint Conference on Artificial Intelligence},
  year      = {2017},
  pages     = {2400-2406},
  doi       = {10.24963/IJCAI.2017/334},
  url       = {https://mlanthology.org/ijcai/2017/lu2017ijcai-sampling/}
}