Sampling for Approximate Maximum Search in Factorized Tensor
Abstract
Factorization models have been extensively used for recovering the missing entries of a matrix or tensor. However, directly computing all of the entries using the learned factorization models is prohibitive when the size of the matrix/tensor is large. On the other hand, in many applications, such as collaborative filtering, we are only interested in a few entries that are the largest among them. In this work, we propose a sampling-based approach for finding the top entries of a tensor which is decomposed by the CANDECOMP/PARAFAC model. We develop an algorithm to sample the entries with probabilities proportional to their values. We further extend it to make the sampling proportional to the $k$-th power of the values, amplifying the focus on the top ones. We provide theoretical analysis of the sampling algorithm and evaluate its performance on several real-world data sets. Experimental results indicate that the proposed approach is orders of magnitude faster than exhaustive computing. When applied to the special case of searching in a matrix, it also requires fewer samples than the other state-of-the-art method.
Cite
Text
Lu et al. "Sampling for Approximate Maximum Search in Factorized Tensor." International Joint Conference on Artificial Intelligence, 2017. doi:10.24963/IJCAI.2017/334Markdown
[Lu et al. "Sampling for Approximate Maximum Search in Factorized Tensor." International Joint Conference on Artificial Intelligence, 2017.](https://mlanthology.org/ijcai/2017/lu2017ijcai-sampling/) doi:10.24963/IJCAI.2017/334BibTeX
@inproceedings{lu2017ijcai-sampling,
title = {{Sampling for Approximate Maximum Search in Factorized Tensor}},
author = {Lu, Zhi and Hu, Yang and Zeng, Bing},
booktitle = {International Joint Conference on Artificial Intelligence},
year = {2017},
pages = {2400-2406},
doi = {10.24963/IJCAI.2017/334},
url = {https://mlanthology.org/ijcai/2017/lu2017ijcai-sampling/}
}