Subquadratic Kronecker Regression with Applications to Tensor Decomposition
Abstract
Kronecker regression is a highly-structured least squares problem $\min_{\mathbf{x}} \lVert \mathbf{K}\mathbf{x} - \mathbf{b} \rVert_{2}^2$, where the design matrix $\mathbf{K} = \mathbf{A}^{(1)} \otimes \cdots \otimes \mathbf{A}^{(N)}$ is a Kronecker product of factor matrices. This regression problem arises in each step of the widely-used alternating least squares (ALS) algorithm for computing the Tucker decomposition of a tensor. We present the first subquadratic-time algorithm for solving Kronecker regression to a $(1+\varepsilon)$-approximation that avoids the exponential term $O(\varepsilon^{-N})$ in the running time. Our techniques combine leverage score sampling and iterative methods. By extending our approach to block-design matrices where one block is a Kronecker product, we also achieve subquadratic-time algorithms for (1) Kronecker ridge regression and (2) updating the factor matrix of a Tucker decomposition in ALS, which is not a pure Kronecker regression problem, thereby improving the running time of all steps of Tucker ALS. We demonstrate the speed and accuracy of this Kronecker regression algorithm on synthetic data and real-world image tensors.
Cite
Text
Fahrbach et al. "Subquadratic Kronecker Regression with Applications to Tensor Decomposition." Neural Information Processing Systems, 2022.Markdown
[Fahrbach et al. "Subquadratic Kronecker Regression with Applications to Tensor Decomposition." Neural Information Processing Systems, 2022.](https://mlanthology.org/neurips/2022/fahrbach2022neurips-subquadratic/)BibTeX
@inproceedings{fahrbach2022neurips-subquadratic,
title = {{Subquadratic Kronecker Regression with Applications to Tensor Decomposition}},
author = {Fahrbach, Matthew and Fu, Gang and Ghadiri, Mehrdad},
booktitle = {Neural Information Processing Systems},
year = {2022},
url = {https://mlanthology.org/neurips/2022/fahrbach2022neurips-subquadratic/}
}