Improved Generalization Bound and Learning of Sparsity Patterns for Data-Driven Low-Rank Approximation

Abstract

Learning sketching matrices for fast and accurate low-rank approximation (LRA) has gained increasing attention. Recently, Bartlett, Indyk, and Wagner (COLT 2022) presented a generalization bound for the learning-based LRA. Specifically, for rank-$k$ approximation using an $m \times n$ learned sketching matrix with $s$ non-zeros in each column, they proved an $\tilde O(nsm)$ bound on the fat shattering dimension ($\tilde O$ hides logarithmic factors). We build on their work and make two contributions. (1) We present a better $\tilde O(nsk)$ bound ($k \le m$). En route to obtaining this result, we give a low-complexity Goldberg–Jerrum algorithm for computing pseudo-inverse matrices, which would be of independent interest. (2) We alleviate an assumption of the previous study that sketching matrices have a fixed sparsity pattern. We prove that learning positions of non-zeros increases the fat shattering dimension only by $O(ns\log n)$. In addition, experiments confirm the practical benefit of learning sparsity patterns.

Cite

Text

Sakaue and Oki. "Improved Generalization Bound and Learning of Sparsity Patterns for Data-Driven Low-Rank Approximation." Artificial Intelligence and Statistics, 2023.

Markdown

[Sakaue and Oki. "Improved Generalization Bound and Learning of Sparsity Patterns for Data-Driven Low-Rank Approximation." Artificial Intelligence and Statistics, 2023.](https://mlanthology.org/aistats/2023/sakaue2023aistats-improved/)

BibTeX

@inproceedings{sakaue2023aistats-improved,
  title     = {{Improved Generalization Bound and Learning of Sparsity Patterns for Data-Driven Low-Rank Approximation}},
  author    = {Sakaue, Shinsaku and Oki, Taihei},
  booktitle = {Artificial Intelligence and Statistics},
  year      = {2023},
  pages     = {1-10},
  volume    = {206},
  url       = {https://mlanthology.org/aistats/2023/sakaue2023aistats-improved/}
}