Adaptive Reduced Rank Regression

Abstract

We study the low rank regression problem y = Mx + ε, where x and y are d1 and d2 dimensional vectors respectively. We consider the extreme high-dimensional setting where the number of observations n is less than d1 + d2. Existing algorithms are designed for settings where n is typically as large as rank(M)(d1+d2). This work provides an efficient algorithm which only involves two SVD, and establishes statistical guarantees on its performance. The algorithm decouples the problem by first estimating the precision matrix of the features, and then solving the matrix denoising problem. To complement the upper bound, we introduce new techniques for establishing lower bounds on the performance of any algorithm for this problem. Our preliminary experiments confirm that our algorithm often out-performs existing baseline, and is always at least competitive.

Cite

Text

Wu et al. "Adaptive Reduced Rank Regression." Neural Information Processing Systems, 2020.

Markdown

[Wu et al. "Adaptive Reduced Rank Regression." Neural Information Processing Systems, 2020.](https://mlanthology.org/neurips/2020/wu2020neurips-adaptive/)

BibTeX

@inproceedings{wu2020neurips-adaptive,
  title     = {{Adaptive Reduced Rank Regression}},
  author    = {Wu, Qiong and Wong, Felix MF and Li, Yanhua and Liu, Zhenming and Kanade, Varun},
  booktitle = {Neural Information Processing Systems},
  year      = {2020},
  url       = {https://mlanthology.org/neurips/2020/wu2020neurips-adaptive/}
}