Understanding Deflation Process in Over-Parametrized Tensor Decomposition

Abstract

In this paper we study the training dynamics for gradient flow on over-parametrized tensor decomposition problems. Empirically, such training process often first fits larger components and then discovers smaller components, which is similar to a tensor deflation process that is commonly used in tensor decomposition algorithms. We prove that for orthogonally decomposable tensor, a slightly modified version of gradient flow would follow a tensor deflation process and recover all the tensor components. Our proof suggests that for orthogonal tensors, gradient flow dynamics works similarly as greedy low-rank learning in the matrix setting, which is a first step towards understanding the implicit regularization effect of over-parametrized models for low-rank tensors.

Cite

Text

Ge et al. "Understanding Deflation Process in Over-Parametrized Tensor Decomposition." Neural Information Processing Systems, 2021.

Markdown

[Ge et al. "Understanding Deflation Process in Over-Parametrized Tensor Decomposition." Neural Information Processing Systems, 2021.](https://mlanthology.org/neurips/2021/ge2021neurips-understanding/)

BibTeX

@inproceedings{ge2021neurips-understanding,
  title     = {{Understanding Deflation Process in Over-Parametrized Tensor Decomposition}},
  author    = {Ge, Rong and Ren, Yunwei and Wang, Xiang and Zhou, Mo},
  booktitle = {Neural Information Processing Systems},
  year      = {2021},
  url       = {https://mlanthology.org/neurips/2021/ge2021neurips-understanding/}
}