Neural Harmonics: Bridging Spectral Embedding and Matrix Completion in Self-Supervised Learning
Abstract
Self-supervised methods received tremendous attention thanks to their seemingly heuristic approach to learning representations that respect the semantics of the data without any apparent supervision in the form of labels. A growing body of literature is already being published in an attempt to build a coherent and theoretically grounded understanding of the workings of a zoo of losses used in modern self-supervised representation learning methods. In this paper, we attempt to provide an understanding from the perspective of a Laplace operator and connect the inductive bias stemming from the augmentation process to a low-rank matrix completion problem.To this end, we leverage the results from low-rank matrix completion to provide theoretical analysis on the convergence of modern SSL methods and a key property that affects their downstream performance.
Cite
Text
Munkhoeva and Oseledets. "Neural Harmonics: Bridging Spectral Embedding and Matrix Completion in Self-Supervised Learning." Neural Information Processing Systems, 2023.Markdown
[Munkhoeva and Oseledets. "Neural Harmonics: Bridging Spectral Embedding and Matrix Completion in Self-Supervised Learning." Neural Information Processing Systems, 2023.](https://mlanthology.org/neurips/2023/munkhoeva2023neurips-neural/)BibTeX
@inproceedings{munkhoeva2023neurips-neural,
title = {{Neural Harmonics: Bridging Spectral Embedding and Matrix Completion in Self-Supervised Learning}},
author = {Munkhoeva, Marina and Oseledets, Ivan},
booktitle = {Neural Information Processing Systems},
year = {2023},
url = {https://mlanthology.org/neurips/2023/munkhoeva2023neurips-neural/}
}