Self-Supervised Representation Learning from Random Data Projectors

Abstract

Self-supervised representation learning SSRL has advanced considerably by exploiting the transformation invariance assumption under artificially designed data augmentations. While augmentation-based SSRL algorithms push the boundaries of performance in computer vision and natural language processing, they are often not directly applicable to other data modalities such as tabular and time-series data. This paper presents an SSRL approach that can be applied to these data modalities because it does not rely on augmentations or masking. Specifically, we show that high-quality data representations can be learned by reconstructing random data projections. We evaluate the proposed approach on real-world applications with tabular and time-series data. We show that it outperforms multiple state-of-the-art SSRL baselines and is competitive with methods built on domain-specific knowledge. Due to its wide applicability and strong empirical results, we argue that learning from randomness is a fruitful research direction worthy of attention and further study.

Cite

Text

Sui et al. "Self-Supervised Representation Learning from Random Data Projectors." NeurIPS 2023 Workshops: TRL, 2023.

Markdown

[Sui et al. "Self-Supervised Representation Learning from Random Data Projectors." NeurIPS 2023 Workshops: TRL, 2023.](https://mlanthology.org/neuripsw/2023/sui2023neuripsw-selfsupervised/)

BibTeX

@inproceedings{sui2023neuripsw-selfsupervised,
  title     = {{Self-Supervised Representation Learning from Random Data Projectors}},
  author    = {Sui, Yi and Wu, Tongzi and Cresswell, Jesse and Wu, Ga and Stein, George and Huang, Xiao Shi and Zhang, Xiaochen and Volkovs, Maksims},
  booktitle = {NeurIPS 2023 Workshops: TRL},
  year      = {2023},
  url       = {https://mlanthology.org/neuripsw/2023/sui2023neuripsw-selfsupervised/}
}