VIME: Extending the Success of Self- and Semi-Supervised Learning to Tabular Domain

Abstract

Self- and semi-supervised learning frameworks have made significant progress in training machine learning models with limited labeled data in image and language domains. These methods heavily rely on the unique structure in the domain datasets (such as spatial relationships in images or semantic relationships in language). They are not adaptable to general tabular data which does not have the same explicit structure as image and language data. In this paper, we fill this gap by proposing novel self- and semi-supervised learning frameworks for tabular data, which we refer to collectively as VIME (Value Imputation and Mask Estimation). We create a novel pretext task of estimating mask vectors from corrupted tabular data in addition to the reconstruction pretext task for self-supervised learning. We also introduce a novel tabular data augmentation method for self- and semi-supervised learning frameworks. In experiments, we evaluate the proposed framework in multiple tabular datasets from various application domains, such as genomics and clinical data. VIME exceeds state-of-the-art performance in comparison to the existing baseline methods.

Cite

Text

Yoon et al. "VIME: Extending the Success of Self- and Semi-Supervised Learning to Tabular Domain." Neural Information Processing Systems, 2020.

Markdown

[Yoon et al. "VIME: Extending the Success of Self- and Semi-Supervised Learning to Tabular Domain." Neural Information Processing Systems, 2020.](https://mlanthology.org/neurips/2020/yoon2020neurips-vime/)

BibTeX

@inproceedings{yoon2020neurips-vime,
  title     = {{VIME: Extending the Success of Self- and Semi-Supervised Learning to Tabular Domain}},
  author    = {Yoon, Jinsung and Zhang, Yao and Jordon, James and van der Schaar, Mihaela},
  booktitle = {Neural Information Processing Systems},
  year      = {2020},
  url       = {https://mlanthology.org/neurips/2020/yoon2020neurips-vime/}
}