Accurate Imputation and Efficient Data Acquisitionwith Transformer-Based VAEs
Abstract
Predicting missing values in tabular data, with uncertainty, is an essential task by itself as well as for downstream tasks such as personalized data acquisition. It is not clear whether state-of-the-art deep generative models for these tasks are well equipped to model the complex relationships that may exist between different features, especially when the subset of observed data are treated as a set. In this work we propose new attention-based models for estimating the joint conditional distribution of randomly missing values in mixed-type tabular data. The models improve on the state-of-the-art Partial Variational Autoencoder (Ma et al. 2019) on a range of imputation and information acquisition tasks.
Cite
Text
Lewis et al. "Accurate Imputation and Efficient Data Acquisitionwith Transformer-Based VAEs." NeurIPS 2021 Workshops: DGMs_Applications, 2021.Markdown
[Lewis et al. "Accurate Imputation and Efficient Data Acquisitionwith Transformer-Based VAEs." NeurIPS 2021 Workshops: DGMs_Applications, 2021.](https://mlanthology.org/neuripsw/2021/lewis2021neuripsw-accurate/)BibTeX
@inproceedings{lewis2021neuripsw-accurate,
title = {{Accurate Imputation and Efficient Data Acquisitionwith Transformer-Based VAEs}},
author = {Lewis, Sarah and Matejovicova, Tatiana and Li, Yingzhen and Lamb, Angus and Zaykov, Yordan and Allamanis, Miltiadis and Zhang, Cheng},
booktitle = {NeurIPS 2021 Workshops: DGMs_Applications},
year = {2021},
url = {https://mlanthology.org/neuripsw/2021/lewis2021neuripsw-accurate/}
}