Maximum Reconstruction Estimation for Generative Latent-Variable Models

Abstract

Generative latent-variable models are important for natural language processing due to their capability of providing compact representations of data. As conventional maximum likelihood estimation (MLE) is prone to focus on explaining irrelevant but common correlations in data, we apply maximum reconstruction estimation (MRE) to learning generative latent-variable models alternatively, which aims to find model parameters that maximize the probability of reconstructing the observed data. We develop tractable algorithms to directly learn hidden Markov models and IBM translation models using the MRE criterion, without the need to introduce a separate reconstruction model to facilitate efficient inference. Experiments on unsupervised part-of-speech induction and unsupervised word alignment show that our approach enables generative latent-variable models to better discover intended correlations in data and outperforms maximum likelihood estimators significantly.

Cite

Text

Cheng et al. "Maximum Reconstruction Estimation for Generative Latent-Variable Models." AAAI Conference on Artificial Intelligence, 2017. doi:10.1609/AAAI.V31I1.10971

Markdown

[Cheng et al. "Maximum Reconstruction Estimation for Generative Latent-Variable Models." AAAI Conference on Artificial Intelligence, 2017.](https://mlanthology.org/aaai/2017/cheng2017aaai-maximum/) doi:10.1609/AAAI.V31I1.10971

BibTeX

@inproceedings{cheng2017aaai-maximum,
  title     = {{Maximum Reconstruction Estimation for Generative Latent-Variable Models}},
  author    = {Cheng, Yong and Liu, Yang and Xu, Wei},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2017},
  pages     = {3173-3179},
  doi       = {10.1609/AAAI.V31I1.10971},
  url       = {https://mlanthology.org/aaai/2017/cheng2017aaai-maximum/}
}