Methods of Moments for Learning Stochastic Languages: Unified Presentation and Empirical Comparison
Abstract
Probabilistic latent-variable models are a powerful tool for modelling structured data. However, traditional expectation-maximization methods of learning such models are both computationally expensive and prone to local-minima. In contrast to these traditional methods, recently developed learning algorithms based upon the method of moments are both computationally efficient and provide strong statistical guarantees. In this work, we provide a unified presentation and empirical comparison of three general moment-based methods in the context of modelling stochastic languages. By rephrasing these methods upon a common theoretical ground, introducing novel theoretical results where necessary, we provide a clear comparison, making explicit the statistical assumptions upon which each method relies. With this theoretical grounding, we then provide an in-depth empirical analysis of the methods on both real and synthetic data with the goal of elucidating performance trends and highlighting important implementation details.
Cite
Text
Balle et al. "Methods of Moments for Learning Stochastic Languages: Unified Presentation and Empirical Comparison." International Conference on Machine Learning, 2014.Markdown
[Balle et al. "Methods of Moments for Learning Stochastic Languages: Unified Presentation and Empirical Comparison." International Conference on Machine Learning, 2014.](https://mlanthology.org/icml/2014/balle2014icml-methods/)BibTeX
@inproceedings{balle2014icml-methods,
title = {{Methods of Moments for Learning Stochastic Languages: Unified Presentation and Empirical Comparison}},
author = {Balle, Borja and Hamilton, William and Pineau, Joelle},
booktitle = {International Conference on Machine Learning},
year = {2014},
pages = {1386-1394},
volume = {32},
url = {https://mlanthology.org/icml/2014/balle2014icml-methods/}
}