Online Inference of Topics with Latent Dirichlet Allocation
Abstract
Inference algorithms for topic models are typically designed to be run over an entire collection of documents after they have been observed. However, in many applications of these models, the collection grows over time, making it infeasible to run batch algorithms repeatedly. This problem can be addressed by using online algorithms, which update estimates of the topics as each document is observed. We introduce two related Rao-Blackwellized online inference algorithms for the latent Dirichlet allocation (LDA) model – incremental Gibbs samplers and particle filters – and compare their runtime and performance to that of existing algorithms.
Cite
Text
Canini et al. "Online Inference of Topics with Latent Dirichlet Allocation." Proceedings of the Twelfth International Conference on Artificial Intelligence and Statistics, 2009.Markdown
[Canini et al. "Online Inference of Topics with Latent Dirichlet Allocation." Proceedings of the Twelfth International Conference on Artificial Intelligence and Statistics, 2009.](https://mlanthology.org/aistats/2009/canini2009aistats-online/)BibTeX
@inproceedings{canini2009aistats-online,
title = {{Online Inference of Topics with Latent Dirichlet Allocation}},
author = {Canini, Kevin and Shi, Lei and Griffiths, Thomas},
booktitle = {Proceedings of the Twelfth International Conference on Artificial Intelligence and Statistics},
year = {2009},
pages = {65-72},
volume = {5},
url = {https://mlanthology.org/aistats/2009/canini2009aistats-online/}
}