Fast Inference and Learning in Large-State-Space HMMs
Abstract
For Hidden Markov Models (HMMs) with fully connected transition models, the three fundamental problems of evaluating the likelihood of an observation sequence, estimating an optimal state sequence for the observations, and learning the model parameters, all have quadratic time complexity in the number of states. We introduce a novel class of non-sparse Markov transition matrices called Dense-Mostly-Constant (DMC) transition matrices that allow us to derive new algorithms for solving the basic HMM problems in sub-quadratic time. We describe the DMC HMM model and algorithms and attempt to convey some intuition for their usage. Empirical results for these algorithms show dramatic speedups for all three problems. In terms of accuracy, the DMC model yields strong results and outperforms the baseline algorithms even in domains known to violate the DMC assumption.
Cite
Text
Siddiqi and Moore. "Fast Inference and Learning in Large-State-Space HMMs." International Conference on Machine Learning, 2005. doi:10.1145/1102351.1102452Markdown
[Siddiqi and Moore. "Fast Inference and Learning in Large-State-Space HMMs." International Conference on Machine Learning, 2005.](https://mlanthology.org/icml/2005/siddiqi2005icml-fast/) doi:10.1145/1102351.1102452BibTeX
@inproceedings{siddiqi2005icml-fast,
title = {{Fast Inference and Learning in Large-State-Space HMMs}},
author = {Siddiqi, Sajid M. and Moore, Andrew W.},
booktitle = {International Conference on Machine Learning},
year = {2005},
pages = {800-807},
doi = {10.1145/1102351.1102452},
url = {https://mlanthology.org/icml/2005/siddiqi2005icml-fast/}
}