Boltzmann Chains and Hidden Markov Models
Abstract
We propose a statistical mechanical framework for the modeling of discrete time series. Maximum likelihood estimation is done via Boltzmann learning in one-dimensional networks with tied weights. We call these networks Boltzmann chains and show that they contain hidden Markov models (HMMs) as a special case. Our framework also motivates new architectures that address partic(cid:173) ular shortcomings of HMMs. We look at two such architectures: parallel chains that model feature sets with disparate time scales, and looped networks that model long-term dependencies between hidden states. For these networks, we show how to implement the Boltzmann learning rule exactly, in polynomial time, without resort to simulated or mean-field annealing. The necessary com(cid:173) putations are done by exact decimation procedures from statistical mechanics.
Cite
Text
Saul and Jordan. "Boltzmann Chains and Hidden Markov Models." Neural Information Processing Systems, 1994.Markdown
[Saul and Jordan. "Boltzmann Chains and Hidden Markov Models." Neural Information Processing Systems, 1994.](https://mlanthology.org/neurips/1994/saul1994neurips-boltzmann/)BibTeX
@inproceedings{saul1994neurips-boltzmann,
title = {{Boltzmann Chains and Hidden Markov Models}},
author = {Saul, Lawrence K. and Jordan, Michael I.},
booktitle = {Neural Information Processing Systems},
year = {1994},
pages = {435-442},
url = {https://mlanthology.org/neurips/1994/saul1994neurips-boltzmann/}
}