Boltzmann Chains and Hidden Markov Models

Abstract

We propose a statistical mechanical framework for the modeling of discrete time series. Maximum likelihood estimation is done via Boltzmann learning in one-dimensional networks with tied weights. We call these networks Boltzmann chains and show that they contain hidden Markov models (HMMs) as a special case. Our framework also motivates new architectures that address partic(cid:173) ular shortcomings of HMMs. We look at two such architectures: parallel chains that model feature sets with disparate time scales, and looped networks that model long-term dependencies between hidden states. For these networks, we show how to implement the Boltzmann learning rule exactly, in polynomial time, without resort to simulated or mean-field annealing. The necessary com(cid:173) putations are done by exact decimation procedures from statistical mechanics.

Cite

Text

Saul and Jordan. "Boltzmann Chains and Hidden Markov Models." Neural Information Processing Systems, 1994.

Markdown

[Saul and Jordan. "Boltzmann Chains and Hidden Markov Models." Neural Information Processing Systems, 1994.](https://mlanthology.org/neurips/1994/saul1994neurips-boltzmann/)

BibTeX

@inproceedings{saul1994neurips-boltzmann,
  title     = {{Boltzmann Chains and Hidden Markov Models}},
  author    = {Saul, Lawrence K. and Jordan, Michael I.},
  booktitle = {Neural Information Processing Systems},
  year      = {1994},
  pages     = {435-442},
  url       = {https://mlanthology.org/neurips/1994/saul1994neurips-boltzmann/}
}