Smooth On-Line Learning Algorithms for Hidden Markov Models

Abstract

A simple learning algorithm for Hidden Markov Models (HMMs) is presented together with a number of variations. Unlike other classical algorithms such as the Baum-Welch algorithm, the algorithms described are smooth and can be used on-line (after each example presentation) or in batch mode, with or without the usual Viterbi most likely path approximation. The algorithms have simple expressions that result from using a normalized-exponential representation for the HMM parameters. All the algorithms presented are proved to be exact or approximate gradient optimization algorithms with respect to likelihood, log-likelihood, or cross-entropy functions, and as such are usually convergent. These algorithms can also be casted in the more general EM (Expectation-Maximization) framework where they can be viewed as exact or approximate GEM (Generalized Expectation-Maximization) algorithms. The mathematical properties of the algorithms are derived in the appendix.

Cite

Text

Baldi and Chauvin. "Smooth On-Line Learning Algorithms for Hidden Markov Models." Neural Computation, 1994. doi:10.1162/NECO.1994.6.2.307

Markdown

[Baldi and Chauvin. "Smooth On-Line Learning Algorithms for Hidden Markov Models." Neural Computation, 1994.](https://mlanthology.org/neco/1994/baldi1994neco-smooth/) doi:10.1162/NECO.1994.6.2.307

BibTeX

@article{baldi1994neco-smooth,
  title     = {{Smooth On-Line Learning Algorithms for Hidden Markov Models}},
  author    = {Baldi, Pierre and Chauvin, Yves},
  journal   = {Neural Computation},
  year      = {1994},
  pages     = {307-318},
  doi       = {10.1162/NECO.1994.6.2.307},
  volume    = {6},
  url       = {https://mlanthology.org/neco/1994/baldi1994neco-smooth/}
}