Training Algorithms for Hidden Markov Models Using Entropy Based Distance Functions

Abstract

We present new algorithms for parameter estimation of HMMs. By adapting a framework used for supervised learning, we construct iterative algorithms that maximize the likelihood of the observations while also attempting to stay "close" to the current estimated parameters. We use a bound on the relative entropy between the two HMMs as a distance mea(cid:173) sure between them. The result is new iterative training algorithms which are similar to the EM (Baum-Welch) algorithm for training HMMs. The proposed algorithms are composed of a step similar to the expectation step of Baum-Welch and a new update of the parameters which replaces the maximization (re-estimation) step. The algorithm takes only negligi(cid:173) bly more time per iteration and an approximated version uses the same expectation step as Baum-Welch. We evaluate experimentally the new algorithms on synthetic and natural speech pronunciation data. For sparse models, i.e. models with relatively small number of non-zero parameters, the proposed algorithms require significantly fewer iterations.

Cite

Text

Singer and Warmuth. "Training Algorithms for Hidden Markov Models Using Entropy Based Distance Functions." Neural Information Processing Systems, 1996.

Markdown

[Singer and Warmuth. "Training Algorithms for Hidden Markov Models Using Entropy Based Distance Functions." Neural Information Processing Systems, 1996.](https://mlanthology.org/neurips/1996/singer1996neurips-training/)

BibTeX

@inproceedings{singer1996neurips-training,
  title     = {{Training Algorithms for Hidden Markov Models Using Entropy Based Distance Functions}},
  author    = {Singer, Yoram and Warmuth, Manfred K.},
  booktitle = {Neural Information Processing Systems},
  year      = {1996},
  pages     = {641-647},
  url       = {https://mlanthology.org/neurips/1996/singer1996neurips-training/}
}