Hierarchical Mixtures of Experts and the EM Algorithm
Abstract
We present a tree-structured architecture for supervised learning. The statistical model underlying the architecture is a hierarchical mixture model in which both the mixture coefficients and the mixture components are generalized linear models (GLIM's). Learning is treated as a maximum likelihood problem; in particular, we present an Expectation-Maximization (EM) algorithm for adjusting the parameters of the architecture. We also develop an on-line learning algorithm in which the parameters are updated incrementally. Comparative simulation results are presented in the robot dynamics domain.
Cite
Text
Jordan and Jacobs. "Hierarchical Mixtures of Experts and the EM Algorithm." Neural Computation, 1994. doi:10.1162/NECO.1994.6.2.181Markdown
[Jordan and Jacobs. "Hierarchical Mixtures of Experts and the EM Algorithm." Neural Computation, 1994.](https://mlanthology.org/neco/1994/jordan1994neco-hierarchical/) doi:10.1162/NECO.1994.6.2.181BibTeX
@article{jordan1994neco-hierarchical,
title = {{Hierarchical Mixtures of Experts and the EM Algorithm}},
author = {Jordan, Michael I. and Jacobs, Robert A.},
journal = {Neural Computation},
year = {1994},
pages = {181-214},
doi = {10.1162/NECO.1994.6.2.181},
volume = {6},
url = {https://mlanthology.org/neco/1994/jordan1994neco-hierarchical/}
}