Bias/Variance Analyses of Mixtures-of-Experts Architectures

Abstract

This article investigates the bias and variance of mixtures-of-experts (ME) architectures. The variance of an ME architecture can be expressed as the sum of two terms: the first term is related to the variances of the expert networks that comprise the architecture and the second term is related to the expert networks' covariances. One goal of this article is to study and quantify a number of properties of ME architectures via the metrics of bias and variance. A second goal is to clarify the relationships between this class of systems and other systems that have recently been proposed. It is shown that in contrast to systems that produce unbiased experts whose estimation errors are uncorrelated, ME architectures produce biased experts whose estimates are negatively correlated.

Cite

Text

Jacobs. "Bias/Variance Analyses of Mixtures-of-Experts Architectures." Neural Computation, 1997. doi:10.1162/NECO.1997.9.2.369

Markdown

[Jacobs. "Bias/Variance Analyses of Mixtures-of-Experts Architectures." Neural Computation, 1997.](https://mlanthology.org/neco/1997/jacobs1997neco-bias/) doi:10.1162/NECO.1997.9.2.369

BibTeX

@article{jacobs1997neco-bias,
  title     = {{Bias/Variance Analyses of Mixtures-of-Experts Architectures}},
  author    = {Jacobs, Robert A.},
  journal   = {Neural Computation},
  year      = {1997},
  pages     = {369-383},
  doi       = {10.1162/NECO.1997.9.2.369},
  volume    = {9},
  url       = {https://mlanthology.org/neco/1997/jacobs1997neco-bias/}
}