Linearly Combining Density Estimators via Stacking
Abstract
This paper presents experimental results with both real and artificial data combining unsupervised learning algorithms using stacking. Specifically, stacking is used to form a linear combination of finite mixture model and kernel density estimators for non-parametric multivariate density estimation. The method outperforms other strategies such as choosing the single best model based on cross-validation, combining with uniform weights, and even using the single best model chosen by “Cheating” and examining the test set. We also investigate (1) how the utility of stacking changes when one of the models being combined is the model that generated the data, (2) how the stacking coefficients of the models compare to the relative frequencies with which cross-validation chooses among the models, (3) visualization of combined “effective” kernels, and (4) the sensitivity of stacking to overfitting as model complexity increases.
Cite
Text
Smyth and Wolpert. "Linearly Combining Density Estimators via Stacking." Machine Learning, 1999. doi:10.1023/A:1007511322260Markdown
[Smyth and Wolpert. "Linearly Combining Density Estimators via Stacking." Machine Learning, 1999.](https://mlanthology.org/mlj/1999/smyth1999mlj-linearly/) doi:10.1023/A:1007511322260BibTeX
@article{smyth1999mlj-linearly,
title = {{Linearly Combining Density Estimators via Stacking}},
author = {Smyth, Padhraic and Wolpert, David H.},
journal = {Machine Learning},
year = {1999},
pages = {59-83},
doi = {10.1023/A:1007511322260},
volume = {36},
url = {https://mlanthology.org/mlj/1999/smyth1999mlj-linearly/}
}