Stacked Generalizations: When Does It Work?
Abstract
Stacked generalization is a general method of using a high-level model to combine lower-level models to achieve greater predictive accuracy. In this paper we resolve two crucial issues which have been considered to be a ‘black art’ in classification tasks ever since the introduction of stacked generalization in 1992 by Wolpert: the type of generalizer that is suitable to derive the higher-level model, and the kind of attributes that should be used as its input. \nWe demonstrate the effectiveness of stacked generalization for combining three different types of learning algorithms, and also for combining models of the same type derived from a single learning algorithm in a multiple-data-batches scenario. We also compare the performance of stacked generalization with published results of arcing and bagging.
Cite
Text
Ting and Witten. "Stacked Generalizations: When Does It Work?." International Joint Conference on Artificial Intelligence, 1997.Markdown
[Ting and Witten. "Stacked Generalizations: When Does It Work?." International Joint Conference on Artificial Intelligence, 1997.](https://mlanthology.org/ijcai/1997/ting1997ijcai-stacked/)BibTeX
@inproceedings{ting1997ijcai-stacked,
title = {{Stacked Generalizations: When Does It Work?}},
author = {Ting, Kai Ming and Witten, Ian H.},
booktitle = {International Joint Conference on Artificial Intelligence},
year = {1997},
pages = {866-873},
url = {https://mlanthology.org/ijcai/1997/ting1997ijcai-stacked/}
}