Nonstationary Kernel Combination
Abstract
The power and popularity of kernel methods stem in part from their ability to handle diverse forms of structured inputs, including vectors, graphs and strings. Recently, several methods have been proposed for combining kernels from heterogeneous data sources. However, all of these methods produce stationary combinations; i.e., the relative weights of the various kernels do not vary among input examples. This article proposes a method for combining multiple kernels in a nonstationary fashion. The approach uses a large-margin latent-variable generative model within the maximum entropy discrimination (MED) framework. Latent parameter estimation is rendered tractable by variational bounds and an iterative optimization procedure. The classifier we use is a log-ratio of Gaussian mixtures, in which each component is implicitly mapped via a Mercer kernel function. We show that the support vector machine is a special case of this model. In this approach, discriminative parameter estimation is feasible via a fast sequential minimal optimization algorithm. Empirical results are presented on synthetic data, several benchmarks, and on a protein function annotation task.
Cite
Text
Lewis et al. "Nonstationary Kernel Combination." International Conference on Machine Learning, 2006. doi:10.1145/1143844.1143914Markdown
[Lewis et al. "Nonstationary Kernel Combination." International Conference on Machine Learning, 2006.](https://mlanthology.org/icml/2006/lewis2006icml-nonstationary/) doi:10.1145/1143844.1143914BibTeX
@inproceedings{lewis2006icml-nonstationary,
title = {{Nonstationary Kernel Combination}},
author = {Lewis, Darrin P. and Jebara, Tony and Noble, William Stafford},
booktitle = {International Conference on Machine Learning},
year = {2006},
pages = {553-560},
doi = {10.1145/1143844.1143914},
url = {https://mlanthology.org/icml/2006/lewis2006icml-nonstationary/}
}