Kernelizing PLS, Degrees of Freedom, and Efficient Model Selection

Abstract

Kernelizing partial least squares (PLS), an algorithm which has been particularly popular in chemometrics, leads to kernel PLS which has several interesting properties, including a sub-cubic runtime for learning, and an iterative construction of directions which are relevant for predicting the outputs. We show that the kernelization of PLS introduces interesting properties not found in ordinary PLS, giving novel insights into the workings of kernel PLS and the connections to kernel ridge regression and conjugate gradient descent methods. Furthermore, we show how to correctly define the degrees of freedom for kernel PLS and how to efficiently compute an unbiased estimate. Finally, we address the practical problem of model selection. We demonstrate how to use the degrees of freedom estimate to perform effective model selection, and discuss how to implement crossvalidation schemes efficiently.

Cite

Text

Krämer and Braun. "Kernelizing PLS, Degrees of Freedom, and Efficient Model Selection." International Conference on Machine Learning, 2007. doi:10.1145/1273496.1273552

Markdown

[Krämer and Braun. "Kernelizing PLS, Degrees of Freedom, and Efficient Model Selection." International Conference on Machine Learning, 2007.](https://mlanthology.org/icml/2007/kramer2007icml-kernelizing/) doi:10.1145/1273496.1273552

BibTeX

@inproceedings{kramer2007icml-kernelizing,
  title     = {{Kernelizing PLS, Degrees of Freedom, and Efficient Model Selection}},
  author    = {Krämer, Nicole and Braun, Mikio L.},
  booktitle = {International Conference on Machine Learning},
  year      = {2007},
  pages     = {441-448},
  doi       = {10.1145/1273496.1273552},
  url       = {https://mlanthology.org/icml/2007/kramer2007icml-kernelizing/}
}