Preventing Over-Fitting During Model Selection via Bayesian Regularisation of the Hyper-Parameters

Abstract

While the model parameters of a kernel machine are typically given by the solution of a convex optimisation problem, with a single global optimum, the selection of good values for the regularisation and kernel parameters is much less straightforward. Fortunately the leave-one-out cross-validation procedure can be performed or a least approximated very efficiently in closed form for a wide variety of kernel learning methods, providing a convenient means for model selection. Leave-one-out cross-validation based estimates of performance, however, generally exhibit a relatively high variance and are therefore prone to over-fitting. In this paper, we investigate the novel use of Bayesian regularisation at the second level of inference, adding a regularisation term to the model selection criterion corresponding to a prior over the hyper-parameter values, where the additional regularisation parameters are integrated out analytically. Results obtained on a suite of thirteen real-world and synthetic benchmark data sets clearly demonstrate the benefit of this approach.

Cite

Text

Cawley and Talbot. "Preventing Over-Fitting During Model Selection via Bayesian Regularisation of the Hyper-Parameters." Journal of Machine Learning Research, 2007.

Markdown

[Cawley and Talbot. "Preventing Over-Fitting During Model Selection via Bayesian Regularisation of the Hyper-Parameters." Journal of Machine Learning Research, 2007.](https://mlanthology.org/jmlr/2007/cawley2007jmlr-preventing/)

BibTeX

@article{cawley2007jmlr-preventing,
  title     = {{Preventing Over-Fitting During Model Selection via Bayesian Regularisation of the Hyper-Parameters}},
  author    = {Cawley, Gavin C. and Talbot, Nicola L. C.},
  journal   = {Journal of Machine Learning Research},
  year      = {2007},
  pages     = {841-861},
  volume    = {8},
  url       = {https://mlanthology.org/jmlr/2007/cawley2007jmlr-preventing/}
}