Analysis of Sparse Bayesian Learning

Abstract

The recent introduction of the 'relevance vector machine' has effec(cid:173) tively demonstrated how sparsity may be obtained in generalised linear models within a Bayesian framework. Using a particular form of Gaussian parameter prior, 'learning' is the maximisation, with respect to hyperparameters, of the marginal likelihood of the data. This paper studies the properties of that objective func(cid:173) tion, and demonstrates that conditioned on an individual hyper(cid:173) parameter, the marginal likelihood has a unique maximum which is computable in closed form. It is further shown that if a derived 'sparsity criterion' is satisfied, this maximum is exactly equivalent to 'pruning' the corresponding parameter from the model.

Cite

Text

Faul and Tipping. "Analysis of Sparse Bayesian Learning." Neural Information Processing Systems, 2001.

Markdown

[Faul and Tipping. "Analysis of Sparse Bayesian Learning." Neural Information Processing Systems, 2001.](https://mlanthology.org/neurips/2001/faul2001neurips-analysis/)

BibTeX

@inproceedings{faul2001neurips-analysis,
  title     = {{Analysis of Sparse Bayesian Learning}},
  author    = {Faul, Anita C. and Tipping, Michael E.},
  booktitle = {Neural Information Processing Systems},
  year      = {2001},
  pages     = {383-389},
  url       = {https://mlanthology.org/neurips/2001/faul2001neurips-analysis/}
}