Constructing Bayesian formulations of sparse kernel learning methods

Gavin C. Cawley, Nicola L. C. Talbot

Research output: Contribution to journalArticlepeer-review

4 Citations (Scopus)

Abstract

We present here a simple technique that simplifies the construction of Bayesian treatments of a variety of sparse kernel learning algorithms. An incomplete Cholesky factorisation is employed to modify the dual parameter space, such that the Gaussian prior over the dual model parameters is whitened. The regularisation term then corresponds to the usual weight-decay regulariser, allowing the Bayesian analysis to proceed via the evidence framework of MacKay. There is in addition a useful by-product associated with the incomplete Cholesky factorisation algorithm, it also identifies a subset of the training data forming an approximate basis for the entire dataset in the kernel-induced feature space, resulting in a sparse model. Bayesian treatments of the kernel ridge regression (KRR) algorithm, with both constant and heteroscedastic (input dependent) variance structures, and kernel logistic regression (KLR) are provided as illustrative examples of the proposed method, which we hope will be more widely applicable.
Original languageEnglish
Pages (from-to)674-683
Number of pages10
JournalNeural Networks
Volume18
Issue number5-6
DOIs
Publication statusPublished - 2005

Cite this