Bayesian Generalized Kernel Mixed Models Zhihua Zhang, Guang Dai and Michael I. Jordan JMLR 2011.

Post on 14-Dec-2015

215 views 0 download

Tags:

Transcript of Bayesian Generalized Kernel Mixed Models Zhihua Zhang, Guang Dai and Michael I. Jordan JMLR 2011.

Bayesian Generalized Kernel Mixed Models

Zhihua Zhang, Guang Dai and Michael I. Jordan

JMLR 2011

Summary of contributions

• Propose generalized kernel models (GKMs) as a framework in which sparsity can be given an explicit treatment and in which a fully Bayesian methodology can be carried out

• Data augmentation methodology to develop a MCMC algorithm for inference

• Approach shown to be related Gaussian processes and provide a flexible approximation method for GPs

Bayesian approach for kernel supervised learning

• The form of the regressor or classifier is given by

• For a Mercer kernel, there exists a corresponding mapping (say ), from the input space , such that

• This provides an equivalent representation in the feature space, where,

Generalized Kernel Models

Prior for regression coefficients

Sparse models

• Recall that the number of active vectors is the number of non-zero components of– We are thus interested in a prior for which

allows some components of to be zero

Methodology

For the indicator vector

Graphical model

Inference

• Gibbs for most parameters• MH for kernel parameters• Reversible jump Markov Chain for – takes 2^n distinct values– For small n, posterior may be obtained by calculating

the normalizing constant by summing over all possible values of

– For large n, a reversible jump MC sampler may be employed to identify high posterior probability models

Automatic choice of active vectors

• We generate a proposal from the current value of by one of the three possible moves:

Prediction :

Sparse Gaussian process for classification

Given a function , then is a Gaussian process with zero mean and covariance function and vice versa.

Also,

Sparse GP classification

Results