University of Cambridge > > CMI Student Seminar Series > Dimension-Robust Function Space MCMC With Neural Network Priors

Dimension-Robust Function Space MCMC With Neural Network Priors

Add to your list(s) Download to your calendar using vCal

If you have a question about this talk, please contact Neil Deo.

At the beginning of this talk, two popular priors defined on function spaces are discussed: Gaussian priors, which come with a set of orthogonal basis functions, and Bayesian Neural Networks (BNNs), which are popular in the machine learning community. I argue that both priors come with disadvantages, and propose a new class of BNN priors that alleviate them. The resulting posteriors are amenable to sampling using Hilbert space Markov chain Monte Carlo methods (unlike standard BNNs), and scale more favourably in the dimension of the function’s domain (unlike most Gaussian measures). Some theoretical results as well as numerical illustrations are presented, and my talk will end by posing future research directions. This talk is loosely based on the following preprint:

This talk is part of the CMI Student Seminar Series series.

Tell a friend about this talk:

This talk is included in these lists:

Note that ex-directory lists are not shown.


© 2006-2023, University of Cambridge. Contact Us | Help and Documentation | Privacy and Publicity