University of Cambridge > > Machine Learning @ CUED > Two Approximate Sampling Methods for Bayesian Deep Learning

Two Approximate Sampling Methods for Bayesian Deep Learning

Add to your list(s) Download to your calendar using vCal

If you have a question about this talk, please contact Adrià Garriga Alonso.

Deep neural networks have become popular for many tasks, especially object classification, in computer vision and machine learning. However, these classes of models are known to be have poor uncertainty representations – e.g. they do not know what they do not know. To address this challenge, we propose two Bayesian approaches to approximate the posterior distribution of the models’ parameters. The first, termed stochastic weight averaging Gaussian (SWAG), fits a Gaussian approximation around the iterates of the stochastic gradient descent trajectory from standard training of DNNs. The second, subspace inference, instead reduces the high dimensionality of DNNs to very low dimensions, before performing Bayesian model averaging in that low dimensional subspace. Both methods draw on existing theory and are demonstrated to have strong empirical results on both regression and classification, scaling to even ImageNet-sized datasets.

This talk is part of the Machine Learning @ CUED series.

Tell a friend about this talk:

This talk is included in these lists:

Note that ex-directory lists are not shown.


© 2006-2023, University of Cambridge. Contact Us | Help and Documentation | Privacy and Publicity