University of Cambridge > Talks.cam > Isaac Newton Institute Seminar Series > Can Bayesian neural networks make confident predictions?

Can Bayesian neural networks make confident predictions?

Add to your list(s) Download to your calendar using vCal

If you have a question about this talk, please contact nobody.

RCLW02 - Calibrating prediction uncertainty : statistics and machine learning perspectives

Bayesian neural networks (BNN) promise a principled approach to quantifying uncertainty in overparameterized models. Evaluating this promise is a challenge because in most practical settings, BNN predictive distributions can only be accessed through approximate inference. To systematically investigate the calibration of BNN predictive distributions, we argue for the use of discrete priors on interior parameters. We demonstrate that networks can be reverse engineered to determine which parameter ‘candidates’ should be given prior weight. This approach reveals that multimodal distributions in parameter space map to multimodal distributions in prediction space which are often only partially captured by approximate methods. We also find that uncertainty metrics have non-intuitive dependence on network dimensions, including cases where network capacity increases but uncertainty decreases. These results raise questions of whether some approximate methods may perform ‘better’ than the true BNN predictive distribution. Co-author: Youssef Marzouk

This talk is part of the Isaac Newton Institute Seminar Series series.

Tell a friend about this talk:

This talk is included in these lists:

Note that ex-directory lists are not shown.

 

© 2006-2025 Talks.cam, University of Cambridge. Contact Us | Help and Documentation | Privacy and Publicity