COOKIES: By using this website you agree that we can place Google Analytics Cookies on your device for performance monitoring. |
University of Cambridge > Talks.cam > Isaac Newton Institute Seminar Series > Random Feature Expansions for Deep Gaussian Processes
Random Feature Expansions for Deep Gaussian ProcessesAdd to your list(s) Download to your calendar using vCal
If you have a question about this talk, please contact INI IT. UNQW02 - Surrogate models for UQ in complex systems Drawing meaningful conclusions on the way complex real life phenomena work and being able to predict the behavior of systems of interest require developing accurate and highly interpretable mathematical models whose parameters need to be estimated from observations. In modern applications, however, we are often challenged with the lack of such models, and even when these are available they are too computational demanding to be suitable for standard parameter optimization/inference methods. While probabilistic models based on Deep Gaussian Processes (DGPs) offer attractive tools to tackle these challenges in a principled way and to allow for a sound quantification of uncertainty, carrying out inference for these models poses huge computational challenges that arguably hinder their wide adoption. In this talk, I will present our contribution to the development of practical and scalable inference for DGPs, which can exploit distributed and GPU computing. In particular, I will introduce a formulation of DGPs based on random features that we infer using stochastic variational inference. Through a series of experiments, I will illustrate how our proposal enables scalable deep probabilistic nonparametric modeling and significantly advances the state-of-the-art on inference methods for DGPs. This talk is part of the Isaac Newton Institute Seminar Series series. This talk is included in these lists:
Note that ex-directory lists are not shown. |
Other listsChallenging Neoliberalism Nonlinear Waves Clare Hall ColloquiumOther talksMaking Refuge: Flight Modularity, criticality and evolvability of a developmental GRN Intrinsically Motivating Teachers;STIR's use of Data Driven Insight to Iterate, Pivot and (where necessary) Fail Fast Quantifying Uncertainty in Turbulent Flow Predictions based on RANS/LES Closures Fundamental Limits to Volcanic Cooling and its Implications for Past Climate on Earth MicroRNAs as circulating biomarkers in cancer |