COOKIES: By using this website you agree that we can place Google Analytics Cookies on your device for performance monitoring. |
University of Cambridge > Talks.cam > Artificial Intelligence Research Group Talks (Computer Laboratory) > A Geometrical Perspective on Deep Neural Networks
A Geometrical Perspective on Deep Neural NetworksAdd to your list(s) Download to your calendar using vCal
If you have a question about this talk, please contact Ben Day. Deep neural networks trained with gradient descent have been extremely successful at learning solutions to a broad suite of difficult problems across a wide range of domains such as vision, gameplay, and natural language, many of which had previously been considered to require intelligence. Despite their tremendous success, however, we still do not have a detailed, predictive understanding of how these systems work. In this talk, I will focus on recent insights into the structure of neural network loss landscapes and how they are navigated by gradient descent during training. In particular, I will discuss a phenomenological approach to modelling their large-scale structure [1,2], and its consequences for ensembling, calibration and Bayesian methods in general [3]. In addition, I will make a connection to empirical observations about loss gradients and Hessians [4,5]. I will conclude with an outlook on several interesting open questions in understanding deep networks.
This talk is part of the Artificial Intelligence Research Group Talks (Computer Laboratory) series. This talk is included in these lists:
Note that ex-directory lists are not shown. |
Other listsInfant Sorrow CMS Special Lectures Cambridge Victorian Studies GroupOther talksAir pollution and human health. Lessons learnt and challenges ahead. Learning Tensors and Random Matrix Theory Materials, Membranes and Motivations for Research Networks of Nonsmooth Oscillators & Applications in Neuroscience |