COOKIES: By using this website you agree that we can place Google Analytics Cookies on your device for performance monitoring. |
University of Cambridge > Talks.cam > Isaac Newton Institute Seminar Series > Optimal Approximation with Sparsely Connected Deep Neural Networks
Optimal Approximation with Sparsely Connected Deep Neural NetworksAdd to your list(s) Download to your calendar using vCal
If you have a question about this talk, please contact INI IT. VMVW02 - Generative models, parameter learning and sparsity Despite the outstanding success of deep neural networks in real-world applications, most of the related research is empirically driven and a mathematical foundation is almost completely missing. One central task of a neural network is to approximate a function, which for instance encodes a classification task. In this talk, we will be concerned with the question, how well a function can be approximated by a neural network with sparse connectivity. Using methods from approximation theory and applied harmonic analysis, we will derive a fundamental lower bound on the sparsity of a neural network. By explicitly constructing neural networks based on certain representation systems, so-called $\alpha$-shearlets, we will then demonstrate that this lower bound can in fact be attained. Finally, we present numerical experiments, which surprisingly show that already the standard backpropagation algorithm generates deep neural networks obeying those optimal approximation rates. This is joint work with H. Bölcskei (ETH Zurich), P. Grohs (Uni Vienna), and P. Petersen (TU Berlin). This talk is part of the Isaac Newton Institute Seminar Series series. This talk is included in these lists:
Note that ex-directory lists are not shown. |
Other listsEngineering Department Structures Research Seminars J CSaP Professional Development Policy SeminarsOther talksBiosensor Technologies (Biacore SPR, Switchsense, Octet) Determining structures in situ using cryo-electron tomography:enveloped viruses and coated vesicles Yikes! Why did past-me say he'd give a talk on future discounting? Visual hallucinations in Parkinson’s disease - imbalances in top-down vs. bottom up information processing Adaptation in log-concave density estimation |