COOKIES: By using this website you agree that we can place Google Analytics Cookies on your device for performance monitoring. |

## Fundamental limits of generative AIAdd to your list(s) Download to your calendar using vCal - Helmut Bölcskei - ETH Zurich
- Friday 28 April 2023, 14:00-15:00
- Centre for Mathematical Sciences MR12, CMS.
If you have a question about this talk, please contact Randolf Altmeyer. Generative AI has seen tremendous successes recently, most notably the chatbot ChatGPT and the DALLE2 software creating realistic images and artwork from text descriptions. Underlying these and other generative AI systems are usually neural networks trained to produce text, images, audio, or video from text inputs. The aim of this talk is to develop an understanding of the fundamental capabilities of generative neural networks. Specifically and mathematically speaking, we consider the realization of high-dimensional random vectors from one-dimensional random variables through deep neural networks. The resulting random vectors follow prescribed conditional probability distributions, where the conditioning represents the text input of the generative system and its output can be text, images, audio, or video. It is shown that every d-dimensional probability distribution can be generated through deep ReLU networks out of a 1-dimensional uniform input distribution. What is more, this is possible without incurring a cost—in terms of approximation error as measured in Wasserstein-distance—relative to generating the d-dimensional target distribution from d independent random variables. This is enabled by a space-filling approach which realizes a Wasserstein-optimal transport map and elicits the importance of network depth in driving the Wasserstein distance between the target distribution and its neural network approximation to zero. Finally, we show that the number of bits needed to encode the corresponding generative networks equals the fundamental limit for encoding probability distributions (by any method) as dictated by quantization theory according to Graf and Luschgy. This result also characterizes the minimum amount of information that needs to be extracted from training data so as to be able to generate a desired output at a prescribed accuracy and establishes that generative ReLU networks can attain this minimum. This is joint work with D. Perekrestenko and L. Eberhard This talk is part of the CCIMI Seminars series. ## This talk is included in these lists:- Centre for Mathematical Sciences MR12, CMS
- All CMS events
- All Talks (aka the CURE list)
- CCIMI
- CCIMI Seminars
- CMS Events
- Cambridge Centre for Data-Driven Discovery (C2D3)
- Cambridge Forum of Science and Humanities
- Cambridge Language Sciences
- Cambridge talks
- Chris Davis' list
- DPMMS Lists
- DPMMS info aggregator
- DPMMS lists
- Guy Emerson's list
- Hanchen DaDaDash
- Interested Talks
- Machine Learning
- School of Physical Sciences
- Statistical Laboratory info aggregator
- Statistics
- Statistics Group
- Trust & Technology Initiative - interesting events
- bld31
- custom
- ndk22's list
- ob366-ai4er
- rp587
Note that ex-directory lists are not shown. |
## Other listsCPERG: Cambridge Peace and Education Research Group## Other talksLabour History Lunch: Worker Resistance to Scientific Management in Sweden 1930-1950 Close The dynamics of the transition from the beam-like internal tides to shorter nonlinear waves on the continental shelf Graph Neural Networks for Geometric Graphs The recently discovered ties between teeth and infants in human evolution Optimizing scalar transport using branching flows |