Creativity in diffusion models: Insights from statistical physics and compositional grammars (Part 2)
- đ¤ Speaker: Alessandro Favero (University of Cambridge)
- đ Date & Time: Tuesday 04 November 2025, 14:00 - 15:00
- đ Venue: DAMTP, MR11
Abstract
How do generative AI systems, such as diffusion models, learn to create new data? I will argue that natural data â such as images or text â can be described as hierarchical compositions of features, which generative models learn to recombine in novel ways. To analyze this mechanism, I will introduce ensembles of synthetic grammars as models of structured data. Within this framework, I will present a theory of composition that predicts a phase transition in the generative dynamics of diffusion models, confirmed in modern architectures trained on real-world datasets. In the second part of the talk, I will discuss how such grammars can be learned: what statistical correlations a learner can exploit to infer grammatical structure and the sample complexity required to do so. This analysis shows that diffusion models progressively build internal representations corresponding to increasingly abstract latent variables, a procedure reminiscent of the renormalization group in physics.
Series This talk is part of the DAMTP Data Intensive Science Seminar series.
Included in Lists
Note: Ex-directory lists are not shown.
![[Talks.cam]](/static/images/talkslogosmall.gif)


Tuesday 04 November 2025, 14:00-15:00