University of Cambridge > Talks.cam > Isaac Newton Institute Seminar Series > Learning and Regularizing Score-Based Diffusion Models

Learning and Regularizing Score-Based Diffusion Models

Add to your list(s) Download to your calendar using vCal

  • UserRicardo Baptista (CALTECH (California Institute of Technology))
  • ClockTuesday 16 July 2024, 10:00-11:00
  • HouseExternal.

If you have a question about this talk, please contact nobody.

DMLW01 - International workshop on diffusions in machine learning: foundations, generative models, and optimisation

Diffusion models have emerged as a powerful framework for generative modeling that relies on score matching to learn gradients of the data distribution’s log-density. A key element for the success of diffusion models is that the optimal score function is not identified when solving the denoising score matching problem. In fact, the optimal score in both unconditioned and conditioned settings leads to a diffusion model that returns to the training samples and effectively memorizes the data distribution. In this presentation, we study the dynamical system associated with the optimal score and describe its long-term behavior relative to the training samples. Lastly, we show the effect of three forms of score function regularization on avoiding memorization: restricting the score’s approximation space, early stopping of the training process, and early stopping of the diffusion process during sample generation. Moreover, we establish a connection between early stopping of the diffusion and explicit Tikhonov regularization of the score matching problem. These results are numerically validated using distributions with and without densities including image-based problems.

This talk is part of the Isaac Newton Institute Seminar Series series.

Tell a friend about this talk:

This talk is included in these lists:

Note that ex-directory lists are not shown.

 

© 2006-2024 Talks.cam, University of Cambridge. Contact Us | Help and Documentation | Privacy and Publicity