Quand

19 février 2026    
11h00 - 12h00

IAS
Bât. 121 rue Jean Teillac, Orsay, 91400

Type d’évènement

Par Tony Bonnaire (IAS)

Abstract : 

Diffusion models have emerged as powerful generative tools, capable of producing highly realistic images, videos, and sounds by learning a stochastic mapping starting from a simple Gaussian distribution. Despite their empirical success, the underlying reasons for their effectiveness remain poorly understood. In this talk, we will give a brief primer on the formalism of diffusion models and then dive into the analysis of a well-defined high-dimensional data distribution-a mixture of two Gaussians-under the assumption of an optimally-trained model. In particular, this reveals the existence of a ‘memorization’ transition where the trajectories are inevitably attracted to one of the training points and reproduce it exactly. Interestingly, we will provide an explanation on why diffusion models can still generalize despite this curse thanks to an implicit regularization mechanism in their training dynamics. This creates a generalization window that grows linearly with the training set size, enabling diffusion models to generalize. These findings are supported by both analytical results on simplified models and large-scale numerical experiments on realistic models and datasets. If time permits, we will discuss ongoing applications of such generative models in physics, especially in disordered systems and cosmology.

Lieu : IAS, bât. 121, salle 4-5 (4ème étage)