Caltech Home > PMA Home > Calendar > Special CMX Lunch Seminar
open search form
Tuesday, December 09, 2025
11:00 AM - 12:00 PM
Annenberg 105

Special CMX Lunch Seminar

Understanding Generalization of Deep Generative Models Requires Rethinking Underlying Low-dimensional Structures
Qing Qu, Assistant Professor, Department of Electrical Engineering and Computer Science, University of Michigan,
Speaker's Bio:
Qing Qu is an Assistant Professor in Electrical Engineering and Computer Science at the University of Michigan. He works at the intersection of the foundations of machine learning, numerical optimization, and signal/image processing, with a current focus on the theory of deep generative models and representation learning. Prior to joining Michigan in 2021, he was a Moore–Sloan Data Science Fellow at the Center for Data Science, New York University (2018–2020). He received his Ph.D. in Electrical Engineering from Columbia University in October 2018 and his B.Eng. in Electrical and Computer Engineering from Tsinghua University in July 2011. His work has been recognized with multiple honors, including the Best Student Paper Award at SPARS 2015, a Microsoft PhD Fellowship in Machine Learning (2016), the Best Paper Award at the NeurIPS Diffusion Models Workshop (2023), an NSF CAREER Award (2022), an Amazon Research Award (AWS AI, 2023), a UM CHS Junior Faculty Award (2025), and a Google Research Scholar Award (2025). He was one of the founding organizers and Program Chair for the new Conference on Parsimony & Learning (CPAL), regularly serves as an Area Chair for NeurIPS, ICML, and ICLR, and is an Action Editor for TMLR.

Diffusion models represent a remarkable new class of deep generative models, yet the mathematical principles underlying their generalization from finite training data are poorly understood. This talk offers novel theoretical insights into diffusion model generalization through the lens of "model reproducibility," revealing a surprising phase transition from memorization to generalization during training, notably occurring without the curse of dimensionality. Our theoretical framework hinges on two crucial observations: (i) the intrinsic low dimensionality of image datasets and (ii) the emergent low-rank property of the denoising autoencoder within trained neural networks. Under simplified settings, we rigorously establish that optimizing the training loss of diffusion models is mathematically equivalent to solving a canonical subspace clustering problem. This insight quantifies the minimal sample requirements for learning low-dimensional distributions, scaling linearly with the intrinsic dimension. Furthermore, by investigating this under a nonlinear two-layer network, we fully explain the memorization-to-generalization transition, highlighting inductive biases in learning dynamics and the models' strong representation learning ability. These theoretical insights have profound practical implications, enabling various applications for generation control and safety, including concept steering, watermarking, and memorization detection. This work not only advances theoretical understanding but also stimulates numerous directions for many applications in engineering and science.

For more information, please contact Jolene Brink by phone at (626)395-2813 or by email at [email protected] or visit CMX Website.