A precise asymptotic analysis of learning diffusion models: theory and insights

📅 2025-01-07
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work investigates the learning dynamics and failure boundaries of two-layer autoencoders on high-dimensional manifold data. Motivated by the prevalence of mode collapse—and even full model collapse—in diffusion-based generative modeling on low-dimensional manifolds, we develop a high-dimensional asymptotic analysis coupled with online stochastic gradient descent (SGD) dynamical modeling. We establish, for the first time, a tight asymptotic characterization of the low-dimensional projection of generated samples, explicitly quantifying its dependence on the training sample size. Our theory uncovers the sequential failure pathway from mode collapse to model collapse, rigorously identifying insufficient sample size as the primary trigger. Moreover, we derive an analytical critical sample-size threshold that demarcates the onset of model collapse. This framework provides the first rigorous theoretical foundation for risk assessment in synthetic-data retraining pipelines.

Technology Category

Application Category

📝 Abstract
In this manuscript, we consider the problem of learning a flow or diffusion-based generative model parametrized by a two-layer auto-encoder, trained with online stochastic gradient descent, on a high-dimensional target density with an underlying low-dimensional manifold structure. We derive a tight asymptotic characterization of low-dimensional projections of the distribution of samples generated by the learned model, ascertaining in particular its dependence on the number of training samples. Building on this analysis, we discuss how mode collapse can arise, and lead to model collapse when the generative model is re-trained on generated synthetic data.
Problem

Research questions and friction points this paper is trying to address.

Two-layer Autoencoder Model
Complex High-dimensional Data
Model Evaluation
Innovation

Methods, ideas, or system contributions that make the work stand out.

Two-layer Autoencoder Model
Online Stochastic Gradient Descent
Visualization of Model Learning State
🔎 Similar Papers
No similar papers found.
Hugo Cui
Hugo Cui
Postdoctoral fellow, Harvard CMSA
Machine LearningStatistical Physics
C
C. Pehlevan
Center for Brain Science, Harvard University; Kempner Institute for the Study of Natural and Artificial Intelligence, Harvard University; John A. Paulson School of Engineering and Applied Sciences, Harvard University
Yue M. Lu
Yue M. Lu
Gordon McKay Professor of Electrical Engineering and of Applied Mathematics, Harvard University
Signal and information processing