On the Statistical Capacity of Deep Generative Models

📅 2025-01-14
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work exposes a fundamental statistical limitation of deep generative models—including VAEs, GANs, and diffusion models—in modeling heavy-tailed high-dimensional data: when latent variables follow Gaussian or more general log-concave distributions, such models can only produce light-tailed samples and provably fail to universally approximate arbitrary heavy-tailed target distributions. We develop the first unified theoretical framework linking generative capacity bounds to concentration phenomena, convex geometry, the Gromov–Lévy inequality, and Riemannian probability theory under Ricci curvature constraints. Our analysis rigorously refutes the widespread assumption that “larger models and bigger datasets suffice to approximate any continuous distribution.” Through rigorous proofs and empirical validation on financial time series and synthetic heavy-tailed data, we precisely characterize the statistical capacity boundaries of mainstream architectures. This provides the first systematic, theoretically grounded criterion for model selection and architectural design in heavy-tailed regimes.

Technology Category

Application Category

📝 Abstract
Deep generative models are routinely used in generating samples from complex, high-dimensional distributions. Despite their apparent successes, their statistical properties are not well understood. A common assumption is that with enough training data and sufficiently large neural networks, deep generative model samples will have arbitrarily small errors in sampling from any continuous target distribution. We set up a unifying framework that debunks this belief. We demonstrate that broad classes of deep generative models, including variational autoencoders and generative adversarial networks, are not universal generators. Under the predominant case of Gaussian latent variables, these models can only generate concentrated samples that exhibit light tails. Using tools from concentration of measure and convex geometry, we give analogous results for more general log-concave and strongly log-concave latent variable distributions. We extend our results to diffusion models via a reduction argument. We use the Gromov--Levy inequality to give similar guarantees when the latent variables lie on manifolds with positive Ricci curvature. These results shed light on the limited capacity of common deep generative models to handle heavy tails. We illustrate the empirical relevance of our work with simulations and financial data.
Problem

Research questions and friction points this paper is trying to address.

Deep Generative Models
Tail Features
High-Dimensional Data
Innovation

Methods, ideas, or system contributions that make the work stand out.

Deep Generative Models
Tail Features
Statistical Characteristics
🔎 Similar Papers
No similar papers found.
Edric Tam
Edric Tam
Stanford University
Bayesian StatisticsComputational BiologyProbabilityMachine LearningEconometrics
D
David B. Dunson
Department of Statistical Science and Department of Mathematics, Duke University, Box 90251 Durham, North Carolina 27708, U.S.A.