Compositional Abilities Emerge Multiplicatively: Exploring Diffusion Models on a Synthetic Task

📅 2023-10-13
🏛️ Neural Information Processing Systems
📈 Citations: 65
Influential: 2
📄 PDF
🤖 AI Summary
While generative models produce high-fidelity samples, they remain unreliable for compositional generalization—i.e., synthesizing novel concept combinations outside the training distribution. This work systematically investigates the compositional generalization of conditional diffusion models in a controlled synthetic setting. We decouple and independently manipulate three key data properties—frequency, structural composition, and factor disentanglement—to construct a multi-dimensional out-of-distribution (OOD) evaluation framework. Our core findings reveal a multiplicative emergence of compositional capability: overall performance equals the product of subtask accuracies; the order of capability emergence is governed by data generative structure; and low-frequency compositions incur substantially higher optimization costs. This is the first empirical demonstration of nonlinear, emergent compositional generalization in diffusion models, establishing quantitative links among data structure, capability emergence, and optimization cost—providing a data-centric theoretical foundation and design principles for trustworthy generative compositional reasoning.
📝 Abstract
Modern generative models exhibit unprecedented capabilities to generate extremely realistic data. However, given the inherent compositionality of the real world, reliable use of these models in practical applications requires that they exhibit the capability to compose a novel set of concepts to generate outputs not seen in the training data set. Prior work demonstrates that recent diffusion models do exhibit intriguing compositional generalization abilities, but also fail unpredictably. Motivated by this, we perform a controlled study for understanding compositional generalization in conditional diffusion models in a synthetic setting, varying different attributes of the training data and measuring the model's ability to generate samples out-of-distribution. Our results show: (i) the order in which the ability to generate samples from a concept and compose them emerges is governed by the structure of the underlying data-generating process; (ii) performance on compositional tasks exhibits a sudden"emergence"due to multiplicative reliance on the performance of constituent tasks, partially explaining emergent phenomena seen in generative models; and (iii) composing concepts with lower frequency in the training data to generate out-of-distribution samples requires considerably more optimization steps compared to generating in-distribution samples. Overall, our study lays a foundation for understanding capabilities and compositionality in generative models from a data-centric perspective.
Problem

Research questions and friction points this paper is trying to address.

Understanding compositional generalization in diffusion models
Exploring emergence of multiplicative abilities in generative models
Analyzing data requirements for out-of-distribution generation
Innovation

Methods, ideas, or system contributions that make the work stand out.

Controlled study on compositional generalization in diffusion models
Multiplicative emergence of compositional abilities in generative models
Data-centric approach to understanding generative model capabilities
🔎 Similar Papers
No similar papers found.
M
Maya Okawa
Center for Brain Science, Harvard University, Cambridge, MA, USA; Physics & Informatics Laboratories, NTT Research, Inc., Sunnyvale, CA, USA
Ekdeep Singh Lubana
Ekdeep Singh Lubana
Goodfire AI
AIMachine LearningDeep Learning
R
Robert P. Dick
EECS Department, University of Michigan, Ann Arbor, MI, USA
H
Hidenori Tanaka
Center for Brain Science, Harvard University, Cambridge, MA, USA; Physics & Informatics Laboratories, NTT Research, Inc., Sunnyvale, CA, USA