🤖 AI Summary
This work addresses the limitations of flow-based generative models in modeling complex distributions—namely, inaccurate density estimation, unreliable likelihood evaluation, and lack of theoretical guarantees. Methodologically, we reformulate normalizing flows as Wasserstein gradient flows driven by neural ordinary differential equations (ODEs), unifying invertible transformations, density evolution, and sampling within a single geometric framework. We establish, for the first time, rigorous convergence guarantees for such flows, bridging optimal transport theory, ODE dynamics, and generative learning. By integrating differential-geometric modeling with Wasserstein metric analysis, our approach enables exact likelihood computation, efficient deterministic sampling, and interpretable generation mechanisms. Empirically, the framework significantly improves modeling stability and generalization across image and biosignal generation tasks, demonstrating both theoretical soundness and practical efficacy.
📝 Abstract
Generative AI (GenAI) has revolutionized data-driven modeling by enabling the synthesis of high-dimensional data across various applications, including image generation, language modeling, biomedical signal processing, and anomaly detection. Flow-based generative models provide a powerful framework for capturing complex probability distributions, offering exact likelihood estimation, efficient sampling, and deterministic transformations between distributions. These models leverage invertible mappings governed by Ordinary Differential Equations (ODEs), enabling precise density estimation and likelihood evaluation. This tutorial presents an intuitive mathematical framework for flow-based generative models, formulating them as neural network-based representations of continuous probability densities. We explore key theoretical principles, including the Wasserstein metric, gradient flows, and density evolution governed by ODEs, to establish convergence guarantees and bridge empirical advancements with theoretical insights. By providing a rigorous yet accessible treatment, we aim to equip researchers and practitioners with the necessary tools to effectively apply flow-based generative models in signal processing and machine learning.