🤖 AI Summary
While denoising diffusion probabilistic models (DDPMs) achieve strong image generation performance, their requirement of ~1000 sampling steps hinders deployment on resource-constrained edge devices. Method: This work systematically compares the geometric properties and inference efficiency of DDPMs and Rectified Flows (RFs) on MNIST, introducing an “efficiency frontier” analytical framework. We employ time-conditioned U-Nets, ODE solvers (Euler and RK4), and curvature quantification to characterize vector field linearity and numerical robustness. Contribution/Results: Experiments reveal that RF learns near-straight optimal transport paths (curvature ≈ 1.02) and maintains high-fidelity generation with only 10 Euler steps—whereas DDPMs degrade significantly under the same step budget. RF’s vector field exhibits markedly higher linearity and stability across solvers. This study provides the first systematic evidence of RF’s structural advantages for low-overhead, low-latency generative modeling, establishing a new paradigm for real-time, edge-deployable synthesis.
📝 Abstract
Denoising Diffusion Probabilistic Models (DDPMs) have established a new state-of-the-art in generative image synthesis, yet their deployment is hindered by significant computational overhead during inference, often requiring up to 1,000 iterative steps. This study presents a rigorous comparative analysis of DDPMs against the emerging Flow Matching (Rectified Flow) paradigm, specifically isolating their geometric and efficiency properties on low-resource hardware. By implementing both frameworks on a shared Time-Conditioned U-Net backbone using the MNIST dataset, we demonstrate that Flow Matching significantly outperforms Diffusion in efficiency. Our geometric analysis reveals that Flow Matching learns a highly rectified transport path (Curvature $mathcal{C} approx 1.02$), which is near-optimal, whereas Diffusion trajectories remain stochastic and tortuous ($mathcal{C} approx 3.45$). Furthermore, we establish an ``efficiency frontier'' at $N=10$ function evaluations, where Flow Matching retains high fidelity while Diffusion collapses. Finally, we show via numerical sensitivity analysis that the learned vector field is sufficiently linear to render high-order ODE solvers (Runge-Kutta 4) unnecessary, validating the use of lightweight Euler solvers for edge deployment. extbf{This work concludes that Flow Matching is the superior algorithmic choice for real-time, resource-constrained generative tasks.}