🤖 AI Summary
Contemporary continuous-time generative models typically exhibit a “V-shaped” transport: samples evolve independently from the prior to data points via straight-line trajectories, ignoring intrinsic shared structures among data. To address this, we propose Y-Flow—a Y-shaped generative flow that explicitly models hierarchical data relationships through a shared latent path: mass first moves collectively along a common trajectory, then bifurcates toward individual targets. Our key innovations include (i) a sublinear exponential velocity field that reduces transport cost and promotes efficient collective mass transport, breaking the V-shaped paradigm; and (ii) a scalable flow-matching objective built upon neural ordinary differential equations. Experiments on synthetic, image, and single-cell biological datasets demonstrate that Y-Flow significantly outperforms mainstream baselines—achieving superior distributional fidelity (e.g., lower FID and MMD) and improved sampling efficiency (reducing ODE integration steps by over 50%).
📝 Abstract
Modern continuous-time generative models often induce V-shaped transport: each sample travels independently along nearly straight trajectories from prior to data, overlooking shared structure. We introduce Y-shaped generative flows, which move probability mass together along shared pathways before branching to target-specific endpoints. Our formulation is based on novel velocity-powered transport cost with a sublinear exponent (between zero and one). this concave dependence rewards joint and fast mass movement. Practically, we instantiate the idea in a scalable neural ODE training objective. On synthetic, image, and biology datasets, Y-flows recover hierarchy-aware structure, improve distributional metrics over strong flow-based baselines, and reach targets with fewer integration steps.