🤖 AI Summary
This work addresses the issue of spurious solutions in neural optimal transport within infinite-dimensional Hilbert spaces, which often impedes accurate approximation of the target distribution. The study provides the first rigorous characterization of the underlying cause and introduces an analytical framework grounded in regularity measure theory. To resolve ill-posedness, the authors propose a Gaussian smoothing mechanism that leverages the kernel of the covariance operator—implemented via Brownian motion—to smooth the input distribution. This regularization ensures well-posedness of the semi-dual formulation of neural optimal transport and recovers the unique Monge map. Empirical evaluations on both synthetic functions and time-series data demonstrate that the method effectively suppresses spurious solutions and significantly outperforms existing baselines.
📝 Abstract
We study Neural Optimal Transport in infinite-dimensional Hilbert spaces. In non-regular settings, Semi-dual Neural OT often generates spurious solutions that fail to accurately capture target distributions. We analytically characterize this spurious solution problem using the framework of regular measures, which generalize Lebesgue absolute continuity in finite dimensions. To resolve ill-posedness, we extend the semi-dual framework via a Gaussian smoothing strategy based on Brownian motion. Our primary theoretical contribution proves that under a regular source measure, the formulation is well-posed and recovers a unique Monge map. Furthermore, we establish a sharp characterization for the regularity of smoothed measures, proving that the success of smoothing depends strictly on the kernel of the covariance operator. Empirical results on synthetic functional data and time-series datasets demonstrate that our approach effectively suppresses spurious solutions and outperforms existing baselines.