Efficient Training of Neural Stochastic Differential Equations by Matching Finite Dimensional Distributions

๐Ÿ“… 2024-10-04
๐Ÿ›๏ธ arXiv.org
๐Ÿ“ˆ Citations: 0
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
Neural stochastic differential equations (Neural SDEs) suffer from training instability, mode collapse, and high computational complexityโ€”e.g., signature kernel methods incur $O(D^2)$ cost per iteration. To address these issues, this paper proposes a novel training paradigm based on finite-dimensional distribution matching (FDM). Its core innovation is the first construction of a strictly proper scoring rule for continuous-time Markov processes, leveraging the Markov property of SDEs to design a differentiable, unbiased loss function. Unlike adversarial GAN-style training, our approach eliminates gradient oscillations and instability while reducing per-iteration complexity to $O(D)$. Extensive experiments on benchmark time-series generation tasks demonstrate that our method achieves superior sample quality and training stability compared to state-of-the-art GAN-based and signature kernel-based approaches.

Technology Category

Application Category

๐Ÿ“ Abstract
Neural Stochastic Differential Equations (Neural SDEs) have emerged as powerful mesh-free generative models for continuous stochastic processes, with critical applications in fields such as finance, physics, and biology. Previous state-of-the-art methods have relied on adversarial training, such as GANs, or on minimizing distance measures between processes using signature kernels. However, GANs suffer from issues like instability, mode collapse, and the need for specialized training techniques, while signature kernel-based methods require solving linear PDEs and backpropagating gradients through the solver, whose computational complexity scales quadratically with the discretization steps. In this paper, we identify a novel class of strictly proper scoring rules for comparing continuous Markov processes. This theoretical finding naturally leads to a novel approach called Finite Dimensional Matching (FDM) for training Neural SDEs. Our method leverages the Markov property of SDEs to provide a computationally efficient training objective. This scoring rule allows us to bypass the computational overhead associated with signature kernels and reduces the training complexity from $O(D^2)$ to $O(D)$ per epoch, where $D$ represents the number of discretization steps of the process. We demonstrate that FDM achieves superior performance, consistently outperforming existing methods in terms of both computational efficiency and generative quality.
Problem

Research questions and friction points this paper is trying to address.

Efficient training of Neural SDEs without adversarial methods
Reducing computational complexity from O(Dยฒ) to O(D)
Improving generative quality and stability in Neural SDEs
Innovation

Methods, ideas, or system contributions that make the work stand out.

Novel scoring rules for Markov processes
Finite Dimensional Matching training method
Reduces training complexity to linear scale
๐Ÿ”Ž Similar Papers
No similar papers found.
Jianxin Zhang
Jianxin Zhang
Research Scientist, Meta
Machine Learning
J
Josh Viktorov
Cisco Systems
D
Doosan Jung
Cisco Systems
Emily Pitler
Emily Pitler
Google