🤖 AI Summary
Training stiff Neural ODEs suffers from severe gradient vanishing, hindering optimization.
Method: We analyze the stability functions of A-stable and L-stable numerical integrators—standard for stiff systems—and model parameter sensitivity propagation. Using explicit gradient derivations for mainstream stiff solvers (e.g., BDF, implicit Runge–Kutta), we rigorously characterize how parameter gradients decay.
Contribution/Results: We prove that any A-stable or L-stable integrator necessarily induces parameter sensitivity decay at rate $O(|z|^{-1})$ for large eigenvalues $z$, establishing gradient vanishing as an intrinsic property of stable integration—not an implementation artifact. This provides the first unified numerical-analytic explanation for training failures across diverse implicit solvers, establishes a fundamental theoretical limit on the trainability of stiff Neural ODEs, and furnishes principled guidance for designing gradient-aware parameterizations of stiff differential equations.
📝 Abstract
Gradient-based optimization of neural differential equations and other parameterized dynamical systems fundamentally relies on the ability to differentiate numerical solutions with respect to model parameters. In stiff systems, it has been observed that sensitivities to parameters controlling fast-decaying modes become vanishingly small during training, leading to optimization difficulties. In this paper, we show that this vanishing gradient phenomenon is not an artifact of any particular method, but a universal feature of all A-stable and L-stable stiff numerical integration schemes. We analyze the rational stability function for general stiff integration schemes and demonstrate that the relevant parameter sensitivities, governed by the derivative of the stability function, decay to zero for large stiffness. Explicit formulas for common stiff integration schemes are provided, which illustrate the mechanism in detail. Finally, we rigorously prove that the slowest possible rate of decay for the derivative of the stability function is $O(|z|^{-1})$, revealing a fundamental limitation: all A-stable time-stepping methods inevitably suppress parameter gradients in stiff regimes, posing a significant barrier for training and parameter identification in stiff neural ODEs.