🤖 AI Summary
Discrete neural samplers are notoriously difficult to train due to the absence of gradients and combinatorial complexity, rendering existing continuous-domain stochastic optimal control methods inapplicable. This work proposes a unified discrete ASBS framework by generalizing the adjoint matching (AM) mechanism and Schrödinger bridge theory to discrete state spaces for the first time. By revealing the state-space-agnostic nature of the AM mechanism and modeling the discrete state space via a cyclic group structure, we derive optimality conditions for the discrete Schrödinger bridge problem. The resulting approach significantly enhances training efficiency and scalability while preserving high sample quality.
📝 Abstract
Learning discrete neural samplers is challenging due to the lack of gradients and combinatorial complexity. While stochastic optimal control (SOC) and Schr\"odinger bridge (SB) provide principled solutions, efficient SOC solvers like adjoint matching (AM), which excel in continuous domains, remain unexplored for discrete spaces. We bridge this gap by revealing that the core mechanism of AM is $\mathit{state}\text{-}\mathit{space~agnostic}$, and introduce $\mathbf{discrete~ASBS}$, a unified framework that extends AM and adjoint Schr\"odinger bridge sampler (ASBS) to discrete spaces. Theoretically, we analyze the optimality conditions of the discrete SB problem and its connection to SOC, identifying a necessary cyclic group structure on the state space to enable this extension. Empirically, discrete ASBS achieves competitive sample quality with significant advantages in training efficiency and scalability.