🤖 AI Summary
Computing similarity between probability measures on path space via signature kernels is computationally expensive, especially for non-smooth stochastic processes such as Lévy processes.
Method: This paper introduces an efficient kernel method based on the expected signature kernel. It extends the Log-PDE approach—previously developed for smooth rough paths—to inhomogeneous Lévy processes, deriving a system of partial differential equations (PDEs) governing their expected signatures. By integrating absolute-continuous path representations in extended tensor algebras, signature cumulants, and the generalized Magnus expansion, the authors construct a tractable PDE framework.
Contribution/Results: In the Gaussian martingale case, the kernel function is shown to satisfy a Goursat-type PDE, admitting an analytical solution. The framework substantially improves computational efficiency for signature kernels of non-smooth processes—particularly Lévy-type processes—and provides both theoretical foundations and practical tools for pathwise statistical learning.
📝 Abstract
The expected signature kernel arises in statistical learning tasks as a similarity measure of probability measures on path space. Computing this kernel for known classes of stochastic processes is an important problem that, in particular, can help reduce computational costs. Building on the representation of the expected signature of (inhomogeneous) Lévy processes with absolutely continuous characteristics as the development of an absolutely continuous path in the extended tensor algebra [F.-H.-Tapia, Forum of Mathematics: Sigma (2022), "Unified signature cumulants and generalized Magnus expansions"], we extend the arguments developed for smooth rough paths in [Lemercier-Lyons-Salvi, "Log-PDE Methods for Rough Signature Kernels"] to derive a PDE system for the expected signature of inhomogeneous Lévy processes. As a specific example, we see that the expected signature kernel of Gaussian martingales satisfies a Goursat PDE.