🤖 AI Summary
This paper addresses Bayesian smoothing for partially observed Gaussian–Markov models. It identifies a fundamental limitation of conventional two-filter smoothers: their reliance on information-form parameterizations induces numerical instability and lacks a rigorous probabilistic interpretation. To overcome this, we propose a purely likelihood-domain two-filter framework: the log-quadratic likelihood function replaces the information vector, and the backward recursion is explicitly defined over likelihoods—not pseudo-distributions—thereby eliminating information-form representations entirely. Building on this, we derive a numerically stable square-root implementation and an explicit forward Markov representation of the path posterior distribution. The resulting method achieves both theoretical rigor—grounded in proper likelihood algebra—and computational robustness, significantly reducing numerical errors and algorithmic complexity. This work establishes a more natural, interpretable, and implementable paradigm for Gaussian smoothing.
📝 Abstract
In this article, the two filter formula is re-examined in the setting of partially observed Gauss--Markov models. It is traditionally formulated as a filter running backward in time, where the Gaussian density is parametrized in ``information form''. However, the quantity in the backward recursion is strictly speaking not a distribution, but a likelihood. Taking this observation seriously, a recursion over log-quadratic likelihoods is formulated instead, which obviates the need for ``information'' parametrization. In particular, it greatly simplifies the square-root formulation of the algorithm. Furthermore, formulae are given for producing the forward Markov representation of the a posteriori distribution over paths from the proposed likelihood representation.