🤖 AI Summary
This work addresses the limited structural generalization capability in physical system modeling by proposing a machine learning framework grounded in non-Euclidean geometry—specifically, the manifold of symmetric positive-definite matrices—thereby circumventing manual incorporation of physical priors in model-free architectures. Methodologically, it formalizes generalization as a topological mapping from state space to parameter space, uncovering the geometric essence of symmetry, invariance, and uniqueness; it leverages differential geometry, Lie group/Lie algebra representations, and manifold optimization to achieve geometric parameterization of linear time-invariant (LTI) systems. Contributions include the first rigorous geometric characterization of generalization for dynamical systems and a principled, intrinsic parameterization that respects the underlying physical structure. Experiments demonstrate substantial improvements in extrapolation performance and physical consistency, overcoming structural mismatch inherent in conventional Euclidean embeddings.
📝 Abstract
Mathematical descriptions of dynamical systems are deeply rooted in topological spaces defined by non-Euclidean geometry. This paper proposes leveraging structure-rich geometric spaces for machine learning to achieve structural generalization when modeling physical systems from data, in contrast to embedding physics bias within model-free architectures. We consider model generalization to be a function of symmetry, invariance and uniqueness, defined as a topological mapping from state space dynamics to the parameter space. We illustrate this view through the machine learning of linear time-invariant dynamical systems, whose dynamics reside on the symmetric positive definite manifold.