Warm Starting State-Space Models with Automata Learning

📅 2026-03-05
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the data inefficiency of state space models (SSMs) in learning systems with symbolic structure and their difficulty in recovering the underlying state dynamics. It establishes, for the first time, a rigorous proof that Moore machines can be exactly represented by SSMs. Building on this result, the authors propose a hybrid approach that injects symbolic inductive bias into continuous learning by initializing SSMs with Moore machines learned via symbolic automata learning—encompassing both active and passive learning paradigms—as a warm start. Evaluated on the SYNTCOMP arbiter benchmark, this method substantially improves performance: warm-started SSMs converge 2–5 times faster than randomly initialized counterparts and achieve higher test accuracy.

Technology Category

Application Category

📝 Abstract
We prove that Moore machines can be exactly realized as state-space models (SSMs), establishing a formal correspondence between symbolic automata and these continuous machine learning architectures. These Moore-SSMs preserve both the complete symbolic structure and input-output behavior of the original Moore machine, but operate in Euclidean space. With this correspondence, we compare the training of SSMs with both passive and active automata learning. In recovering automata from the SYNTCOMP benchmark, we show that SSMs require orders of magnitude more data than symbolic methods and fail to learn state structure. This suggests that symbolic structure provides a strong inductive bias for learning these systems. We leverage this insight to combine the strengths of both automata learning and SSMs in order to learn complex systems efficiently. We learn an adaptive arbitration policy on a suite of arbiters from SYNTCOMP and show that initializing SSMs with symbolically-learned approximations learn both faster and better. We see 2-5 times faster convergence compared to randomly initialized models and better overall model accuracies on test data. Our work lifts automata learning out of purely discrete spaces, enabling principled exploitation of symbolic structure in continuous domains for efficiently learning in complex settings.
Problem

Research questions and friction points this paper is trying to address.

state-space models
automata learning
Moore machines
inductive bias
symbolic structure
Innovation

Methods, ideas, or system contributions that make the work stand out.

state-space models
Moore machines
automata learning
warm starting
inductive bias
🔎 Similar Papers
No similar papers found.
W
William Fishell
Columbia University
S
Sam Nicholas Kouteili
Yale University
Mark Santolucito
Mark Santolucito
Barnard College, Columbia University
Program SynthesisReactive SynthesisFormal MethodsComputer Music