🤖 AI Summary
To address low exploration efficiency in automated synthesis of Linear Temporal Logic (LTS) controllers, this paper proposes a reinforcement learning (RL) framework integrated with Graph Neural Networks (GNNs). Unlike conventional methods relying on instantaneous states or handcrafted heuristics, our approach models the state-transition history as a dynamic graph structure, leveraging a trajectory-based graph construction and update mechanism to capture non-local, context-aware exploration cues. This constitutes the first application of graph-structured historical encoding to RL-based exploration in controller synthesis, enabling globally contextualized guidance. Evaluated across five benchmark domains, the method achieves significant improvements in four: average convergence speed increases by 37%, and policy transfer success rate improves by 22%; performance remains comparable only in the highly symmetric, locally interactive domain.
📝 Abstract
Controller synthesis is a formal method approach for automatically generating Labeled Transition System (LTS) controllers that satisfy specified properties. The efficiency of the synthesis process, however, is critically dependent on exploration policies. These policies often rely on fixed rules or strategies learned through reinforcement learning (RL) that consider only a limited set of current features. To address this limitation, this paper introduces GCRL, an approach that enhances RL-based methods by integrating Graph Neural Networks (GNNs). GCRL encodes the history of LTS exploration into a graph structure, allowing it to capture a broader, non-current-based context. In a comparative experiment against state-of-the-art methods, GCRL exhibited superior learning efficiency and generalization across four out of five benchmark domains, except one particular domain characterized by high symmetry and strictly local interactions.