Inferring Causal Graph Temporal Logic Formulas to Expedite Reinforcement Learning in Temporally Extended Tasks

📅 2026-01-06
🏛️ arXiv.org
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the limitations of existing black-box reinforcement learning methods in graph-structured temporal tasks, which often neglect the propagation mechanisms of local perturbations through the network, resulting in low sample efficiency and uninterpretable policies. To overcome this, the authors propose GTL-CIRL, a novel closed-loop framework that integrates Causal Graph Temporal Logic (Causal GTL) with reinforcement learning for the first time. The approach jointly learns policies and formally verifiable causal specifications through counterexample-guided reward shaping and robustness constraints, while leveraging Gaussian process-driven Bayesian optimization to fine-tune logical template parameters. By explicitly modeling spatiotemporal dependencies, GTL-CIRL significantly enhances exploration efficiency. Experiments on gene regulatory and power grid tasks demonstrate that the method not only accelerates convergence compared to baselines but also yields clear, interpretable, and formally verifiable behavioral policies.

Technology Category

Application Category

📝 Abstract
Decision-making tasks often unfold on graphs with spatial-temporal dynamics. Black-box reinforcement learning often overlooks how local changes spread through network structure, limiting sample efficiency and interpretability. We present GTL-CIRL, a closed-loop framework that simultaneously learns policies and mines Causal Graph Temporal Logic (Causal GTL) specifications. The method shapes rewards with robustness, collects counterexamples when effects fail, and uses Gaussian Process (GP) driven Bayesian optimization to refine parameterized cause templates. The GP models capture spatial and temporal correlations in the system dynamics, enabling efficient exploration of complex parameter spaces. Case studies in gene and power networks show faster learning and clearer, verifiable behavior compared to standard RL baselines.
Problem

Research questions and friction points this paper is trying to address.

reinforcement learning
causal inference
graph temporal logic
sample efficiency
interpretability
Innovation

Methods, ideas, or system contributions that make the work stand out.

Causal Graph Temporal Logic
Gaussian Process
Bayesian Optimization
Reinforcement Learning
Spatio-temporal Dynamics
🔎 Similar Papers
No similar papers found.
H
Hadi Partovi Aria
School for Engineering of Matter, Transport and Energy, Arizona State University, USA
Zhe Xu
Zhe Xu
Assistant Professor, Arizona State University
Cyber-Physical SystemsControl TheoryReinforcement LearningFormal MethodsRobotics