Symbolic State Partitioning for Reinforcement Learning

📅 2024-09-25
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address weak generalization, low sample efficiency, and approximation errors induced by manual state-space partitioning in continuous-state reinforcement learning (RL), this paper proposes an automated state-space partitioning method grounded in symbolic execution of environment dynamics. It is the first work to integrate symbolic execution into RL-based state abstraction, enabling structural awareness and dynamics-driven coarse-grained partitioning—thereby avoiding biases inherent in grid-based or prior-driven discretization, especially in settings with nonlinear interdependencies among state variables. Coupled with tabular RL and state abstraction, the framework significantly enhances policy reliability and state coverage in sparse-reward tasks. Benchmark evaluations demonstrate an 1.8× improvement in state coverage over conventional methods, while preserving both accuracy and scalability.

Technology Category

Application Category

📝 Abstract
Tabular reinforcement learning methods cannot operate directly on continuous state spaces. One solution for this problem is to partition the state space. A good partitioning enables generalization during learning and more efficient exploitation of prior experiences. Consequently, the learning process becomes faster and produces more reliable policies. However, partitioning introduces approximation, which is particularly harmful in the presence of nonlinear relations between state components. An ideal partition should be as coarse as possible, while capturing the key structure of the state space for the given problem. This work extracts partitions from the environment dynamics by symbolic execution. We show that symbolic partitioning improves state space coverage with respect to environmental behavior and allows reinforcement learning to perform better for sparse rewards. We evaluate symbolic state space partitioning with respect to precision, scalability, learning agent performance and state space coverage for the learnt policies.
Problem

Research questions and friction points this paper is trying to address.

Partitioning continuous state spaces
Improving reinforcement learning efficiency
Enhancing policy reliability through symbolic execution
Innovation

Methods, ideas, or system contributions that make the work stand out.

Symbolic execution for partitioning
Improves state space coverage
Enhances reinforcement learning performance
🔎 Similar Papers
No similar papers found.
Mohsen Ghaffari
Mohsen Ghaffari
Associate Professor, MIT
M
M. Varshosaz
Department of Computer Science, IT-University, Copenhagen, Denmark
E
E. Johnsen
Department of Informatics, University of Oslo, Oslo, Norway
A
Andrzej Wkasowski
Department of Computer Science, IT-University, Copenhagen, Denmark