Goal Discovery with Causal Capacity for Efficient Reinforcement Learning

πŸ“… 2025-08-13
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
In reinforcement learning, the difficulty of quantifying causal relationships impedes efficient exploration. Method: This paper proposes the β€œGoal Discovery and Causal Capacity” framework, which formally defines and computes the *causal capacity* of state-action pairs to quantify the causal influence of agent actions on state transitions. Leveraging this metric, the framework automatically identifies high-causal-capacity states as semantically coherent subgoals, enabling goal-directed, targeted exploration. The approach integrates causal inference with Monte Carlo estimation, supporting both discrete and high-dimensional continuous state spaces, and seamlessly interfaces with mainstream RL algorithms. Contribution/Results: Extensive evaluation across multi-task benchmarks demonstrates that the discovered subgoals strongly align with human-defined priors, and the method achieves significantly higher task success rates compared to state-of-the-art baselines.

Technology Category

Application Category

πŸ“ Abstract
Causal inference is crucial for humans to explore the world, which can be modeled to enable an agent to efficiently explore the environment in reinforcement learning. Existing research indicates that establishing the causality between action and state transition will enhance an agent to reason how a policy affects its future trajectory, thereby promoting directed exploration. However, it is challenging to measure the causality due to its intractability in the vast state-action space of complex scenarios. In this paper, we propose a novel Goal Discovery with Causal Capacity (GDCC) framework for efficient environment exploration. Specifically, we first derive a measurement of causality in state space, emph{i.e.,} causal capacity, which represents the highest influence of an agent's behavior on future trajectories. After that, we present a Monte Carlo based method to identify critical points in discrete state space and further optimize this method for continuous high-dimensional environments. Those critical points are used to uncover where the agent makes important decisions in the environment, which are then regarded as our subgoals to guide the agent to make exploration more purposefully and efficiently. Empirical results from multi-objective tasks demonstrate that states with high causal capacity align with our expected subgoals, and our GDCC achieves significant success rate improvements compared to baselines.
Problem

Research questions and friction points this paper is trying to address.

Measure causality in state space for efficient exploration
Identify critical decision points as subgoals in environments
Improve reinforcement learning success rates via directed exploration
Innovation

Methods, ideas, or system contributions that make the work stand out.

Proposes Goal Discovery with Causal Capacity framework
Measures causality via causal capacity in state space
Uses critical points as subgoals for efficient exploration
πŸ”Ž Similar Papers
No similar papers found.