On the Equilibrium between Feasible Zone and Uncertain Model in Safe Exploration

📅 2026-01-31
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Achieving safe exploration with zero constraint violations in reinforcement learning remains highly challenging. This work proposes the Safe Equilibrium Exploration (SEE) framework, which formalizes the objective of safe exploration as a dynamic equilibrium between the feasible region and environmental model uncertainty, and efficiently attains this equilibrium through alternating optimization of the two components. By integrating graph-structured uncertainty modeling, dynamic expansion of the feasible region, and an alternating optimization strategy, SEE achieves zero constraint violations in classical control tasks and rapidly converges to the equilibrium within only a few iterations, substantially improving the efficiency of safe exploration.

Technology Category

Application Category

📝 Abstract
Ensuring the safety of environmental exploration is a critical problem in reinforcement learning (RL). While limiting exploration to a feasible zone has become widely accepted as a way to ensure safety, key questions remain unresolved: what is the maximum feasible zone achievable through exploration, and how can it be identified? This paper, for the first time, answers these questions by revealing that the goal of safe exploration is to find the equilibrium between the feasible zone and the environment model. This conclusion is based on the understanding that these two components are interdependent: a larger feasible zone leads to a more accurate environment model, and a more accurate model, in turn, enables exploring a larger zone. We propose the first equilibrium-oriented safe exploration framework called safe equilibrium exploration (SEE), which alternates between finding the maximum feasible zone and the least uncertain model. Using a graph formulation of the uncertain model, we prove that the uncertain model obtained by SEE is monotonically refined, the feasible zones monotonically expand, and both converge to the equilibrium of safe exploration. Experiments on classic control tasks show that our algorithm successfully expands the feasible zones with zero constraint violation, and achieves the equilibrium of safe exploration within a few iterations.
Problem

Research questions and friction points this paper is trying to address.

safe exploration
feasible zone
uncertain model
reinforcement learning
equilibrium
Innovation

Methods, ideas, or system contributions that make the work stand out.

safe exploration
feasible zone
uncertain model
equilibrium
reinforcement learning
🔎 Similar Papers
No similar papers found.
Yujie Yang
Yujie Yang
Tsinghua University
safe reinforcement learningautonomous driving
Z
Zhilong Zheng
School of Vehicle and Mobility and State Key Lab of Intelligent Green Vehicle and Mobility, Tsinghua University, Beijing, 100084, China
S
Shengbo Eben Li
School of Vehicle and Mobility and State Key Lab of Intelligent Green Vehicle and Mobility, Tsinghua University, Beijing, 100084, China