Conformal Safety Shielding for Imperfect-Perception Agents

📅 2025-06-12
🏛️ arXiv.org
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Ensuring runtime safety for discrete autonomous agents—such as airport taxiway guidance systems—under imperfect perception, where high-dimensional observations induce state estimation errors. Method: We propose a safety-critical control framework integrating conformal prediction with runtime action shielding. Specifically, we introduce conformal prediction into safety shielding for the first time, designing an action pruning strategy grounded in a state estimate set; safety is formalized via a Markov decision process to guarantee local deterministic safety. Additionally, we derive, for the first time, the global safety probability bound for an ideal (perfect-perception) shielder. Results: Experiments on a real-world taxiway guidance system demonstrate that our approach significantly reduces boundary violation risk caused by perception errors, achieving reliable runtime safety at 95% confidence.

Technology Category

Application Category

📝 Abstract
We consider the problem of safe control in discrete autonomous agents that use learned components for imperfect perception (or more generally, state estimation) from high-dimensional observations. We propose a shield construction that provides run-time safety guarantees under perception errors by restricting the actions available to an agent, modeled as a Markov decision process, as a function of the state estimates. Our construction uses conformal prediction for the perception component, which guarantees that for each observation, the predicted set of estimates includes the actual state with a user-specified probability. The shield allows an action only if it is allowed for all the estimates in the predicted set, resulting in a local safety guarantee. We also articulate and prove a global safety property of existing shield constructions for perfect-perception agents bounding the probability of reaching unsafe states if the agent always chooses actions prescribed by the shield. We illustrate our approach with a case-study of an experimental autonomous system that guides airplanes on taxiways using high-dimensional perception DNNs.
Problem

Research questions and friction points this paper is trying to address.

Safe control for agents with imperfect perception
Run-time safety guarantees under perception errors
Shielding actions based on conformal prediction sets
Innovation

Methods, ideas, or system contributions that make the work stand out.

Conformal prediction for perception safety
Run-time shield restricts unsafe actions
Global safety proof for perfect-perception agents
W
William Scarbro
Colorado State University, USA
C
Calum Imrie
University of York, UK
S
Sinem Getir Yaman
University of York, UK
K
Kavan Fatehi
University of York, UK
C
Corina S. Păsăreanu
Carnegie Mellon University, USA
R
R. Calinescu
University of York, UK
Ravi Mangal
Ravi Mangal
Assistant Professor, Colorado State University
Trustworthy AIFormal MethodsMachine LearningSafe AutonomyProgram Verification