Deceptive Sequential Decision-Making via Regularized Policy Optimization

📅 2025-01-30
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
In adversarial environments, autonomous systems risk exposing their intent and reward functions through inverse reinforcement learning (IRL) inference by adversaries. Method: This paper proposes a deceptive sequential decision-making framework based on structured policy regularization. It introduces two novel deception mechanisms—diversionary and targeted—integrated within a Markov decision process (MDP) formulation that jointly incorporates IRL-based adversarial analysis, robust policy optimization, and multi-agent game-theoretic simulation. Contribution/Results: Theoretically, we establish the first upper bound on cumulative reward loss induced by deception and achieve controllable misdirection of the adversary’s IRL inference. Empirically, diversionary deception causes adversaries to misclassify critical agents as least important, achieving 98.83% of the performance of the optimal non-deceptive policy; targeted deception enables arbitrary designation of decoys as “most important,” attaining 99.25% performance.

Technology Category

Application Category

📝 Abstract
Autonomous systems are increasingly expected to operate in the presence of adversaries, though an adversary may infer sensitive information simply by observing a system, without even needing to interact with it. Therefore, in this work we present a deceptive decision-making framework that not only conceals sensitive information, but in fact actively misleads adversaries about it. We model autonomous systems as Markov decision processes, and we consider adversaries that attempt to infer their reward functions using inverse reinforcement learning. To counter such efforts, we present two regularization strategies for policy synthesis problems that actively deceive an adversary about a system's underlying rewards. The first form of deception is ``diversionary'', and it leads an adversary to draw any false conclusion about what the system's reward function is. The second form of deception is ``targeted'', and it leads an adversary to draw a specific false conclusion about what the system's reward function is. We then show how each form of deception can be implemented in policy optimization problems, and we analytically bound the loss in total accumulated reward that is induced by deception. Next, we evaluate these developments in a multi-agent sequential decision-making problem with one real agent and multiple decoys. We show that diversionary deception can cause the adversary to believe that the most important agent is the least important, while attaining a total accumulated reward that is $98.83%$ of its optimal, non-deceptive value. Similarly, we show that targeted deception can make any decoy appear to be the most important agent, while still attaining a total accumulated reward that is $99.25%$ of its optimal, non-deceptive value.
Problem

Research questions and friction points this paper is trying to address.

Autonomous Systems
Deceptive Decision Making
Inverse Reinforcement Learning
Innovation

Methods, ideas, or system contributions that make the work stand out.

Deception Strategies
Reward Misdirection
Autonomous Systems
🔎 Similar Papers
Y
Yerin Kim
Georgia Institute of Technology, Atlanta, GA, USA
Alexander Benvenuti
Alexander Benvenuti
PhD Student, Georiga Institute of Technology
Differential PrivacyOptimizationMarkov Processes
B
Bo Chen
Georgia Institute of Technology, Atlanta, GA, USA
Mustafa O. Karabag
Mustafa O. Karabag
Postdoctoral Fellow, The University of Texas at Austin
A
Abhishek Kulkarni
University of Texas at Austin, Austin, TX, USA
Nathaniel D. Bastian
Nathaniel D. Bastian
United States Military Academy
artificial intelligenceoperations researchdata sciencesystems engineeringapplied economics
U
U. Topcu
University of Texas at Austin, Austin, TX, USA
M
Matthew T. Hale
Georgia Institute of Technology, Atlanta, GA, USA