ClawLess: A Security Model of AI Agents

πŸ“… 2026-04-07
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
Autonomous AI agents pose significant security risks when executing complex tasks due to their ability to autonomously acquire information and execute code, yet existing approaches lack fundamental safeguards. This work proposes the first formal, dynamically adaptive security framework specifically designed for AI agents. Operating under the worst-case adversarial threat model, the framework enforces fine-grained security policies that dynamically constrain agent behavior. Enforcement is achieved through a user-space kernel combined with eBPF-based system call interception, ensuring strict compliance with predefined policies regardless of the agent’s internal design. By bridging the gap between theoretical policy specification and runtime enforcement, the framework provides end-to-end security guarantees, effectively closing the loop between policy intent and actual execution.
πŸ“ Abstract
Autonomous AI agents powered by Large Language Models can reason, plan, and execute complex tasks, but their ability to autonomously retrieve information and run code introduces significant security risks. Existing approaches attempt to regulate agent behavior through training or prompting, which does not offer fundamental security guarantees. We present ClawLess, a security framework that enforces formally verified policies on AI agents under a worst-case threat model where the agent itself may be adversarial. ClawLess formalizes a fine-grained security model over system entities, trust scopes, and permissions to express dynamic policies that adapt to agents' runtime behavior. These policies are translated into concrete security rules and enforced through a user-space kernel augmented with BPF-based syscall interception. This approach bridges the formal security model with practical enforcement, ensuring security regardless of the agent's internal design.
Problem

Research questions and friction points this paper is trying to address.

AI agents
security risks
Large Language Models
autonomous execution
formal security guarantees
Innovation

Methods, ideas, or system contributions that make the work stand out.

formal verification
AI agent security
BPF syscall interception
fine-grained access control
adversarial agent model
πŸ”Ž Similar Papers
No similar papers found.
H
Hongyi Lu
Research Institute of Trustworthy Autonomous Systems, Southern University of Science and Technology
N
Nian Liu
Department of Computer Science and Engineering, Southern University of Science and Technology
Shuai Wang
Shuai Wang
The Hong Kong University of Science and Technology
Computer SecuritySoftware Engineering
Fengwei Zhang
Fengwei Zhang
Associate Professor, Southern University of Science and Technology
Systems Security