ARMOR: Agentic Reasoning for Methods Orchestration and Reparameterization for Robust Adversarial Attacks

📅 2026-01-26
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work proposes an adaptive adversarial attack framework based on collaborative agent reasoning, addressing the limitations of existing methods that rely on static strategy ensembles and lack semantic awareness or dynamic adaptability. The framework leverages a vision-language model (VLM) to identify semantic vulnerabilities and orchestrates three attack primitives—CW, JSMA, and STA—to generate and fuse perturbations on a shared “mixing board.” A large language model (LLM) enables real-time, closed-loop reparameterization of the attack strategy. By synergistically integrating VLM and LLM for the first time in adversarial attacks, this approach transcends conventional static ensemble paradigms, significantly improving cross-model transferability and attack success rates on standard benchmarks. It delivers fused attacks against black-box targets and adaptively selects either optimal or fused strategies for white-box targets, achieving a balanced trade-off between effectiveness and robustness.

Technology Category

Application Category

📝 Abstract
Existing automated attack suites operate as static ensembles with fixed sequences, lacking strategic adaptation and semantic awareness. This paper introduces the Agentic Reasoning for Methods Orchestration and Reparameterization (ARMOR) framework to address these limitations. ARMOR orchestrates three canonical adversarial primitives, Carlini-Wagner (CW), Jacobian-based Saliency Map Attack (JSMA), and Spatially Transformed Attacks (STA) via Vision Language Models (VLM)-guided agents that collaboratively generate and synthesize perturbations through a shared ``Mixing Desk". Large Language Models (LLMs) adaptively tune and reparameterize parallel attack agents in a real-time, closed-loop system that exploits image-specific semantic vulnerabilities. On standard benchmarks, ARMOR achieves improved cross-architecture transfer and reliably fools both settings, delivering a blended output for blind targets and selecting the best attack or blended attacks for white-box targets using a confidence-and-SSIM score.
Problem

Research questions and friction points this paper is trying to address.

adversarial attacks
static ensembles
strategic adaptation
semantic awareness
cross-architecture transfer
Innovation

Methods, ideas, or system contributions that make the work stand out.

Agentic Reasoning
Methods Orchestration
Reparameterization
Vision Language Models
Adversarial Attacks
🔎 Similar Papers
No similar papers found.
G
Gabriel Jun Rong Lee
Infocomm Technology Cluster, Singapore Institute of Technology, Singapore
C
Christos Korgialas
Dept. of Informatics, Aristotle University of Thessaloniki, Greece
D
Dion Jia Xu Ho
Dept. of Applied Physics and Applied Mathematics, Columbia University, USA
P
Pai Chet Ng
Infocomm Technology Cluster, Singapore Institute of Technology, Singapore
Xiaoxiao Miao
Xiaoxiao Miao
Duke Kunshan University
Speech PrivacySpeaker and Language IdentificationSpeech Synthesis
K
Konstantinos N. Plataniotis
Dept. of Electrical and Computer Engineering, University of Toronto, Canada