ExtendAttack: Attacking Servers of LRMs via Extending Reasoning

📅 2025-06-16
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This paper exposes a novel resource-exhaustion attack targeting Large Reasoning Model (LRM) servers: adversaries implicitly elongate inference chains to induce DDoS-like service overload or failure. To realize this, we propose the first character-level multi-radix ASCII obfuscation mechanism for stealthy inference expansion—preserving both user query semantics and answer correctness while significantly increasing computational load. Our approach integrates multi-base ASCII encoding obfuscation, inference-path steering, and lightweight prompt perturbation, achieving high stealthiness and low observability. Evaluated on the o3 model using the HumanEval benchmark, our method increases response length by over 2.5× without degrading functional correctness. This work establishes a critical empirical foundation for LRM service security assessment and introduces a new adversarial paradigm for probing inference-time vulnerabilities.

Technology Category

Application Category

📝 Abstract
Large Reasoning Models (LRMs) have demonstrated promising performance in complex tasks. However, the resource-consuming reasoning processes may be exploited by attackers to maliciously occupy the resources of the servers, leading to a crash, like the DDoS attack in cyber. To this end, we propose a novel attack method on LRMs termed ExtendAttack to maliciously occupy the resources of servers by stealthily extending the reasoning processes of LRMs. Concretely, we systematically obfuscate characters within a benign prompt, transforming them into a complex, poly-base ASCII representation. This compels the model to perform a series of computationally intensive decoding sub-tasks that are deeply embedded within the semantic structure of the query itself. Extensive experiments demonstrate the effectiveness of our proposed ExtendAttack. Remarkably, it increases the length of the model's response by over 2.5 times for the o3 model on the HumanEval benchmark. Besides, it preserves the original meaning of the query and achieves comparable answer accuracy, showing the stealthiness.
Problem

Research questions and friction points this paper is trying to address.

Exploiting LRM reasoning to maliciously occupy server resources
Stealthily extending reasoning processes to crash servers
Obfuscating prompts to force intensive decoding tasks
Innovation

Methods, ideas, or system contributions that make the work stand out.

ExtendAttack exploits LRM reasoning processes
Obfuscates prompts into complex ASCII representations
Increases response length while preserving query meaning
🔎 Similar Papers
No similar papers found.
Z
Zhenhao Zhu
Tsinghua University
Y
Yue Liu
National University of Singapore
Yingwei Ma
Yingwei Ma
Moonshot AI
LLMCoding Agent
Hongcheng Gao
Hongcheng Gao
University of Chinese Academy of Sciences
Natural Language ProcessingLarge Language ModelsVision Language Models
N
Nuo Chen
National University of Singapore
Y
Yanpei Guo
National University of Singapore
Wenjie Qu
Wenjie Qu
National University of Singapore
Applied CryptographyLLM Security
H
Huiying Xu
Zhejiang Normal University
X
Xinzhong Zhu
Zhejiang Normal University
Jiaheng Zhang
Jiaheng Zhang
Assistant Professor, National University of Singapore.
Zero-knowledge proofsAI safetyApplied cryptographyBlockchain