How Vulnerable Are Edge LLMs?

📅 2026-03-24
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the security risks associated with query-based knowledge extraction from quantized large language models (LLMs) deployed at the edge, such as INT8/INT4 Qwen. Despite their efficiency-oriented design, these models remain vulnerable to such attacks, and this study is the first to systematically demonstrate that quantization alone offers insufficient defense. To this end, the authors propose CLIQ (Clustered Instruction Querying), a framework that leverages structured clustering to generate highly covering yet non-redundant instruction queries, enabling efficient extraction of internal semantic knowledge under limited query budgets. Experimental results show that CLIQ substantially outperforms baseline methods in exposing model knowledge on quantized Qwen variants, revealing a critical privacy vulnerability in current edge-deployed LLMs.

Technology Category

Application Category

📝 Abstract
Large language models (LLMs) are increasingly deployed on edge devices under strict computation and quantization constraints, yet their security implications remain unclear. We study query-based knowledge extraction from quantized edge-deployed LLMs under realistic query budgets and show that, although quantization introduces noise, it does not remove the underlying semantic knowledge, allowing substantial behavioral recovery through carefully designed queries. To systematically analyze this risk, we propose \textbf{CLIQ} (\textbf{Cl}ustered \textbf{I}nstruction \textbf{Q}uerying), a structured query construction framework that improves semantic coverage while reducing redundancy. Experiments on quantized Qwen models (INT8/INT4) demonstrate that CLIQ consistently outperforms original queries across BERTScore, BLEU, and ROUGE, enabling more efficient extraction under limited budgets. These results indicate that quantization alone does not provide effective protection against query-based extraction, highlighting a previously underexplored security risk in edge-deployed LLMs.
Problem

Research questions and friction points this paper is trying to address.

edge LLMs
quantization
query-based extraction
security risk
knowledge extraction
Innovation

Methods, ideas, or system contributions that make the work stand out.

edge LLMs
quantization
query-based extraction
CLIQ
security risk
🔎 Similar Papers
No similar papers found.
A
Ao Ding
China University of Geoscience Beijing
H
Hongzong Li
Hong Kong University of Science and Technology
Zi Liang
Zi Liang
Hong Kong Polytechnic University
Natural Language ProcessingAI Security
Z
Zhanpeng Shi
Jilin University
S
Shuxin Zhuang
City University of Hong Kong, Chinese Academy of Sciences
Shiqin Tang
Shiqin Tang
Center for AI and Robotics, Chinese Academy of Sciences
Machine Learning
R
Rong Feng
City University of Hong Kong, Chinese Academy of Sciences
P
Ping Lu
City University of Hong Kong (Dongguan)