SeDA: Secure and Efficient DNN Accelerators with Hardware/Software Synergy

📅 2025-08-26
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the substantial hardware overhead and frequent off-chip memory accesses incurred by confidentiality and integrity protection in DNN accelerators for safety-critical applications—such as autonomous driving, healthcare, and finance—this paper proposes a software-hardware co-designed secure acceleration architecture. We introduce three key innovations: (1) a bandwidth-aware encryption mechanism that adapts encryption granularity to memory bandwidth constraints; (2) an optimal tiling strategy jointly optimizing intra-layer and inter-layer partitioning to minimize data movement; and (3) a lightweight, multi-level integrity verification scheme leveraging hierarchical checksums and Merkle trees. Our approach significantly reduces memory bandwidth pressure and hardware resource utilization while preserving strong security guarantees. Evaluated on both server-grade and edge-class NPUs, it achieves ≥12% end-to-end inference speedup over baseline secure accelerators, with excellent scalability and practical deployability for high-assurance DNN inference scenarios.

Technology Category

Application Category

📝 Abstract
Ensuring the confidentiality and integrity of DNN accelerators is paramount across various scenarios spanning autonomous driving, healthcare, and finance. However, current security approaches typically require extensive hardware resources, and incur significant off-chip memory access overheads. This paper introduces SeDA, which utilizes 1) a bandwidth-aware encryption mechanism to improve hardware resource efficiency, 2) optimal block granularity through intra-layer and inter-layer tiling patterns, and 3) a multi-level integrity verification mechanism that minimizes, or even eliminates, memory access overheads. Experimental results show that SeDA decreases performance overhead by over 12% for both server and edge neural processing units (NPUs), while ensuring robust scalability.
Problem

Research questions and friction points this paper is trying to address.

Securing DNN accelerators with minimal hardware overhead
Reducing off-chip memory access costs in security mechanisms
Ensuring confidentiality and integrity across diverse application scenarios
Innovation

Methods, ideas, or system contributions that make the work stand out.

Bandwidth-aware encryption for resource efficiency
Optimal block granularity via tiling patterns
Multi-level integrity verification minimizing memory overhead
🔎 Similar Papers
No similar papers found.
W
Wei Xuan
ACCESS – AI Chip Center for Emerging Smart Systems, InnoHK Centers, Hong Kong Science Park, Hong Kong, China
Z
Zhongrui Wang
Southern University of Science and Technology, Shenzhen, Guangdong, China
Lang Feng
Lang Feng
Nanyang Technological University
Reinforcement Learning
Ning Lin
Ning Lin
Princeton University
HurricanesStorm SurgeClimate AdaptationCoastal ResilienceRisk Analysis
Z
Zihao Xuan
ACCESS – AI Chip Center for Emerging Smart Systems, InnoHK Centers, Hong Kong Science Park, Hong Kong, China
R
Rongliang Fu
The Chinese University of Hong Kong, Hong Kong, China
Tsung-Yi Ho
Tsung-Yi Ho
Chinese University of Hong Kong
Electronic Design AutomationMicrofluidicsTrustworthy Machine Learning
Y
Yuzhong Jiao
ACCESS – AI Chip Center for Emerging Smart Systems, InnoHK Centers, Hong Kong Science Park, Hong Kong, China
L
Luhong Liang
ACCESS – AI Chip Center for Emerging Smart Systems, InnoHK Centers, Hong Kong Science Park, Hong Kong, China