Secure and Storage-Efficient Deep Learning Models for Edge AI Using Automatic Weight Generation

πŸ“… 2025-07-08
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
Edge AI faces critical challenges of high neural network weight storage overhead and weak security guarantees. To address these, this paper proposes WINGsβ€”a novel framework that introduces a dynamic weight generation mechanism integrating Principal Component Analysis (PCA) for dimensionality reduction and lightweight Support Vector Regression (SVR), thereby eliminating explicit storage of fully connected layer weights. Additionally, leveraging sensitivity analysis, WINGs applies selective weight compression only to low-sensitivity layers in CNNs. This design jointly optimizes storage efficiency and robustness: it achieves 53Γ— and 18Γ— model compression on MNIST and CIFAR-10, respectively, significantly reducing memory footprint, improving inference throughput, lowering energy consumption, and incurring only 1–2% accuracy degradation. Moreover, WINGs enhances detection capability against adversarial weight tampering attacks. Overall, WINGs establishes a new paradigm for efficient and secure neural inference under stringent resource constraints.

Technology Category

Application Category

πŸ“ Abstract
Complex neural networks require substantial memory to store a large number of synaptic weights. This work introduces WINGs (Automatic Weight Generator for Secure and Storage-Efficient Deep Learning Models), a novel framework that dynamically generates layer weights in a fully connected neural network (FC) and compresses the weights in convolutional neural networks (CNNs) during inference, significantly reducing memory requirements without sacrificing accuracy. WINGs framework uses principal component analysis (PCA) for dimensionality reduction and lightweight support vector regression (SVR) models to predict layer weights in the FC networks, removing the need for storing full-weight matrices and achieving substantial memory savings. It also preferentially compresses the weights in low-sensitivity layers of CNNs using PCA and SVR with sensitivity analysis. The sensitivity-aware design also offers an added level of security, as any bit-flip attack with weights in compressed layers has an amplified and readily detectable effect on accuracy. WINGs achieves 53x compression for the FC layers and 28x for AlexNet with MNIST dataset, and 18x for Alexnet with CIFAR-10 dataset with 1-2% accuracy loss. This significant reduction in memory results in higher throughput and lower energy for DNN inference, making it attractive for resource-constrained edge applications.
Problem

Research questions and friction points this paper is trying to address.

Reducing memory usage in deep learning models for edge AI
Dynamically generating and compressing neural network weights
Enhancing security and efficiency in resource-constrained edge applications
Innovation

Methods, ideas, or system contributions that make the work stand out.

Dynamic weight generation in FC networks
PCA and SVR for CNN weight compression
Sensitivity-aware design enhances security
πŸ”Ž Similar Papers
No similar papers found.
H
Habibur Rahaman
School of Electrical and Computer Engineering, University of Florida
A
Atri Chatterjee
School of Electrical and Computer Engineering, University of Florida
Swarup Bhunia
Swarup Bhunia
University of Florida
IoT SecurityHardware SecurityEnergy-Efficient ElectronicsFood/Medicine Safety