Privacy-Aware Joint DNN Model Deployment and Partition Optimization for Delay-Efficient Collaborative Edge Inference

๐Ÿ“… 2025-02-22
๐Ÿ“ˆ Citations: 0
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
This paper addresses the joint deployment optimization of deep neural networks (DNNs) for edge collaborative inference under resource constraints and privacy requirements, aiming to minimize long-term average inference latency. We propose the first integrated framework combining Lyapunov optimization, coalition game modeling, and greedy heuristics: Lyapunov optimization enables dynamic resourceโ€“latency trade-offs; coalition game theory jointly models server association and model partitioning decisions; and implicit differential privacy constraints enforce end-to-end privacy budget adherence. The framework simultaneously achieves low latency, strong privacy guarantees, and scalability in dynamic edge environments. Simulation results demonstrate a 23.6%โ€“38.1% reduction in inference latency compared to baseline methods, strict compliance with the prescribed privacy budget, and robust performance under highly variable workloads.

Technology Category

Application Category

๐Ÿ“ Abstract
Edge inference (EI) is a key solution to address the growing challenges of delayed response times, limited scalability, and privacy concerns in cloud-based Deep Neural Network (DNN) inference. However, deploying DNN models on resource-constrained edge devices faces more severe challenges, such as model storage limitations, dynamic service requests, and privacy risks. This paper proposes a novel framework for privacy-aware joint DNN model deployment and partition optimization to minimize long-term average inference delay under resource and privacy constraints. Specifically, the problem is formulated as a complex optimization problem considering model deployment, user-server association, and model partition strategies. To handle the NP-hardness and future uncertainties, a Lyapunov-based approach is introduced to transform the long-term optimization into a single-time-slot problem, ensuring system performance. Additionally, a coalition formation game model is proposed for edge server association, and a greedy-based algorithm is developed for model deployment within each coalition to efficiently solve the problem. Extensive simulations show that the proposed algorithms effectively reduce inference delay while satisfying privacy constraints, outperforming baseline approaches in various scenarios.
Problem

Research questions and friction points this paper is trying to address.

Optimize DNN model deployment
Minimize inference delay
Ensure privacy constraints
Innovation

Methods, ideas, or system contributions that make the work stand out.

Privacy-aware DNN model deployment
Lyapunov-based optimization approach
Coalition formation game model
๐Ÿ”Ž Similar Papers
No similar papers found.
Zhipeng Cheng
Zhipeng Cheng
Soochow University
Edge IntelligenceFederated LearningUAV NetworksService Computing
Xiaoyu Xia
Xiaoyu Xia
School of Computing Technologies, RMIT University
Parallel and Distributed ComputingSystem SecurityEdge ComputingSustainable Computing
H
Hong Wang
School of Future Science and Engineering, Soochow University, Suzhou 215006, China
M
Minghui Liwang
Department of Control Science and Engineering, The National Key Laboratory of Autonomous Intelligent Unmanned Systems, Tongji University, Shanghai 201804, China, and also with the Frontiers Science Center for Intelligent Autonomous Systems, Ministry of Education, Tongji University, Shanghai 201804, China
N
Ning Chen
Department of Automation, China University of Petroleum (East China), Qingdao, China
X
Xuwei Fan
College of Computer and Information Sciences, Fujian Agriculture and Forestry University, Fuzhou 350002, China
X
Xianbin Wang
Department of Electrical and Computer Engineering, Western University, Ontario, Canada