Energy Optimization of Multi-task DNN Inference in MEC-assisted XR Devices: A Lyapunov-Guided Reinforcement Learning Approach

📅 2025-01-05
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address high energy consumption, severe resource contention, and coupled queue interference in multi-task DNN inference for lightweight XR devices under MEC assistance, this paper establishes a distributed queuing model and proposes a dual-timescale joint optimization framework that jointly determines model partitioning and edge-terminal resource allocation. We innovatively design Lyapunov-guided Proximal Policy Optimization (LyaPPO), a reinforcement learning algorithm that directly embeds queue stability constraints into the policy gradient update, enabling tight coupling between energy-efficiency minimization and strong system stability. Experimental results demonstrate that the proposed approach reduces XR-device energy consumption by 24.29%–56.62%, achieves system-level energy savings of 24.79%–46.14%, and significantly outperforms baseline methods.

Technology Category

Application Category

📝 Abstract
Extended reality (XR), blending virtual and real worlds, is a key application of future networks. While AI advancements enhance XR capabilities, they also impose significant computational and energy challenges on lightweight XR devices. In this paper, we developed a distributed queue model for multi-task DNN inference, addressing issues of resource competition and queue coupling. In response to the challenges posed by the high energy consumption and limited resources of XR devices, we designed a dual time-scale joint optimization strategy for model partitioning and resource allocation, formulated as a bi-level optimization problem. This strategy aims to minimize the total energy consumption of XR devices while ensuring queue stability and adhering to computational and communication resource constraints. To tackle this problem, we devised a Lyapunov-guided Proximal Policy Optimization algorithm, named LyaPPO. Numerical results demonstrate that the LyaPPO algorithm outperforms the baselines, achieving energy conservation of 24.79% to 46.14% under varying resource capacities. Specifically, the proposed algorithm reduces the energy consumption of XR devices by 24.29% to 56.62% compared to baseline algorithms.
Problem

Research questions and friction points this paper is trying to address.

Energy Efficiency
Resource Management
Multi-task Deep Neural Networks
Innovation

Methods, ideas, or system contributions that make the work stand out.

Lyapunov-based Reinforcement Learning
Resource Allocation for XR Devices
Energy Efficiency Optimization
🔎 Similar Papers
No similar papers found.
Y
Yanzan Sun
School of Communication and Information Engineering, Shanghai University, Shanghai 200444, China
J
Jiacheng Qiu
School of Communication and Information Engineering, Shanghai University, Shanghai 200444, China
Guangjin Pan
Guangjin Pan
Postdoc, Chalmers University of Technology
Semantic communicationsRadio localization and sensingAI-native networks
Shugong Xu
Shugong Xu
Professor at Xi'an Jiaotong-Liverpool University, IEEE Fellow
Machine LearningPattern RecognitionWireless Systems
Shunqing Zhang
Shunqing Zhang
Professor, SHU; Previous in Intel & Huawei
Green CommunicationsRelayInterference Management5G Wireless NetworksNon-orthogonal Waveform Design
X
Xiaoyun Wang
China Mobile Research Institute, Beijing 100053, China
S
Shuangfeng Han
China Mobile Research Institute, Beijing 100053, China