From Simulation to Reality: Practical Deep Reinforcement Learning-based Link Adaptation for Cellular Networks

📅 2026-02-28
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the performance degradation of existing deep reinforcement learning (DRL)-based link adaptation methods when deployed in real-world systems, as they are typically evaluated under idealized simulation conditions that neglect practical impairments such as ACK/NACK feedback delays, HARQ retransmissions, and DRL inference latency. To bridge this gap, the authors propose the Decoupling-DQN framework (DC-DQN-LA), which uniquely decouples DRL training and inference into a non-real-time training module and a real-time inference module. The design explicitly incorporates feedback delay and parallel HARQ mechanisms into the state, action, and reward formulations. Validated on the srsRAN 5G protocol stack with a USRP software-defined radio platform, the approach achieves 40%–70% higher throughput than baseline algorithms under mobility scenarios while maintaining comparable block error rates and demonstrating rapid adaptation to dynamic channel conditions, thereby significantly enhancing real-world deployment feasibility.

Technology Category

Application Category

📝 Abstract
Link Adaptation (LA) that dynamically adjusts the Modulation and Coding Schemes (MCS) to accommodate time-varying channels is crucial and challenging in cellular networks. Deep reinforcement learning (DRL)-based LA that learns to make decision through the interaction with the environment is a promising approach to improve throughput. However, existing DRL-based LA algorithms are typically evaluated in simplified simulation environments, neglecting practical issues such as ACK/NACK feedback delay, retransmission and parallel hybrid automatic repeat request (HARQ). Moreover, these algorithms overlook the impact of DRL execution latency, which can significantly degrade system performance. To address these challenges, we propose Decoupling-DQN (DC-DQN), a new DRL framework that separates traditional DRL's coupled training and inference processes into two modules based on Deep Q Networks (DQN): a real-time inference module and an out-of-decision-loop training module. Based on this framework, we introduce a novel DRL-based LA algorithm, DC-DQN-LA. The algorithm incorporates practical considerations by designing state, action, and reward functions that account for feedback delays, parallel HARQ, and retransmissions. We implemented a prototype using USRP software-defined radios and srsRAN software. Experimental results demonstrate that DC-DQN-LA improves throughput by 40\% to 70\% in mobile scenario compared with baseline LA algorithms, while maintaining comparable block error rates, and can quickly adapt to environment changes in mobile-to-static scenario. These results highlight the efficiency and practicality of the proposed DRL-based LA algorithm.
Problem

Research questions and friction points this paper is trying to address.

Link Adaptation
Deep Reinforcement Learning
Feedback Delay
HARQ
Execution Latency
Innovation

Methods, ideas, or system contributions that make the work stand out.

Decoupling-DQN
Link Adaptation
Deep Reinforcement Learning
HARQ
Practical Wireless Implementation
🔎 Similar Papers
No similar papers found.
Lizhao You
Lizhao You
School of Informatics, Xiamen University
Wireless NetworksComputer Networks
N
Nanqing Zhou
School of Informatics, Xiamen University, China; ArrayComm Wireless Technologies Co., Ltd., China
G
Guanglong Pang
School of Informatics, Xiamen University, China
J
Jiajie Huang
School of Informatics, Xiamen University, China
Yulin Shao
Yulin Shao
University of Hong Kong
Coding and ModulationMachine LearningStochastic Control
Liqun Fu
Liqun Fu
Full Professor, Xiamen University
wireless communication networks