Learning-Based vs Human-Derived Congestion Control: An In-Depth Experimental Study

📅 2025-10-28
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work systematically evaluates the practical performance and deployment bottlenecks of learning-based congestion control algorithms—particularly reinforcement learning (RL) approaches—in dynamic network environments. Addressing critical gaps in prior work—including lack of reproducible benchmarking, insufficient robustness evaluation, and poor generalization of fairness across network conditions—we establish a unified experimental framework for rigorous comparison among TCP Cubic, BBR, and state-of-the-art learning-based methods. Our key methodological innovation is the explicit incorporation of fairness into the RL reward function. Results reveal that while learning-based algorithms achieve high bandwidth utilization in low-latency settings, their performance degrades significantly under joint bandwidth–delay fluctuations and non-congestion-induced packet loss; moreover, fairness fails to generalize across diverse network topologies. To foster transparency and reproducibility, we fully open-source all code, datasets, and evaluation pipelines, advancing accountable AI-driven networking research.

Technology Category

Application Category

📝 Abstract
Learning-based congestion control (CC), including Reinforcement-Learning, promises efficient CC in a fast-changing networking landscape, where evolving communication technologies, applications and traffic workloads pose severe challenges to human-derived, static CC algorithms. Learning-based CC is in its early days and substantial research is required to understand existing limitations, identify research challenges and, eventually, yield deployable solutions for real-world networks. In this paper, we extend our prior work and present a reproducible and systematic study of learning-based CC with the aim to highlight strengths and uncover fundamental limitations of the state-of-the-art. We directly contrast said approaches with widely deployed, human-derived CC algorithms, namely TCP Cubic and BBR (version 3). We identify challenges in evaluating learning-based CC, establish a methodology for studying said approaches and perform large-scale experimentation with learning-based CC approaches that are publicly available. We show that embedding fairness directly into reward functions is effective; however, the fairness properties do not generalise into unseen conditions. We then show that RL learning-based approaches existing approaches can acquire all available bandwidth while largely maintaining low latency. Finally, we highlight that existing the latest learning-based CC approaches under-perform when the available bandwidth and end-to-end latency dynamically change while remaining resistant to non-congestive loss. As with our initial study, our experimentation codebase and datasets are publicly available with the aim to galvanise the research community towards transparency and reproducibility, which have been recognised as crucial for researching and evaluating machine-generated policies.
Problem

Research questions and friction points this paper is trying to address.

Evaluating learning-based congestion control against human-derived algorithms
Identifying limitations of fairness generalization in unseen network conditions
Assessing performance under dynamic bandwidth and latency changes
Innovation

Methods, ideas, or system contributions that make the work stand out.

Learning-based congestion control uses reinforcement learning
Directly embedding fairness into reward functions
Acquiring bandwidth while maintaining low latency
🔎 Similar Papers
No similar papers found.
M
Mihai Mazilu
School of Engineering and Informatics, University of Sussex
L
Luca Giacomoni
School of Engineering and Informatics, University of Sussex
George Parisis
George Parisis
Professor of Computer Networks, University of Sussex
network protocol designnetwork managementnetwork verificationopportunistic networksinformation-centric networks