🤖 AI Summary
Achieving ultra-low latency and high-throughput congestion control in L4S networks remains challenging for conventional active queue management (AQM) schemes. Method: This paper pioneers the integration of large language model (LLM) distillation into AQM, proposing a lightweight LLM-AQM controller that combines few-shot learning, context-awareness, and speculative decoding, with reinforcement learning optimizing the distillation process. Contribution/Results: We introduce the first open-source, kernel-level FreeBSD-14 L4S-AQM experimental platform supporting LLM integration—designed to advance IETF standardization efforts. Extensive experiments demonstrate that our approach reduces end-to-end latency by 37% on average, significantly mitigates queue buildup and packet loss, and enhances throughput stability. These results validate the efficacy, low computational overhead, and strong generalization capability of LLM distillation for real-time network control.
📝 Abstract
The growing complexity of network traffic and demand for ultra-low latency communication require smarter packet traffic management. Existing Deep Learning-based queuing approaches struggle with dynamic network scenarios and demand high engineering effort. We propose AQM-LLM, distilling Large Language Models (LLMs) with few-shot learning, contextual understanding, and pattern recognition to improve Active Queue Management (AQM) [RFC 9330] with minimal manual effort. We consider a specific case where AQM is Low Latency, Low Loss, and Scalable Throughput (L4S) and our design of AQM-LLM builds on speculative decoding and reinforcement-based distilling of LLM by tackling congestion prevention in the L4S architecture using Explicit Congestion Notification (ECN) [RFC 9331] and periodic packet dropping. We develop a new open-source experimental platform by executing L4S-AQM on FreeBSD-14, providing interoperable modules to support LLM integration and facilitate IETF recognition through wider testing. Our extensive evaluations show L4S-LLM enhances queue management, prevents congestion, reduces latency, and boosts network performance, showcasing LLMs' adaptability and efficiency in uplifting AQM systems.