🤖 AI Summary
In self-hosted environments, static batching configurations for CodeLLM services struggle to handle request-rate fluctuations and heterogeneous workloads, leading to frequent SLA violations and unstable performance. To address this, we propose a dynamic request scheduling framework. Its core innovation is a lightweight SLA feasibility predictor that enables per-request latency estimation and real-time, adaptive batching decisions—eliminating the need for manual hyperparameter tuning. The framework jointly optimizes throughput, latency stability, and resource utilization under constrained hardware conditions. Experimental results demonstrate that, compared to the optimal static configuration, our approach achieves up to 26% higher effective throughput and reduces latency standard deviation by up to 45%. This significantly improves service robustness and SLA compliance, particularly under dynamic and resource-constrained deployment scenarios.
📝 Abstract
Code Large Language Models (CodeLLMs) are increasingly integrated into modern software development workflows, yet efficiently serving them in resource-constrained, self-hosted environments remains a significant challenge. Existing LLM serving systems employs Continuous Batching for throughput improvement. However, they rely on static batch size configurations that cannot adapt to fluctuating request rates or heterogeneous workloads, leading to frequent SLA (Service Level Agreement) violations and unstable performance. In this study, We propose SABER, a dynamic batching strategy that predicts per-request SLA feasibility and adjusts decisions in real time. SABER improves goodput by up to 26% over the best static configurations and reduces latency variability by up to 45%, all without manual tuning or service restarts. Our results demonstrate that SLA-aware, adaptive scheduling is key to robust, high-performance CodeLLM serving.