🤖 AI Summary
This work addresses the inefficiency of post-training reinforcement learning (RL) fine-tuning in commodity networks—particularly wide-area and standard Ethernet—where limited bandwidth for parameter synchronization severely hampers scalability. To overcome this, the authors propose SparrowRL, a distributed RL fine-tuning system that, for the first time, exploits the high sparsity of parameter updates while preserving bit-level update accuracy. SparrowRL integrates lossless sparse delta encoding, multi-stream parallel transmission, overlap of communication with rollout generation, bandwidth-aware scheduling, and a lease-based fault tolerance mechanism. Experiments on the Qwen3-8B model demonstrate that SparrowRL reduces communication volume by 79×, achieves 2.4–9.5× higher throughput over WANs, reaches 91.09% of RDMA-based single-datacenter performance, and improves cost efficiency by 1.21–1.59× per dollar spent.
📝 Abstract
LLM post-training with reinforcement learning (RL) requires frequent synchronization of large model parameters between the trainer and distributed rollout actors. High-throughput RL post-training therefore relies on dedicated RDMA HPC clusters, an infrastructure cost most organizations cannot absorb. A natural alternative is to aggregate loosely-coupled GPUs over standard Ethernet and WAN links, but this commodity connectivity cannot sustain full-weight broadcasts: synchronizing an 8B model can take over 100~seconds on bandwidth-limited links, while rollout generation typically takes tens of seconds. Toward making RL practical in this regime, we observe that RL fine-tuning yields highly sparse per-step updates, with only around 1\% of parameter elements changing. Atop this insight, we present SparrowRL, a novel high-performance RL training system that preserves bit-exact updates without dropping or quantizing information, designed for commodity-networked, loosely-coupled GPU resources. SparrowRL represents each step as a sparse delta checkpoint, pipelines delta extraction with multi-stream transmission, overlaps transfer with rollout generation, and coordinates heterogeneous workers with throughput- and bandwidth-aware scheduling plus lease-based fault tolerance. On Qwen3 models from 4B to 14B deployed across up to four geographic regions, SparrowRL reduces per-step transfer payload by 79$\times$ for Qwen3-8B and improves throughput by 2.4--9.5$\times$ over full-weight broadcast across WAN, narrowing the throughput gap relative to an ideal RDMA single-datacenter baseline to within 8.91\%. By leveraging on-demand, cross-cloud GPUs over commodity links, SparrowRL delivers 1.21--1.59$\times$ higher tokens per dollar than reserved RDMA clusters at comparable throughput.