🤖 AI Summary
To address three key challenges in reinforcement learning (RL)-enhanced reasoning for large language models (LLMs)—outdated policy data, RL algorithm performance degradation, and low AI accelerator utilization—this paper proposes an efficient online RL training framework tailored for long-sequence generation. Our method introduces: (1) an “in-flight weight update” mechanism that dynamically loads the latest policy weights during sequence generation, substantially improving data freshness; and (2) an asynchronous pipeline-parallel architecture that decouples data generation from model training, integrating real-time weight synchronization and fine-grained GPU scheduling. Experiments on a 128-H100 cluster demonstrate that our approach achieves approximately 2× speedup over conventional RL training while strictly preserving high online-ness and sustaining >92% GPU utilization—effectively alleviating throughput bottlenecks inherent in RLHF-style training.
📝 Abstract
Reinforcement Learning (RL) is increasingly utilized to enhance the reasoning capabilities of Large Language Models (LLMs). However, effectively scaling these RL methods presents significant challenges, primarily due to the difficulty in maintaining high AI accelerator utilization without generating stale, off-policy data that harms common RL algorithms. This paper introduces PipelineRL, an approach designed to achieve a superior trade-off between hardware efficiency and data on-policyness for LLM training. PipelineRL employs concurrent asynchronous data generation and model training, distinguished by the novel in-flight weight updates. This mechanism allows the LLM generation engine to receive updated model weights with minimal interruption during the generation of token sequences, thereby maximizing both the accelerator utilization and the freshness of training data. Experiments conducted on long-form reasoning tasks using 128 H100 GPUs demonstrate that PipelineRL achieves approximately $sim 2x$ faster learning compared to conventional RL baselines while maintaining highly on-policy training data. A scalable and modular open-source implementation of PipelineRL is also released as a key contribution.