🤖 AI Summary
This work addresses the high cost and safety risks associated with reinforcement learning fine-tuning of vision-language-action (VLA) models in real-world environments, as well as limitations of existing world model–based approaches—such as pixel-level modeling inefficiencies, multi-view inconsistencies, and error accumulation under sparse rewards. To overcome these challenges, the authors propose the VLA-MBPO framework, which leverages a unified multimodal model for efficient world modeling, introduces an interleaved view decoding mechanism to ensure multi-view consistency, and incorporates a chunked branch rollback strategy to mitigate error propagation in long-horizon predictions. Evaluated on both simulated and real robotic tasks, the method demonstrates substantial improvements in policy performance and sample efficiency, highlighting its robustness and practical deployment potential.
📝 Abstract
Vision-Language-Action (VLA) models show strong generalization for robotic control, but finetuning them with reinforcement learning (RL) is constrained by the high cost and safety risks of real-world interaction. Training VLA models in interactive world models avoids these issues but introduces several challenges, including pixel-level world modeling, multi-view consistency, and compounding errors under sparse rewards. Building on recent advances across large multimodal models and model-based RL, we propose VLA-MBPO, a practical framework to tackle these problems in VLA finetuning. Our approach has three key design choices: (i) adapting unified multimodal models (UMMs) for data-efficient world modeling; (ii) an interleaved view decoding mechanism to enforce multi-view consistency; and (iii) chunk-level branched rollout to mitigate error compounding. Theoretical analysis and experiments across simulation and real-world tasks demonstrate that VLA-MBPO significantly improves policy performance and sample efficiency, underscoring its robustness and scalability for real-world robotic deployment.