ALOE: Action-Level Off-Policy Evaluation for Vision-Language-Action Model Post-Training

📅 2026-02-13
📈 Citations: 0
Influential: 0
📄 PDF

Technology Category

Application Category

📝 Abstract
We study how to improve large foundation vision-language-action (VLA) systems through online reinforcement learning (RL) in real-world settings. Central to this process is the value function, which provides learning signals to guide VLA learning from experience. In practice, the value function is estimated from trajectory fragments collected from different data sources, including historical policies and intermittent human interventions. Estimating the value function of current behavior quality from the mixture data is inherently an off-policy evaluation problem. However, prior work often adopts conservative on-policy estimation for stability, which avoids direct evaluation of the current high-capacity policy and limits learning effectiveness. In this paper, we propose ALOE, an action-level off-policy evaluation framework for VLA post-training. ALOE applies chunking-based temporal-difference bootstrapping to evaluate individual action sequences instead of predicting final task outcomes. This design improves effective credit assignment to critical action chunks under sparse rewards and supports stable policy improvement. We evaluate our method on three real-world manipulation tasks, including smartphone packing as a high-precision task, laundry folding as a long-horizon deformable-object task, and bimanual pick-and-place involving multi-object perception. Across all tasks, ALOE improves learning efficiency without compromising execution speed, showing that off-policy RL can be reintroduced in a reliable manner for real-world VLA post-training. Videos and additional materials are available at our project website.
Problem

Research questions and friction points this paper is trying to address.

off-policy evaluation
vision-language-action model
reinforcement learning
value function estimation
real-world robotics
Innovation

Methods, ideas, or system contributions that make the work stand out.

off-policy evaluation
vision-language-action models
action-level credit assignment
temporal-difference bootstrapping
real-world reinforcement learning
🔎 Similar Papers
No similar papers found.
Rushuai Yang
Rushuai Yang
Hong Kong University of Science and Technology
Reinforcement LearningEmbodied AI
H
Hecheng Wang
AgiBot
C
Chiming Liu
AgiBot
X
Xiaohan Yan
AgiBot
Y
Yunlong Wang
AgiBot
X
Xuan Du
AgiBot
S
Shuoyu Yue
AgiBot
Y
Yongcheng Liu
AgiBot
C
Chuheng Zhang
Independent Researcher
L
Lizhe Qi
Fudan University
Yi Chen
Yi Chen
Hong Kong University of Science and Technology
W
Wei Shan
AgiBot
Maoqing Yao
Maoqing Yao
Google