π€ AI Summary
This work addresses the challenges of supervision conflicts and inefficient learning in offline heterogeneous robotic datasets, which arise from discrepancies in embodiment, camera configurations, and demonstration quality. To this end, the authors propose a reward-free conservative offline policy learning method that introduces a novel posterior-uniform ratio as sample weights. By leveraging latent goal encoding, candidate goal pool matching, and a transition scorer to estimate the posterior distribution, the approach enables adaptive credit assignment without requiring policy likelihoods, and is compatible with both diffusion models and flow-matching action heads. Combined with clipped mixture weighting and self-normalized weighted regression, the method significantly enhances the policyβs conservative adaptability to heterogeneous data, effectively mitigating performance degradation caused by low-quality or conflicting demonstrations.
π Abstract
Offline post-training adapts a pretrained robot policy to a target dataset by supervised regression on recorded actions. In practice, robot datasets are heterogeneous: they mix embodiments, camera setups, and demonstrations of varying quality, so many trajectories reflect recovery behavior, inconsistent operator skill, or weakly informative supervision. Uniform post-training gives equal credit to all samples and can therefore average over conflicting or low-attribution data. We propose Posterior-Transition Reweighting (PTR), a reward-free and conservative post-training method that decides how much each training sample should influence the supervised update. For each sample, PTR encodes the observed post-action consequence as a latent target, inserts it into a candidate pool of mismatched targets, and uses a separate transition scorer to estimate a softmax identification posterior over target indices. The posterior-to-uniform ratio defines the PTR score, which is converted into a clipped-and-mixed weight and applied to the original action objective through self-normalized weighted regression. This construction requires no tractable policy likelihood and is compatible with both diffusion and flow-matching action heads. Rather than uniformly trusting all recorded supervision, PTR reallocates credit according to how attributable each sample's post-action consequence is under the current representation, improving conservative offline adaptation to heterogeneous robot data.