π€ AI Summary
Vision-language-action (VLA) foundation models suffer from poor confidence calibration, undermining decision reliability and hindering risk-aware deployment.
Method: We propose the first systematic calibration framework for VLA models, integrating Bayesian-inspired prompt ensembling, action-dimension-wise Platt scaling, and temporal confidence analysis.
Contribution/Results: On multiple VLA models (e.g., RT-2, OpenVLA) and benchmark tasks, we empirically uncoverβ for the first timeβa dynamic drift in model confidence over task execution, enabling time-sensitive, risk-aware intervention. Experiments show our method reduces expected calibration error (ECE) by 38% on average while preserving high task success rates. It significantly improves confidence reliability without sacrificing performance, demonstrating that accuracy and calibration can be jointly optimized. This work establishes a new paradigm and practical toolkit for safe, trustworthy VLA model deployment.
π Abstract
Trustworthy robot behavior requires not only high levels of task success but also that the robot can reliably quantify how likely it is to succeed. To this end, we present the first systematic study of confidence calibration in vision-language-action (VLA) foundation models, which map visual observations and natural-language instructions to low-level robot motor commands. We begin with extensive benchmarking to understand the critical relationship between task success and calibration error across multiple datasets and VLA variants, finding that task performance and calibration are not in tension. Next, we introduce prompt ensembles for VLAs, a lightweight, Bayesian-inspired algorithm that averages confidence across paraphrased instructions and consistently improves calibration. We further analyze calibration over the task time horizon, showing that confidence is often most reliable after making some progress, suggesting natural points for risk-aware intervention. Finally, we reveal differential miscalibration across action dimensions and propose action-wise Platt scaling, a method to recalibrate each action dimension independently to produce better confidence estimates. Our aim in this study is to begin to develop the tools and conceptual understanding necessary to render VLAs both highly performant and highly trustworthy via reliable uncertainty quantification.