CARE Drive A Framework for Evaluating Reason-Responsiveness of Vision Language Models in Automated Driving

πŸ“… 2026-02-17
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
Current evaluation methods for vision-language models in autonomous driving focus solely on output outcomes, making it difficult to assess whether the models’ decisions genuinely respond to human-interpretable reasons and thus risking misjudgment of their explanatory capabilities. This work proposes CARE Drive, the first model-agnostic framework for systematic evaluation: it employs a two-stage design that first stabilizes outputs through prompt calibration and then applies controlled contextual perturbations to quantify the model’s causal responsiveness to normative reasons such as safe following distance and social pressure. Experiments in cyclist-overtaking scenarios demonstrate that explicitly incorporating human-aligned reasons significantly improves alignment with expert driving behavior; however, the models exhibit uneven sensitivity across different reasons, revealing persistent limitations in their capacity to reason responsively.

Technology Category

Application Category

πŸ“ Abstract
Foundation models, including vision language models, are increasingly used in automated driving to interpret scenes, recommend actions, and generate natural language explanations. However, existing evaluation methods primarily assess outcome based performance, such as safety and trajectory accuracy, without determining whether model decisions reflect human relevant considerations. As a result, it remains unclear whether explanations produced by such models correspond to genuine reason responsive decision making or merely post hoc rationalizations. This limitation is especially significant in safety critical domains because it can create false confidence. To address this gap, we propose CARE Drive, Context Aware Reasons Evaluation for Driving, a model agnostic framework for evaluating reason responsiveness in vision language models applied to automated driving. CARE Drive compares baseline and reason augmented model decisions under controlled contextual variation to assess whether human reasons causally influence decision behavior. The framework employs a two stage evaluation process. Prompt calibration ensures stable outputs. Systematic contextual perturbation then measures decision sensitivity to human reasons such as safety margins, social pressure, and efficiency constraints. We demonstrate CARE Drive in a cyclist overtaking scenario involving competing normative considerations. Results show that explicit human reasons significantly influence model decisions, improving alignment with expert recommended behavior. However, responsiveness varies across contextual factors, indicating uneven sensitivity to different types of reasons. These findings provide empirical evidence that reason responsiveness in foundation models can be systematically evaluated without modifying model parameters.
Problem

Research questions and friction points this paper is trying to address.

reason-responsiveness
vision language models
automated driving
model evaluation
human-aligned reasoning
Innovation

Methods, ideas, or system contributions that make the work stand out.

reason-responsiveness
vision language models
automated driving
contextual perturbation
model-agnostic evaluation
πŸ”Ž Similar Papers
No similar papers found.
Lucas Elbert Suryana
Lucas Elbert Suryana
PhD researcher, TU Delft
control systemsautomated vehiclemeaningful human controlresponsible AI
F
Farah Bierenga
Faculty of Mechanical Engineering, Delft University of Technology, Delft, The Netherlands
S
Sanne van Buuren
Faculty of Mechanical Engineering, Delft University of Technology, Delft, The Netherlands
P
Pepijn Kooij
Faculty of Mechanical Engineering, Delft University of Technology, Delft, The Netherlands
E
Elsefien Tulleners
Faculty of Mechanical Engineering, Delft University of Technology, Delft, The Netherlands
F
Federico Scari
Department of Cognitive Robotics, Faculty of Mechanical Engineering, Delft University of Technology; Centre for Meaningful Human Control, Delft University of Technology
S
Simeon Calvert
Department of Transport & Planning, Faculty of Civil Engineering and Geosciences, Delft University of Technology; Centre for Meaningful Human Control, Delft University of Technology
Bart van Arem
Bart van Arem
Delft University of Technology
Intelligent Transport Modelling
Arkady Zgonnikov
Arkady Zgonnikov
Assistant Professor, Delft University of Technology
Cognitive modelingDriver behaviorAutomated drivingMeaningful human controlAI