The Trust Calibration Maturity Model for Characterizing and Communicating Trustworthiness of AI Systems

📅 2025-01-28
🏛️ arXiv.org
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Amid growing concerns over insufficient evaluation metrics for AI system trustworthiness and difficulties in calibrating user trust in the large language model era, this paper proposes the Trustworthiness Calibration Maturity Model (TCMM)—the first multidimensional maturity framework explicitly designed for user trust calibration. TCMM establishes a structured, measurable, communicable, and evolvable metric system across five dimensions: performance characterization, bias and robustness, transparency, safety and security, and usability—thereby enabling systematic quantification and articulation of abstract trustworthiness. Through maturity modeling, multidimensional trust assessment, and governance framework design, TCMM delivers theoretical formalization and standardized definitions. Empirical validation on two AI system–task pairs demonstrates its practical efficacy, significantly improving users’ accuracy in perceiving AI capability boundaries and enhancing usage rationality.

Technology Category

Application Category

📝 Abstract
The proliferation of powerful AI capabilities and systems necessitates a commensurate focus on user trust. We introduce the Trust Calibration Maturity Model (TCMM) to capture and communicate the maturity of AI system trustworthiness. The TCMM scores maturity along 5 dimensions that drive user trust: Performance Characterization, Bias&Robustness Quantification, Transparency, Safety&Security, and Usability. Information captured in the TCMM can be presented along with system performance information to help a user to appropriately calibrate trust, to compare requirements with current states of development, and to clarify trustworthiness needs. We present the TCMM and demonstrate its use on two AI system-target task pairs.
Problem

Research questions and friction points this paper is trying to address.

Characterizing trustworthiness of AI systems for user calibration
Addressing inaccessibility of trust evaluation in large-scale AI
Providing a maturity model to assess AI trust dimensions
Innovation

Methods, ideas, or system contributions that make the work stand out.

Proposes Trust Calibration Maturity Model (TCMM)
Incorporates five dimensions of analytic maturity
Demonstrates TCMM on ChatGPT and PhaseNet
S
Scott Steinmetz
Sandia National Labs
A
Asmeret Naugle
Sandia National Labs
P
Paul Schutte
Sandia National Labs
M
Matt Sweitzer
Sandia National Labs
Alex Washburne
Alex Washburne
Sandia National Labs
L
Lisa Linville
Sandia National Labs
D
Daniel Krofcheck
Sandia National Labs
Michal Kucer
Michal Kucer
Staff Scientist, Los Alamos National Laboratory
Computer VisionDeep LearningMachine Learning
S
Samuel Myren
Los Alamos National Labs, Virginia Tech