Human-aligned AI Model Cards with Weighted Hierarchy Architecture

πŸ“… 2025-10-08
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
The rapid proliferation of large language models (LLMs) has exacerbated challenges in discovering and adopting domain-specific models, primarily due to inconsistent, qualitative-only documentation across platforms and the absence of cross-model quantitative comparability. To address this, we propose the Comprehensive Responsible AI Model Card Framework for LLMs (CRAI-MCF)β€”the first weighted hierarchical framework integrating technical, ethical, and operational dimensions. Grounded in value-sensitive design, CRAI-MCF synthesizes 217 parameters empirically derived from 240 open-source projects, organized into an eight-module, layered architecture. It introduces quantitative sufficiency criteria to enable computationally tractable, cross-model evaluation. By shifting model documentation from static disclosure to a dynamic, actionable, and human-centered paradigm, CRAI-MCF significantly improves assessment efficiency, adoption confidence, and responsible deployment capability.

Technology Category

Application Category

πŸ“ Abstract
The proliferation of Large Language Models (LLMs) has led to a burgeoning ecosystem of specialized, domain-specific models. While this rapid growth accelerates innovation, it has simultaneously created significant challenges in model discovery and adoption. Users struggle to navigate this landscape due to inconsistent, incomplete, and imbalanced documentation across platforms. Existing documentation frameworks, such as Model Cards and FactSheets, attempt to standardize reporting but are often static, predominantly qualitative, and lack the quantitative mechanisms needed for rigorous cross-model comparison. This gap exacerbates model underutilization and hinders responsible adoption. To address these shortcomings, we introduce the Comprehensive Responsible AI Model Card Framework (CRAI-MCF), a novel approach that transitions from static disclosures to actionable, human-aligned documentation. Grounded in Value Sensitive Design (VSD), CRAI-MCF is built upon an empirical analysis of 240 open-source projects, distilling 217 parameters into an eight-module, value-aligned architecture. Our framework introduces a quantitative sufficiency criterion to operationalize evaluation and enables rigorous cross-model comparison under a unified scheme. By balancing technical, ethical, and operational dimensions, CRAI-MCF empowers practitioners to efficiently assess, select, and adopt LLMs with greater confidence and operational integrity.
Problem

Research questions and friction points this paper is trying to address.

Addresses inconsistent documentation hindering model discovery and adoption
Introduces quantitative framework for rigorous cross-model comparison
Balances technical ethical operational dimensions for responsible AI selection
Innovation

Methods, ideas, or system contributions that make the work stand out.

Framework uses weighted hierarchy for human-aligned AI
Quantitative sufficiency criterion enables cross-model comparison
Balances technical, ethical, operational dimensions for assessment
πŸ”Ž Similar Papers
No similar papers found.
P
Pengyue Yang
University of Sydney
Haolin Jin
Haolin Jin
University of Sydney
Q
Qingwen Zeng
University of Sydney
J
Jiawen Wen
University of Sydney
H
Harry Rao
Bytedance
Huaming Chen
Huaming Chen
The University of Sydney
Trustworthy MLApplied Machine LearningData MiningService Computing