On the Trustworthiness of Generative Foundation Models: Guideline, Assessment, and Perspective

πŸ“… 2025-02-20
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
The widespread deployment of generative foundation models (GenFMs) necessitates systematic, trustworthy evaluation and enhancement mechanisms. Method: We propose the first full-stack framework for GenFM trustworthiness, integrating multidimensional governance guidelines, a dynamic evaluation platform (TrustGen), and an evolution roadmap. Our approach uniquely unifies technical, ethical, legal, and societal perspectives to formulate trustworthiness principles; introduces TrustGenβ€”a modular, multimodal, iteratively updatable platform supporting metadata governance, test-case generation, and contextual perturbation for fine-grained trust assessment; and grounds design in global AI policy analysis and interdisciplinary co-modeling. Contribution/Results: We systematically characterize advances and bottlenecks across robustness, fairness, and explainability dimensions; open-source the TrustGen toolkit; and deliver a reusable, extensible infrastructure for dynamic, evidence-based GenFM trust evaluation.

Technology Category

Application Category

πŸ“ Abstract
Generative Foundation Models (GenFMs) have emerged as transformative tools. However, their widespread adoption raises critical concerns regarding trustworthiness across dimensions. This paper presents a comprehensive framework to address these challenges through three key contributions. First, we systematically review global AI governance laws and policies from governments and regulatory bodies, as well as industry practices and standards. Based on this analysis, we propose a set of guiding principles for GenFMs, developed through extensive multidisciplinary collaboration that integrates technical, ethical, legal, and societal perspectives. Second, we introduce TrustGen, the first dynamic benchmarking platform designed to evaluate trustworthiness across multiple dimensions and model types, including text-to-image, large language, and vision-language models. TrustGen leverages modular components--metadata curation, test case generation, and contextual variation--to enable adaptive and iterative assessments, overcoming the limitations of static evaluation methods. Using TrustGen, we reveal significant progress in trustworthiness while identifying persistent challenges. Finally, we provide an in-depth discussion of the challenges and future directions for trustworthy GenFMs, which reveals the complex, evolving nature of trustworthiness, highlighting the nuanced trade-offs between utility and trustworthiness, and consideration for various downstream applications, identifying persistent challenges and providing a strategic roadmap for future research. This work establishes a holistic framework for advancing trustworthiness in GenAI, paving the way for safer and more responsible integration of GenFMs into critical applications. To facilitate advancement in the community, we release the toolkit for dynamic evaluation.
Problem

Research questions and friction points this paper is trying to address.

Addresses trustworthiness concerns in Generative Foundation Models.
Introduces TrustGen platform for dynamic trustworthiness evaluation.
Proposes guiding principles integrating technical, ethical, legal perspectives.
Innovation

Methods, ideas, or system contributions that make the work stand out.

Dynamic benchmarking platform TrustGen
Multidisciplinary guiding principles
Modular components for adaptive assessments
πŸ”Ž Similar Papers
No similar papers found.
Y
Yue Huang
University of Notre Dame
Chujie Gao
Chujie Gao
iSURE visiting student, University of Notre Dame
S
Siyuan Wu
University of Waterloo
H
Haoran Wang
Emory University
X
Xiangqi Wang
University of Notre Dame
Yujun Zhou
Yujun Zhou
University of Notre Dame
Trustworthy LLMLLM ReasoninngAdversarial Machine Learning
Y
Yanbo Wang
Mohamed bin Zayed University of Artificial Intelligence
Jiayi Ye
Jiayi Ye
Master student in Shanghaitech University
Embodied AIComputer Vision
Jiawen Shi
Jiawen Shi
Huazhong University of Science and Technology
AI Security
Qihui Zhang
Qihui Zhang
Peking University
Human AlignmentMulti-ModalityLarge Language Model
Y
Yuan Li
Carnegie Mellon University
H
Han Bao
University of Queensland
Z
Zhaoyi Liu
University of Illinois Urbana-Champaign
Tianrui Guan
Tianrui Guan
Waymo
Computer VisionPerceptionRoboticsVLM
D
Dongping Chen
University of Washington
Ruoxi Chen
Ruoxi Chen
Zhejiang University of Technology
Trustworthy AIMultimodal Models
Kehan Guo
Kehan Guo
University of Notre Dame
LLMMachine ReasoningGenerative ModelsXAIAI for Science
Andy Zou
Andy Zou
PhD Student, Carnegie Mellon University
ML SafetyAI Safety
B
Bryan Hooi Kuen-Yew
National University of Singapore
Caiming Xiong
Caiming Xiong
Salesforce Research
Machine LearningNLPComputer VisionMultimediaData Mining
Elias Stengel-Eskin
Elias Stengel-Eskin
Assistant Professor, University of Texas at Austin
Natural language processingcomputational semanticscomputational linguistics
Hongyang Zhang
Hongyang Zhang
Assistant Professor of Computer Science, University of Waterloo
Machine LearningInference AccelerationAI Security
Hongzhi Yin
Hongzhi Yin
Professor and ARC Future Fellow, University of Queensland
Recommender SystemGraph LearningSpatial-temporal PredictionEdge IntelligenceLLM
H
Huan Zhang
University of Illinois Urbana-Champaign
Huaxiu Yao
Huaxiu Yao
Assistant Professor of Computer Science and Data Science, UNC Chapel Hill
Machine LearningFoundation ModelsAI AlignmentAI AgentRobot Learning
J
Jaehong Yoon
UNC Chapel Hill
Jieyu Zhang
Jieyu Zhang
University of Washington
Data-Centric AIAgentic AIMultimodal ModelsMachine LearningComputer Vision
Kai Shu
Kai Shu
Assistant Professor of Computer Science, Emory University
Data MiningTrustworthy AISocial ComputingMachine LearningAI Safety
Kaijie Zhu
Kaijie Zhu
University of California, Santa Barbara
Ranjay Krishna
Ranjay Krishna
University of Washington, Allen Institute for AI
Computer VisionNatural Language ProcessingMachine LearningHuman Computer Interaction
Swabha Swayamdipta
Swabha Swayamdipta
University of Southern California
Natural Language ProcessingMachine Learning
Taiwei Shi
Taiwei Shi
University of Southern California
Natural Language ProcessingComputational Social ScienceMachine Learning
Weijia Shi
Weijia Shi
University of Washington
Natural Language ProcessingMachine Learning
X
Xiang Li
Massachusetts General Hospital
Y
Yiwei Li
University of Georgia
Yuexing Hao
Yuexing Hao
Research Fellow
Human Computer InteractionHealth Intelligence
Zhihao Jia
Zhihao Jia
Assistant Professor of Computer Science, Carnegie Mellon University
Computer SystemsMachine LearningDeep Neural Networks
Zhize Li
Zhize Li
Assistant Professor, Singapore Management University
OptimizationFederated LearningAI PrivacyMachine Learning
Xiuying Chen
Xiuying Chen
MBZUAI
Trustworthy NLPHuman-Centered NLPComputational Social Science
Zhengzhong Tu
Zhengzhong Tu
Texas A&M University, Google Research, University of Texas at Austin
Agentic AITrustworthy AIEmbodied AI
Xiyang Hu
Xiyang Hu
PhD, Carnegie Mellon University
Machine LearningTrustworthyHuman-AIGenerative AIOut of Distribution
T
Tianyi Zhou
University of Maryland
Jieyu Zhao
Jieyu Zhao
Assistant Professor at USC
Natural Language ProcessingMachine LearningFairness in AI
L
Lichao Sun
Lehigh University
Furong Huang
Furong Huang
Associate Professor of Computer Science, University of Maryland
Trustworthy AI/MLReinforcement LearningGenerative AI
O
Or Cohen Sasson
University of Miami
P
P. Sattigeri
IBM Research
Anka Reuel
Anka Reuel
CS Ph.D. Candidate, Stanford University
AI GovernanceResponsible AIAI EthicsAI Safety
Max Lamparth
Max Lamparth
Research Fellow, Stanford University
Machine LearningUncertainty QuantificationInterpretabilityAI SafetyResponsible AI
Y
Yue Zhao
University of Southern California
Nouha Dziri
Nouha Dziri
Allen Institute for AI (Ai2)
Artificial IntelligenceNatural Language Processing
Y
Yu Su
Ohio State University
Huan Sun
Huan Sun
Endowed CoE Innovation Scholar and Associate Professor, The Ohio State University
AgentsLarge Language ModelsNatural Language ProcessingAI
Heng Ji
Heng Ji
Professor of Computer Science, AICE Director, ASKS Director, UIUC, Amazon Scholar
Natural Language ProcessingLarge Language Models
Chaowei Xiao
Chaowei Xiao
University of Wisconsin - Madison/NVIDIA
Trustworthy Machine LearningAdversarial Machine LearningAI SafetyRobust AISecurity
Mohit Bansal
Mohit Bansal
Parker Distinguished Professor, Computer Science, UNC Chapel Hill
Natural Language ProcessingComputer VisionMachine LearningMultimodal AI
N
N. V. Chawla
University of Notre Dame
Jian Pei
Jian Pei
Arthur S. Pearse Distinguished Professor, Duke University
Data miningbig data analyticsdatabase systemsinformation retrieval
J
Jianfeng Gao
Microsoft Research
Michael Backes
Michael Backes
Chairman and Founding Director of the CISPA Helmholtz Center for Information Security
SecurityprivacycryptographyAI
Philip S. Yu
Philip S. Yu
Professor of Computer Science, University of Illinons at Chicago
Data miningDatabasePrivacy
Neil Zhenqiang Gong
Neil Zhenqiang Gong
Associate Professor, Duke University
SecurityAI Security/SafetySocial Networks SecurityGenerative AI
Pin-Yu Chen
Pin-Yu Chen
Principal Research Scientist, IBM Research AI; MIT-IBM Watson AI Lab; RPI-IBM AIRC
AI SafetyGenerative AITrustworthy Machine LearningAdversarial Machine Learning
B
Bo Li
University of Chicago
Xiangliang Zhang
Xiangliang Zhang
Leonard C. Bettex Collegiate Professor, Computer Science and Engineering, University of Notre Dame
Machine LearningAI for Science