Never Compromise to Vulnerabilities: A Comprehensive Survey on AI Governance

📅 2025-08-12
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
AI safety vulnerabilities—including algorithmic bias and adversarial fragility—fuel misinformation, inequity, security threats, and eroded public trust, exposing critical gaps in existing governance frameworks. To address this, we propose a unified three-dimensional governance framework: “intrinsic safety,” “derivative safety,” and “socio-ethical alignment.” This is the first approach to holistically integrate technical defenses (e.g., robustness enhancement, fairness-aware modeling, adversarial detection), real-world risk assessment (leveraging emerging evaluation benchmarks), and cross-disciplinary policy coordination. Through a systematic review of over 300 studies, we identify three core challenges: generalization gaps, insufficient evaluation rigor, and regulatory fragmentation—arguing for proactive, lifecycle-integrated governance rather than reactive remediation. Our work delivers a tripartite output: actionable technical guidelines, standardized evaluation metrics, evidence-based policy recommendations, and an open-source tool suite to advance global trustworthy AI ecosystems.

Technology Category

Application Category

📝 Abstract
The rapid advancement of AI has expanded its capabilities across domains, yet introduced critical technical vulnerabilities, such as algorithmic bias and adversarial sensitivity, that pose significant societal risks, including misinformation, inequity, security breaches, physical harm, and eroded public trust. These challenges highlight the urgent need for robust AI governance. We propose a comprehensive framework integrating technical and societal dimensions, structured around three interconnected pillars: Intrinsic Security (system reliability), Derivative Security (real-world harm mitigation), and Social Ethics (value alignment and accountability). Uniquely, our approach unifies technical methods, emerging evaluation benchmarks, and policy insights to promote transparency, accountability, and trust in AI systems. Through a systematic review of over 300 studies, we identify three core challenges: (1) the generalization gap, where defenses fail against evolving threats; (2) inadequate evaluation protocols that overlook real-world risks; and (3) fragmented regulations leading to inconsistent oversight. These shortcomings stem from treating governance as an afterthought, rather than a foundational design principle, resulting in reactive, siloed efforts that fail to address the interdependence of technical integrity and societal trust. To overcome this, we present an integrated research agenda that bridges technical rigor with social responsibility. Our framework offers actionable guidance for researchers, engineers, and policymakers to develop AI systems that are not only robust and secure but also ethically aligned and publicly trustworthy. The accompanying repository is available at https://github.com/ZTianle/Awesome-AI-SG.
Problem

Research questions and friction points this paper is trying to address.

Addressing AI vulnerabilities like bias and adversarial attacks
Bridging technical security with societal ethics in AI governance
Overcoming fragmented regulations and inadequate evaluation protocols
Innovation

Methods, ideas, or system contributions that make the work stand out.

Integrates technical and societal governance dimensions
Unifies methods, benchmarks, and policy for transparency
Bridges technical rigor with social responsibility
🔎 Similar Papers
No similar papers found.
Yuchu Jiang
Yuchu Jiang
Southeast University
Large Language ModelsComputer Vision
J
Jian Zhao
Institute of Artificial Intelligence (TeleAI), China Telecom; Northwestern Polytechnical University
Y
Yuchen Yuan
Institute of Artificial Intelligence (TeleAI), China Telecom
T
Tianle Zhang
Institute of Artificial Intelligence (TeleAI), China Telecom
Yao Huang
Yao Huang
Institute of Artificial Intelligence, Beihang University
Trustworthy MLMultimodal Learning
Yanghao Zhang
Yanghao Zhang
Imperial College London | Safe Intelligence
RobustnessAI safetyTrustworthy AI
Y
Yan Wang
Institute of Artificial Intelligence (TeleAI), China Telecom; University of Edinburgh
Yanshu Li
Yanshu Li
Brown University
NLPMultimodal Learning
X
Xizhong Guo
Institute of Artificial Intelligence (TeleAI), China Telecom; University of Chinese Academy of Sciences
Y
Yusheng Zhao
Institute of Artificial Intelligence (TeleAI), China Telecom; University of Science and Technology of China
J
Jun Zhang
Institute of Artificial Intelligence (TeleAI), China Telecom; University of Chinese Academy of Sciences
Z
Zhi Zhang
University of Amsterdam
X
Xiaojian Lin
Southeast University
Y
Yixiu Zou
Institute of Artificial Intelligence (TeleAI), China Telecom
Haoxuan Ma
Haoxuan Ma
University of California, Los Angeles
Intelligent Transportation SystemsMachine LearningAutomated Vehicle
Y
Yuhu Shang
Beijing University of Posts and Telecommunications
Y
Yuzhi Hu
University of Washington
K
Keshu Cai
University of Washington
Ruochen Zhang
Ruochen Zhang
Brown University
Multilingual NLPInterpretabilityCode-Switching
B
Boyuan Chen
Peking University
Y
Yilan Gao
Northwestern Polytechnical University
Z
Ziheng Jiao
Northwestern Polytechnical University
Yi Qin
Yi Qin
Chongqing University
signal processingfault diagnosisartificial intelligencemeasurement
S
Shuangjun Du
University of Chinese Academy of Sciences
X
Xiao Tong
Beijing Jiaotong University