🤖 AI Summary
This work addresses the inadequacy of existing AI governance frameworks, which are rooted in static, centralized von Neumann architectures and fail to accommodate NeuroAI systems driven by neuromorphic hardware and spiking neural networks. For the first time, it extends AI governance into the domain of neuromorphic computing, arguing that governance mechanisms must co-evolve with the physical characteristics, learning dynamics, and embodied efficiency inherent to brain-inspired computation. By integrating spiking neural network modeling, neuromorphic hardware simulation, AI governance theory, and technical auditing methodologies, the study develops a verifiable and auditable regulatory framework tailored to NeuroAI. It elucidates the fundamental reasons for the failure of conventional governance approaches in this context and establishes technical governance principles aligned with this emerging computational paradigm.
📝 Abstract
Current AI governance frameworks, including regulatory benchmarks for accuracy, latency, and energy efficiency, are built for static, centrally trained artificial neural networks on von Neumann hardware. NeuroAI systems, embodied in neuromorphic hardware and implemented via spiking neural networks, break these assumptions. This paper examines the limitations of current AI governance frameworks for NeuroAI, arguing that assurance and audit methods must co-evolve with these architectures, aligning traditional regulatory metrics with the physics, learning dynamics, and embodied efficiency of brain-inspired computation to enable technically grounded assurance.