🤖 AI Summary
This paper systematically exposes the multidimensional threats posed by malicious AI swarms to democratic institutions—specifically, covert coordination, community infiltration, and persistent A/B testing to fabricate artificial grassroots consensus, fracture shared reality, enable micro-targeted voter suppression/mobilization, poison training data, and erode institutional trust.
Method: It introduces the first formal threat model of “malicious AI swarms” and proposes a three-tiered defense architecture: platform-level (AI Influence Observatory), model-level (digital watermarking and provenance authentication), and system-level (multi-agent simulation coupled with real-time detection dashboards).
Contribution/Results: The authors develop a deployable toolchain for swarm detection and adversarial stress testing, enabling transparent platform auditing and end-user protection. The framework bridges theory and practice, delivering a policy-technology co-design pathway for global AI governance that is both theoretically rigorous and operationally feasible.
📝 Abstract
Advances in AI portend a new era of sophisticated disinformation operations. While individual AI systems already create convincing -- and at times misleading -- information, an imminent development is the emergence of malicious AI swarms. These systems can coordinate covertly, infiltrate communities, evade traditional detectors, and run continuous A/B tests, with round-the-clock persistence. The result can include fabricated grassroots consensus, fragmented shared reality, mass harassment, voter micro-suppression or mobilization, contamination of AI training data, and erosion of institutional trust. With democratic processes worldwide increasingly vulnerable, we urge a three-pronged response: (1) platform-side defenses -- always-on swarm-detection dashboards, pre-election high-fidelity swarm-simulation stress-tests, transparency audits, and optional client-side"AI shields"for users; (2) model-side safeguards -- standardized persuasion-risk tests, provenance-authenticating passkeys, and watermarking; and (3) system-level oversight -- a UN-backed AI Influence Observatory.