🤖 AI Summary
Existing deepfake detection methods struggle to distinguish benign content from adversarial deepfakes deliberately designed to manipulate public perception—especially in realistic social media settings. This paper proposes a harm-oriented detection paradigm grounded in social context, introduces SocialDF—the first publicly available benchmark dataset covering multiple platforms and attack types—and designs a multimodal detection framework leveraging LLM-based multi-agent collaborative reasoning. The framework integrates face recognition, automatic speech recognition (ASR), cross-modal alignment, and contextual consistency analysis, enabling fine-grained harm assessment via multi-factor cross-validation. It achieves 92.7% accuracy on real-world social deepfake samples, significantly outperforming unimodal baselines. Key contributions are: (1) the first social-context-driven harm discrimination paradigm; (2) the open-sourced SocialDF benchmark—the first harm-oriented deepfake benchmark; and (3) empirical validation of LLM-based multi-agent collaborative reasoning for robust deepfake detection.
📝 Abstract
The rapid advancement of deep generative models has significantly improved the realism of synthetic media, presenting both opportunities and security challenges. While deepfake technology has valuable applications in entertainment and accessibility, it has emerged as a potent vector for misinformation campaigns, particularly on social media. Existing detection frameworks struggle to distinguish between benign and adversarially generated deepfakes engineered to manipulate public perception. To address this challenge, we introduce SocialDF, a curated dataset reflecting real-world deepfake challenges on social media platforms. This dataset encompasses high-fidelity deepfakes sourced from various online ecosystems, ensuring broad coverage of manipulative techniques. We propose a novel LLM-based multi-factor detection approach that combines facial recognition, automated speech transcription, and a multi-agent LLM pipeline to cross-verify audio-visual cues. Our methodology emphasizes robust, multi-modal verification techniques that incorporate linguistic, behavioral, and contextual analysis to effectively discern synthetic media from authentic content.