Differential Privacy in Machine Learning: From Symbolic AI to LLMs

📅 2025-06-13
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the vulnerability of machine learning models to sensitive data leakage, this paper establishes a unified differential privacy (DP) analysis framework spanning the entire AI stack—from symbolic systems to large language models (LLMs). Methodologically, it integrates DP-SGD, Laplace/Gaussian mechanisms, privacy budget allocation strategies, and tools for privacy-utility trade-off evaluation and adversarial robustness analysis. Innovatively, it proposes a novel DP adaptation paradigm for LLMs, accompanied by feasibility assessment criteria; it further identifies and systematizes the applicability boundaries of twelve mainstream DP-ML approaches—the first such comprehensive taxonomy. The work bridges foundational DP theory with generative AI practice, delivering dual theoretical rigor and engineering practicality to support secure, compliant, and verifiable AI systems. (132 words)

Technology Category

Application Category

📝 Abstract
Machine learning models should not reveal particular information that is not otherwise accessible. Differential privacy provides a formal framework to mitigate privacy risks by ensuring that the inclusion or exclusion of any single data point does not significantly alter the output of an algorithm, thus limiting the exposure of private information. This survey paper explores the foundational definitions of differential privacy, reviews its original formulations and tracing its evolution through key research contributions. It then provides an in-depth examination of how DP has been integrated into machine learning models, analyzing existing proposals and methods to preserve privacy when training ML models. Finally, it describes how DP-based ML techniques can be evaluated in practice. %Finally, it discusses the broader implications of DP, highlighting its potential for public benefit, its real-world applications, and the challenges it faces, including vulnerabilities to adversarial attacks. By offering a comprehensive overview of differential privacy in machine learning, this work aims to contribute to the ongoing development of secure and responsible AI systems.
Problem

Research questions and friction points this paper is trying to address.

Ensuring machine learning models do not reveal private information
Integrating differential privacy into ML training methods
Evaluating DP-based ML techniques in practical scenarios
Innovation

Methods, ideas, or system contributions that make the work stand out.

Differential privacy framework for machine learning
Integration of DP in ML model training
Evaluation methods for DP-based ML techniques
F
Francisco Aguilera-Martinez
Department of Computer Science and Artificial Intelligence, University of Granada, Spain
Fernando Berzal
Fernando Berzal
Associate Professor of Computer Science and A.I., University of Granada
Data MiningSoftware EngineeringComplex NetworksMachine LearningDeep Learning