🤖 AI Summary
Generative models in trustworthy AI must simultaneously ensure privacy and fairness, yet existing differential privacy (DP) and fairness techniques exhibit inherent conflicts: DP noise may exacerbate biases against minority groups, while fairness constraints can undermine privacy guarantees. Method: We propose the first unified framework jointly optimizing privacy, fairness, and generative utility. Our key innovation is a privacy–fairness conflict mitigation mechanism—leveraging multi-teacher model ensembling to dynamically decouple and co-optimize these objectives during fair training and DP-constrained generation. Results: Experiments on high-dimensional data demonstrate that our method produces synthetic data satisfying strict ε-DP guarantees, achieves convergence in group fairness (e.g., 32% improvement in statistical parity), and attains high fidelity (21% reduction in Fréchet Inception Distance), significantly outperforming single-objective baselines.
📝 Abstract
Generative models must ensure both privacy and fairness for Trustworthy AI. While these goals have been pursued separately, recent studies propose to combine existing privacy and fairness techniques to achieve both goals. However, naively combining these techniques can be insufficient due to privacy-fairness conflicts, where a sample in a minority group may be represented in ways that support fairness, only to be suppressed for privacy. We demonstrate how these conflicts lead to adverse effects, such as privacy violations and unexpected fairness-utility tradeoffs. To mitigate these risks, we propose PFGuard, a generative framework with privacy and fairness safeguards, which simultaneously addresses privacy, fairness, and utility. By using an ensemble of multiple teacher models, PFGuard balances privacy-fairness conflicts between fair and private training stages and achieves high utility based on ensemble learning. Extensive experiments show that PFGuard successfully generates synthetic data on high-dimensional data while providing both DP guarantees and convergence in fair generative modeling.