Sy-FAR: Symmetry-based Fair Adversarial Robustness

📅 2025-09-16
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses *unfair robustness*—a critical issue in security-sensitive tasks such as face recognition—where adversarial vulnerability varies significantly across classes or demographic groups. Instead of pursuing conventional fairness objectives, we propose *inter-class symmetry* as a novel fairness criterion, extending individual-level symmetry theory to the group level for the first time, and identifying and mitigating the previously overlooked phenomenon of *target-class vulnerability*. Methodologically, we introduce a symmetry-regularized term into the adversarial training framework, enforcing symmetric attack-defense responses across all classes. Extensive experiments on five benchmark datasets demonstrate that our approach significantly improves fair adversarial robustness: it converges faster, yields more stable results, and notably strengthens protection for high-risk target classes—outperforming state-of-the-art methods.

Technology Category

Application Category

📝 Abstract
Security-critical machine-learning (ML) systems, such as face-recognition systems, are susceptible to adversarial examples, including real-world physically realizable attacks. Various means to boost ML's adversarial robustness have been proposed; however, they typically induce unfair robustness: It is often easier to attack from certain classes or groups than from others. Several techniques have been developed to improve adversarial robustness while seeking perfect fairness between classes. Yet, prior work has focused on settings where security and fairness are less critical. Our insight is that achieving perfect parity in realistic fairness-critical tasks, such as face recognition, is often infeasible -- some classes may be highly similar, leading to more misclassifications between them. Instead, we suggest that seeking symmetry -- i.e., attacks from class $i$ to $j$ would be as successful as from $j$ to $i$ -- is more tractable. Intuitively, symmetry is a desirable because class resemblance is a symmetric relation in most domains. Additionally, as we prove theoretically, symmetry between individuals induces symmetry between any set of sub-groups, in contrast to other fairness notions where group-fairness is often elusive. We develop Sy-FAR, a technique to encourage symmetry while also optimizing adversarial robustness and extensively evaluate it using five datasets, with three model architectures, including against targeted and untargeted realistic attacks. The results show Sy-FAR significantly improves fair adversarial robustness compared to state-of-the-art methods. Moreover, we find that Sy-FAR is faster and more consistent across runs. Notably, Sy-FAR also ameliorates another type of unfairness we discover in this work -- target classes that adversarial examples are likely to be classified into become significantly less vulnerable after inducing symmetry.
Problem

Research questions and friction points this paper is trying to address.

Achieving fair adversarial robustness across classes in ML systems
Addressing symmetry in attack success between different classes
Improving consistency and speed in adversarial robustness methods
Innovation

Methods, ideas, or system contributions that make the work stand out.

Symmetry-based adversarial robustness optimization
Encourages symmetric attack success rates
Improves fairness across classes and subgroups
🔎 Similar Papers
No similar papers found.