Divisive Decisions: Improving Salience-Based Training for Generalization in Binary Classification Tasks

๐Ÿ“… 2025-07-22
๐Ÿ“ˆ Citations: 0
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
Existing saliency-guided training solely aligns ground-truth class CAMs with human-annotated saliency maps, neglecting discriminative information from erroneous-class CAMs. To address insufficient generalization in binary classification tasks, this work introduces erroneous-class CAMs for the first time and proposes a โ€œTrue-vs.-False Class Activation Differenceโ€ mechanism. By modeling the contrastive responses of true and false class CAMs within human-identified critical feature regions, we devise three novel fused training strategies and a post-hoc explainability analysis tool. Our approach integrates CAMs, saliency-aware loss, and contrastive learning, supervised by human-annotated saliency maps. Extensive experiments on face forgery detection, biometric attack identification, and chest X-ray abnormality classification demonstrate that our method significantly outperforms state-of-the-art saliency-guided approaches, achieving superior generalization and enhanced focus on task-critical features.

Technology Category

Application Category

๐Ÿ“ Abstract
Existing saliency-guided training approaches improve model generalization by incorporating a loss term that compares the model's class activation map (CAM) for a sample's true-class ({it i.e.}, correct-label class) against a human reference saliency map. However, prior work has ignored the false-class CAM(s), that is the model's saliency obtained for incorrect-label class. We hypothesize that in binary tasks the true and false CAMs should diverge on the important classification features identified by humans (and reflected in human saliency maps). We use this hypothesis to motivate three new saliency-guided training methods incorporating both true- and false-class model's CAM into the training strategy and a novel post-hoc tool for identifying important features. We evaluate all introduced methods on several diverse binary close-set and open-set classification tasks, including synthetic face detection, biometric presentation attack detection, and classification of anomalies in chest X-ray scans, and find that the proposed methods improve generalization capabilities of deep learning models over traditional (true-class CAM only) saliency-guided training approaches. We offer source codes and model weightsfootnote{GitHub repository link removed to preserve anonymity} to support reproducible research.
Problem

Research questions and friction points this paper is trying to address.

Improving model generalization using true and false class activation maps
Enhancing binary classification with saliency-guided training methods
Evaluating methods on diverse tasks like face detection and X-ray analysis
Innovation

Methods, ideas, or system contributions that make the work stand out.

Incorporates true and false class CAMs
Introduces three new saliency-guided methods
Improves model generalization in binary tasks
๐Ÿ”Ž Similar Papers
No similar papers found.
J
Jacob Piland
Department of Computer Science and Engineering, University of Notre Dame
C
Christopher Sweet
Center for Research Computing, University of Notre Dame
Adam Czajka
Adam Czajka
University of Notre Dame
BiometricsComputer VisionIris RecognitionPresentation Attack DetectionPost-mortem Biometrics